commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
f97d30c16de2b0e57a3ffb96c803bcafc3b0dca7
emission/core/wrapper/modestattimesummary.py
emission/core/wrapper/modestattimesummary.py
import logging import emission.core.wrapper.wrapperbase as ecwb import emission.core.wrapper.motionactivity as ecwm # Used for various metrics such as count, distance, mean speed calorie consumption, # median speed calorie consumption # Should come later: carbon footprint # Optimal doesn't look like it fits this, because it is not per mode class ModeStatTimeSummary(ecwb.WrapperBase): # We will end up with props like # { # MotionTypes.IN_VEHICLE: ecwb.WrapperBase.Access.WORM # } # Each distance will have # # props = dict([(t.name, ecwb.WrapperBase.Access.WORM) for t in ecwm.MotionTypes]) props.update( {'ts': ecwb.WrapperBase.Access.WORM, # YYYY-MM-DD 'local_dt': ecwb.WrapperBase.Access.WORM, 'fmt_time': ecwb.WrapperBase.Access.WORM} # YYYY-MM-DD ) enums = {} geojson = [] nullable = [] local_dates = ['end_local_dt'] def _populateDependencies(self): pass
import logging import emission.core.wrapper.wrapperbase as ecwb import emission.core.wrapper.motionactivity as ecwm # Used for various metrics such as count, distance, mean speed calorie consumption, # median speed calorie consumption # Should come later: carbon footprint # Optimal doesn't look like it fits this, because it is not per mode class ModeStatTimeSummary(ecwb.WrapperBase): # We will end up with props like # { # MotionTypes.IN_VEHICLE: ecwb.WrapperBase.Access.WORM # } # Each distance will have # # props = dict([(t.name, ecwb.WrapperBase.Access.WORM) for t in ecwm.MotionTypes]) props.update( {'ts': ecwb.WrapperBase.Access.WORM, # YYYY-MM-DD 'local_dt': ecwb.WrapperBase.Access.WORM, 'fmt_time': ecwb.WrapperBase.Access.WORM} # YYYY-MM-DD ) enums = {} geojson = [] nullable = [] local_dates = ['local_dt'] def _populateDependencies(self): pass
Fix the list of local_date objects in the ModeStatTimeSummary
Fix the list of local_date objects in the ModeStatTimeSummary
Python
bsd-3-clause
yw374cornell/e-mission-server,shankari/e-mission-server,e-mission/e-mission-server,yw374cornell/e-mission-server,sunil07t/e-mission-server,sunil07t/e-mission-server,e-mission/e-mission-server,yw374cornell/e-mission-server,shankari/e-mission-server,sunil07t/e-mission-server,sunil07t/e-mission-server,shankari/e-mission-server,shankari/e-mission-server,e-mission/e-mission-server,e-mission/e-mission-server,yw374cornell/e-mission-server
import logging import emission.core.wrapper.wrapperbase as ecwb import emission.core.wrapper.motionactivity as ecwm # Used for various metrics such as count, distance, mean speed calorie consumption, # median speed calorie consumption # Should come later: carbon footprint # Optimal doesn't look like it fits this, because it is not per mode class ModeStatTimeSummary(ecwb.WrapperBase): # We will end up with props like # { # MotionTypes.IN_VEHICLE: ecwb.WrapperBase.Access.WORM # } # Each distance will have # # props = dict([(t.name, ecwb.WrapperBase.Access.WORM) for t in ecwm.MotionTypes]) props.update( {'ts': ecwb.WrapperBase.Access.WORM, # YYYY-MM-DD 'local_dt': ecwb.WrapperBase.Access.WORM, 'fmt_time': ecwb.WrapperBase.Access.WORM} # YYYY-MM-DD ) enums = {} geojson = [] nullable = [] local_dates = ['end_local_dt'] def _populateDependencies(self): pass Fix the list of local_date objects in the ModeStatTimeSummary
import logging import emission.core.wrapper.wrapperbase as ecwb import emission.core.wrapper.motionactivity as ecwm # Used for various metrics such as count, distance, mean speed calorie consumption, # median speed calorie consumption # Should come later: carbon footprint # Optimal doesn't look like it fits this, because it is not per mode class ModeStatTimeSummary(ecwb.WrapperBase): # We will end up with props like # { # MotionTypes.IN_VEHICLE: ecwb.WrapperBase.Access.WORM # } # Each distance will have # # props = dict([(t.name, ecwb.WrapperBase.Access.WORM) for t in ecwm.MotionTypes]) props.update( {'ts': ecwb.WrapperBase.Access.WORM, # YYYY-MM-DD 'local_dt': ecwb.WrapperBase.Access.WORM, 'fmt_time': ecwb.WrapperBase.Access.WORM} # YYYY-MM-DD ) enums = {} geojson = [] nullable = [] local_dates = ['local_dt'] def _populateDependencies(self): pass
<commit_before>import logging import emission.core.wrapper.wrapperbase as ecwb import emission.core.wrapper.motionactivity as ecwm # Used for various metrics such as count, distance, mean speed calorie consumption, # median speed calorie consumption # Should come later: carbon footprint # Optimal doesn't look like it fits this, because it is not per mode class ModeStatTimeSummary(ecwb.WrapperBase): # We will end up with props like # { # MotionTypes.IN_VEHICLE: ecwb.WrapperBase.Access.WORM # } # Each distance will have # # props = dict([(t.name, ecwb.WrapperBase.Access.WORM) for t in ecwm.MotionTypes]) props.update( {'ts': ecwb.WrapperBase.Access.WORM, # YYYY-MM-DD 'local_dt': ecwb.WrapperBase.Access.WORM, 'fmt_time': ecwb.WrapperBase.Access.WORM} # YYYY-MM-DD ) enums = {} geojson = [] nullable = [] local_dates = ['end_local_dt'] def _populateDependencies(self): pass <commit_msg>Fix the list of local_date objects in the ModeStatTimeSummary<commit_after>
import logging import emission.core.wrapper.wrapperbase as ecwb import emission.core.wrapper.motionactivity as ecwm # Used for various metrics such as count, distance, mean speed calorie consumption, # median speed calorie consumption # Should come later: carbon footprint # Optimal doesn't look like it fits this, because it is not per mode class ModeStatTimeSummary(ecwb.WrapperBase): # We will end up with props like # { # MotionTypes.IN_VEHICLE: ecwb.WrapperBase.Access.WORM # } # Each distance will have # # props = dict([(t.name, ecwb.WrapperBase.Access.WORM) for t in ecwm.MotionTypes]) props.update( {'ts': ecwb.WrapperBase.Access.WORM, # YYYY-MM-DD 'local_dt': ecwb.WrapperBase.Access.WORM, 'fmt_time': ecwb.WrapperBase.Access.WORM} # YYYY-MM-DD ) enums = {} geojson = [] nullable = [] local_dates = ['local_dt'] def _populateDependencies(self): pass
import logging import emission.core.wrapper.wrapperbase as ecwb import emission.core.wrapper.motionactivity as ecwm # Used for various metrics such as count, distance, mean speed calorie consumption, # median speed calorie consumption # Should come later: carbon footprint # Optimal doesn't look like it fits this, because it is not per mode class ModeStatTimeSummary(ecwb.WrapperBase): # We will end up with props like # { # MotionTypes.IN_VEHICLE: ecwb.WrapperBase.Access.WORM # } # Each distance will have # # props = dict([(t.name, ecwb.WrapperBase.Access.WORM) for t in ecwm.MotionTypes]) props.update( {'ts': ecwb.WrapperBase.Access.WORM, # YYYY-MM-DD 'local_dt': ecwb.WrapperBase.Access.WORM, 'fmt_time': ecwb.WrapperBase.Access.WORM} # YYYY-MM-DD ) enums = {} geojson = [] nullable = [] local_dates = ['end_local_dt'] def _populateDependencies(self): pass Fix the list of local_date objects in the ModeStatTimeSummaryimport logging import emission.core.wrapper.wrapperbase as ecwb import emission.core.wrapper.motionactivity as ecwm # Used for various metrics such as count, distance, mean speed calorie consumption, # median speed calorie consumption # Should come later: carbon footprint # Optimal doesn't look like it fits this, because it is not per mode class ModeStatTimeSummary(ecwb.WrapperBase): # We will end up with props like # { # MotionTypes.IN_VEHICLE: ecwb.WrapperBase.Access.WORM # } # Each distance will have # # props = dict([(t.name, ecwb.WrapperBase.Access.WORM) for t in ecwm.MotionTypes]) props.update( {'ts': ecwb.WrapperBase.Access.WORM, # YYYY-MM-DD 'local_dt': ecwb.WrapperBase.Access.WORM, 'fmt_time': ecwb.WrapperBase.Access.WORM} # YYYY-MM-DD ) enums = {} geojson = [] nullable = [] local_dates = ['local_dt'] def _populateDependencies(self): pass
<commit_before>import logging import emission.core.wrapper.wrapperbase as ecwb import emission.core.wrapper.motionactivity as ecwm # Used for various metrics such as count, distance, mean speed calorie consumption, # median speed calorie consumption # Should come later: carbon footprint # Optimal doesn't look like it fits this, because it is not per mode class ModeStatTimeSummary(ecwb.WrapperBase): # We will end up with props like # { # MotionTypes.IN_VEHICLE: ecwb.WrapperBase.Access.WORM # } # Each distance will have # # props = dict([(t.name, ecwb.WrapperBase.Access.WORM) for t in ecwm.MotionTypes]) props.update( {'ts': ecwb.WrapperBase.Access.WORM, # YYYY-MM-DD 'local_dt': ecwb.WrapperBase.Access.WORM, 'fmt_time': ecwb.WrapperBase.Access.WORM} # YYYY-MM-DD ) enums = {} geojson = [] nullable = [] local_dates = ['end_local_dt'] def _populateDependencies(self): pass <commit_msg>Fix the list of local_date objects in the ModeStatTimeSummary<commit_after>import logging import emission.core.wrapper.wrapperbase as ecwb import emission.core.wrapper.motionactivity as ecwm # Used for various metrics such as count, distance, mean speed calorie consumption, # median speed calorie consumption # Should come later: carbon footprint # Optimal doesn't look like it fits this, because it is not per mode class ModeStatTimeSummary(ecwb.WrapperBase): # We will end up with props like # { # MotionTypes.IN_VEHICLE: ecwb.WrapperBase.Access.WORM # } # Each distance will have # # props = dict([(t.name, ecwb.WrapperBase.Access.WORM) for t in ecwm.MotionTypes]) props.update( {'ts': ecwb.WrapperBase.Access.WORM, # YYYY-MM-DD 'local_dt': ecwb.WrapperBase.Access.WORM, 'fmt_time': ecwb.WrapperBase.Access.WORM} # YYYY-MM-DD ) enums = {} geojson = [] nullable = [] local_dates = ['local_dt'] def _populateDependencies(self): pass
e4a3d3c273a1b7e26e9fdcf7f8da060bf127f27e
examples/django_project/django_project/tests.py
examples/django_project/django_project/tests.py
import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): # Note this tests should fail unless the snapshot-update command line # option is specified. Run `python manage.py test --snapshot-update`. now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
Add note for test_api_me django-project test
Add note for test_api_me django-project test
Python
mit
syrusakbary/snapshottest
import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main() Add note for test_api_me django-project test
import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): # Note this tests should fail unless the snapshot-update command line # option is specified. Run `python manage.py test --snapshot-update`. now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
<commit_before>import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main() <commit_msg>Add note for test_api_me django-project test<commit_after>
import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): # Note this tests should fail unless the snapshot-update command line # option is specified. Run `python manage.py test --snapshot-update`. now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main() Add note for test_api_me django-project testimport unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): # Note this tests should fail unless the snapshot-update command line # option is specified. Run `python manage.py test --snapshot-update`. now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
<commit_before>import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main() <commit_msg>Add note for test_api_me django-project test<commit_after>import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): # Note this tests should fail unless the snapshot-update command line # option is specified. Run `python manage.py test --snapshot-update`. now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
c1d906da71571552de92f028c91b3fc2ec2d28de
test/unit/registration/test_ants.py
test/unit/registration/test_ants.py
from nose.tools import * import os, shutil import logging logger = logging.getLogger(__name__) import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', '..', '..')) from qipipe.registration import ants # The test parent directory. ROOT = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) # The test fixture. FIXTURE = os.path.join(ROOT, 'fixtures', 'registration', 'breast', 'subject03', 'session01') # The test results. RESULT = os.path.join(ROOT, 'results', 'registration', 'ants') WORK = os.path.join(RESULT, 'work') OUTPUT = os.path.join(RESULT, 'registered') class TestANTS: """ANTS registration unit tests.""" def setup(self): shutil.rmtree(RESULT, True) def teardown(self): shutil.rmtree(RESULT, True) def test_registration(self): rdict = ants.register(FIXTURE, output=OUTPUT, work=WORK) # Verify that each input is registered. for fn in os.listdir(FIXTURE): f = os.path.join(FIXTURE, fn) rfn = fn.replace('.dcm', 'Registered.nii.gz') assert_equal(rfn, rdict[fn], "Missing registration mapping: %s" % rfn) if __name__ == "__main__": import nose nose.main(defaultTest=__name__)
from nose.tools import * import os, shutil import logging logger = logging.getLogger(__name__) import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', '..', '..')) from qipipe.registration import ants # The test parent directory. ROOT = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) # The test fixture. FIXTURE = os.path.join(ROOT, 'fixtures', 'registration', 'breast', 'Breast03', 'Session01') # The test results. RESULT = os.path.join(ROOT, 'results', 'registration', 'ants') WORK = os.path.join(RESULT, 'work') OUTPUT = os.path.join(RESULT, 'registered') class TestANTS: """ANTS registration unit tests.""" def setup(self): shutil.rmtree(RESULT, True) def teardown(self): shutil.rmtree(RESULT, True) def test_registration(self): rdict = ants.register(FIXTURE, output=OUTPUT, work=WORK) # Verify that each input is registered. for fn in os.listdir(FIXTURE): f = os.path.join(FIXTURE, fn) rfn = fn.replace('.dcm', 'Registered.nii.gz') assert_equal(rfn, rdict[fn], "Missing registration mapping: %s" % rfn) if __name__ == "__main__": import nose nose.main(defaultTest=__name__)
Standardize the fixture directory names.
Standardize the fixture directory names.
Python
bsd-2-clause
ohsu-qin/qipipe
from nose.tools import * import os, shutil import logging logger = logging.getLogger(__name__) import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', '..', '..')) from qipipe.registration import ants # The test parent directory. ROOT = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) # The test fixture. FIXTURE = os.path.join(ROOT, 'fixtures', 'registration', 'breast', 'subject03', 'session01') # The test results. RESULT = os.path.join(ROOT, 'results', 'registration', 'ants') WORK = os.path.join(RESULT, 'work') OUTPUT = os.path.join(RESULT, 'registered') class TestANTS: """ANTS registration unit tests.""" def setup(self): shutil.rmtree(RESULT, True) def teardown(self): shutil.rmtree(RESULT, True) def test_registration(self): rdict = ants.register(FIXTURE, output=OUTPUT, work=WORK) # Verify that each input is registered. for fn in os.listdir(FIXTURE): f = os.path.join(FIXTURE, fn) rfn = fn.replace('.dcm', 'Registered.nii.gz') assert_equal(rfn, rdict[fn], "Missing registration mapping: %s" % rfn) if __name__ == "__main__": import nose nose.main(defaultTest=__name__) Standardize the fixture directory names.
from nose.tools import * import os, shutil import logging logger = logging.getLogger(__name__) import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', '..', '..')) from qipipe.registration import ants # The test parent directory. ROOT = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) # The test fixture. FIXTURE = os.path.join(ROOT, 'fixtures', 'registration', 'breast', 'Breast03', 'Session01') # The test results. RESULT = os.path.join(ROOT, 'results', 'registration', 'ants') WORK = os.path.join(RESULT, 'work') OUTPUT = os.path.join(RESULT, 'registered') class TestANTS: """ANTS registration unit tests.""" def setup(self): shutil.rmtree(RESULT, True) def teardown(self): shutil.rmtree(RESULT, True) def test_registration(self): rdict = ants.register(FIXTURE, output=OUTPUT, work=WORK) # Verify that each input is registered. for fn in os.listdir(FIXTURE): f = os.path.join(FIXTURE, fn) rfn = fn.replace('.dcm', 'Registered.nii.gz') assert_equal(rfn, rdict[fn], "Missing registration mapping: %s" % rfn) if __name__ == "__main__": import nose nose.main(defaultTest=__name__)
<commit_before>from nose.tools import * import os, shutil import logging logger = logging.getLogger(__name__) import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', '..', '..')) from qipipe.registration import ants # The test parent directory. ROOT = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) # The test fixture. FIXTURE = os.path.join(ROOT, 'fixtures', 'registration', 'breast', 'subject03', 'session01') # The test results. RESULT = os.path.join(ROOT, 'results', 'registration', 'ants') WORK = os.path.join(RESULT, 'work') OUTPUT = os.path.join(RESULT, 'registered') class TestANTS: """ANTS registration unit tests.""" def setup(self): shutil.rmtree(RESULT, True) def teardown(self): shutil.rmtree(RESULT, True) def test_registration(self): rdict = ants.register(FIXTURE, output=OUTPUT, work=WORK) # Verify that each input is registered. for fn in os.listdir(FIXTURE): f = os.path.join(FIXTURE, fn) rfn = fn.replace('.dcm', 'Registered.nii.gz') assert_equal(rfn, rdict[fn], "Missing registration mapping: %s" % rfn) if __name__ == "__main__": import nose nose.main(defaultTest=__name__) <commit_msg>Standardize the fixture directory names.<commit_after>
from nose.tools import * import os, shutil import logging logger = logging.getLogger(__name__) import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', '..', '..')) from qipipe.registration import ants # The test parent directory. ROOT = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) # The test fixture. FIXTURE = os.path.join(ROOT, 'fixtures', 'registration', 'breast', 'Breast03', 'Session01') # The test results. RESULT = os.path.join(ROOT, 'results', 'registration', 'ants') WORK = os.path.join(RESULT, 'work') OUTPUT = os.path.join(RESULT, 'registered') class TestANTS: """ANTS registration unit tests.""" def setup(self): shutil.rmtree(RESULT, True) def teardown(self): shutil.rmtree(RESULT, True) def test_registration(self): rdict = ants.register(FIXTURE, output=OUTPUT, work=WORK) # Verify that each input is registered. for fn in os.listdir(FIXTURE): f = os.path.join(FIXTURE, fn) rfn = fn.replace('.dcm', 'Registered.nii.gz') assert_equal(rfn, rdict[fn], "Missing registration mapping: %s" % rfn) if __name__ == "__main__": import nose nose.main(defaultTest=__name__)
from nose.tools import * import os, shutil import logging logger = logging.getLogger(__name__) import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', '..', '..')) from qipipe.registration import ants # The test parent directory. ROOT = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) # The test fixture. FIXTURE = os.path.join(ROOT, 'fixtures', 'registration', 'breast', 'subject03', 'session01') # The test results. RESULT = os.path.join(ROOT, 'results', 'registration', 'ants') WORK = os.path.join(RESULT, 'work') OUTPUT = os.path.join(RESULT, 'registered') class TestANTS: """ANTS registration unit tests.""" def setup(self): shutil.rmtree(RESULT, True) def teardown(self): shutil.rmtree(RESULT, True) def test_registration(self): rdict = ants.register(FIXTURE, output=OUTPUT, work=WORK) # Verify that each input is registered. for fn in os.listdir(FIXTURE): f = os.path.join(FIXTURE, fn) rfn = fn.replace('.dcm', 'Registered.nii.gz') assert_equal(rfn, rdict[fn], "Missing registration mapping: %s" % rfn) if __name__ == "__main__": import nose nose.main(defaultTest=__name__) Standardize the fixture directory names.from nose.tools import * import os, shutil import logging logger = logging.getLogger(__name__) import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', '..', '..')) from qipipe.registration import ants # The test parent directory. ROOT = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) # The test fixture. FIXTURE = os.path.join(ROOT, 'fixtures', 'registration', 'breast', 'Breast03', 'Session01') # The test results. RESULT = os.path.join(ROOT, 'results', 'registration', 'ants') WORK = os.path.join(RESULT, 'work') OUTPUT = os.path.join(RESULT, 'registered') class TestANTS: """ANTS registration unit tests.""" def setup(self): shutil.rmtree(RESULT, True) def teardown(self): shutil.rmtree(RESULT, True) def test_registration(self): rdict = ants.register(FIXTURE, output=OUTPUT, work=WORK) # Verify that each input is registered. for fn in os.listdir(FIXTURE): f = os.path.join(FIXTURE, fn) rfn = fn.replace('.dcm', 'Registered.nii.gz') assert_equal(rfn, rdict[fn], "Missing registration mapping: %s" % rfn) if __name__ == "__main__": import nose nose.main(defaultTest=__name__)
<commit_before>from nose.tools import * import os, shutil import logging logger = logging.getLogger(__name__) import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', '..', '..')) from qipipe.registration import ants # The test parent directory. ROOT = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) # The test fixture. FIXTURE = os.path.join(ROOT, 'fixtures', 'registration', 'breast', 'subject03', 'session01') # The test results. RESULT = os.path.join(ROOT, 'results', 'registration', 'ants') WORK = os.path.join(RESULT, 'work') OUTPUT = os.path.join(RESULT, 'registered') class TestANTS: """ANTS registration unit tests.""" def setup(self): shutil.rmtree(RESULT, True) def teardown(self): shutil.rmtree(RESULT, True) def test_registration(self): rdict = ants.register(FIXTURE, output=OUTPUT, work=WORK) # Verify that each input is registered. for fn in os.listdir(FIXTURE): f = os.path.join(FIXTURE, fn) rfn = fn.replace('.dcm', 'Registered.nii.gz') assert_equal(rfn, rdict[fn], "Missing registration mapping: %s" % rfn) if __name__ == "__main__": import nose nose.main(defaultTest=__name__) <commit_msg>Standardize the fixture directory names.<commit_after>from nose.tools import * import os, shutil import logging logger = logging.getLogger(__name__) import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', '..', '..')) from qipipe.registration import ants # The test parent directory. ROOT = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) # The test fixture. FIXTURE = os.path.join(ROOT, 'fixtures', 'registration', 'breast', 'Breast03', 'Session01') # The test results. RESULT = os.path.join(ROOT, 'results', 'registration', 'ants') WORK = os.path.join(RESULT, 'work') OUTPUT = os.path.join(RESULT, 'registered') class TestANTS: """ANTS registration unit tests.""" def setup(self): shutil.rmtree(RESULT, True) def teardown(self): shutil.rmtree(RESULT, True) def test_registration(self): rdict = ants.register(FIXTURE, output=OUTPUT, work=WORK) # Verify that each input is registered. for fn in os.listdir(FIXTURE): f = os.path.join(FIXTURE, fn) rfn = fn.replace('.dcm', 'Registered.nii.gz') assert_equal(rfn, rdict[fn], "Missing registration mapping: %s" % rfn) if __name__ == "__main__": import nose nose.main(defaultTest=__name__)
857ff014ea7e889f75f741b191d82fadb09cd48d
integration_tests/telecommand/program_upload.py
integration_tests/telecommand/program_upload.py
import struct from telecommand import Telecommand class EraseBootTableEntry(Telecommand): def apid(self): return 0xB0 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return [mask] def __init__(self, entries): self._entries = entries class WriteProgramPart(Telecommand): MAX_PART_SIZE = Telecommand.MAX_PAYLOAD_SIZE - 3 def apid(self): return 0xB1 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BI', mask, self._offset)) + list(self._content) def __init__(self, entries, offset, content): self._offset = offset self._content = content self._entries = entries class FinalizeProgramEntry(Telecommand): def apid(self): return 0xB2 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BIH', mask, self._length, self._expected_crc)) + list(self._name) def __init__(self, entries, length, expected_crc, name): self._entries = entries self._length = length self._expected_crc = expected_crc self._name = name
import struct from telecommand import Telecommand class EraseBootTableEntry(Telecommand): def apid(self): return 0xB0 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return [mask] def __init__(self, entries): self._entries = entries class WriteProgramPart(Telecommand): MAX_PART_SIZE = Telecommand.MAX_PAYLOAD_SIZE - 5 def apid(self): return 0xB1 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BI', mask, self._offset)) + list(self._content) def __init__(self, entries, offset, content): self._offset = offset self._content = content self._entries = entries class FinalizeProgramEntry(Telecommand): def apid(self): return 0xB2 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BIH', mask, self._length, self._expected_crc)) + list(self._name) def __init__(self, entries, length, expected_crc, name): self._entries = entries self._length = length self._expected_crc = expected_crc self._name = name
Fix write program telecommand part size
Fix write program telecommand part size
Python
agpl-3.0
PW-Sat2/PWSat2OBC,PW-Sat2/PWSat2OBC,PW-Sat2/PWSat2OBC,PW-Sat2/PWSat2OBC
import struct from telecommand import Telecommand class EraseBootTableEntry(Telecommand): def apid(self): return 0xB0 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return [mask] def __init__(self, entries): self._entries = entries class WriteProgramPart(Telecommand): MAX_PART_SIZE = Telecommand.MAX_PAYLOAD_SIZE - 3 def apid(self): return 0xB1 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BI', mask, self._offset)) + list(self._content) def __init__(self, entries, offset, content): self._offset = offset self._content = content self._entries = entries class FinalizeProgramEntry(Telecommand): def apid(self): return 0xB2 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BIH', mask, self._length, self._expected_crc)) + list(self._name) def __init__(self, entries, length, expected_crc, name): self._entries = entries self._length = length self._expected_crc = expected_crc self._name = name Fix write program telecommand part size
import struct from telecommand import Telecommand class EraseBootTableEntry(Telecommand): def apid(self): return 0xB0 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return [mask] def __init__(self, entries): self._entries = entries class WriteProgramPart(Telecommand): MAX_PART_SIZE = Telecommand.MAX_PAYLOAD_SIZE - 5 def apid(self): return 0xB1 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BI', mask, self._offset)) + list(self._content) def __init__(self, entries, offset, content): self._offset = offset self._content = content self._entries = entries class FinalizeProgramEntry(Telecommand): def apid(self): return 0xB2 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BIH', mask, self._length, self._expected_crc)) + list(self._name) def __init__(self, entries, length, expected_crc, name): self._entries = entries self._length = length self._expected_crc = expected_crc self._name = name
<commit_before>import struct from telecommand import Telecommand class EraseBootTableEntry(Telecommand): def apid(self): return 0xB0 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return [mask] def __init__(self, entries): self._entries = entries class WriteProgramPart(Telecommand): MAX_PART_SIZE = Telecommand.MAX_PAYLOAD_SIZE - 3 def apid(self): return 0xB1 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BI', mask, self._offset)) + list(self._content) def __init__(self, entries, offset, content): self._offset = offset self._content = content self._entries = entries class FinalizeProgramEntry(Telecommand): def apid(self): return 0xB2 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BIH', mask, self._length, self._expected_crc)) + list(self._name) def __init__(self, entries, length, expected_crc, name): self._entries = entries self._length = length self._expected_crc = expected_crc self._name = name <commit_msg>Fix write program telecommand part size<commit_after>
import struct from telecommand import Telecommand class EraseBootTableEntry(Telecommand): def apid(self): return 0xB0 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return [mask] def __init__(self, entries): self._entries = entries class WriteProgramPart(Telecommand): MAX_PART_SIZE = Telecommand.MAX_PAYLOAD_SIZE - 5 def apid(self): return 0xB1 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BI', mask, self._offset)) + list(self._content) def __init__(self, entries, offset, content): self._offset = offset self._content = content self._entries = entries class FinalizeProgramEntry(Telecommand): def apid(self): return 0xB2 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BIH', mask, self._length, self._expected_crc)) + list(self._name) def __init__(self, entries, length, expected_crc, name): self._entries = entries self._length = length self._expected_crc = expected_crc self._name = name
import struct from telecommand import Telecommand class EraseBootTableEntry(Telecommand): def apid(self): return 0xB0 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return [mask] def __init__(self, entries): self._entries = entries class WriteProgramPart(Telecommand): MAX_PART_SIZE = Telecommand.MAX_PAYLOAD_SIZE - 3 def apid(self): return 0xB1 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BI', mask, self._offset)) + list(self._content) def __init__(self, entries, offset, content): self._offset = offset self._content = content self._entries = entries class FinalizeProgramEntry(Telecommand): def apid(self): return 0xB2 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BIH', mask, self._length, self._expected_crc)) + list(self._name) def __init__(self, entries, length, expected_crc, name): self._entries = entries self._length = length self._expected_crc = expected_crc self._name = name Fix write program telecommand part sizeimport struct from telecommand import Telecommand class EraseBootTableEntry(Telecommand): def apid(self): return 0xB0 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return [mask] def __init__(self, entries): self._entries = entries class WriteProgramPart(Telecommand): MAX_PART_SIZE = Telecommand.MAX_PAYLOAD_SIZE - 5 def apid(self): return 0xB1 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BI', mask, self._offset)) + list(self._content) def __init__(self, entries, offset, content): self._offset = offset self._content = content self._entries = entries class FinalizeProgramEntry(Telecommand): def apid(self): return 0xB2 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BIH', mask, self._length, self._expected_crc)) + list(self._name) def __init__(self, entries, length, expected_crc, name): self._entries = entries self._length = length self._expected_crc = expected_crc self._name = name
<commit_before>import struct from telecommand import Telecommand class EraseBootTableEntry(Telecommand): def apid(self): return 0xB0 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return [mask] def __init__(self, entries): self._entries = entries class WriteProgramPart(Telecommand): MAX_PART_SIZE = Telecommand.MAX_PAYLOAD_SIZE - 3 def apid(self): return 0xB1 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BI', mask, self._offset)) + list(self._content) def __init__(self, entries, offset, content): self._offset = offset self._content = content self._entries = entries class FinalizeProgramEntry(Telecommand): def apid(self): return 0xB2 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BIH', mask, self._length, self._expected_crc)) + list(self._name) def __init__(self, entries, length, expected_crc, name): self._entries = entries self._length = length self._expected_crc = expected_crc self._name = name <commit_msg>Fix write program telecommand part size<commit_after>import struct from telecommand import Telecommand class EraseBootTableEntry(Telecommand): def apid(self): return 0xB0 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return [mask] def __init__(self, entries): self._entries = entries class WriteProgramPart(Telecommand): MAX_PART_SIZE = Telecommand.MAX_PAYLOAD_SIZE - 5 def apid(self): return 0xB1 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BI', mask, self._offset)) + list(self._content) def __init__(self, entries, offset, content): self._offset = offset self._content = content self._entries = entries class FinalizeProgramEntry(Telecommand): def apid(self): return 0xB2 def payload(self): mask = 0 for e in self._entries: mask |= 1 << e return list(struct.pack('<BIH', mask, self._length, self._expected_crc)) + list(self._name) def __init__(self, entries, length, expected_crc, name): self._entries = entries self._length = length self._expected_crc = expected_crc self._name = name
c43fbdf1c0a4e976297f9fa75f97a2cfe4a0bee3
app/__init__.py
app/__init__.py
from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager from flask_oauthlib.client import OAuth from config import config db = SQLAlchemy() oauth = OAuth() lm = LoginManager() lm.login_view = "views.login" from app.models import User @lm.user_loader def load_user(id): return User.query.get(int(id)) def create_app(config_name): app = Flask(__name__) app.config.from_object(config[config_name]) db.init_app(app) lm.init_app(app) oauth.init_app(app) from app.views import views app.register_blueprint(views) return app
from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager from flask_oauthlib.client import OAuth from config import config db = SQLAlchemy() oa = OAuth() lm = LoginManager() lm.login_view = "views.login" from app.models import User @lm.user_loader def load_user(id): return User.query.get(int(id)) def create_app(config_name): app = Flask(__name__) app.config.from_object(config[config_name]) db.init_app(app) lm.init_app(app) oa.init_app(app) from app.views import views app.register_blueprint(views) return app
Rename OAuth var to avoid naming conflicts
Rename OAuth var to avoid naming conflicts
Python
mit
Encrylize/MyDictionary,Encrylize/MyDictionary,Encrylize/MyDictionary
from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager from flask_oauthlib.client import OAuth from config import config db = SQLAlchemy() oauth = OAuth() lm = LoginManager() lm.login_view = "views.login" from app.models import User @lm.user_loader def load_user(id): return User.query.get(int(id)) def create_app(config_name): app = Flask(__name__) app.config.from_object(config[config_name]) db.init_app(app) lm.init_app(app) oauth.init_app(app) from app.views import views app.register_blueprint(views) return appRename OAuth var to avoid naming conflicts
from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager from flask_oauthlib.client import OAuth from config import config db = SQLAlchemy() oa = OAuth() lm = LoginManager() lm.login_view = "views.login" from app.models import User @lm.user_loader def load_user(id): return User.query.get(int(id)) def create_app(config_name): app = Flask(__name__) app.config.from_object(config[config_name]) db.init_app(app) lm.init_app(app) oa.init_app(app) from app.views import views app.register_blueprint(views) return app
<commit_before>from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager from flask_oauthlib.client import OAuth from config import config db = SQLAlchemy() oauth = OAuth() lm = LoginManager() lm.login_view = "views.login" from app.models import User @lm.user_loader def load_user(id): return User.query.get(int(id)) def create_app(config_name): app = Flask(__name__) app.config.from_object(config[config_name]) db.init_app(app) lm.init_app(app) oauth.init_app(app) from app.views import views app.register_blueprint(views) return app<commit_msg>Rename OAuth var to avoid naming conflicts<commit_after>
from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager from flask_oauthlib.client import OAuth from config import config db = SQLAlchemy() oa = OAuth() lm = LoginManager() lm.login_view = "views.login" from app.models import User @lm.user_loader def load_user(id): return User.query.get(int(id)) def create_app(config_name): app = Flask(__name__) app.config.from_object(config[config_name]) db.init_app(app) lm.init_app(app) oa.init_app(app) from app.views import views app.register_blueprint(views) return app
from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager from flask_oauthlib.client import OAuth from config import config db = SQLAlchemy() oauth = OAuth() lm = LoginManager() lm.login_view = "views.login" from app.models import User @lm.user_loader def load_user(id): return User.query.get(int(id)) def create_app(config_name): app = Flask(__name__) app.config.from_object(config[config_name]) db.init_app(app) lm.init_app(app) oauth.init_app(app) from app.views import views app.register_blueprint(views) return appRename OAuth var to avoid naming conflictsfrom flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager from flask_oauthlib.client import OAuth from config import config db = SQLAlchemy() oa = OAuth() lm = LoginManager() lm.login_view = "views.login" from app.models import User @lm.user_loader def load_user(id): return User.query.get(int(id)) def create_app(config_name): app = Flask(__name__) app.config.from_object(config[config_name]) db.init_app(app) lm.init_app(app) oa.init_app(app) from app.views import views app.register_blueprint(views) return app
<commit_before>from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager from flask_oauthlib.client import OAuth from config import config db = SQLAlchemy() oauth = OAuth() lm = LoginManager() lm.login_view = "views.login" from app.models import User @lm.user_loader def load_user(id): return User.query.get(int(id)) def create_app(config_name): app = Flask(__name__) app.config.from_object(config[config_name]) db.init_app(app) lm.init_app(app) oauth.init_app(app) from app.views import views app.register_blueprint(views) return app<commit_msg>Rename OAuth var to avoid naming conflicts<commit_after>from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager from flask_oauthlib.client import OAuth from config import config db = SQLAlchemy() oa = OAuth() lm = LoginManager() lm.login_view = "views.login" from app.models import User @lm.user_loader def load_user(id): return User.query.get(int(id)) def create_app(config_name): app = Flask(__name__) app.config.from_object(config[config_name]) db.init_app(app) lm.init_app(app) oa.init_app(app) from app.views import views app.register_blueprint(views) return app
18a133cc3b56fdc1a5185bf3da3bed001e1b5c64
dictionary.py
dictionary.py
# Import non-standard modules import json def definition(word): '''This function returns the available definitions(s) of the input''' return data[word] # Load dictionary data from data.json to python dictionary data = json.load(open('data.json', 'r')) while True: ip = input("Enter word:") if ip in {'!q', '!Q'}: break elif data.__contains__(ip): print(definition(ip)) else: print("Please enter a valid word! \nEnter '!q' to quit!!!\n")
# Import non-standard modules import json from difflib import get_close_matches def definition(word): '''This function returns the available definitions(s) of the input''' return data[word] # Load dictionary data from data.json to python dictionary data = json.load(open('data.json', 'r')) # Infinite loop for processing while True: # Accept case-insensitive input from user ip = str(input("Enter word:")).lower() # Exit from program - user choice if ip == '!q': break # Check dictionary for definition elif data.__contains__(ip): print(definition(ip)) # If exact definition is not found, provide suggestion elif len(get_close_matches(ip, data.keys(), cutoff=0.8)) > 0: print("Did you mean to type", get_close_matches(ip, data.keys(), cutoff=0.8)[0], "?(y/n):") choice = str(input()).lower() # Provide output if generated suggestion is accepted if choice == 'y': ip = get_close_matches(ip, data.keys(), cutoff=0.8)[0] print(definition(ip)) # No suggestion or definition found else: print("No such word exists!! \nEnter '!q' to quit!!!")
Add feature to process user input and provide automatic suggestions
Add feature to process user input and provide automatic suggestions
Python
mit
jojenreed/Python-CLI-Dictionary
# Import non-standard modules import json def definition(word): '''This function returns the available definitions(s) of the input''' return data[word] # Load dictionary data from data.json to python dictionary data = json.load(open('data.json', 'r')) while True: ip = input("Enter word:") if ip in {'!q', '!Q'}: break elif data.__contains__(ip): print(definition(ip)) else: print("Please enter a valid word! \nEnter '!q' to quit!!!\n") Add feature to process user input and provide automatic suggestions
# Import non-standard modules import json from difflib import get_close_matches def definition(word): '''This function returns the available definitions(s) of the input''' return data[word] # Load dictionary data from data.json to python dictionary data = json.load(open('data.json', 'r')) # Infinite loop for processing while True: # Accept case-insensitive input from user ip = str(input("Enter word:")).lower() # Exit from program - user choice if ip == '!q': break # Check dictionary for definition elif data.__contains__(ip): print(definition(ip)) # If exact definition is not found, provide suggestion elif len(get_close_matches(ip, data.keys(), cutoff=0.8)) > 0: print("Did you mean to type", get_close_matches(ip, data.keys(), cutoff=0.8)[0], "?(y/n):") choice = str(input()).lower() # Provide output if generated suggestion is accepted if choice == 'y': ip = get_close_matches(ip, data.keys(), cutoff=0.8)[0] print(definition(ip)) # No suggestion or definition found else: print("No such word exists!! \nEnter '!q' to quit!!!")
<commit_before># Import non-standard modules import json def definition(word): '''This function returns the available definitions(s) of the input''' return data[word] # Load dictionary data from data.json to python dictionary data = json.load(open('data.json', 'r')) while True: ip = input("Enter word:") if ip in {'!q', '!Q'}: break elif data.__contains__(ip): print(definition(ip)) else: print("Please enter a valid word! \nEnter '!q' to quit!!!\n") <commit_msg>Add feature to process user input and provide automatic suggestions<commit_after>
# Import non-standard modules import json from difflib import get_close_matches def definition(word): '''This function returns the available definitions(s) of the input''' return data[word] # Load dictionary data from data.json to python dictionary data = json.load(open('data.json', 'r')) # Infinite loop for processing while True: # Accept case-insensitive input from user ip = str(input("Enter word:")).lower() # Exit from program - user choice if ip == '!q': break # Check dictionary for definition elif data.__contains__(ip): print(definition(ip)) # If exact definition is not found, provide suggestion elif len(get_close_matches(ip, data.keys(), cutoff=0.8)) > 0: print("Did you mean to type", get_close_matches(ip, data.keys(), cutoff=0.8)[0], "?(y/n):") choice = str(input()).lower() # Provide output if generated suggestion is accepted if choice == 'y': ip = get_close_matches(ip, data.keys(), cutoff=0.8)[0] print(definition(ip)) # No suggestion or definition found else: print("No such word exists!! \nEnter '!q' to quit!!!")
# Import non-standard modules import json def definition(word): '''This function returns the available definitions(s) of the input''' return data[word] # Load dictionary data from data.json to python dictionary data = json.load(open('data.json', 'r')) while True: ip = input("Enter word:") if ip in {'!q', '!Q'}: break elif data.__contains__(ip): print(definition(ip)) else: print("Please enter a valid word! \nEnter '!q' to quit!!!\n") Add feature to process user input and provide automatic suggestions# Import non-standard modules import json from difflib import get_close_matches def definition(word): '''This function returns the available definitions(s) of the input''' return data[word] # Load dictionary data from data.json to python dictionary data = json.load(open('data.json', 'r')) # Infinite loop for processing while True: # Accept case-insensitive input from user ip = str(input("Enter word:")).lower() # Exit from program - user choice if ip == '!q': break # Check dictionary for definition elif data.__contains__(ip): print(definition(ip)) # If exact definition is not found, provide suggestion elif len(get_close_matches(ip, data.keys(), cutoff=0.8)) > 0: print("Did you mean to type", get_close_matches(ip, data.keys(), cutoff=0.8)[0], "?(y/n):") choice = str(input()).lower() # Provide output if generated suggestion is accepted if choice == 'y': ip = get_close_matches(ip, data.keys(), cutoff=0.8)[0] print(definition(ip)) # No suggestion or definition found else: print("No such word exists!! \nEnter '!q' to quit!!!")
<commit_before># Import non-standard modules import json def definition(word): '''This function returns the available definitions(s) of the input''' return data[word] # Load dictionary data from data.json to python dictionary data = json.load(open('data.json', 'r')) while True: ip = input("Enter word:") if ip in {'!q', '!Q'}: break elif data.__contains__(ip): print(definition(ip)) else: print("Please enter a valid word! \nEnter '!q' to quit!!!\n") <commit_msg>Add feature to process user input and provide automatic suggestions<commit_after># Import non-standard modules import json from difflib import get_close_matches def definition(word): '''This function returns the available definitions(s) of the input''' return data[word] # Load dictionary data from data.json to python dictionary data = json.load(open('data.json', 'r')) # Infinite loop for processing while True: # Accept case-insensitive input from user ip = str(input("Enter word:")).lower() # Exit from program - user choice if ip == '!q': break # Check dictionary for definition elif data.__contains__(ip): print(definition(ip)) # If exact definition is not found, provide suggestion elif len(get_close_matches(ip, data.keys(), cutoff=0.8)) > 0: print("Did you mean to type", get_close_matches(ip, data.keys(), cutoff=0.8)[0], "?(y/n):") choice = str(input()).lower() # Provide output if generated suggestion is accepted if choice == 'y': ip = get_close_matches(ip, data.keys(), cutoff=0.8)[0] print(definition(ip)) # No suggestion or definition found else: print("No such word exists!! \nEnter '!q' to quit!!!")
cc838a311e891294da10ca7465782d71e622b5ef
dodge.py
dodge.py
import platform class OSXDodger(object): allowed_version = "10.12.1" def __init__(self, applications_dir): self.app_dir = applications_dir def load_applications(self): """ Read all applications in the `/Applications/` dir """ pass def select_applications(self): """ Allow user to select an application they want not to appear on the Dock """ pass def load_dodger_filer(self): """ Load the file to modify for the application chosen by the user in `select_applications` The file to be loaded for is `info.plist` """ pass def dodge_application(self): """ Remive the application from the Dock """ pass @classmethod def pc_is_macintosh(cls): """ Check if it is an `Apple Computer` i.e a Mac @return bool """ system = platform.system().lower() sys_version = int((platform.mac_ver())[0].replace(".", "")) allowed_version = int(cls.allowed_version.replace(".", "")) if (system == "darwin") and (sys_version >= allowed_version): return True else: print("\nSorry :(") print("FAILED. OsX-dock-dodger is only applicable to computers " + "running OS X {} or higher".format(cls.allowed_version)) return False dodge = OSXDodger("/Applications/") dodge.pc_is_macintosh()
import platform class OSXDodger(object): allowed_version = "10.12.1" def __init__(self, applications_dir): self.app_dir = applications_dir def load_applications(self): """ Read all applications in the `/Applications/` dir """ self.pc_is_macintosh() def select_applications(self): """ Allow user to select an application they want not to appear on the Dock """ pass def load_dodger_filer(self): """ Load the file to modify for the application chosen by the user in `select_applications` The file to be loaded for is `info.plist` """ pass def dodge_application(self): """ Remive the application from the Dock """ pass @classmethod def pc_is_macintosh(cls): """ Check if it is an `Apple Computer` i.e a Mac @return bool """ system = platform.system().lower() sys_version = int((platform.mac_ver())[0].replace(".", "")) allowed_version = int(cls.allowed_version.replace(".", "")) if (system == "darwin") and (sys_version >= allowed_version): return True else: print("\nSorry :(") print("FAILED. OsX-dock-dodger is only applicable to computers " + "running OS X {} or higher".format(cls.allowed_version)) return False dodge = OSXDodger("/Applications/") dodge.load_applications()
Add class method to check if PC is a Mac
Add class method to check if PC is a Mac
Python
mit
yoda-yoda/osx-dock-dodger,denisKaranja/osx-dock-dodger
import platform class OSXDodger(object): allowed_version = "10.12.1" def __init__(self, applications_dir): self.app_dir = applications_dir def load_applications(self): """ Read all applications in the `/Applications/` dir """ pass def select_applications(self): """ Allow user to select an application they want not to appear on the Dock """ pass def load_dodger_filer(self): """ Load the file to modify for the application chosen by the user in `select_applications` The file to be loaded for is `info.plist` """ pass def dodge_application(self): """ Remive the application from the Dock """ pass @classmethod def pc_is_macintosh(cls): """ Check if it is an `Apple Computer` i.e a Mac @return bool """ system = platform.system().lower() sys_version = int((platform.mac_ver())[0].replace(".", "")) allowed_version = int(cls.allowed_version.replace(".", "")) if (system == "darwin") and (sys_version >= allowed_version): return True else: print("\nSorry :(") print("FAILED. OsX-dock-dodger is only applicable to computers " + "running OS X {} or higher".format(cls.allowed_version)) return False dodge = OSXDodger("/Applications/") dodge.pc_is_macintosh() Add class method to check if PC is a Mac
import platform class OSXDodger(object): allowed_version = "10.12.1" def __init__(self, applications_dir): self.app_dir = applications_dir def load_applications(self): """ Read all applications in the `/Applications/` dir """ self.pc_is_macintosh() def select_applications(self): """ Allow user to select an application they want not to appear on the Dock """ pass def load_dodger_filer(self): """ Load the file to modify for the application chosen by the user in `select_applications` The file to be loaded for is `info.plist` """ pass def dodge_application(self): """ Remive the application from the Dock """ pass @classmethod def pc_is_macintosh(cls): """ Check if it is an `Apple Computer` i.e a Mac @return bool """ system = platform.system().lower() sys_version = int((platform.mac_ver())[0].replace(".", "")) allowed_version = int(cls.allowed_version.replace(".", "")) if (system == "darwin") and (sys_version >= allowed_version): return True else: print("\nSorry :(") print("FAILED. OsX-dock-dodger is only applicable to computers " + "running OS X {} or higher".format(cls.allowed_version)) return False dodge = OSXDodger("/Applications/") dodge.load_applications()
<commit_before>import platform class OSXDodger(object): allowed_version = "10.12.1" def __init__(self, applications_dir): self.app_dir = applications_dir def load_applications(self): """ Read all applications in the `/Applications/` dir """ pass def select_applications(self): """ Allow user to select an application they want not to appear on the Dock """ pass def load_dodger_filer(self): """ Load the file to modify for the application chosen by the user in `select_applications` The file to be loaded for is `info.plist` """ pass def dodge_application(self): """ Remive the application from the Dock """ pass @classmethod def pc_is_macintosh(cls): """ Check if it is an `Apple Computer` i.e a Mac @return bool """ system = platform.system().lower() sys_version = int((platform.mac_ver())[0].replace(".", "")) allowed_version = int(cls.allowed_version.replace(".", "")) if (system == "darwin") and (sys_version >= allowed_version): return True else: print("\nSorry :(") print("FAILED. OsX-dock-dodger is only applicable to computers " + "running OS X {} or higher".format(cls.allowed_version)) return False dodge = OSXDodger("/Applications/") dodge.pc_is_macintosh() <commit_msg>Add class method to check if PC is a Mac<commit_after>
import platform class OSXDodger(object): allowed_version = "10.12.1" def __init__(self, applications_dir): self.app_dir = applications_dir def load_applications(self): """ Read all applications in the `/Applications/` dir """ self.pc_is_macintosh() def select_applications(self): """ Allow user to select an application they want not to appear on the Dock """ pass def load_dodger_filer(self): """ Load the file to modify for the application chosen by the user in `select_applications` The file to be loaded for is `info.plist` """ pass def dodge_application(self): """ Remive the application from the Dock """ pass @classmethod def pc_is_macintosh(cls): """ Check if it is an `Apple Computer` i.e a Mac @return bool """ system = platform.system().lower() sys_version = int((platform.mac_ver())[0].replace(".", "")) allowed_version = int(cls.allowed_version.replace(".", "")) if (system == "darwin") and (sys_version >= allowed_version): return True else: print("\nSorry :(") print("FAILED. OsX-dock-dodger is only applicable to computers " + "running OS X {} or higher".format(cls.allowed_version)) return False dodge = OSXDodger("/Applications/") dodge.load_applications()
import platform class OSXDodger(object): allowed_version = "10.12.1" def __init__(self, applications_dir): self.app_dir = applications_dir def load_applications(self): """ Read all applications in the `/Applications/` dir """ pass def select_applications(self): """ Allow user to select an application they want not to appear on the Dock """ pass def load_dodger_filer(self): """ Load the file to modify for the application chosen by the user in `select_applications` The file to be loaded for is `info.plist` """ pass def dodge_application(self): """ Remive the application from the Dock """ pass @classmethod def pc_is_macintosh(cls): """ Check if it is an `Apple Computer` i.e a Mac @return bool """ system = platform.system().lower() sys_version = int((platform.mac_ver())[0].replace(".", "")) allowed_version = int(cls.allowed_version.replace(".", "")) if (system == "darwin") and (sys_version >= allowed_version): return True else: print("\nSorry :(") print("FAILED. OsX-dock-dodger is only applicable to computers " + "running OS X {} or higher".format(cls.allowed_version)) return False dodge = OSXDodger("/Applications/") dodge.pc_is_macintosh() Add class method to check if PC is a Macimport platform class OSXDodger(object): allowed_version = "10.12.1" def __init__(self, applications_dir): self.app_dir = applications_dir def load_applications(self): """ Read all applications in the `/Applications/` dir """ self.pc_is_macintosh() def select_applications(self): """ Allow user to select an application they want not to appear on the Dock """ pass def load_dodger_filer(self): """ Load the file to modify for the application chosen by the user in `select_applications` The file to be loaded for is `info.plist` """ pass def dodge_application(self): """ Remive the application from the Dock """ pass @classmethod def pc_is_macintosh(cls): """ Check if it is an `Apple Computer` i.e a Mac @return bool """ system = platform.system().lower() sys_version = int((platform.mac_ver())[0].replace(".", "")) allowed_version = int(cls.allowed_version.replace(".", "")) if (system == "darwin") and (sys_version >= allowed_version): return True else: print("\nSorry :(") print("FAILED. OsX-dock-dodger is only applicable to computers " + "running OS X {} or higher".format(cls.allowed_version)) return False dodge = OSXDodger("/Applications/") dodge.load_applications()
<commit_before>import platform class OSXDodger(object): allowed_version = "10.12.1" def __init__(self, applications_dir): self.app_dir = applications_dir def load_applications(self): """ Read all applications in the `/Applications/` dir """ pass def select_applications(self): """ Allow user to select an application they want not to appear on the Dock """ pass def load_dodger_filer(self): """ Load the file to modify for the application chosen by the user in `select_applications` The file to be loaded for is `info.plist` """ pass def dodge_application(self): """ Remive the application from the Dock """ pass @classmethod def pc_is_macintosh(cls): """ Check if it is an `Apple Computer` i.e a Mac @return bool """ system = platform.system().lower() sys_version = int((platform.mac_ver())[0].replace(".", "")) allowed_version = int(cls.allowed_version.replace(".", "")) if (system == "darwin") and (sys_version >= allowed_version): return True else: print("\nSorry :(") print("FAILED. OsX-dock-dodger is only applicable to computers " + "running OS X {} or higher".format(cls.allowed_version)) return False dodge = OSXDodger("/Applications/") dodge.pc_is_macintosh() <commit_msg>Add class method to check if PC is a Mac<commit_after>import platform class OSXDodger(object): allowed_version = "10.12.1" def __init__(self, applications_dir): self.app_dir = applications_dir def load_applications(self): """ Read all applications in the `/Applications/` dir """ self.pc_is_macintosh() def select_applications(self): """ Allow user to select an application they want not to appear on the Dock """ pass def load_dodger_filer(self): """ Load the file to modify for the application chosen by the user in `select_applications` The file to be loaded for is `info.plist` """ pass def dodge_application(self): """ Remive the application from the Dock """ pass @classmethod def pc_is_macintosh(cls): """ Check if it is an `Apple Computer` i.e a Mac @return bool """ system = platform.system().lower() sys_version = int((platform.mac_ver())[0].replace(".", "")) allowed_version = int(cls.allowed_version.replace(".", "")) if (system == "darwin") and (sys_version >= allowed_version): return True else: print("\nSorry :(") print("FAILED. OsX-dock-dodger is only applicable to computers " + "running OS X {} or higher".format(cls.allowed_version)) return False dodge = OSXDodger("/Applications/") dodge.load_applications()
967f9363872ff64847e4b93a7c1ea75869eaabd9
benchmarks/benchmarks.py
benchmarks/benchmarks.py
# Write the benchmarking functions here. # See "Writing benchmarks" in the asv docs for more information. import os.path as osp import numpy as np from tempfile import TemporaryDirectory import h5py class TimeSuite: """ An example benchmark that times the performance of various kinds of iterating over dictionaries in Python. """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') with h5py.File(path, 'w') as f: f['a'] = np.arange(100000) self.f = h5py.File(path, 'r') def teardown(self): self.f.close() self._td.cleanup() def time_many_small_reads(self): ds = self.f['a'] for i in range(10000): arr = ds[i * 10:(i + 1) * 10]
# Write the benchmarking functions here. # See "Writing benchmarks" in the asv docs for more information. import os.path as osp import numpy as np from tempfile import TemporaryDirectory import h5py class TimeSuite: """ An example benchmark that times the performance of various kinds of iterating over dictionaries in Python. """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') with h5py.File(path, 'w') as f: f['a'] = np.arange(100000) self.f = h5py.File(path, 'r') def teardown(self): self.f.close() self._td.cleanup() def time_many_small_reads(self): ds = self.f['a'] for i in range(10000): arr = ds[i * 10:(i + 1) * 10] class WritingTimeSuite: """Based on example in GitHub issue 492: https://github.com/h5py/h5py/issues/492 """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') self.f = h5py.File(path, 'w') self.shape = shape = (128, 1024, 512) self.f.create_dataset( 'a', shape=shape, dtype=np.float32, chunks=(1, shape[1], 64) ) def teardown(self): self.f.close() self._td.cleanup() def time_write_index_last_axis(self): ds = self.f['a'] data = np.zeros(self.shape[:2]) for i in range(self.shape[2]): ds[..., i] = data def time_write_slice_last_axis(self): ds = self.f['a'] data = np.zeros(self.shape[:2]) for i in range(self.shape[2]): ds[..., i:i+1] = data[..., np.newaxis]
Add benchmark for writing with index of last dimension
Add benchmark for writing with index of last dimension
Python
bsd-3-clause
h5py/h5py,h5py/h5py,h5py/h5py
# Write the benchmarking functions here. # See "Writing benchmarks" in the asv docs for more information. import os.path as osp import numpy as np from tempfile import TemporaryDirectory import h5py class TimeSuite: """ An example benchmark that times the performance of various kinds of iterating over dictionaries in Python. """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') with h5py.File(path, 'w') as f: f['a'] = np.arange(100000) self.f = h5py.File(path, 'r') def teardown(self): self.f.close() self._td.cleanup() def time_many_small_reads(self): ds = self.f['a'] for i in range(10000): arr = ds[i * 10:(i + 1) * 10] Add benchmark for writing with index of last dimension
# Write the benchmarking functions here. # See "Writing benchmarks" in the asv docs for more information. import os.path as osp import numpy as np from tempfile import TemporaryDirectory import h5py class TimeSuite: """ An example benchmark that times the performance of various kinds of iterating over dictionaries in Python. """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') with h5py.File(path, 'w') as f: f['a'] = np.arange(100000) self.f = h5py.File(path, 'r') def teardown(self): self.f.close() self._td.cleanup() def time_many_small_reads(self): ds = self.f['a'] for i in range(10000): arr = ds[i * 10:(i + 1) * 10] class WritingTimeSuite: """Based on example in GitHub issue 492: https://github.com/h5py/h5py/issues/492 """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') self.f = h5py.File(path, 'w') self.shape = shape = (128, 1024, 512) self.f.create_dataset( 'a', shape=shape, dtype=np.float32, chunks=(1, shape[1], 64) ) def teardown(self): self.f.close() self._td.cleanup() def time_write_index_last_axis(self): ds = self.f['a'] data = np.zeros(self.shape[:2]) for i in range(self.shape[2]): ds[..., i] = data def time_write_slice_last_axis(self): ds = self.f['a'] data = np.zeros(self.shape[:2]) for i in range(self.shape[2]): ds[..., i:i+1] = data[..., np.newaxis]
<commit_before># Write the benchmarking functions here. # See "Writing benchmarks" in the asv docs for more information. import os.path as osp import numpy as np from tempfile import TemporaryDirectory import h5py class TimeSuite: """ An example benchmark that times the performance of various kinds of iterating over dictionaries in Python. """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') with h5py.File(path, 'w') as f: f['a'] = np.arange(100000) self.f = h5py.File(path, 'r') def teardown(self): self.f.close() self._td.cleanup() def time_many_small_reads(self): ds = self.f['a'] for i in range(10000): arr = ds[i * 10:(i + 1) * 10] <commit_msg>Add benchmark for writing with index of last dimension<commit_after>
# Write the benchmarking functions here. # See "Writing benchmarks" in the asv docs for more information. import os.path as osp import numpy as np from tempfile import TemporaryDirectory import h5py class TimeSuite: """ An example benchmark that times the performance of various kinds of iterating over dictionaries in Python. """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') with h5py.File(path, 'w') as f: f['a'] = np.arange(100000) self.f = h5py.File(path, 'r') def teardown(self): self.f.close() self._td.cleanup() def time_many_small_reads(self): ds = self.f['a'] for i in range(10000): arr = ds[i * 10:(i + 1) * 10] class WritingTimeSuite: """Based on example in GitHub issue 492: https://github.com/h5py/h5py/issues/492 """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') self.f = h5py.File(path, 'w') self.shape = shape = (128, 1024, 512) self.f.create_dataset( 'a', shape=shape, dtype=np.float32, chunks=(1, shape[1], 64) ) def teardown(self): self.f.close() self._td.cleanup() def time_write_index_last_axis(self): ds = self.f['a'] data = np.zeros(self.shape[:2]) for i in range(self.shape[2]): ds[..., i] = data def time_write_slice_last_axis(self): ds = self.f['a'] data = np.zeros(self.shape[:2]) for i in range(self.shape[2]): ds[..., i:i+1] = data[..., np.newaxis]
# Write the benchmarking functions here. # See "Writing benchmarks" in the asv docs for more information. import os.path as osp import numpy as np from tempfile import TemporaryDirectory import h5py class TimeSuite: """ An example benchmark that times the performance of various kinds of iterating over dictionaries in Python. """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') with h5py.File(path, 'w') as f: f['a'] = np.arange(100000) self.f = h5py.File(path, 'r') def teardown(self): self.f.close() self._td.cleanup() def time_many_small_reads(self): ds = self.f['a'] for i in range(10000): arr = ds[i * 10:(i + 1) * 10] Add benchmark for writing with index of last dimension# Write the benchmarking functions here. # See "Writing benchmarks" in the asv docs for more information. import os.path as osp import numpy as np from tempfile import TemporaryDirectory import h5py class TimeSuite: """ An example benchmark that times the performance of various kinds of iterating over dictionaries in Python. """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') with h5py.File(path, 'w') as f: f['a'] = np.arange(100000) self.f = h5py.File(path, 'r') def teardown(self): self.f.close() self._td.cleanup() def time_many_small_reads(self): ds = self.f['a'] for i in range(10000): arr = ds[i * 10:(i + 1) * 10] class WritingTimeSuite: """Based on example in GitHub issue 492: https://github.com/h5py/h5py/issues/492 """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') self.f = h5py.File(path, 'w') self.shape = shape = (128, 1024, 512) self.f.create_dataset( 'a', shape=shape, dtype=np.float32, chunks=(1, shape[1], 64) ) def teardown(self): self.f.close() self._td.cleanup() def time_write_index_last_axis(self): ds = self.f['a'] data = np.zeros(self.shape[:2]) for i in range(self.shape[2]): ds[..., i] = data def time_write_slice_last_axis(self): ds = self.f['a'] data = np.zeros(self.shape[:2]) for i in range(self.shape[2]): ds[..., i:i+1] = data[..., np.newaxis]
<commit_before># Write the benchmarking functions here. # See "Writing benchmarks" in the asv docs for more information. import os.path as osp import numpy as np from tempfile import TemporaryDirectory import h5py class TimeSuite: """ An example benchmark that times the performance of various kinds of iterating over dictionaries in Python. """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') with h5py.File(path, 'w') as f: f['a'] = np.arange(100000) self.f = h5py.File(path, 'r') def teardown(self): self.f.close() self._td.cleanup() def time_many_small_reads(self): ds = self.f['a'] for i in range(10000): arr = ds[i * 10:(i + 1) * 10] <commit_msg>Add benchmark for writing with index of last dimension<commit_after># Write the benchmarking functions here. # See "Writing benchmarks" in the asv docs for more information. import os.path as osp import numpy as np from tempfile import TemporaryDirectory import h5py class TimeSuite: """ An example benchmark that times the performance of various kinds of iterating over dictionaries in Python. """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') with h5py.File(path, 'w') as f: f['a'] = np.arange(100000) self.f = h5py.File(path, 'r') def teardown(self): self.f.close() self._td.cleanup() def time_many_small_reads(self): ds = self.f['a'] for i in range(10000): arr = ds[i * 10:(i + 1) * 10] class WritingTimeSuite: """Based on example in GitHub issue 492: https://github.com/h5py/h5py/issues/492 """ def setup(self): self._td = TemporaryDirectory() path = osp.join(self._td.name, 'test.h5') self.f = h5py.File(path, 'w') self.shape = shape = (128, 1024, 512) self.f.create_dataset( 'a', shape=shape, dtype=np.float32, chunks=(1, shape[1], 64) ) def teardown(self): self.f.close() self._td.cleanup() def time_write_index_last_axis(self): ds = self.f['a'] data = np.zeros(self.shape[:2]) for i in range(self.shape[2]): ds[..., i] = data def time_write_slice_last_axis(self): ds = self.f['a'] data = np.zeros(self.shape[:2]) for i in range(self.shape[2]): ds[..., i:i+1] = data[..., np.newaxis]
3b0432edc1a9e55e6e7da5955a1ecbeea8000371
openpassword/keychain.py
openpassword/keychain.py
from openpassword.exceptions import InvalidPasswordException class Keychain: def __init__(self, encryption_key): self.encryption_key = encryption_key self._locked = True def unlock(self, password): try: self.encryption_key.decrypt(password) self._locked = False except InvalidPasswordException as e: self._locked = True raise(e) def lock(self): self._locked = True def is_locked(self): return self._locked
from openpassword.exceptions import InvalidPasswordException class Keychain: def __init__(self, encryption_key): self.encryption_key = encryption_key self._locked = True def unlock(self, password): try: self.encryption_key.decrypt(password) self._locked = False except InvalidPasswordException as e: self._locked = True raise e def lock(self): self._locked = True def is_locked(self): return self._locked
Remove unecessary parentesis from raise call
Remove unecessary parentesis from raise call
Python
mit
openpassword/blimey,openpassword/blimey
from openpassword.exceptions import InvalidPasswordException class Keychain: def __init__(self, encryption_key): self.encryption_key = encryption_key self._locked = True def unlock(self, password): try: self.encryption_key.decrypt(password) self._locked = False except InvalidPasswordException as e: self._locked = True raise(e) def lock(self): self._locked = True def is_locked(self): return self._locked Remove unecessary parentesis from raise call
from openpassword.exceptions import InvalidPasswordException class Keychain: def __init__(self, encryption_key): self.encryption_key = encryption_key self._locked = True def unlock(self, password): try: self.encryption_key.decrypt(password) self._locked = False except InvalidPasswordException as e: self._locked = True raise e def lock(self): self._locked = True def is_locked(self): return self._locked
<commit_before>from openpassword.exceptions import InvalidPasswordException class Keychain: def __init__(self, encryption_key): self.encryption_key = encryption_key self._locked = True def unlock(self, password): try: self.encryption_key.decrypt(password) self._locked = False except InvalidPasswordException as e: self._locked = True raise(e) def lock(self): self._locked = True def is_locked(self): return self._locked <commit_msg>Remove unecessary parentesis from raise call<commit_after>
from openpassword.exceptions import InvalidPasswordException class Keychain: def __init__(self, encryption_key): self.encryption_key = encryption_key self._locked = True def unlock(self, password): try: self.encryption_key.decrypt(password) self._locked = False except InvalidPasswordException as e: self._locked = True raise e def lock(self): self._locked = True def is_locked(self): return self._locked
from openpassword.exceptions import InvalidPasswordException class Keychain: def __init__(self, encryption_key): self.encryption_key = encryption_key self._locked = True def unlock(self, password): try: self.encryption_key.decrypt(password) self._locked = False except InvalidPasswordException as e: self._locked = True raise(e) def lock(self): self._locked = True def is_locked(self): return self._locked Remove unecessary parentesis from raise callfrom openpassword.exceptions import InvalidPasswordException class Keychain: def __init__(self, encryption_key): self.encryption_key = encryption_key self._locked = True def unlock(self, password): try: self.encryption_key.decrypt(password) self._locked = False except InvalidPasswordException as e: self._locked = True raise e def lock(self): self._locked = True def is_locked(self): return self._locked
<commit_before>from openpassword.exceptions import InvalidPasswordException class Keychain: def __init__(self, encryption_key): self.encryption_key = encryption_key self._locked = True def unlock(self, password): try: self.encryption_key.decrypt(password) self._locked = False except InvalidPasswordException as e: self._locked = True raise(e) def lock(self): self._locked = True def is_locked(self): return self._locked <commit_msg>Remove unecessary parentesis from raise call<commit_after>from openpassword.exceptions import InvalidPasswordException class Keychain: def __init__(self, encryption_key): self.encryption_key = encryption_key self._locked = True def unlock(self, password): try: self.encryption_key.decrypt(password) self._locked = False except InvalidPasswordException as e: self._locked = True raise e def lock(self): self._locked = True def is_locked(self): return self._locked
7613fc5c3ef1cc552d39e5fde533e8f9af7cc931
form_designer/tests/test_cms_plugin.py
form_designer/tests/test_cms_plugin.py
import django from django.contrib.auth.models import AnonymousUser from django.utils.crypto import get_random_string import pytest from cms import api from cms.page_rendering import render_page from form_designer.contrib.cms_plugins.form_designer_form.cms_plugins import FormDesignerPlugin from form_designer.models import FormDefinition, FormDefinitionField @pytest.mark.django_db def test_cms_plugin_renders_in_cms_page(rf): if django.VERSION >= (1, 10): pytest.xfail('This test is broken in Django 1.10+') fd = FormDefinition.objects.create( mail_to='test@example.com', mail_subject='Someone sent you a greeting: {{ test }}' ) field = FormDefinitionField.objects.create( form_definition=fd, name='test', label=get_random_string(), field_class='django.forms.CharField', ) page = api.create_page("test", "page.html", "en") ph = page.get_placeholders()[0] api.add_plugin(ph, FormDesignerPlugin, "en", form_definition=fd) request = rf.get("/") request.user = AnonymousUser() request.current_page = page response = render_page(request, page, "fi", "test") response.render() content = response.content.decode("utf8") assert field.label in content assert "<form" in content
from django.contrib.auth.models import AnonymousUser from django.utils.crypto import get_random_string import pytest from cms import api from cms.page_rendering import render_page from form_designer.contrib.cms_plugins.form_designer_form.cms_plugins import FormDesignerPlugin from form_designer.models import FormDefinition, FormDefinitionField @pytest.mark.django_db def test_cms_plugin_renders_in_cms_page(rf): fd = FormDefinition.objects.create( mail_to='test@example.com', mail_subject='Someone sent you a greeting: {{ test }}' ) field = FormDefinitionField.objects.create( form_definition=fd, name='test', label=get_random_string(), field_class='django.forms.CharField', ) page = api.create_page("test", "page.html", "en") ph = page.get_placeholders()[0] api.add_plugin(ph, FormDesignerPlugin, "en", form_definition=fd) request = rf.get("/") request.user = AnonymousUser() request.current_page = page response = render_page(request, page, "fi", "test") response.render() content = response.content.decode("utf8") assert field.label in content assert "<form" in content
Revert "Disable Django-CMS test on Django 1.10+"
Revert "Disable Django-CMS test on Django 1.10+" Django CMS tests should work now with Django 1.10 and 1.11 too, since the Django CMS version 3.4.5 supports them. This reverts commit fcfe2513fc8532dc2212a254da42d75048e76de7.
Python
bsd-3-clause
kcsry/django-form-designer,andersinno/django-form-designer-ai,kcsry/django-form-designer,andersinno/django-form-designer,andersinno/django-form-designer-ai,andersinno/django-form-designer
import django from django.contrib.auth.models import AnonymousUser from django.utils.crypto import get_random_string import pytest from cms import api from cms.page_rendering import render_page from form_designer.contrib.cms_plugins.form_designer_form.cms_plugins import FormDesignerPlugin from form_designer.models import FormDefinition, FormDefinitionField @pytest.mark.django_db def test_cms_plugin_renders_in_cms_page(rf): if django.VERSION >= (1, 10): pytest.xfail('This test is broken in Django 1.10+') fd = FormDefinition.objects.create( mail_to='test@example.com', mail_subject='Someone sent you a greeting: {{ test }}' ) field = FormDefinitionField.objects.create( form_definition=fd, name='test', label=get_random_string(), field_class='django.forms.CharField', ) page = api.create_page("test", "page.html", "en") ph = page.get_placeholders()[0] api.add_plugin(ph, FormDesignerPlugin, "en", form_definition=fd) request = rf.get("/") request.user = AnonymousUser() request.current_page = page response = render_page(request, page, "fi", "test") response.render() content = response.content.decode("utf8") assert field.label in content assert "<form" in content Revert "Disable Django-CMS test on Django 1.10+" Django CMS tests should work now with Django 1.10 and 1.11 too, since the Django CMS version 3.4.5 supports them. This reverts commit fcfe2513fc8532dc2212a254da42d75048e76de7.
from django.contrib.auth.models import AnonymousUser from django.utils.crypto import get_random_string import pytest from cms import api from cms.page_rendering import render_page from form_designer.contrib.cms_plugins.form_designer_form.cms_plugins import FormDesignerPlugin from form_designer.models import FormDefinition, FormDefinitionField @pytest.mark.django_db def test_cms_plugin_renders_in_cms_page(rf): fd = FormDefinition.objects.create( mail_to='test@example.com', mail_subject='Someone sent you a greeting: {{ test }}' ) field = FormDefinitionField.objects.create( form_definition=fd, name='test', label=get_random_string(), field_class='django.forms.CharField', ) page = api.create_page("test", "page.html", "en") ph = page.get_placeholders()[0] api.add_plugin(ph, FormDesignerPlugin, "en", form_definition=fd) request = rf.get("/") request.user = AnonymousUser() request.current_page = page response = render_page(request, page, "fi", "test") response.render() content = response.content.decode("utf8") assert field.label in content assert "<form" in content
<commit_before>import django from django.contrib.auth.models import AnonymousUser from django.utils.crypto import get_random_string import pytest from cms import api from cms.page_rendering import render_page from form_designer.contrib.cms_plugins.form_designer_form.cms_plugins import FormDesignerPlugin from form_designer.models import FormDefinition, FormDefinitionField @pytest.mark.django_db def test_cms_plugin_renders_in_cms_page(rf): if django.VERSION >= (1, 10): pytest.xfail('This test is broken in Django 1.10+') fd = FormDefinition.objects.create( mail_to='test@example.com', mail_subject='Someone sent you a greeting: {{ test }}' ) field = FormDefinitionField.objects.create( form_definition=fd, name='test', label=get_random_string(), field_class='django.forms.CharField', ) page = api.create_page("test", "page.html", "en") ph = page.get_placeholders()[0] api.add_plugin(ph, FormDesignerPlugin, "en", form_definition=fd) request = rf.get("/") request.user = AnonymousUser() request.current_page = page response = render_page(request, page, "fi", "test") response.render() content = response.content.decode("utf8") assert field.label in content assert "<form" in content <commit_msg>Revert "Disable Django-CMS test on Django 1.10+" Django CMS tests should work now with Django 1.10 and 1.11 too, since the Django CMS version 3.4.5 supports them. This reverts commit fcfe2513fc8532dc2212a254da42d75048e76de7.<commit_after>
from django.contrib.auth.models import AnonymousUser from django.utils.crypto import get_random_string import pytest from cms import api from cms.page_rendering import render_page from form_designer.contrib.cms_plugins.form_designer_form.cms_plugins import FormDesignerPlugin from form_designer.models import FormDefinition, FormDefinitionField @pytest.mark.django_db def test_cms_plugin_renders_in_cms_page(rf): fd = FormDefinition.objects.create( mail_to='test@example.com', mail_subject='Someone sent you a greeting: {{ test }}' ) field = FormDefinitionField.objects.create( form_definition=fd, name='test', label=get_random_string(), field_class='django.forms.CharField', ) page = api.create_page("test", "page.html", "en") ph = page.get_placeholders()[0] api.add_plugin(ph, FormDesignerPlugin, "en", form_definition=fd) request = rf.get("/") request.user = AnonymousUser() request.current_page = page response = render_page(request, page, "fi", "test") response.render() content = response.content.decode("utf8") assert field.label in content assert "<form" in content
import django from django.contrib.auth.models import AnonymousUser from django.utils.crypto import get_random_string import pytest from cms import api from cms.page_rendering import render_page from form_designer.contrib.cms_plugins.form_designer_form.cms_plugins import FormDesignerPlugin from form_designer.models import FormDefinition, FormDefinitionField @pytest.mark.django_db def test_cms_plugin_renders_in_cms_page(rf): if django.VERSION >= (1, 10): pytest.xfail('This test is broken in Django 1.10+') fd = FormDefinition.objects.create( mail_to='test@example.com', mail_subject='Someone sent you a greeting: {{ test }}' ) field = FormDefinitionField.objects.create( form_definition=fd, name='test', label=get_random_string(), field_class='django.forms.CharField', ) page = api.create_page("test", "page.html", "en") ph = page.get_placeholders()[0] api.add_plugin(ph, FormDesignerPlugin, "en", form_definition=fd) request = rf.get("/") request.user = AnonymousUser() request.current_page = page response = render_page(request, page, "fi", "test") response.render() content = response.content.decode("utf8") assert field.label in content assert "<form" in content Revert "Disable Django-CMS test on Django 1.10+" Django CMS tests should work now with Django 1.10 and 1.11 too, since the Django CMS version 3.4.5 supports them. This reverts commit fcfe2513fc8532dc2212a254da42d75048e76de7.from django.contrib.auth.models import AnonymousUser from django.utils.crypto import get_random_string import pytest from cms import api from cms.page_rendering import render_page from form_designer.contrib.cms_plugins.form_designer_form.cms_plugins import FormDesignerPlugin from form_designer.models import FormDefinition, FormDefinitionField @pytest.mark.django_db def test_cms_plugin_renders_in_cms_page(rf): fd = FormDefinition.objects.create( mail_to='test@example.com', mail_subject='Someone sent you a greeting: {{ test }}' ) field = FormDefinitionField.objects.create( form_definition=fd, name='test', label=get_random_string(), field_class='django.forms.CharField', ) page = api.create_page("test", "page.html", "en") ph = page.get_placeholders()[0] api.add_plugin(ph, FormDesignerPlugin, "en", form_definition=fd) request = rf.get("/") request.user = AnonymousUser() request.current_page = page response = render_page(request, page, "fi", "test") response.render() content = response.content.decode("utf8") assert field.label in content assert "<form" in content
<commit_before>import django from django.contrib.auth.models import AnonymousUser from django.utils.crypto import get_random_string import pytest from cms import api from cms.page_rendering import render_page from form_designer.contrib.cms_plugins.form_designer_form.cms_plugins import FormDesignerPlugin from form_designer.models import FormDefinition, FormDefinitionField @pytest.mark.django_db def test_cms_plugin_renders_in_cms_page(rf): if django.VERSION >= (1, 10): pytest.xfail('This test is broken in Django 1.10+') fd = FormDefinition.objects.create( mail_to='test@example.com', mail_subject='Someone sent you a greeting: {{ test }}' ) field = FormDefinitionField.objects.create( form_definition=fd, name='test', label=get_random_string(), field_class='django.forms.CharField', ) page = api.create_page("test", "page.html", "en") ph = page.get_placeholders()[0] api.add_plugin(ph, FormDesignerPlugin, "en", form_definition=fd) request = rf.get("/") request.user = AnonymousUser() request.current_page = page response = render_page(request, page, "fi", "test") response.render() content = response.content.decode("utf8") assert field.label in content assert "<form" in content <commit_msg>Revert "Disable Django-CMS test on Django 1.10+" Django CMS tests should work now with Django 1.10 and 1.11 too, since the Django CMS version 3.4.5 supports them. This reverts commit fcfe2513fc8532dc2212a254da42d75048e76de7.<commit_after>from django.contrib.auth.models import AnonymousUser from django.utils.crypto import get_random_string import pytest from cms import api from cms.page_rendering import render_page from form_designer.contrib.cms_plugins.form_designer_form.cms_plugins import FormDesignerPlugin from form_designer.models import FormDefinition, FormDefinitionField @pytest.mark.django_db def test_cms_plugin_renders_in_cms_page(rf): fd = FormDefinition.objects.create( mail_to='test@example.com', mail_subject='Someone sent you a greeting: {{ test }}' ) field = FormDefinitionField.objects.create( form_definition=fd, name='test', label=get_random_string(), field_class='django.forms.CharField', ) page = api.create_page("test", "page.html", "en") ph = page.get_placeholders()[0] api.add_plugin(ph, FormDesignerPlugin, "en", form_definition=fd) request = rf.get("/") request.user = AnonymousUser() request.current_page = page response = render_page(request, page, "fi", "test") response.render() content = response.content.decode("utf8") assert field.label in content assert "<form" in content
e1da1e7e8fedd288e9ac3a41b529c5bfe3e0612b
orchestrator/__init__.py
orchestrator/__init__.py
from __future__ import absolute_import from celery.signals import setup_logging import orchestrator.logger __version__ = '0.3.10' __author__ = 'sukrit' orchestrator.logger.init_logging() setup_logging.connect(orchestrator.logger.init_celery_logging)
from __future__ import absolute_import from celery.signals import setup_logging import orchestrator.logger __version__ = '0.4.0' __author__ = 'sukrit' orchestrator.logger.init_logging() setup_logging.connect(orchestrator.logger.init_celery_logging)
Prepare for next development version
Prepare for next development version
Python
mit
totem/cluster-orchestrator,totem/cluster-orchestrator,totem/cluster-orchestrator
from __future__ import absolute_import from celery.signals import setup_logging import orchestrator.logger __version__ = '0.3.10' __author__ = 'sukrit' orchestrator.logger.init_logging() setup_logging.connect(orchestrator.logger.init_celery_logging) Prepare for next development version
from __future__ import absolute_import from celery.signals import setup_logging import orchestrator.logger __version__ = '0.4.0' __author__ = 'sukrit' orchestrator.logger.init_logging() setup_logging.connect(orchestrator.logger.init_celery_logging)
<commit_before>from __future__ import absolute_import from celery.signals import setup_logging import orchestrator.logger __version__ = '0.3.10' __author__ = 'sukrit' orchestrator.logger.init_logging() setup_logging.connect(orchestrator.logger.init_celery_logging) <commit_msg>Prepare for next development version<commit_after>
from __future__ import absolute_import from celery.signals import setup_logging import orchestrator.logger __version__ = '0.4.0' __author__ = 'sukrit' orchestrator.logger.init_logging() setup_logging.connect(orchestrator.logger.init_celery_logging)
from __future__ import absolute_import from celery.signals import setup_logging import orchestrator.logger __version__ = '0.3.10' __author__ = 'sukrit' orchestrator.logger.init_logging() setup_logging.connect(orchestrator.logger.init_celery_logging) Prepare for next development versionfrom __future__ import absolute_import from celery.signals import setup_logging import orchestrator.logger __version__ = '0.4.0' __author__ = 'sukrit' orchestrator.logger.init_logging() setup_logging.connect(orchestrator.logger.init_celery_logging)
<commit_before>from __future__ import absolute_import from celery.signals import setup_logging import orchestrator.logger __version__ = '0.3.10' __author__ = 'sukrit' orchestrator.logger.init_logging() setup_logging.connect(orchestrator.logger.init_celery_logging) <commit_msg>Prepare for next development version<commit_after>from __future__ import absolute_import from celery.signals import setup_logging import orchestrator.logger __version__ = '0.4.0' __author__ = 'sukrit' orchestrator.logger.init_logging() setup_logging.connect(orchestrator.logger.init_celery_logging)
39a1212508c27a5c21f8b027fef3fb409a28657f
app/commands.py
app/commands.py
from flask import current_app def list_routes(): """List URLs of all application routes.""" for rule in sorted(current_app.url_map.iter_rules(), key=lambda r: r.rule): print("{:10} {}".format(", ".join(rule.methods - set(['OPTIONS', 'HEAD'])), rule.rule)) # noqa def setup_commands(application): application.cli.command('list-routes')(list_routes)
import click from flask import current_app from flask.cli import with_appcontext @click.command('list-routes') @with_appcontext def list_routes(): """List URLs of all application routes.""" for rule in sorted(current_app.url_map.iter_rules(), key=lambda r: r.rule): print("{:10} {}".format(", ".join(rule.methods - set(['OPTIONS', 'HEAD'])), rule.rule)) # noqa def setup_commands(application): application.cli.add_command(list_routes)
Switch existing command to standard approach
Switch existing command to standard approach This is the suggested approach in the documentation [1] and using it makes it clearer what's going on and to add other commands with arguments, which we'll do in the next commit. [1]: https://flask.palletsprojects.com/en/2.0.x/cli/#custom-commands
Python
mit
alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin
from flask import current_app def list_routes(): """List URLs of all application routes.""" for rule in sorted(current_app.url_map.iter_rules(), key=lambda r: r.rule): print("{:10} {}".format(", ".join(rule.methods - set(['OPTIONS', 'HEAD'])), rule.rule)) # noqa def setup_commands(application): application.cli.command('list-routes')(list_routes) Switch existing command to standard approach This is the suggested approach in the documentation [1] and using it makes it clearer what's going on and to add other commands with arguments, which we'll do in the next commit. [1]: https://flask.palletsprojects.com/en/2.0.x/cli/#custom-commands
import click from flask import current_app from flask.cli import with_appcontext @click.command('list-routes') @with_appcontext def list_routes(): """List URLs of all application routes.""" for rule in sorted(current_app.url_map.iter_rules(), key=lambda r: r.rule): print("{:10} {}".format(", ".join(rule.methods - set(['OPTIONS', 'HEAD'])), rule.rule)) # noqa def setup_commands(application): application.cli.add_command(list_routes)
<commit_before>from flask import current_app def list_routes(): """List URLs of all application routes.""" for rule in sorted(current_app.url_map.iter_rules(), key=lambda r: r.rule): print("{:10} {}".format(", ".join(rule.methods - set(['OPTIONS', 'HEAD'])), rule.rule)) # noqa def setup_commands(application): application.cli.command('list-routes')(list_routes) <commit_msg>Switch existing command to standard approach This is the suggested approach in the documentation [1] and using it makes it clearer what's going on and to add other commands with arguments, which we'll do in the next commit. [1]: https://flask.palletsprojects.com/en/2.0.x/cli/#custom-commands<commit_after>
import click from flask import current_app from flask.cli import with_appcontext @click.command('list-routes') @with_appcontext def list_routes(): """List URLs of all application routes.""" for rule in sorted(current_app.url_map.iter_rules(), key=lambda r: r.rule): print("{:10} {}".format(", ".join(rule.methods - set(['OPTIONS', 'HEAD'])), rule.rule)) # noqa def setup_commands(application): application.cli.add_command(list_routes)
from flask import current_app def list_routes(): """List URLs of all application routes.""" for rule in sorted(current_app.url_map.iter_rules(), key=lambda r: r.rule): print("{:10} {}".format(", ".join(rule.methods - set(['OPTIONS', 'HEAD'])), rule.rule)) # noqa def setup_commands(application): application.cli.command('list-routes')(list_routes) Switch existing command to standard approach This is the suggested approach in the documentation [1] and using it makes it clearer what's going on and to add other commands with arguments, which we'll do in the next commit. [1]: https://flask.palletsprojects.com/en/2.0.x/cli/#custom-commandsimport click from flask import current_app from flask.cli import with_appcontext @click.command('list-routes') @with_appcontext def list_routes(): """List URLs of all application routes.""" for rule in sorted(current_app.url_map.iter_rules(), key=lambda r: r.rule): print("{:10} {}".format(", ".join(rule.methods - set(['OPTIONS', 'HEAD'])), rule.rule)) # noqa def setup_commands(application): application.cli.add_command(list_routes)
<commit_before>from flask import current_app def list_routes(): """List URLs of all application routes.""" for rule in sorted(current_app.url_map.iter_rules(), key=lambda r: r.rule): print("{:10} {}".format(", ".join(rule.methods - set(['OPTIONS', 'HEAD'])), rule.rule)) # noqa def setup_commands(application): application.cli.command('list-routes')(list_routes) <commit_msg>Switch existing command to standard approach This is the suggested approach in the documentation [1] and using it makes it clearer what's going on and to add other commands with arguments, which we'll do in the next commit. [1]: https://flask.palletsprojects.com/en/2.0.x/cli/#custom-commands<commit_after>import click from flask import current_app from flask.cli import with_appcontext @click.command('list-routes') @with_appcontext def list_routes(): """List URLs of all application routes.""" for rule in sorted(current_app.url_map.iter_rules(), key=lambda r: r.rule): print("{:10} {}".format(", ".join(rule.methods - set(['OPTIONS', 'HEAD'])), rule.rule)) # noqa def setup_commands(application): application.cli.add_command(list_routes)
79a2f5ddc1d6d05dca0d44ee6586b2d8809ee1c0
deep_q_rl/ale_run.py
deep_q_rl/ale_run.py
"""This script launches all of the processes necessary to train a deep Q-network on an ALE game. Usage: ale_run.py [--glue_port GLUE_PORT] All unrecognized command line arguments will be passed on to rl_glue_ale_agent.py """ import subprocess import sys import os import argparse ROM_PATH = "/home/spragunr/neural_rl_libraries/roms/breakout.bin" # Check for glue_port command line argument and set it up... parser = argparse.ArgumentParser(description='Neural rl agent.') parser.add_argument('--glue_port', type=str, default="4096", help='rlglue port (default 4096)') args, unknown = parser.parse_known_args() my_env = os.environ.copy() my_env["RLGLUE_PORT"] = args.glue_port # Start the necessary processes: p1 = subprocess.Popen(['rl_glue'], env=my_env) ale_string = ("ale -game_controller rlglue -frame_skip 4 " "-restricted_action_set true ") p2 = subprocess.Popen(ale_string + ROM_PATH, shell=True, env=my_env) p3 = subprocess.Popen(['./rl_glue_ale_experiment.py'], env=my_env) p4 = subprocess.Popen(['./rl_glue_ale_agent.py'] + sys.argv[1:], env=my_env) p1.wait() p2.wait() p3.wait() p4.wait()
"""This script launches all of the processes necessary to train a deep Q-network on an ALE game. Usage: ale_run.py [--glue_port GLUE_PORT] All unrecognized command line arguments will be passed on to rl_glue_ale_agent.py """ import subprocess import sys import os import argparse # Put your binaries under the directory 'deep_q_rl/roms' ROM_PATH = "../roms/breakout.bin" # Check for glue_port command line argument and set it up... parser = argparse.ArgumentParser(description='Neural rl agent.') parser.add_argument('--glue_port', type=str, default="4096", help='rlglue port (default 4096)') args, unknown = parser.parse_known_args() my_env = os.environ.copy() my_env["RLGLUE_PORT"] = args.glue_port # Start the necessary processes: p1 = subprocess.Popen(['rl_glue'], env=my_env) ale_string = ("ale -game_controller rlglue -frame_skip 4 " "-restricted_action_set true ") p2 = subprocess.Popen(ale_string + ROM_PATH, shell=True, env=my_env) p3 = subprocess.Popen(['./rl_glue_ale_experiment.py'], env=my_env) p4 = subprocess.Popen(['./rl_glue_ale_agent.py'] + sys.argv[1:], env=my_env) p1.wait() p2.wait() p3.wait() p4.wait()
Make the position of the roms work for everybody.
Make the position of the roms work for everybody. Here you previously used a directory that mentions your particular setting. Substitute it by a relative directory so it works for everybody. Also, suggest at the read file to create a 'rome' directory with 'breakout.bin' in it.
Python
bsd-3-clause
r0k3/deep_q_rl,aaannndddyyy/deep_q_rl,davidsj/deep_q_rl,spragunr/deep_q_rl,gogobebe2/deep_q_rl,aaannndddyyy/deep_q_rl,omnivert/deep_q_rl,npow/deep_q_rl,vvw/deep_q_rl,codeaudit/deep_q_rl,alito/deep_q_rl,jleni/deep_q_rl,peterzcc/deep_q_rl,davidsj/deep_q_rl,spragunr/deep_q_rl,codeaudit/deep_q_rl,udibr/deep_q_rl,udibr/deep_q_rl,sygi/deep_q_rl,r0k3/deep_q_rl,tmylk/deep_q_rl,jcatw/deep_q_rl,gogobebe2/deep_q_rl,vvw/deep_q_rl,jleni/deep_q_rl,tmylk/deep_q_rl,peterzcc/deep_q_rl,jcatw/deep_q_rl,alito/deep_q_rl,sygi/deep_q_rl,omnivert/deep_q_rl
"""This script launches all of the processes necessary to train a deep Q-network on an ALE game. Usage: ale_run.py [--glue_port GLUE_PORT] All unrecognized command line arguments will be passed on to rl_glue_ale_agent.py """ import subprocess import sys import os import argparse ROM_PATH = "/home/spragunr/neural_rl_libraries/roms/breakout.bin" # Check for glue_port command line argument and set it up... parser = argparse.ArgumentParser(description='Neural rl agent.') parser.add_argument('--glue_port', type=str, default="4096", help='rlglue port (default 4096)') args, unknown = parser.parse_known_args() my_env = os.environ.copy() my_env["RLGLUE_PORT"] = args.glue_port # Start the necessary processes: p1 = subprocess.Popen(['rl_glue'], env=my_env) ale_string = ("ale -game_controller rlglue -frame_skip 4 " "-restricted_action_set true ") p2 = subprocess.Popen(ale_string + ROM_PATH, shell=True, env=my_env) p3 = subprocess.Popen(['./rl_glue_ale_experiment.py'], env=my_env) p4 = subprocess.Popen(['./rl_glue_ale_agent.py'] + sys.argv[1:], env=my_env) p1.wait() p2.wait() p3.wait() p4.wait() Make the position of the roms work for everybody. Here you previously used a directory that mentions your particular setting. Substitute it by a relative directory so it works for everybody. Also, suggest at the read file to create a 'rome' directory with 'breakout.bin' in it.
"""This script launches all of the processes necessary to train a deep Q-network on an ALE game. Usage: ale_run.py [--glue_port GLUE_PORT] All unrecognized command line arguments will be passed on to rl_glue_ale_agent.py """ import subprocess import sys import os import argparse # Put your binaries under the directory 'deep_q_rl/roms' ROM_PATH = "../roms/breakout.bin" # Check for glue_port command line argument and set it up... parser = argparse.ArgumentParser(description='Neural rl agent.') parser.add_argument('--glue_port', type=str, default="4096", help='rlglue port (default 4096)') args, unknown = parser.parse_known_args() my_env = os.environ.copy() my_env["RLGLUE_PORT"] = args.glue_port # Start the necessary processes: p1 = subprocess.Popen(['rl_glue'], env=my_env) ale_string = ("ale -game_controller rlglue -frame_skip 4 " "-restricted_action_set true ") p2 = subprocess.Popen(ale_string + ROM_PATH, shell=True, env=my_env) p3 = subprocess.Popen(['./rl_glue_ale_experiment.py'], env=my_env) p4 = subprocess.Popen(['./rl_glue_ale_agent.py'] + sys.argv[1:], env=my_env) p1.wait() p2.wait() p3.wait() p4.wait()
<commit_before>"""This script launches all of the processes necessary to train a deep Q-network on an ALE game. Usage: ale_run.py [--glue_port GLUE_PORT] All unrecognized command line arguments will be passed on to rl_glue_ale_agent.py """ import subprocess import sys import os import argparse ROM_PATH = "/home/spragunr/neural_rl_libraries/roms/breakout.bin" # Check for glue_port command line argument and set it up... parser = argparse.ArgumentParser(description='Neural rl agent.') parser.add_argument('--glue_port', type=str, default="4096", help='rlglue port (default 4096)') args, unknown = parser.parse_known_args() my_env = os.environ.copy() my_env["RLGLUE_PORT"] = args.glue_port # Start the necessary processes: p1 = subprocess.Popen(['rl_glue'], env=my_env) ale_string = ("ale -game_controller rlglue -frame_skip 4 " "-restricted_action_set true ") p2 = subprocess.Popen(ale_string + ROM_PATH, shell=True, env=my_env) p3 = subprocess.Popen(['./rl_glue_ale_experiment.py'], env=my_env) p4 = subprocess.Popen(['./rl_glue_ale_agent.py'] + sys.argv[1:], env=my_env) p1.wait() p2.wait() p3.wait() p4.wait() <commit_msg>Make the position of the roms work for everybody. Here you previously used a directory that mentions your particular setting. Substitute it by a relative directory so it works for everybody. Also, suggest at the read file to create a 'rome' directory with 'breakout.bin' in it.<commit_after>
"""This script launches all of the processes necessary to train a deep Q-network on an ALE game. Usage: ale_run.py [--glue_port GLUE_PORT] All unrecognized command line arguments will be passed on to rl_glue_ale_agent.py """ import subprocess import sys import os import argparse # Put your binaries under the directory 'deep_q_rl/roms' ROM_PATH = "../roms/breakout.bin" # Check for glue_port command line argument and set it up... parser = argparse.ArgumentParser(description='Neural rl agent.') parser.add_argument('--glue_port', type=str, default="4096", help='rlglue port (default 4096)') args, unknown = parser.parse_known_args() my_env = os.environ.copy() my_env["RLGLUE_PORT"] = args.glue_port # Start the necessary processes: p1 = subprocess.Popen(['rl_glue'], env=my_env) ale_string = ("ale -game_controller rlglue -frame_skip 4 " "-restricted_action_set true ") p2 = subprocess.Popen(ale_string + ROM_PATH, shell=True, env=my_env) p3 = subprocess.Popen(['./rl_glue_ale_experiment.py'], env=my_env) p4 = subprocess.Popen(['./rl_glue_ale_agent.py'] + sys.argv[1:], env=my_env) p1.wait() p2.wait() p3.wait() p4.wait()
"""This script launches all of the processes necessary to train a deep Q-network on an ALE game. Usage: ale_run.py [--glue_port GLUE_PORT] All unrecognized command line arguments will be passed on to rl_glue_ale_agent.py """ import subprocess import sys import os import argparse ROM_PATH = "/home/spragunr/neural_rl_libraries/roms/breakout.bin" # Check for glue_port command line argument and set it up... parser = argparse.ArgumentParser(description='Neural rl agent.') parser.add_argument('--glue_port', type=str, default="4096", help='rlglue port (default 4096)') args, unknown = parser.parse_known_args() my_env = os.environ.copy() my_env["RLGLUE_PORT"] = args.glue_port # Start the necessary processes: p1 = subprocess.Popen(['rl_glue'], env=my_env) ale_string = ("ale -game_controller rlglue -frame_skip 4 " "-restricted_action_set true ") p2 = subprocess.Popen(ale_string + ROM_PATH, shell=True, env=my_env) p3 = subprocess.Popen(['./rl_glue_ale_experiment.py'], env=my_env) p4 = subprocess.Popen(['./rl_glue_ale_agent.py'] + sys.argv[1:], env=my_env) p1.wait() p2.wait() p3.wait() p4.wait() Make the position of the roms work for everybody. Here you previously used a directory that mentions your particular setting. Substitute it by a relative directory so it works for everybody. Also, suggest at the read file to create a 'rome' directory with 'breakout.bin' in it."""This script launches all of the processes necessary to train a deep Q-network on an ALE game. Usage: ale_run.py [--glue_port GLUE_PORT] All unrecognized command line arguments will be passed on to rl_glue_ale_agent.py """ import subprocess import sys import os import argparse # Put your binaries under the directory 'deep_q_rl/roms' ROM_PATH = "../roms/breakout.bin" # Check for glue_port command line argument and set it up... parser = argparse.ArgumentParser(description='Neural rl agent.') parser.add_argument('--glue_port', type=str, default="4096", help='rlglue port (default 4096)') args, unknown = parser.parse_known_args() my_env = os.environ.copy() my_env["RLGLUE_PORT"] = args.glue_port # Start the necessary processes: p1 = subprocess.Popen(['rl_glue'], env=my_env) ale_string = ("ale -game_controller rlglue -frame_skip 4 " "-restricted_action_set true ") p2 = subprocess.Popen(ale_string + ROM_PATH, shell=True, env=my_env) p3 = subprocess.Popen(['./rl_glue_ale_experiment.py'], env=my_env) p4 = subprocess.Popen(['./rl_glue_ale_agent.py'] + sys.argv[1:], env=my_env) p1.wait() p2.wait() p3.wait() p4.wait()
<commit_before>"""This script launches all of the processes necessary to train a deep Q-network on an ALE game. Usage: ale_run.py [--glue_port GLUE_PORT] All unrecognized command line arguments will be passed on to rl_glue_ale_agent.py """ import subprocess import sys import os import argparse ROM_PATH = "/home/spragunr/neural_rl_libraries/roms/breakout.bin" # Check for glue_port command line argument and set it up... parser = argparse.ArgumentParser(description='Neural rl agent.') parser.add_argument('--glue_port', type=str, default="4096", help='rlglue port (default 4096)') args, unknown = parser.parse_known_args() my_env = os.environ.copy() my_env["RLGLUE_PORT"] = args.glue_port # Start the necessary processes: p1 = subprocess.Popen(['rl_glue'], env=my_env) ale_string = ("ale -game_controller rlglue -frame_skip 4 " "-restricted_action_set true ") p2 = subprocess.Popen(ale_string + ROM_PATH, shell=True, env=my_env) p3 = subprocess.Popen(['./rl_glue_ale_experiment.py'], env=my_env) p4 = subprocess.Popen(['./rl_glue_ale_agent.py'] + sys.argv[1:], env=my_env) p1.wait() p2.wait() p3.wait() p4.wait() <commit_msg>Make the position of the roms work for everybody. Here you previously used a directory that mentions your particular setting. Substitute it by a relative directory so it works for everybody. Also, suggest at the read file to create a 'rome' directory with 'breakout.bin' in it.<commit_after>"""This script launches all of the processes necessary to train a deep Q-network on an ALE game. Usage: ale_run.py [--glue_port GLUE_PORT] All unrecognized command line arguments will be passed on to rl_glue_ale_agent.py """ import subprocess import sys import os import argparse # Put your binaries under the directory 'deep_q_rl/roms' ROM_PATH = "../roms/breakout.bin" # Check for glue_port command line argument and set it up... parser = argparse.ArgumentParser(description='Neural rl agent.') parser.add_argument('--glue_port', type=str, default="4096", help='rlglue port (default 4096)') args, unknown = parser.parse_known_args() my_env = os.environ.copy() my_env["RLGLUE_PORT"] = args.glue_port # Start the necessary processes: p1 = subprocess.Popen(['rl_glue'], env=my_env) ale_string = ("ale -game_controller rlglue -frame_skip 4 " "-restricted_action_set true ") p2 = subprocess.Popen(ale_string + ROM_PATH, shell=True, env=my_env) p3 = subprocess.Popen(['./rl_glue_ale_experiment.py'], env=my_env) p4 = subprocess.Popen(['./rl_glue_ale_agent.py'] + sys.argv[1:], env=my_env) p1.wait() p2.wait() p3.wait() p4.wait()
8b16084a4fe72a369ada80969f6b728abf611d8f
marathon_itests/environment.py
marathon_itests/environment.py
import time from itest_utils import wait_for_marathon from itest_utils import print_container_logs def before_all(context): wait_for_marathon() def after_scenario(context, scenario): """If a marathon client object exists in our context, delete any apps in Marathon and wait until they die.""" print_container_logs('zookeeper') print_container_logs('marathon') if context.client: while True: apps = context.client.list_apps() if not apps: break for app in apps: context.client.delete_app(app.id, force=True) time.sleep(0.5) while context.client.list_deployments(): time.sleep(0.5)
import time from itest_utils import wait_for_marathon from itest_utils import print_container_logs def before_all(context): wait_for_marathon() def after_scenario(context, scenario): """If a marathon client object exists in our context, delete any apps in Marathon and wait until they die.""" if scenario.status != 'passed': print "Zookeeper container logs:" print_container_logs('zookeeper') print "Marathon container logs:" print_container_logs('marathon') if context.client: while True: apps = context.client.list_apps() if not apps: break for app in apps: context.client.delete_app(app.id, force=True) time.sleep(0.5) while context.client.list_deployments(): time.sleep(0.5)
Print container logs only in case of failed scenario
Print container logs only in case of failed scenario
Python
apache-2.0
somic/paasta,gstarnberger/paasta,Yelp/paasta,gstarnberger/paasta,somic/paasta,Yelp/paasta
import time from itest_utils import wait_for_marathon from itest_utils import print_container_logs def before_all(context): wait_for_marathon() def after_scenario(context, scenario): """If a marathon client object exists in our context, delete any apps in Marathon and wait until they die.""" print_container_logs('zookeeper') print_container_logs('marathon') if context.client: while True: apps = context.client.list_apps() if not apps: break for app in apps: context.client.delete_app(app.id, force=True) time.sleep(0.5) while context.client.list_deployments(): time.sleep(0.5) Print container logs only in case of failed scenario
import time from itest_utils import wait_for_marathon from itest_utils import print_container_logs def before_all(context): wait_for_marathon() def after_scenario(context, scenario): """If a marathon client object exists in our context, delete any apps in Marathon and wait until they die.""" if scenario.status != 'passed': print "Zookeeper container logs:" print_container_logs('zookeeper') print "Marathon container logs:" print_container_logs('marathon') if context.client: while True: apps = context.client.list_apps() if not apps: break for app in apps: context.client.delete_app(app.id, force=True) time.sleep(0.5) while context.client.list_deployments(): time.sleep(0.5)
<commit_before>import time from itest_utils import wait_for_marathon from itest_utils import print_container_logs def before_all(context): wait_for_marathon() def after_scenario(context, scenario): """If a marathon client object exists in our context, delete any apps in Marathon and wait until they die.""" print_container_logs('zookeeper') print_container_logs('marathon') if context.client: while True: apps = context.client.list_apps() if not apps: break for app in apps: context.client.delete_app(app.id, force=True) time.sleep(0.5) while context.client.list_deployments(): time.sleep(0.5) <commit_msg>Print container logs only in case of failed scenario<commit_after>
import time from itest_utils import wait_for_marathon from itest_utils import print_container_logs def before_all(context): wait_for_marathon() def after_scenario(context, scenario): """If a marathon client object exists in our context, delete any apps in Marathon and wait until they die.""" if scenario.status != 'passed': print "Zookeeper container logs:" print_container_logs('zookeeper') print "Marathon container logs:" print_container_logs('marathon') if context.client: while True: apps = context.client.list_apps() if not apps: break for app in apps: context.client.delete_app(app.id, force=True) time.sleep(0.5) while context.client.list_deployments(): time.sleep(0.5)
import time from itest_utils import wait_for_marathon from itest_utils import print_container_logs def before_all(context): wait_for_marathon() def after_scenario(context, scenario): """If a marathon client object exists in our context, delete any apps in Marathon and wait until they die.""" print_container_logs('zookeeper') print_container_logs('marathon') if context.client: while True: apps = context.client.list_apps() if not apps: break for app in apps: context.client.delete_app(app.id, force=True) time.sleep(0.5) while context.client.list_deployments(): time.sleep(0.5) Print container logs only in case of failed scenarioimport time from itest_utils import wait_for_marathon from itest_utils import print_container_logs def before_all(context): wait_for_marathon() def after_scenario(context, scenario): """If a marathon client object exists in our context, delete any apps in Marathon and wait until they die.""" if scenario.status != 'passed': print "Zookeeper container logs:" print_container_logs('zookeeper') print "Marathon container logs:" print_container_logs('marathon') if context.client: while True: apps = context.client.list_apps() if not apps: break for app in apps: context.client.delete_app(app.id, force=True) time.sleep(0.5) while context.client.list_deployments(): time.sleep(0.5)
<commit_before>import time from itest_utils import wait_for_marathon from itest_utils import print_container_logs def before_all(context): wait_for_marathon() def after_scenario(context, scenario): """If a marathon client object exists in our context, delete any apps in Marathon and wait until they die.""" print_container_logs('zookeeper') print_container_logs('marathon') if context.client: while True: apps = context.client.list_apps() if not apps: break for app in apps: context.client.delete_app(app.id, force=True) time.sleep(0.5) while context.client.list_deployments(): time.sleep(0.5) <commit_msg>Print container logs only in case of failed scenario<commit_after>import time from itest_utils import wait_for_marathon from itest_utils import print_container_logs def before_all(context): wait_for_marathon() def after_scenario(context, scenario): """If a marathon client object exists in our context, delete any apps in Marathon and wait until they die.""" if scenario.status != 'passed': print "Zookeeper container logs:" print_container_logs('zookeeper') print "Marathon container logs:" print_container_logs('marathon') if context.client: while True: apps = context.client.list_apps() if not apps: break for app in apps: context.client.delete_app(app.id, force=True) time.sleep(0.5) while context.client.list_deployments(): time.sleep(0.5)
692dd2d6563f75ff1a421c81037f7654cababafa
tests/pages/internet_health.py
tests/pages/internet_health.py
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from pages.base import BasePage from pages.regions.download_button import DownloadButton class InternetHealthPage(BasePage): URL_TEMPLATE = '/{locale}/internet-health/' _download_button_locator = (By.ID, 'global-nav-download-firefox') @property def download_button(self): el = self.find_element(*self._download_button_locator) return DownloadButton(self, root=el)
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from pages.base import BasePage from pages.regions.download_button import DownloadButton class InternetHealthPage(BasePage): URL_TEMPLATE = '/{locale}/internet-health/' _download_button_locator = (By.ID, 'protocol-nav-download-firefox') @property def download_button(self): el = self.find_element(*self._download_button_locator) return DownloadButton(self, root=el)
Fix failing integration test on /internet-health/
Fix failing integration test on /internet-health/
Python
mpl-2.0
mozilla/bedrock,alexgibson/bedrock,craigcook/bedrock,MichaelKohler/bedrock,sylvestre/bedrock,MichaelKohler/bedrock,pascalchevrel/bedrock,craigcook/bedrock,alexgibson/bedrock,ericawright/bedrock,pascalchevrel/bedrock,sgarrity/bedrock,mozilla/bedrock,alexgibson/bedrock,MichaelKohler/bedrock,hoosteeno/bedrock,flodolo/bedrock,sgarrity/bedrock,kyoshino/bedrock,sylvestre/bedrock,pascalchevrel/bedrock,flodolo/bedrock,flodolo/bedrock,ericawright/bedrock,kyoshino/bedrock,sylvestre/bedrock,alexgibson/bedrock,hoosteeno/bedrock,sylvestre/bedrock,ericawright/bedrock,mozilla/bedrock,hoosteeno/bedrock,kyoshino/bedrock,flodolo/bedrock,kyoshino/bedrock,craigcook/bedrock,sgarrity/bedrock,pascalchevrel/bedrock,MichaelKohler/bedrock,craigcook/bedrock,sgarrity/bedrock,mozilla/bedrock,ericawright/bedrock,hoosteeno/bedrock
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from pages.base import BasePage from pages.regions.download_button import DownloadButton class InternetHealthPage(BasePage): URL_TEMPLATE = '/{locale}/internet-health/' _download_button_locator = (By.ID, 'global-nav-download-firefox') @property def download_button(self): el = self.find_element(*self._download_button_locator) return DownloadButton(self, root=el) Fix failing integration test on /internet-health/
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from pages.base import BasePage from pages.regions.download_button import DownloadButton class InternetHealthPage(BasePage): URL_TEMPLATE = '/{locale}/internet-health/' _download_button_locator = (By.ID, 'protocol-nav-download-firefox') @property def download_button(self): el = self.find_element(*self._download_button_locator) return DownloadButton(self, root=el)
<commit_before># This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from pages.base import BasePage from pages.regions.download_button import DownloadButton class InternetHealthPage(BasePage): URL_TEMPLATE = '/{locale}/internet-health/' _download_button_locator = (By.ID, 'global-nav-download-firefox') @property def download_button(self): el = self.find_element(*self._download_button_locator) return DownloadButton(self, root=el) <commit_msg>Fix failing integration test on /internet-health/<commit_after>
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from pages.base import BasePage from pages.regions.download_button import DownloadButton class InternetHealthPage(BasePage): URL_TEMPLATE = '/{locale}/internet-health/' _download_button_locator = (By.ID, 'protocol-nav-download-firefox') @property def download_button(self): el = self.find_element(*self._download_button_locator) return DownloadButton(self, root=el)
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from pages.base import BasePage from pages.regions.download_button import DownloadButton class InternetHealthPage(BasePage): URL_TEMPLATE = '/{locale}/internet-health/' _download_button_locator = (By.ID, 'global-nav-download-firefox') @property def download_button(self): el = self.find_element(*self._download_button_locator) return DownloadButton(self, root=el) Fix failing integration test on /internet-health/# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from pages.base import BasePage from pages.regions.download_button import DownloadButton class InternetHealthPage(BasePage): URL_TEMPLATE = '/{locale}/internet-health/' _download_button_locator = (By.ID, 'protocol-nav-download-firefox') @property def download_button(self): el = self.find_element(*self._download_button_locator) return DownloadButton(self, root=el)
<commit_before># This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from pages.base import BasePage from pages.regions.download_button import DownloadButton class InternetHealthPage(BasePage): URL_TEMPLATE = '/{locale}/internet-health/' _download_button_locator = (By.ID, 'global-nav-download-firefox') @property def download_button(self): el = self.find_element(*self._download_button_locator) return DownloadButton(self, root=el) <commit_msg>Fix failing integration test on /internet-health/<commit_after># This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from selenium.webdriver.common.by import By from pages.base import BasePage from pages.regions.download_button import DownloadButton class InternetHealthPage(BasePage): URL_TEMPLATE = '/{locale}/internet-health/' _download_button_locator = (By.ID, 'protocol-nav-download-firefox') @property def download_button(self): el = self.find_element(*self._download_button_locator) return DownloadButton(self, root=el)
413c0b7f2df43543fd360bca1a9a6b9de4f6f5e8
integration-tests/features/steps/user_intent.py
integration-tests/features/steps/user_intent.py
"""Basic checks for the server API.""" from behave import then, when from urllib.parse import urljoin import requests from src.authorization_tokens import authorization from src.parsing import parse_token_clause def post_data_to_user_intent_endpoint(context, payload=None): """Post data into the REST API endpoint for user-intent.""" url = "/api/v1/user-intent" if payload is not None: context.response = requests.post(context.coreapi_url + url, headers=authorization(context), data=payload) else: context.response = requests.post(context.coreapi_url + url, headers=authorization(context)) @when('I call user-intent endpoint without any payload') def check_user_intent_without_payload(context): """Post no payload into the REST API endpoint for user-intent.""" post_data_to_user_intent_endpoint(context) @when('I call user-intent endpoint with empty payload') def check_user_intent_with_empty_payload(context): """Post empty into the REST API endpoint for user-intent.""" payload = {} post_data_to_user_intent_endpoint(context, payload)
"""Basic checks for the server API.""" from behave import then, when from urllib.parse import urljoin import requests from src.authorization_tokens import authorization from src.parsing import parse_token_clause def post_data_to_user_intent_endpoint(context, payload=None): """Post data into the REST API endpoint for user-intent.""" url = "/api/v1/user-intent" if payload is not None: context.response = requests.post(context.coreapi_url + url, headers=authorization(context), json=payload) else: context.response = requests.post(context.coreapi_url + url, headers=authorization(context)) @when('I call user-intent endpoint without any payload') def check_user_intent_without_payload(context): """Post no payload into the REST API endpoint for user-intent.""" post_data_to_user_intent_endpoint(context) @when('I call user-intent endpoint with empty payload') def check_user_intent_with_empty_payload(context): """Post empty into the REST API endpoint for user-intent.""" payload = {} post_data_to_user_intent_endpoint(context, payload)
Fix - send JSON to the API, not raw data
Fix - send JSON to the API, not raw data
Python
apache-2.0
tisnik/fabric8-analytics-common,tisnik/fabric8-analytics-common,tisnik/fabric8-analytics-common
"""Basic checks for the server API.""" from behave import then, when from urllib.parse import urljoin import requests from src.authorization_tokens import authorization from src.parsing import parse_token_clause def post_data_to_user_intent_endpoint(context, payload=None): """Post data into the REST API endpoint for user-intent.""" url = "/api/v1/user-intent" if payload is not None: context.response = requests.post(context.coreapi_url + url, headers=authorization(context), data=payload) else: context.response = requests.post(context.coreapi_url + url, headers=authorization(context)) @when('I call user-intent endpoint without any payload') def check_user_intent_without_payload(context): """Post no payload into the REST API endpoint for user-intent.""" post_data_to_user_intent_endpoint(context) @when('I call user-intent endpoint with empty payload') def check_user_intent_with_empty_payload(context): """Post empty into the REST API endpoint for user-intent.""" payload = {} post_data_to_user_intent_endpoint(context, payload) Fix - send JSON to the API, not raw data
"""Basic checks for the server API.""" from behave import then, when from urllib.parse import urljoin import requests from src.authorization_tokens import authorization from src.parsing import parse_token_clause def post_data_to_user_intent_endpoint(context, payload=None): """Post data into the REST API endpoint for user-intent.""" url = "/api/v1/user-intent" if payload is not None: context.response = requests.post(context.coreapi_url + url, headers=authorization(context), json=payload) else: context.response = requests.post(context.coreapi_url + url, headers=authorization(context)) @when('I call user-intent endpoint without any payload') def check_user_intent_without_payload(context): """Post no payload into the REST API endpoint for user-intent.""" post_data_to_user_intent_endpoint(context) @when('I call user-intent endpoint with empty payload') def check_user_intent_with_empty_payload(context): """Post empty into the REST API endpoint for user-intent.""" payload = {} post_data_to_user_intent_endpoint(context, payload)
<commit_before>"""Basic checks for the server API.""" from behave import then, when from urllib.parse import urljoin import requests from src.authorization_tokens import authorization from src.parsing import parse_token_clause def post_data_to_user_intent_endpoint(context, payload=None): """Post data into the REST API endpoint for user-intent.""" url = "/api/v1/user-intent" if payload is not None: context.response = requests.post(context.coreapi_url + url, headers=authorization(context), data=payload) else: context.response = requests.post(context.coreapi_url + url, headers=authorization(context)) @when('I call user-intent endpoint without any payload') def check_user_intent_without_payload(context): """Post no payload into the REST API endpoint for user-intent.""" post_data_to_user_intent_endpoint(context) @when('I call user-intent endpoint with empty payload') def check_user_intent_with_empty_payload(context): """Post empty into the REST API endpoint for user-intent.""" payload = {} post_data_to_user_intent_endpoint(context, payload) <commit_msg>Fix - send JSON to the API, not raw data<commit_after>
"""Basic checks for the server API.""" from behave import then, when from urllib.parse import urljoin import requests from src.authorization_tokens import authorization from src.parsing import parse_token_clause def post_data_to_user_intent_endpoint(context, payload=None): """Post data into the REST API endpoint for user-intent.""" url = "/api/v1/user-intent" if payload is not None: context.response = requests.post(context.coreapi_url + url, headers=authorization(context), json=payload) else: context.response = requests.post(context.coreapi_url + url, headers=authorization(context)) @when('I call user-intent endpoint without any payload') def check_user_intent_without_payload(context): """Post no payload into the REST API endpoint for user-intent.""" post_data_to_user_intent_endpoint(context) @when('I call user-intent endpoint with empty payload') def check_user_intent_with_empty_payload(context): """Post empty into the REST API endpoint for user-intent.""" payload = {} post_data_to_user_intent_endpoint(context, payload)
"""Basic checks for the server API.""" from behave import then, when from urllib.parse import urljoin import requests from src.authorization_tokens import authorization from src.parsing import parse_token_clause def post_data_to_user_intent_endpoint(context, payload=None): """Post data into the REST API endpoint for user-intent.""" url = "/api/v1/user-intent" if payload is not None: context.response = requests.post(context.coreapi_url + url, headers=authorization(context), data=payload) else: context.response = requests.post(context.coreapi_url + url, headers=authorization(context)) @when('I call user-intent endpoint without any payload') def check_user_intent_without_payload(context): """Post no payload into the REST API endpoint for user-intent.""" post_data_to_user_intent_endpoint(context) @when('I call user-intent endpoint with empty payload') def check_user_intent_with_empty_payload(context): """Post empty into the REST API endpoint for user-intent.""" payload = {} post_data_to_user_intent_endpoint(context, payload) Fix - send JSON to the API, not raw data"""Basic checks for the server API.""" from behave import then, when from urllib.parse import urljoin import requests from src.authorization_tokens import authorization from src.parsing import parse_token_clause def post_data_to_user_intent_endpoint(context, payload=None): """Post data into the REST API endpoint for user-intent.""" url = "/api/v1/user-intent" if payload is not None: context.response = requests.post(context.coreapi_url + url, headers=authorization(context), json=payload) else: context.response = requests.post(context.coreapi_url + url, headers=authorization(context)) @when('I call user-intent endpoint without any payload') def check_user_intent_without_payload(context): """Post no payload into the REST API endpoint for user-intent.""" post_data_to_user_intent_endpoint(context) @when('I call user-intent endpoint with empty payload') def check_user_intent_with_empty_payload(context): """Post empty into the REST API endpoint for user-intent.""" payload = {} post_data_to_user_intent_endpoint(context, payload)
<commit_before>"""Basic checks for the server API.""" from behave import then, when from urllib.parse import urljoin import requests from src.authorization_tokens import authorization from src.parsing import parse_token_clause def post_data_to_user_intent_endpoint(context, payload=None): """Post data into the REST API endpoint for user-intent.""" url = "/api/v1/user-intent" if payload is not None: context.response = requests.post(context.coreapi_url + url, headers=authorization(context), data=payload) else: context.response = requests.post(context.coreapi_url + url, headers=authorization(context)) @when('I call user-intent endpoint without any payload') def check_user_intent_without_payload(context): """Post no payload into the REST API endpoint for user-intent.""" post_data_to_user_intent_endpoint(context) @when('I call user-intent endpoint with empty payload') def check_user_intent_with_empty_payload(context): """Post empty into the REST API endpoint for user-intent.""" payload = {} post_data_to_user_intent_endpoint(context, payload) <commit_msg>Fix - send JSON to the API, not raw data<commit_after>"""Basic checks for the server API.""" from behave import then, when from urllib.parse import urljoin import requests from src.authorization_tokens import authorization from src.parsing import parse_token_clause def post_data_to_user_intent_endpoint(context, payload=None): """Post data into the REST API endpoint for user-intent.""" url = "/api/v1/user-intent" if payload is not None: context.response = requests.post(context.coreapi_url + url, headers=authorization(context), json=payload) else: context.response = requests.post(context.coreapi_url + url, headers=authorization(context)) @when('I call user-intent endpoint without any payload') def check_user_intent_without_payload(context): """Post no payload into the REST API endpoint for user-intent.""" post_data_to_user_intent_endpoint(context) @when('I call user-intent endpoint with empty payload') def check_user_intent_with_empty_payload(context): """Post empty into the REST API endpoint for user-intent.""" payload = {} post_data_to_user_intent_endpoint(context, payload)
a8af7cd7918322615a342c2fd662f394866da55f
tests/sentry/metrics/test_datadog.py
tests/sentry/metrics/test_datadog.py
from __future__ import absolute_import import socket from mock import patch from sentry.metrics.datadog import DatadogMetricsBackend from sentry.testutils import TestCase class DatadogMetricsBackendTest(TestCase): def setUp(self): self.backend = DatadogMetricsBackend(prefix='sentrytest.') @patch('datadog.threadstats.base.ThreadStats.increment') def test_incr(self, mock_incr): self.backend.incr('foo', instance='bar') mock_incr.assert_called_once_with( 'sentrytest.foo', 1, tags=['instance:bar'], host=socket.gethostname(), ) @patch('datadog.threadstats.base.ThreadStats.timing') def test_timing(self, mock_timing): self.backend.timing('foo', 30, instance='bar') mock_timing.assert_called_once_with( 'sentrytest.foo', 30, sample_rate=1, tags=['instance:bar'], host=socket.gethostname(), )
from __future__ import absolute_import from mock import patch from datadog.util.hostname import get_hostname from sentry.metrics.datadog import DatadogMetricsBackend from sentry.testutils import TestCase class DatadogMetricsBackendTest(TestCase): def setUp(self): self.backend = DatadogMetricsBackend(prefix='sentrytest.') @patch('datadog.threadstats.base.ThreadStats.increment') def test_incr(self, mock_incr): self.backend.incr('foo', instance='bar') mock_incr.assert_called_once_with( 'sentrytest.foo', 1, tags=['instance:bar'], host=get_hostname(), ) @patch('datadog.threadstats.base.ThreadStats.timing') def test_timing(self, mock_timing): self.backend.timing('foo', 30, instance='bar') mock_timing.assert_called_once_with( 'sentrytest.foo', 30, sample_rate=1, tags=['instance:bar'], host=get_hostname(), )
Test DatadogMetricsBackend against datadog's get_hostname
Test DatadogMetricsBackend against datadog's get_hostname This fixes tests in Travis since the hostname returned is different
Python
bsd-3-clause
pauloschilling/sentry,pauloschilling/sentry,pauloschilling/sentry
from __future__ import absolute_import import socket from mock import patch from sentry.metrics.datadog import DatadogMetricsBackend from sentry.testutils import TestCase class DatadogMetricsBackendTest(TestCase): def setUp(self): self.backend = DatadogMetricsBackend(prefix='sentrytest.') @patch('datadog.threadstats.base.ThreadStats.increment') def test_incr(self, mock_incr): self.backend.incr('foo', instance='bar') mock_incr.assert_called_once_with( 'sentrytest.foo', 1, tags=['instance:bar'], host=socket.gethostname(), ) @patch('datadog.threadstats.base.ThreadStats.timing') def test_timing(self, mock_timing): self.backend.timing('foo', 30, instance='bar') mock_timing.assert_called_once_with( 'sentrytest.foo', 30, sample_rate=1, tags=['instance:bar'], host=socket.gethostname(), ) Test DatadogMetricsBackend against datadog's get_hostname This fixes tests in Travis since the hostname returned is different
from __future__ import absolute_import from mock import patch from datadog.util.hostname import get_hostname from sentry.metrics.datadog import DatadogMetricsBackend from sentry.testutils import TestCase class DatadogMetricsBackendTest(TestCase): def setUp(self): self.backend = DatadogMetricsBackend(prefix='sentrytest.') @patch('datadog.threadstats.base.ThreadStats.increment') def test_incr(self, mock_incr): self.backend.incr('foo', instance='bar') mock_incr.assert_called_once_with( 'sentrytest.foo', 1, tags=['instance:bar'], host=get_hostname(), ) @patch('datadog.threadstats.base.ThreadStats.timing') def test_timing(self, mock_timing): self.backend.timing('foo', 30, instance='bar') mock_timing.assert_called_once_with( 'sentrytest.foo', 30, sample_rate=1, tags=['instance:bar'], host=get_hostname(), )
<commit_before>from __future__ import absolute_import import socket from mock import patch from sentry.metrics.datadog import DatadogMetricsBackend from sentry.testutils import TestCase class DatadogMetricsBackendTest(TestCase): def setUp(self): self.backend = DatadogMetricsBackend(prefix='sentrytest.') @patch('datadog.threadstats.base.ThreadStats.increment') def test_incr(self, mock_incr): self.backend.incr('foo', instance='bar') mock_incr.assert_called_once_with( 'sentrytest.foo', 1, tags=['instance:bar'], host=socket.gethostname(), ) @patch('datadog.threadstats.base.ThreadStats.timing') def test_timing(self, mock_timing): self.backend.timing('foo', 30, instance='bar') mock_timing.assert_called_once_with( 'sentrytest.foo', 30, sample_rate=1, tags=['instance:bar'], host=socket.gethostname(), ) <commit_msg>Test DatadogMetricsBackend against datadog's get_hostname This fixes tests in Travis since the hostname returned is different<commit_after>
from __future__ import absolute_import from mock import patch from datadog.util.hostname import get_hostname from sentry.metrics.datadog import DatadogMetricsBackend from sentry.testutils import TestCase class DatadogMetricsBackendTest(TestCase): def setUp(self): self.backend = DatadogMetricsBackend(prefix='sentrytest.') @patch('datadog.threadstats.base.ThreadStats.increment') def test_incr(self, mock_incr): self.backend.incr('foo', instance='bar') mock_incr.assert_called_once_with( 'sentrytest.foo', 1, tags=['instance:bar'], host=get_hostname(), ) @patch('datadog.threadstats.base.ThreadStats.timing') def test_timing(self, mock_timing): self.backend.timing('foo', 30, instance='bar') mock_timing.assert_called_once_with( 'sentrytest.foo', 30, sample_rate=1, tags=['instance:bar'], host=get_hostname(), )
from __future__ import absolute_import import socket from mock import patch from sentry.metrics.datadog import DatadogMetricsBackend from sentry.testutils import TestCase class DatadogMetricsBackendTest(TestCase): def setUp(self): self.backend = DatadogMetricsBackend(prefix='sentrytest.') @patch('datadog.threadstats.base.ThreadStats.increment') def test_incr(self, mock_incr): self.backend.incr('foo', instance='bar') mock_incr.assert_called_once_with( 'sentrytest.foo', 1, tags=['instance:bar'], host=socket.gethostname(), ) @patch('datadog.threadstats.base.ThreadStats.timing') def test_timing(self, mock_timing): self.backend.timing('foo', 30, instance='bar') mock_timing.assert_called_once_with( 'sentrytest.foo', 30, sample_rate=1, tags=['instance:bar'], host=socket.gethostname(), ) Test DatadogMetricsBackend against datadog's get_hostname This fixes tests in Travis since the hostname returned is differentfrom __future__ import absolute_import from mock import patch from datadog.util.hostname import get_hostname from sentry.metrics.datadog import DatadogMetricsBackend from sentry.testutils import TestCase class DatadogMetricsBackendTest(TestCase): def setUp(self): self.backend = DatadogMetricsBackend(prefix='sentrytest.') @patch('datadog.threadstats.base.ThreadStats.increment') def test_incr(self, mock_incr): self.backend.incr('foo', instance='bar') mock_incr.assert_called_once_with( 'sentrytest.foo', 1, tags=['instance:bar'], host=get_hostname(), ) @patch('datadog.threadstats.base.ThreadStats.timing') def test_timing(self, mock_timing): self.backend.timing('foo', 30, instance='bar') mock_timing.assert_called_once_with( 'sentrytest.foo', 30, sample_rate=1, tags=['instance:bar'], host=get_hostname(), )
<commit_before>from __future__ import absolute_import import socket from mock import patch from sentry.metrics.datadog import DatadogMetricsBackend from sentry.testutils import TestCase class DatadogMetricsBackendTest(TestCase): def setUp(self): self.backend = DatadogMetricsBackend(prefix='sentrytest.') @patch('datadog.threadstats.base.ThreadStats.increment') def test_incr(self, mock_incr): self.backend.incr('foo', instance='bar') mock_incr.assert_called_once_with( 'sentrytest.foo', 1, tags=['instance:bar'], host=socket.gethostname(), ) @patch('datadog.threadstats.base.ThreadStats.timing') def test_timing(self, mock_timing): self.backend.timing('foo', 30, instance='bar') mock_timing.assert_called_once_with( 'sentrytest.foo', 30, sample_rate=1, tags=['instance:bar'], host=socket.gethostname(), ) <commit_msg>Test DatadogMetricsBackend against datadog's get_hostname This fixes tests in Travis since the hostname returned is different<commit_after>from __future__ import absolute_import from mock import patch from datadog.util.hostname import get_hostname from sentry.metrics.datadog import DatadogMetricsBackend from sentry.testutils import TestCase class DatadogMetricsBackendTest(TestCase): def setUp(self): self.backend = DatadogMetricsBackend(prefix='sentrytest.') @patch('datadog.threadstats.base.ThreadStats.increment') def test_incr(self, mock_incr): self.backend.incr('foo', instance='bar') mock_incr.assert_called_once_with( 'sentrytest.foo', 1, tags=['instance:bar'], host=get_hostname(), ) @patch('datadog.threadstats.base.ThreadStats.timing') def test_timing(self, mock_timing): self.backend.timing('foo', 30, instance='bar') mock_timing.assert_called_once_with( 'sentrytest.foo', 30, sample_rate=1, tags=['instance:bar'], host=get_hostname(), )
7c117c4555fdf30b3d98a453c7e28245dca0c9b9
tests/test_stack/test_stack.py
tests/test_stack/test_stack.py
import unittest from aids.stack.stack import Stack class StackTestCase(unittest.TestCase): ''' Unit tests for the Stack data structure ''' def setUp(self): pass def test_stack_initialization(self): test_stack = Stack() self.assertTrue(isinstance(test_stack, Stack)) def test_stack_is_empty(self): test_stack = Stack() self.assertTrue(test_stack.is_empty()) def tearDown(self): pass
import unittest from aids.stack.stack import Stack class StackTestCase(unittest.TestCase): ''' Unit tests for the Stack data structure ''' def setUp(self): self.test_stack = Stack() def test_stack_initialization(self): self.assertTrue(isinstance(self.test_stack, Stack)) def test_stack_is_empty(self): self.assertTrue(self.test_stack.is_empty()) def test_stack_push(self): self.test_stack.push(1) self.assertEqual(len(self.test_stack), 1) def test_stack_peek(self): self.test_stack.push(1) self.assertEqual(self.test_stack.peek(), 1) def test_stack_pop(self): self.test_stack.push(1) self.assertEqual(self.test_stack.pop(), 1) def test_stack_size(self): self.test_stack.push(1) self.assertEqual(self.test_stack.size(), 1) def tearDown(self): pass
Add unit tests for push, peek, pop and size
Add unit tests for push, peek, pop and size
Python
mit
ueg1990/aids
import unittest from aids.stack.stack import Stack class StackTestCase(unittest.TestCase): ''' Unit tests for the Stack data structure ''' def setUp(self): pass def test_stack_initialization(self): test_stack = Stack() self.assertTrue(isinstance(test_stack, Stack)) def test_stack_is_empty(self): test_stack = Stack() self.assertTrue(test_stack.is_empty()) def tearDown(self): pass Add unit tests for push, peek, pop and size
import unittest from aids.stack.stack import Stack class StackTestCase(unittest.TestCase): ''' Unit tests for the Stack data structure ''' def setUp(self): self.test_stack = Stack() def test_stack_initialization(self): self.assertTrue(isinstance(self.test_stack, Stack)) def test_stack_is_empty(self): self.assertTrue(self.test_stack.is_empty()) def test_stack_push(self): self.test_stack.push(1) self.assertEqual(len(self.test_stack), 1) def test_stack_peek(self): self.test_stack.push(1) self.assertEqual(self.test_stack.peek(), 1) def test_stack_pop(self): self.test_stack.push(1) self.assertEqual(self.test_stack.pop(), 1) def test_stack_size(self): self.test_stack.push(1) self.assertEqual(self.test_stack.size(), 1) def tearDown(self): pass
<commit_before>import unittest from aids.stack.stack import Stack class StackTestCase(unittest.TestCase): ''' Unit tests for the Stack data structure ''' def setUp(self): pass def test_stack_initialization(self): test_stack = Stack() self.assertTrue(isinstance(test_stack, Stack)) def test_stack_is_empty(self): test_stack = Stack() self.assertTrue(test_stack.is_empty()) def tearDown(self): pass <commit_msg>Add unit tests for push, peek, pop and size<commit_after>
import unittest from aids.stack.stack import Stack class StackTestCase(unittest.TestCase): ''' Unit tests for the Stack data structure ''' def setUp(self): self.test_stack = Stack() def test_stack_initialization(self): self.assertTrue(isinstance(self.test_stack, Stack)) def test_stack_is_empty(self): self.assertTrue(self.test_stack.is_empty()) def test_stack_push(self): self.test_stack.push(1) self.assertEqual(len(self.test_stack), 1) def test_stack_peek(self): self.test_stack.push(1) self.assertEqual(self.test_stack.peek(), 1) def test_stack_pop(self): self.test_stack.push(1) self.assertEqual(self.test_stack.pop(), 1) def test_stack_size(self): self.test_stack.push(1) self.assertEqual(self.test_stack.size(), 1) def tearDown(self): pass
import unittest from aids.stack.stack import Stack class StackTestCase(unittest.TestCase): ''' Unit tests for the Stack data structure ''' def setUp(self): pass def test_stack_initialization(self): test_stack = Stack() self.assertTrue(isinstance(test_stack, Stack)) def test_stack_is_empty(self): test_stack = Stack() self.assertTrue(test_stack.is_empty()) def tearDown(self): pass Add unit tests for push, peek, pop and sizeimport unittest from aids.stack.stack import Stack class StackTestCase(unittest.TestCase): ''' Unit tests for the Stack data structure ''' def setUp(self): self.test_stack = Stack() def test_stack_initialization(self): self.assertTrue(isinstance(self.test_stack, Stack)) def test_stack_is_empty(self): self.assertTrue(self.test_stack.is_empty()) def test_stack_push(self): self.test_stack.push(1) self.assertEqual(len(self.test_stack), 1) def test_stack_peek(self): self.test_stack.push(1) self.assertEqual(self.test_stack.peek(), 1) def test_stack_pop(self): self.test_stack.push(1) self.assertEqual(self.test_stack.pop(), 1) def test_stack_size(self): self.test_stack.push(1) self.assertEqual(self.test_stack.size(), 1) def tearDown(self): pass
<commit_before>import unittest from aids.stack.stack import Stack class StackTestCase(unittest.TestCase): ''' Unit tests for the Stack data structure ''' def setUp(self): pass def test_stack_initialization(self): test_stack = Stack() self.assertTrue(isinstance(test_stack, Stack)) def test_stack_is_empty(self): test_stack = Stack() self.assertTrue(test_stack.is_empty()) def tearDown(self): pass <commit_msg>Add unit tests for push, peek, pop and size<commit_after>import unittest from aids.stack.stack import Stack class StackTestCase(unittest.TestCase): ''' Unit tests for the Stack data structure ''' def setUp(self): self.test_stack = Stack() def test_stack_initialization(self): self.assertTrue(isinstance(self.test_stack, Stack)) def test_stack_is_empty(self): self.assertTrue(self.test_stack.is_empty()) def test_stack_push(self): self.test_stack.push(1) self.assertEqual(len(self.test_stack), 1) def test_stack_peek(self): self.test_stack.push(1) self.assertEqual(self.test_stack.peek(), 1) def test_stack_pop(self): self.test_stack.push(1) self.assertEqual(self.test_stack.pop(), 1) def test_stack_size(self): self.test_stack.push(1) self.assertEqual(self.test_stack.size(), 1) def tearDown(self): pass
1ef1851e508295f6d4bf01289591f42c21656df7
test/on_yubikey/test_interfaces.py
test/on_yubikey/test_interfaces.py
import unittest from .framework import DestructiveYubikeyTestCase, exactly_one_yubikey_present from yubikit.core.otp import OtpConnection from yubikit.core.fido import FidoConnection from yubikit.core.smartcard import SmartCardConnection from ykman.device import connect_to_device from time import sleep @unittest.skipIf( not exactly_one_yubikey_present(), "Exactly one YubiKey must be present." ) class TestInterfaces(DestructiveYubikeyTestCase): def try_connection(self, conn_type): for _ in range(8): try: conn = connect_to_device(None, [conn_type])[0] conn.close() return except Exception: sleep(0.5) self.fail("Failed connecting to device over %s" % conn_type) def test_switch_interfaces(self): self.try_connection(FidoConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection)
import unittest from .framework import DestructiveYubikeyTestCase, exactly_one_yubikey_present from yubikit.core.otp import OtpConnection from yubikit.core.fido import FidoConnection from yubikit.core.smartcard import SmartCardConnection from ykman.base import YUBIKEY from ykman.device import connect_to_device from time import sleep @unittest.skipIf( not exactly_one_yubikey_present(), "Exactly one YubiKey must be present." ) class TestInterfaces(DestructiveYubikeyTestCase): def try_connection(self, conn_type): if self.key_type == YUBIKEY.NEO and conn_type == SmartCardConnection: sleep(3.5) conn, dev, info = connect_to_device(None, [conn_type]) conn.close() def setUp(self): conn, dev, info = connect_to_device() conn.close() self.key_type = dev.pid.get_type() def test_switch_interfaces(self): self.try_connection(FidoConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(SmartCardConnection) self.try_connection(FidoConnection)
Test handling of reclaim timeout.
Test handling of reclaim timeout.
Python
bsd-2-clause
Yubico/yubikey-manager,Yubico/yubikey-manager
import unittest from .framework import DestructiveYubikeyTestCase, exactly_one_yubikey_present from yubikit.core.otp import OtpConnection from yubikit.core.fido import FidoConnection from yubikit.core.smartcard import SmartCardConnection from ykman.device import connect_to_device from time import sleep @unittest.skipIf( not exactly_one_yubikey_present(), "Exactly one YubiKey must be present." ) class TestInterfaces(DestructiveYubikeyTestCase): def try_connection(self, conn_type): for _ in range(8): try: conn = connect_to_device(None, [conn_type])[0] conn.close() return except Exception: sleep(0.5) self.fail("Failed connecting to device over %s" % conn_type) def test_switch_interfaces(self): self.try_connection(FidoConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) Test handling of reclaim timeout.
import unittest from .framework import DestructiveYubikeyTestCase, exactly_one_yubikey_present from yubikit.core.otp import OtpConnection from yubikit.core.fido import FidoConnection from yubikit.core.smartcard import SmartCardConnection from ykman.base import YUBIKEY from ykman.device import connect_to_device from time import sleep @unittest.skipIf( not exactly_one_yubikey_present(), "Exactly one YubiKey must be present." ) class TestInterfaces(DestructiveYubikeyTestCase): def try_connection(self, conn_type): if self.key_type == YUBIKEY.NEO and conn_type == SmartCardConnection: sleep(3.5) conn, dev, info = connect_to_device(None, [conn_type]) conn.close() def setUp(self): conn, dev, info = connect_to_device() conn.close() self.key_type = dev.pid.get_type() def test_switch_interfaces(self): self.try_connection(FidoConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(SmartCardConnection) self.try_connection(FidoConnection)
<commit_before>import unittest from .framework import DestructiveYubikeyTestCase, exactly_one_yubikey_present from yubikit.core.otp import OtpConnection from yubikit.core.fido import FidoConnection from yubikit.core.smartcard import SmartCardConnection from ykman.device import connect_to_device from time import sleep @unittest.skipIf( not exactly_one_yubikey_present(), "Exactly one YubiKey must be present." ) class TestInterfaces(DestructiveYubikeyTestCase): def try_connection(self, conn_type): for _ in range(8): try: conn = connect_to_device(None, [conn_type])[0] conn.close() return except Exception: sleep(0.5) self.fail("Failed connecting to device over %s" % conn_type) def test_switch_interfaces(self): self.try_connection(FidoConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) <commit_msg>Test handling of reclaim timeout.<commit_after>
import unittest from .framework import DestructiveYubikeyTestCase, exactly_one_yubikey_present from yubikit.core.otp import OtpConnection from yubikit.core.fido import FidoConnection from yubikit.core.smartcard import SmartCardConnection from ykman.base import YUBIKEY from ykman.device import connect_to_device from time import sleep @unittest.skipIf( not exactly_one_yubikey_present(), "Exactly one YubiKey must be present." ) class TestInterfaces(DestructiveYubikeyTestCase): def try_connection(self, conn_type): if self.key_type == YUBIKEY.NEO and conn_type == SmartCardConnection: sleep(3.5) conn, dev, info = connect_to_device(None, [conn_type]) conn.close() def setUp(self): conn, dev, info = connect_to_device() conn.close() self.key_type = dev.pid.get_type() def test_switch_interfaces(self): self.try_connection(FidoConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(SmartCardConnection) self.try_connection(FidoConnection)
import unittest from .framework import DestructiveYubikeyTestCase, exactly_one_yubikey_present from yubikit.core.otp import OtpConnection from yubikit.core.fido import FidoConnection from yubikit.core.smartcard import SmartCardConnection from ykman.device import connect_to_device from time import sleep @unittest.skipIf( not exactly_one_yubikey_present(), "Exactly one YubiKey must be present." ) class TestInterfaces(DestructiveYubikeyTestCase): def try_connection(self, conn_type): for _ in range(8): try: conn = connect_to_device(None, [conn_type])[0] conn.close() return except Exception: sleep(0.5) self.fail("Failed connecting to device over %s" % conn_type) def test_switch_interfaces(self): self.try_connection(FidoConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) Test handling of reclaim timeout.import unittest from .framework import DestructiveYubikeyTestCase, exactly_one_yubikey_present from yubikit.core.otp import OtpConnection from yubikit.core.fido import FidoConnection from yubikit.core.smartcard import SmartCardConnection from ykman.base import YUBIKEY from ykman.device import connect_to_device from time import sleep @unittest.skipIf( not exactly_one_yubikey_present(), "Exactly one YubiKey must be present." ) class TestInterfaces(DestructiveYubikeyTestCase): def try_connection(self, conn_type): if self.key_type == YUBIKEY.NEO and conn_type == SmartCardConnection: sleep(3.5) conn, dev, info = connect_to_device(None, [conn_type]) conn.close() def setUp(self): conn, dev, info = connect_to_device() conn.close() self.key_type = dev.pid.get_type() def test_switch_interfaces(self): self.try_connection(FidoConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(SmartCardConnection) self.try_connection(FidoConnection)
<commit_before>import unittest from .framework import DestructiveYubikeyTestCase, exactly_one_yubikey_present from yubikit.core.otp import OtpConnection from yubikit.core.fido import FidoConnection from yubikit.core.smartcard import SmartCardConnection from ykman.device import connect_to_device from time import sleep @unittest.skipIf( not exactly_one_yubikey_present(), "Exactly one YubiKey must be present." ) class TestInterfaces(DestructiveYubikeyTestCase): def try_connection(self, conn_type): for _ in range(8): try: conn = connect_to_device(None, [conn_type])[0] conn.close() return except Exception: sleep(0.5) self.fail("Failed connecting to device over %s" % conn_type) def test_switch_interfaces(self): self.try_connection(FidoConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) <commit_msg>Test handling of reclaim timeout.<commit_after>import unittest from .framework import DestructiveYubikeyTestCase, exactly_one_yubikey_present from yubikit.core.otp import OtpConnection from yubikit.core.fido import FidoConnection from yubikit.core.smartcard import SmartCardConnection from ykman.base import YUBIKEY from ykman.device import connect_to_device from time import sleep @unittest.skipIf( not exactly_one_yubikey_present(), "Exactly one YubiKey must be present." ) class TestInterfaces(DestructiveYubikeyTestCase): def try_connection(self, conn_type): if self.key_type == YUBIKEY.NEO and conn_type == SmartCardConnection: sleep(3.5) conn, dev, info = connect_to_device(None, [conn_type]) conn.close() def setUp(self): conn, dev, info = connect_to_device() conn.close() self.key_type = dev.pid.get_type() def test_switch_interfaces(self): self.try_connection(FidoConnection) self.try_connection(OtpConnection) self.try_connection(FidoConnection) self.try_connection(SmartCardConnection) self.try_connection(OtpConnection) self.try_connection(SmartCardConnection) self.try_connection(FidoConnection)
300cbd3ff4d0e5021892f7c9940635695cb017a3
integration-test/197-clip-buildings.py
integration-test/197-clip-buildings.py
from shapely.geometry import shape # this is mid way along the High Line in NYC, which is a huge long # "building". we should be clipping it to a buffer of 3x the tile # dimensions. # http://www.openstreetmap.org/way/37054313 with features_in_tile_layer(16, 19295, 24631, 'buildings') as buildings: # max width and height in degress as 3x the size of the above tile max_w = 0.0164794921875 max_h = 0.012484410579673977 # need to check that we at least saw the high line saw_the_high_line = False for building in buildings: bounds = shape(building['geometry']).bounds w = bounds[2] - bounds[0] h = bounds[3] - bounds[1] if building['properties']['id'] == 37054313: saw_the_high_line = True if w > max_w or h > max_h: raise Exception("feature %r is %rx%r, larger than the allowed " "%rx%r." % (building['properties']['id'], w, h, max_w, max_h)) if not saw_the_high_line: raise Exception("Expected to see the High Line in this tile, " "but didn't.")
from shapely.geometry import shape # this is mid way along the High Line in NYC, which is a huge long # "building". we should be clipping it to a buffer of 3x the tile # dimensions. # http://www.openstreetmap.org/relation/7141751 with features_in_tile_layer(16, 19295, 24631, 'buildings') as buildings: # max width and height in degress as 3x the size of the above tile max_w = 0.0164794921875 max_h = 0.012484410579673977 # need to check that we at least saw the high line saw_the_high_line = False for building in buildings: bounds = shape(building['geometry']).bounds w = bounds[2] - bounds[0] h = bounds[3] - bounds[1] if building['properties']['id'] == -7141751: saw_the_high_line = True if w > max_w or h > max_h: raise Exception("feature %r is %rx%r, larger than the allowed " "%rx%r." % (building['properties']['id'], w, h, max_w, max_h)) if not saw_the_high_line: raise Exception("Expected to see the High Line in this tile, " "but didn't.")
Update data for the High Line, NYC
Update data for the High Line, NYC Looks like it was turned into a multipolygon relation in [this changeset](http://www.openstreetmap.org/changeset/47542769).
Python
mit
mapzen/vector-datasource,mapzen/vector-datasource,mapzen/vector-datasource
from shapely.geometry import shape # this is mid way along the High Line in NYC, which is a huge long # "building". we should be clipping it to a buffer of 3x the tile # dimensions. # http://www.openstreetmap.org/way/37054313 with features_in_tile_layer(16, 19295, 24631, 'buildings') as buildings: # max width and height in degress as 3x the size of the above tile max_w = 0.0164794921875 max_h = 0.012484410579673977 # need to check that we at least saw the high line saw_the_high_line = False for building in buildings: bounds = shape(building['geometry']).bounds w = bounds[2] - bounds[0] h = bounds[3] - bounds[1] if building['properties']['id'] == 37054313: saw_the_high_line = True if w > max_w or h > max_h: raise Exception("feature %r is %rx%r, larger than the allowed " "%rx%r." % (building['properties']['id'], w, h, max_w, max_h)) if not saw_the_high_line: raise Exception("Expected to see the High Line in this tile, " "but didn't.") Update data for the High Line, NYC Looks like it was turned into a multipolygon relation in [this changeset](http://www.openstreetmap.org/changeset/47542769).
from shapely.geometry import shape # this is mid way along the High Line in NYC, which is a huge long # "building". we should be clipping it to a buffer of 3x the tile # dimensions. # http://www.openstreetmap.org/relation/7141751 with features_in_tile_layer(16, 19295, 24631, 'buildings') as buildings: # max width and height in degress as 3x the size of the above tile max_w = 0.0164794921875 max_h = 0.012484410579673977 # need to check that we at least saw the high line saw_the_high_line = False for building in buildings: bounds = shape(building['geometry']).bounds w = bounds[2] - bounds[0] h = bounds[3] - bounds[1] if building['properties']['id'] == -7141751: saw_the_high_line = True if w > max_w or h > max_h: raise Exception("feature %r is %rx%r, larger than the allowed " "%rx%r." % (building['properties']['id'], w, h, max_w, max_h)) if not saw_the_high_line: raise Exception("Expected to see the High Line in this tile, " "but didn't.")
<commit_before>from shapely.geometry import shape # this is mid way along the High Line in NYC, which is a huge long # "building". we should be clipping it to a buffer of 3x the tile # dimensions. # http://www.openstreetmap.org/way/37054313 with features_in_tile_layer(16, 19295, 24631, 'buildings') as buildings: # max width and height in degress as 3x the size of the above tile max_w = 0.0164794921875 max_h = 0.012484410579673977 # need to check that we at least saw the high line saw_the_high_line = False for building in buildings: bounds = shape(building['geometry']).bounds w = bounds[2] - bounds[0] h = bounds[3] - bounds[1] if building['properties']['id'] == 37054313: saw_the_high_line = True if w > max_w or h > max_h: raise Exception("feature %r is %rx%r, larger than the allowed " "%rx%r." % (building['properties']['id'], w, h, max_w, max_h)) if not saw_the_high_line: raise Exception("Expected to see the High Line in this tile, " "but didn't.") <commit_msg>Update data for the High Line, NYC Looks like it was turned into a multipolygon relation in [this changeset](http://www.openstreetmap.org/changeset/47542769).<commit_after>
from shapely.geometry import shape # this is mid way along the High Line in NYC, which is a huge long # "building". we should be clipping it to a buffer of 3x the tile # dimensions. # http://www.openstreetmap.org/relation/7141751 with features_in_tile_layer(16, 19295, 24631, 'buildings') as buildings: # max width and height in degress as 3x the size of the above tile max_w = 0.0164794921875 max_h = 0.012484410579673977 # need to check that we at least saw the high line saw_the_high_line = False for building in buildings: bounds = shape(building['geometry']).bounds w = bounds[2] - bounds[0] h = bounds[3] - bounds[1] if building['properties']['id'] == -7141751: saw_the_high_line = True if w > max_w or h > max_h: raise Exception("feature %r is %rx%r, larger than the allowed " "%rx%r." % (building['properties']['id'], w, h, max_w, max_h)) if not saw_the_high_line: raise Exception("Expected to see the High Line in this tile, " "but didn't.")
from shapely.geometry import shape # this is mid way along the High Line in NYC, which is a huge long # "building". we should be clipping it to a buffer of 3x the tile # dimensions. # http://www.openstreetmap.org/way/37054313 with features_in_tile_layer(16, 19295, 24631, 'buildings') as buildings: # max width and height in degress as 3x the size of the above tile max_w = 0.0164794921875 max_h = 0.012484410579673977 # need to check that we at least saw the high line saw_the_high_line = False for building in buildings: bounds = shape(building['geometry']).bounds w = bounds[2] - bounds[0] h = bounds[3] - bounds[1] if building['properties']['id'] == 37054313: saw_the_high_line = True if w > max_w or h > max_h: raise Exception("feature %r is %rx%r, larger than the allowed " "%rx%r." % (building['properties']['id'], w, h, max_w, max_h)) if not saw_the_high_line: raise Exception("Expected to see the High Line in this tile, " "but didn't.") Update data for the High Line, NYC Looks like it was turned into a multipolygon relation in [this changeset](http://www.openstreetmap.org/changeset/47542769).from shapely.geometry import shape # this is mid way along the High Line in NYC, which is a huge long # "building". we should be clipping it to a buffer of 3x the tile # dimensions. # http://www.openstreetmap.org/relation/7141751 with features_in_tile_layer(16, 19295, 24631, 'buildings') as buildings: # max width and height in degress as 3x the size of the above tile max_w = 0.0164794921875 max_h = 0.012484410579673977 # need to check that we at least saw the high line saw_the_high_line = False for building in buildings: bounds = shape(building['geometry']).bounds w = bounds[2] - bounds[0] h = bounds[3] - bounds[1] if building['properties']['id'] == -7141751: saw_the_high_line = True if w > max_w or h > max_h: raise Exception("feature %r is %rx%r, larger than the allowed " "%rx%r." % (building['properties']['id'], w, h, max_w, max_h)) if not saw_the_high_line: raise Exception("Expected to see the High Line in this tile, " "but didn't.")
<commit_before>from shapely.geometry import shape # this is mid way along the High Line in NYC, which is a huge long # "building". we should be clipping it to a buffer of 3x the tile # dimensions. # http://www.openstreetmap.org/way/37054313 with features_in_tile_layer(16, 19295, 24631, 'buildings') as buildings: # max width and height in degress as 3x the size of the above tile max_w = 0.0164794921875 max_h = 0.012484410579673977 # need to check that we at least saw the high line saw_the_high_line = False for building in buildings: bounds = shape(building['geometry']).bounds w = bounds[2] - bounds[0] h = bounds[3] - bounds[1] if building['properties']['id'] == 37054313: saw_the_high_line = True if w > max_w or h > max_h: raise Exception("feature %r is %rx%r, larger than the allowed " "%rx%r." % (building['properties']['id'], w, h, max_w, max_h)) if not saw_the_high_line: raise Exception("Expected to see the High Line in this tile, " "but didn't.") <commit_msg>Update data for the High Line, NYC Looks like it was turned into a multipolygon relation in [this changeset](http://www.openstreetmap.org/changeset/47542769).<commit_after>from shapely.geometry import shape # this is mid way along the High Line in NYC, which is a huge long # "building". we should be clipping it to a buffer of 3x the tile # dimensions. # http://www.openstreetmap.org/relation/7141751 with features_in_tile_layer(16, 19295, 24631, 'buildings') as buildings: # max width and height in degress as 3x the size of the above tile max_w = 0.0164794921875 max_h = 0.012484410579673977 # need to check that we at least saw the high line saw_the_high_line = False for building in buildings: bounds = shape(building['geometry']).bounds w = bounds[2] - bounds[0] h = bounds[3] - bounds[1] if building['properties']['id'] == -7141751: saw_the_high_line = True if w > max_w or h > max_h: raise Exception("feature %r is %rx%r, larger than the allowed " "%rx%r." % (building['properties']['id'], w, h, max_w, max_h)) if not saw_the_high_line: raise Exception("Expected to see the High Line in this tile, " "but didn't.")
86446c6d1b0b8583562e0fccf1745e95ce7003c2
util/__init__.py
util/__init__.py
#!/usr/bin/env python # # dials.util.__init__.py # # Copyright (C) 2013 Diamond Light Source # # Author: James Parkhurst # # This code is distributed under the BSD license, a copy of which is # included in the root directory of this package. from __future__ import division class HalError(RuntimeError): def __init__(self, string=''): # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Dave' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that. {1}'.format( username, string) # Init base class RuntimeError.__init__(self, text) def halraiser(e): ''' Function to re-raise an exception with a Hal message. ''' # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Humanoid' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that.'.format(username) # Append to exception if len(e.args) == 0: e.args = (text,) elif len(e.args) == 1: e.args = (text + ' ' + e.args[0],) else: e.args = (text,) + e.args # Reraise the exception raise
#!/usr/bin/env python # # dials.util.__init__.py # # Copyright (C) 2013 Diamond Light Source # # Author: James Parkhurst # # This code is distributed under the BSD license, a copy of which is # included in the root directory of this package. from __future__ import division class HalError(RuntimeError): def __init__(self, string=''): # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Dave' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that. {1}'.format( username, string) # Init base class RuntimeError.__init__(self, text) def halraiser(e): ''' Function to re-raise an exception with a Hal message. ''' import logging # Get the log and write the error to the log file log = logging.getLogger(__name__) log.error(e) # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Humanoid' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that.'.format(username) # Append to exception if len(e.args) == 0: e.args = (text,) elif len(e.args) == 1: e.args = (text + ' ' + e.args[0],) else: e.args = (text,) + e.args # Reraise the exception raise
Print out errors to log.
Print out errors to log.
Python
bsd-3-clause
dials/dials,dials/dials,dials/dials,dials/dials,dials/dials
#!/usr/bin/env python # # dials.util.__init__.py # # Copyright (C) 2013 Diamond Light Source # # Author: James Parkhurst # # This code is distributed under the BSD license, a copy of which is # included in the root directory of this package. from __future__ import division class HalError(RuntimeError): def __init__(self, string=''): # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Dave' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that. {1}'.format( username, string) # Init base class RuntimeError.__init__(self, text) def halraiser(e): ''' Function to re-raise an exception with a Hal message. ''' # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Humanoid' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that.'.format(username) # Append to exception if len(e.args) == 0: e.args = (text,) elif len(e.args) == 1: e.args = (text + ' ' + e.args[0],) else: e.args = (text,) + e.args # Reraise the exception raise Print out errors to log.
#!/usr/bin/env python # # dials.util.__init__.py # # Copyright (C) 2013 Diamond Light Source # # Author: James Parkhurst # # This code is distributed under the BSD license, a copy of which is # included in the root directory of this package. from __future__ import division class HalError(RuntimeError): def __init__(self, string=''): # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Dave' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that. {1}'.format( username, string) # Init base class RuntimeError.__init__(self, text) def halraiser(e): ''' Function to re-raise an exception with a Hal message. ''' import logging # Get the log and write the error to the log file log = logging.getLogger(__name__) log.error(e) # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Humanoid' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that.'.format(username) # Append to exception if len(e.args) == 0: e.args = (text,) elif len(e.args) == 1: e.args = (text + ' ' + e.args[0],) else: e.args = (text,) + e.args # Reraise the exception raise
<commit_before>#!/usr/bin/env python # # dials.util.__init__.py # # Copyright (C) 2013 Diamond Light Source # # Author: James Parkhurst # # This code is distributed under the BSD license, a copy of which is # included in the root directory of this package. from __future__ import division class HalError(RuntimeError): def __init__(self, string=''): # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Dave' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that. {1}'.format( username, string) # Init base class RuntimeError.__init__(self, text) def halraiser(e): ''' Function to re-raise an exception with a Hal message. ''' # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Humanoid' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that.'.format(username) # Append to exception if len(e.args) == 0: e.args = (text,) elif len(e.args) == 1: e.args = (text + ' ' + e.args[0],) else: e.args = (text,) + e.args # Reraise the exception raise <commit_msg>Print out errors to log.<commit_after>
#!/usr/bin/env python # # dials.util.__init__.py # # Copyright (C) 2013 Diamond Light Source # # Author: James Parkhurst # # This code is distributed under the BSD license, a copy of which is # included in the root directory of this package. from __future__ import division class HalError(RuntimeError): def __init__(self, string=''): # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Dave' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that. {1}'.format( username, string) # Init base class RuntimeError.__init__(self, text) def halraiser(e): ''' Function to re-raise an exception with a Hal message. ''' import logging # Get the log and write the error to the log file log = logging.getLogger(__name__) log.error(e) # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Humanoid' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that.'.format(username) # Append to exception if len(e.args) == 0: e.args = (text,) elif len(e.args) == 1: e.args = (text + ' ' + e.args[0],) else: e.args = (text,) + e.args # Reraise the exception raise
#!/usr/bin/env python # # dials.util.__init__.py # # Copyright (C) 2013 Diamond Light Source # # Author: James Parkhurst # # This code is distributed under the BSD license, a copy of which is # included in the root directory of this package. from __future__ import division class HalError(RuntimeError): def __init__(self, string=''): # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Dave' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that. {1}'.format( username, string) # Init base class RuntimeError.__init__(self, text) def halraiser(e): ''' Function to re-raise an exception with a Hal message. ''' # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Humanoid' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that.'.format(username) # Append to exception if len(e.args) == 0: e.args = (text,) elif len(e.args) == 1: e.args = (text + ' ' + e.args[0],) else: e.args = (text,) + e.args # Reraise the exception raise Print out errors to log.#!/usr/bin/env python # # dials.util.__init__.py # # Copyright (C) 2013 Diamond Light Source # # Author: James Parkhurst # # This code is distributed under the BSD license, a copy of which is # included in the root directory of this package. from __future__ import division class HalError(RuntimeError): def __init__(self, string=''): # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Dave' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that. {1}'.format( username, string) # Init base class RuntimeError.__init__(self, text) def halraiser(e): ''' Function to re-raise an exception with a Hal message. ''' import logging # Get the log and write the error to the log file log = logging.getLogger(__name__) log.error(e) # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Humanoid' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that.'.format(username) # Append to exception if len(e.args) == 0: e.args = (text,) elif len(e.args) == 1: e.args = (text + ' ' + e.args[0],) else: e.args = (text,) + e.args # Reraise the exception raise
<commit_before>#!/usr/bin/env python # # dials.util.__init__.py # # Copyright (C) 2013 Diamond Light Source # # Author: James Parkhurst # # This code is distributed under the BSD license, a copy of which is # included in the root directory of this package. from __future__ import division class HalError(RuntimeError): def __init__(self, string=''): # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Dave' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that. {1}'.format( username, string) # Init base class RuntimeError.__init__(self, text) def halraiser(e): ''' Function to re-raise an exception with a Hal message. ''' # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Humanoid' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that.'.format(username) # Append to exception if len(e.args) == 0: e.args = (text,) elif len(e.args) == 1: e.args = (text + ' ' + e.args[0],) else: e.args = (text,) + e.args # Reraise the exception raise <commit_msg>Print out errors to log.<commit_after>#!/usr/bin/env python # # dials.util.__init__.py # # Copyright (C) 2013 Diamond Light Source # # Author: James Parkhurst # # This code is distributed under the BSD license, a copy of which is # included in the root directory of this package. from __future__ import division class HalError(RuntimeError): def __init__(self, string=''): # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Dave' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that. {1}'.format( username, string) # Init base class RuntimeError.__init__(self, text) def halraiser(e): ''' Function to re-raise an exception with a Hal message. ''' import logging # Get the log and write the error to the log file log = logging.getLogger(__name__) log.error(e) # Get the username try: from getpass import getuser username = getuser() except Exception: username = 'Humanoid' # Put in HAL error text. text = 'I\'m sorry {0}. I\'m afraid I can\'t do that.'.format(username) # Append to exception if len(e.args) == 0: e.args = (text,) elif len(e.args) == 1: e.args = (text + ' ' + e.args[0],) else: e.args = (text,) + e.args # Reraise the exception raise
aa3134912af3e57362310eb486d0f4e1d8660d0c
grains/grains.py
grains/grains.py
# File: grains.py # Purpose: Write a program that calculates the number of grains of wheat # on a chessboard given that the number on each square doubles. # Programmer: Amal Shehu # Course: Exercism # Date: Sunday 18 September 2016, 05:25 PM import itertools square = [x for x in range(1, 65)] grains = [2 ** x for x in range(0, 65)] board = dict(zip(square, grains)) def on_square(num): for k, v in board.iteritems(): if k == num: return v def total_after(num): if num == 1: return 1 else: for k, v in board.iteritems(): if k == num: total_after = sum(map(board.get, itertools.takewhile(lambda key: key != v, board))) return total_after print (board) print (total_after(1)) print(on_square(1))
# File: grains.py # Purpose: Write a program that calculates the number of grains of wheat # on a chessboard given that the number on each square doubles. # Programmer: Amal Shehu # Course: Exercism # Date: Sunday 18 September 2016, 05:25 PM square = [x for x in range(1, 65)] grains = [2 ** x for x in range(0, 65)] board = dict(zip(square, grains)) def on_square(num): for k, v in board.iteritems(): if k == num: return v def total_after(num): total = 0 for i in range(1, num+1): total += on_square(i) return total
Reformat total_after function + Remove itertools
Reformat total_after function + Remove itertools
Python
mit
amalshehu/exercism-python
# File: grains.py # Purpose: Write a program that calculates the number of grains of wheat # on a chessboard given that the number on each square doubles. # Programmer: Amal Shehu # Course: Exercism # Date: Sunday 18 September 2016, 05:25 PM import itertools square = [x for x in range(1, 65)] grains = [2 ** x for x in range(0, 65)] board = dict(zip(square, grains)) def on_square(num): for k, v in board.iteritems(): if k == num: return v def total_after(num): if num == 1: return 1 else: for k, v in board.iteritems(): if k == num: total_after = sum(map(board.get, itertools.takewhile(lambda key: key != v, board))) return total_after print (board) print (total_after(1)) print(on_square(1)) Reformat total_after function + Remove itertools
# File: grains.py # Purpose: Write a program that calculates the number of grains of wheat # on a chessboard given that the number on each square doubles. # Programmer: Amal Shehu # Course: Exercism # Date: Sunday 18 September 2016, 05:25 PM square = [x for x in range(1, 65)] grains = [2 ** x for x in range(0, 65)] board = dict(zip(square, grains)) def on_square(num): for k, v in board.iteritems(): if k == num: return v def total_after(num): total = 0 for i in range(1, num+1): total += on_square(i) return total
<commit_before># File: grains.py # Purpose: Write a program that calculates the number of grains of wheat # on a chessboard given that the number on each square doubles. # Programmer: Amal Shehu # Course: Exercism # Date: Sunday 18 September 2016, 05:25 PM import itertools square = [x for x in range(1, 65)] grains = [2 ** x for x in range(0, 65)] board = dict(zip(square, grains)) def on_square(num): for k, v in board.iteritems(): if k == num: return v def total_after(num): if num == 1: return 1 else: for k, v in board.iteritems(): if k == num: total_after = sum(map(board.get, itertools.takewhile(lambda key: key != v, board))) return total_after print (board) print (total_after(1)) print(on_square(1)) <commit_msg>Reformat total_after function + Remove itertools<commit_after>
# File: grains.py # Purpose: Write a program that calculates the number of grains of wheat # on a chessboard given that the number on each square doubles. # Programmer: Amal Shehu # Course: Exercism # Date: Sunday 18 September 2016, 05:25 PM square = [x for x in range(1, 65)] grains = [2 ** x for x in range(0, 65)] board = dict(zip(square, grains)) def on_square(num): for k, v in board.iteritems(): if k == num: return v def total_after(num): total = 0 for i in range(1, num+1): total += on_square(i) return total
# File: grains.py # Purpose: Write a program that calculates the number of grains of wheat # on a chessboard given that the number on each square doubles. # Programmer: Amal Shehu # Course: Exercism # Date: Sunday 18 September 2016, 05:25 PM import itertools square = [x for x in range(1, 65)] grains = [2 ** x for x in range(0, 65)] board = dict(zip(square, grains)) def on_square(num): for k, v in board.iteritems(): if k == num: return v def total_after(num): if num == 1: return 1 else: for k, v in board.iteritems(): if k == num: total_after = sum(map(board.get, itertools.takewhile(lambda key: key != v, board))) return total_after print (board) print (total_after(1)) print(on_square(1)) Reformat total_after function + Remove itertools# File: grains.py # Purpose: Write a program that calculates the number of grains of wheat # on a chessboard given that the number on each square doubles. # Programmer: Amal Shehu # Course: Exercism # Date: Sunday 18 September 2016, 05:25 PM square = [x for x in range(1, 65)] grains = [2 ** x for x in range(0, 65)] board = dict(zip(square, grains)) def on_square(num): for k, v in board.iteritems(): if k == num: return v def total_after(num): total = 0 for i in range(1, num+1): total += on_square(i) return total
<commit_before># File: grains.py # Purpose: Write a program that calculates the number of grains of wheat # on a chessboard given that the number on each square doubles. # Programmer: Amal Shehu # Course: Exercism # Date: Sunday 18 September 2016, 05:25 PM import itertools square = [x for x in range(1, 65)] grains = [2 ** x for x in range(0, 65)] board = dict(zip(square, grains)) def on_square(num): for k, v in board.iteritems(): if k == num: return v def total_after(num): if num == 1: return 1 else: for k, v in board.iteritems(): if k == num: total_after = sum(map(board.get, itertools.takewhile(lambda key: key != v, board))) return total_after print (board) print (total_after(1)) print(on_square(1)) <commit_msg>Reformat total_after function + Remove itertools<commit_after># File: grains.py # Purpose: Write a program that calculates the number of grains of wheat # on a chessboard given that the number on each square doubles. # Programmer: Amal Shehu # Course: Exercism # Date: Sunday 18 September 2016, 05:25 PM square = [x for x in range(1, 65)] grains = [2 ** x for x in range(0, 65)] board = dict(zip(square, grains)) def on_square(num): for k, v in board.iteritems(): if k == num: return v def total_after(num): total = 0 for i in range(1, num+1): total += on_square(i) return total
5188861fb873ea301eb5ec386f1df65c0707e146
openstack/tests/functional/object/v1/test_container.py
openstack/tests/functional/object/v1/test_container.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import uuid from openstack.object_store.v1 import container from openstack.tests.functional import base class TestContainer(base.BaseFunctionalTest): NAME = uuid.uuid4().hex @classmethod def setUpClass(cls): super(TestContainer, cls).setUpClass() tainer = cls.conn.object_store.create_container(name=cls.NAME) assert isinstance(tainer, container.Container) cls.assertIs(cls.NAME, tainer.name) @classmethod def tearDownClass(cls): pass # TODO(thowe): uncomment this when bug/1451211 fixed # tainer = cls.conn.object_store.delete_container(cls.NAME, # ignore_missing=False) # cls.assertIs(None, tainer) def test_list(self): names = [o.name for o in self.conn.object_store.containers()] self.assertIn(self.NAME, names)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import uuid from openstack.object_store.v1 import container from openstack.tests.functional import base class TestContainer(base.BaseFunctionalTest): NAME = uuid.uuid4().hex @classmethod def setUpClass(cls): super(TestContainer, cls).setUpClass() tainer = cls.conn.object_store.create_container(name=cls.NAME) assert isinstance(tainer, container.Container) cls.assertIs(cls.NAME, tainer.name) @classmethod def tearDownClass(cls): result = cls.conn.object_store.delete_container(cls.NAME, ignore_missing=False) cls.assertIs(None, result) def test_list(self): names = [o.name for o in self.conn.object_store.containers()] self.assertIn(self.NAME, names) def test_get_metadata(self): tainer = self.conn.object_store.get_container_metadata(self.NAME) self.assertEqual(0, tainer.object_count) self.assertEqual(0, tainer.bytes_used) self.assertEqual(self.NAME, tainer.name)
Add functional tests for container metadata and delete
Add functional tests for container metadata and delete Change-Id: Id13073b37d19482ca6ff6e53e315aaa424c3f1b9
Python
apache-2.0
openstack/python-openstacksdk,dtroyer/python-openstacksdk,mtougeron/python-openstacksdk,briancurtin/python-openstacksdk,dudymas/python-openstacksdk,briancurtin/python-openstacksdk,dudymas/python-openstacksdk,openstack/python-openstacksdk,stackforge/python-openstacksdk,dtroyer/python-openstacksdk,mtougeron/python-openstacksdk,stackforge/python-openstacksdk
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import uuid from openstack.object_store.v1 import container from openstack.tests.functional import base class TestContainer(base.BaseFunctionalTest): NAME = uuid.uuid4().hex @classmethod def setUpClass(cls): super(TestContainer, cls).setUpClass() tainer = cls.conn.object_store.create_container(name=cls.NAME) assert isinstance(tainer, container.Container) cls.assertIs(cls.NAME, tainer.name) @classmethod def tearDownClass(cls): pass # TODO(thowe): uncomment this when bug/1451211 fixed # tainer = cls.conn.object_store.delete_container(cls.NAME, # ignore_missing=False) # cls.assertIs(None, tainer) def test_list(self): names = [o.name for o in self.conn.object_store.containers()] self.assertIn(self.NAME, names) Add functional tests for container metadata and delete Change-Id: Id13073b37d19482ca6ff6e53e315aaa424c3f1b9
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import uuid from openstack.object_store.v1 import container from openstack.tests.functional import base class TestContainer(base.BaseFunctionalTest): NAME = uuid.uuid4().hex @classmethod def setUpClass(cls): super(TestContainer, cls).setUpClass() tainer = cls.conn.object_store.create_container(name=cls.NAME) assert isinstance(tainer, container.Container) cls.assertIs(cls.NAME, tainer.name) @classmethod def tearDownClass(cls): result = cls.conn.object_store.delete_container(cls.NAME, ignore_missing=False) cls.assertIs(None, result) def test_list(self): names = [o.name for o in self.conn.object_store.containers()] self.assertIn(self.NAME, names) def test_get_metadata(self): tainer = self.conn.object_store.get_container_metadata(self.NAME) self.assertEqual(0, tainer.object_count) self.assertEqual(0, tainer.bytes_used) self.assertEqual(self.NAME, tainer.name)
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import uuid from openstack.object_store.v1 import container from openstack.tests.functional import base class TestContainer(base.BaseFunctionalTest): NAME = uuid.uuid4().hex @classmethod def setUpClass(cls): super(TestContainer, cls).setUpClass() tainer = cls.conn.object_store.create_container(name=cls.NAME) assert isinstance(tainer, container.Container) cls.assertIs(cls.NAME, tainer.name) @classmethod def tearDownClass(cls): pass # TODO(thowe): uncomment this when bug/1451211 fixed # tainer = cls.conn.object_store.delete_container(cls.NAME, # ignore_missing=False) # cls.assertIs(None, tainer) def test_list(self): names = [o.name for o in self.conn.object_store.containers()] self.assertIn(self.NAME, names) <commit_msg>Add functional tests for container metadata and delete Change-Id: Id13073b37d19482ca6ff6e53e315aaa424c3f1b9<commit_after>
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import uuid from openstack.object_store.v1 import container from openstack.tests.functional import base class TestContainer(base.BaseFunctionalTest): NAME = uuid.uuid4().hex @classmethod def setUpClass(cls): super(TestContainer, cls).setUpClass() tainer = cls.conn.object_store.create_container(name=cls.NAME) assert isinstance(tainer, container.Container) cls.assertIs(cls.NAME, tainer.name) @classmethod def tearDownClass(cls): result = cls.conn.object_store.delete_container(cls.NAME, ignore_missing=False) cls.assertIs(None, result) def test_list(self): names = [o.name for o in self.conn.object_store.containers()] self.assertIn(self.NAME, names) def test_get_metadata(self): tainer = self.conn.object_store.get_container_metadata(self.NAME) self.assertEqual(0, tainer.object_count) self.assertEqual(0, tainer.bytes_used) self.assertEqual(self.NAME, tainer.name)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import uuid from openstack.object_store.v1 import container from openstack.tests.functional import base class TestContainer(base.BaseFunctionalTest): NAME = uuid.uuid4().hex @classmethod def setUpClass(cls): super(TestContainer, cls).setUpClass() tainer = cls.conn.object_store.create_container(name=cls.NAME) assert isinstance(tainer, container.Container) cls.assertIs(cls.NAME, tainer.name) @classmethod def tearDownClass(cls): pass # TODO(thowe): uncomment this when bug/1451211 fixed # tainer = cls.conn.object_store.delete_container(cls.NAME, # ignore_missing=False) # cls.assertIs(None, tainer) def test_list(self): names = [o.name for o in self.conn.object_store.containers()] self.assertIn(self.NAME, names) Add functional tests for container metadata and delete Change-Id: Id13073b37d19482ca6ff6e53e315aaa424c3f1b9# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import uuid from openstack.object_store.v1 import container from openstack.tests.functional import base class TestContainer(base.BaseFunctionalTest): NAME = uuid.uuid4().hex @classmethod def setUpClass(cls): super(TestContainer, cls).setUpClass() tainer = cls.conn.object_store.create_container(name=cls.NAME) assert isinstance(tainer, container.Container) cls.assertIs(cls.NAME, tainer.name) @classmethod def tearDownClass(cls): result = cls.conn.object_store.delete_container(cls.NAME, ignore_missing=False) cls.assertIs(None, result) def test_list(self): names = [o.name for o in self.conn.object_store.containers()] self.assertIn(self.NAME, names) def test_get_metadata(self): tainer = self.conn.object_store.get_container_metadata(self.NAME) self.assertEqual(0, tainer.object_count) self.assertEqual(0, tainer.bytes_used) self.assertEqual(self.NAME, tainer.name)
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import uuid from openstack.object_store.v1 import container from openstack.tests.functional import base class TestContainer(base.BaseFunctionalTest): NAME = uuid.uuid4().hex @classmethod def setUpClass(cls): super(TestContainer, cls).setUpClass() tainer = cls.conn.object_store.create_container(name=cls.NAME) assert isinstance(tainer, container.Container) cls.assertIs(cls.NAME, tainer.name) @classmethod def tearDownClass(cls): pass # TODO(thowe): uncomment this when bug/1451211 fixed # tainer = cls.conn.object_store.delete_container(cls.NAME, # ignore_missing=False) # cls.assertIs(None, tainer) def test_list(self): names = [o.name for o in self.conn.object_store.containers()] self.assertIn(self.NAME, names) <commit_msg>Add functional tests for container metadata and delete Change-Id: Id13073b37d19482ca6ff6e53e315aaa424c3f1b9<commit_after># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import uuid from openstack.object_store.v1 import container from openstack.tests.functional import base class TestContainer(base.BaseFunctionalTest): NAME = uuid.uuid4().hex @classmethod def setUpClass(cls): super(TestContainer, cls).setUpClass() tainer = cls.conn.object_store.create_container(name=cls.NAME) assert isinstance(tainer, container.Container) cls.assertIs(cls.NAME, tainer.name) @classmethod def tearDownClass(cls): result = cls.conn.object_store.delete_container(cls.NAME, ignore_missing=False) cls.assertIs(None, result) def test_list(self): names = [o.name for o in self.conn.object_store.containers()] self.assertIn(self.NAME, names) def test_get_metadata(self): tainer = self.conn.object_store.get_container_metadata(self.NAME) self.assertEqual(0, tainer.object_count) self.assertEqual(0, tainer.bytes_used) self.assertEqual(self.NAME, tainer.name)
6b84688c1b5a7f2e8c9e5007455b88cbaa845e9f
tests/test_track_output/results.py
tests/test_track_output/results.py
#!/usr/bin/env python import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../track.py', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
#!/usr/bin/env python import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
Fix path to script in test_track_output
Fix path to script in test_track_output
Python
mit
mjlong/openmc,wbinventor/openmc,bhermanmit/openmc,wbinventor/openmc,johnnyliu27/openmc,paulromano/openmc,mjlong/openmc,smharper/openmc,lilulu/openmc,johnnyliu27/openmc,amandalund/openmc,samuelshaner/openmc,mit-crpg/openmc,shikhar413/openmc,johnnyliu27/openmc,mit-crpg/openmc,liangjg/openmc,kellyrowland/openmc,amandalund/openmc,walshjon/openmc,lilulu/openmc,liangjg/openmc,liangjg/openmc,johnnyliu27/openmc,samuelshaner/openmc,mit-crpg/openmc,amandalund/openmc,smharper/openmc,shikhar413/openmc,samuelshaner/openmc,mit-crpg/openmc,shikhar413/openmc,paulromano/openmc,liangjg/openmc,smharper/openmc,shikhar413/openmc,samuelshaner/openmc,paulromano/openmc,walshjon/openmc,walshjon/openmc,smharper/openmc,bhermanmit/openmc,wbinventor/openmc,wbinventor/openmc,walshjon/openmc,paulromano/openmc,lilulu/openmc,amandalund/openmc,kellyrowland/openmc
#!/usr/bin/env python import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../track.py', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat') Fix path to script in test_track_output
#!/usr/bin/env python import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
<commit_before>#!/usr/bin/env python import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../track.py', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat') <commit_msg>Fix path to script in test_track_output<commit_after>
#!/usr/bin/env python import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
#!/usr/bin/env python import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../track.py', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat') Fix path to script in test_track_output#!/usr/bin/env python import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
<commit_before>#!/usr/bin/env python import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../track.py', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat') <commit_msg>Fix path to script in test_track_output<commit_after>#!/usr/bin/env python import os import sys import glob import shutil from subprocess import call # If vtk python module is not available, we can't run track.py so skip this # test cwd = os.getcwd() try: import vtk except ImportError: print('----------------Skipping test-------------') shutil.copy('results_true.dat', 'results_test.dat') exit() # Run track processing script call(['../../scripts/openmc-track-to-vtk', '-o', 'poly'] + glob.glob(''.join((cwd, '/track*')))) poly = ''.join((cwd, '/poly.pvtp')) assert os.path.isfile(poly), 'poly.pvtp file not found.' shutil.copy('poly.pvtp', 'results_test.dat')
85cb348dab356386362fe7657dee3e31aa1f92bf
rep.py
rep.py
""" The top level of the APL Read-Evaluate-Print loop UNDER DEVELOPMENT This version adds a dummy evaluate routine. """ import sys # ------------------------------ def evaluate(expression): """ Evaluate an APL expression - dummy version """ return (expression) def read_evaluate_print (prompt): """ Read input, echo input """ try: while True: print(end=prompt) line = input().lstrip() if line: if line[0] == ')': if line[0:4].upper() == ')OFF': apl_exit("Bye bye") print('⎕', evaluate(line)) except EOFError: apl_exit(None) def apl_quit (): """ Quit without clean up """ print () sys.exit(0) def apl_exit (message): """ Clean up and quit """ if message is None: print () else: print (message) sys.exit(0) # EOF
""" The top level of the APL Read-Evaluate-Print loop UNDER DEVELOPMENT This version adds simple APL exception handling """ import sys from apl_exception import APL_Exception as apl_exception # ------------------------------ def evaluate(expression): """ Evaluate an APL expression - dummy version """ return (expression) def read_evaluate_print (prompt): """ Read input, echo input """ try: while True: print(end=prompt) line = input().lstrip() if line: if line[0] == ')': if line[0:4].upper() == ')OFF': apl_exit("Bye bye") try: result = evaluate(line) except apl_exception as e: print(' '*(len(prompt)+len(line)-len(e.line)),end="^\n") result = e.message finally: print('⎕', result) except EOFError: apl_exit(None) def apl_quit (): """ Quit without clean up """ print() sys.exit(0) def apl_exit (message): """ Clean up and quit """ if message is None: print() else: print(message) sys.exit(0) # EOF
Add exception handling to the read-evaluate-loop
Add exception handling to the read-evaluate-loop
Python
apache-2.0
NewForester/apl-py,NewForester/apl-py
""" The top level of the APL Read-Evaluate-Print loop UNDER DEVELOPMENT This version adds a dummy evaluate routine. """ import sys # ------------------------------ def evaluate(expression): """ Evaluate an APL expression - dummy version """ return (expression) def read_evaluate_print (prompt): """ Read input, echo input """ try: while True: print(end=prompt) line = input().lstrip() if line: if line[0] == ')': if line[0:4].upper() == ')OFF': apl_exit("Bye bye") print('⎕', evaluate(line)) except EOFError: apl_exit(None) def apl_quit (): """ Quit without clean up """ print () sys.exit(0) def apl_exit (message): """ Clean up and quit """ if message is None: print () else: print (message) sys.exit(0) # EOF Add exception handling to the read-evaluate-loop
""" The top level of the APL Read-Evaluate-Print loop UNDER DEVELOPMENT This version adds simple APL exception handling """ import sys from apl_exception import APL_Exception as apl_exception # ------------------------------ def evaluate(expression): """ Evaluate an APL expression - dummy version """ return (expression) def read_evaluate_print (prompt): """ Read input, echo input """ try: while True: print(end=prompt) line = input().lstrip() if line: if line[0] == ')': if line[0:4].upper() == ')OFF': apl_exit("Bye bye") try: result = evaluate(line) except apl_exception as e: print(' '*(len(prompt)+len(line)-len(e.line)),end="^\n") result = e.message finally: print('⎕', result) except EOFError: apl_exit(None) def apl_quit (): """ Quit without clean up """ print() sys.exit(0) def apl_exit (message): """ Clean up and quit """ if message is None: print() else: print(message) sys.exit(0) # EOF
<commit_before>""" The top level of the APL Read-Evaluate-Print loop UNDER DEVELOPMENT This version adds a dummy evaluate routine. """ import sys # ------------------------------ def evaluate(expression): """ Evaluate an APL expression - dummy version """ return (expression) def read_evaluate_print (prompt): """ Read input, echo input """ try: while True: print(end=prompt) line = input().lstrip() if line: if line[0] == ')': if line[0:4].upper() == ')OFF': apl_exit("Bye bye") print('⎕', evaluate(line)) except EOFError: apl_exit(None) def apl_quit (): """ Quit without clean up """ print () sys.exit(0) def apl_exit (message): """ Clean up and quit """ if message is None: print () else: print (message) sys.exit(0) # EOF <commit_msg>Add exception handling to the read-evaluate-loop<commit_after>
""" The top level of the APL Read-Evaluate-Print loop UNDER DEVELOPMENT This version adds simple APL exception handling """ import sys from apl_exception import APL_Exception as apl_exception # ------------------------------ def evaluate(expression): """ Evaluate an APL expression - dummy version """ return (expression) def read_evaluate_print (prompt): """ Read input, echo input """ try: while True: print(end=prompt) line = input().lstrip() if line: if line[0] == ')': if line[0:4].upper() == ')OFF': apl_exit("Bye bye") try: result = evaluate(line) except apl_exception as e: print(' '*(len(prompt)+len(line)-len(e.line)),end="^\n") result = e.message finally: print('⎕', result) except EOFError: apl_exit(None) def apl_quit (): """ Quit without clean up """ print() sys.exit(0) def apl_exit (message): """ Clean up and quit """ if message is None: print() else: print(message) sys.exit(0) # EOF
""" The top level of the APL Read-Evaluate-Print loop UNDER DEVELOPMENT This version adds a dummy evaluate routine. """ import sys # ------------------------------ def evaluate(expression): """ Evaluate an APL expression - dummy version """ return (expression) def read_evaluate_print (prompt): """ Read input, echo input """ try: while True: print(end=prompt) line = input().lstrip() if line: if line[0] == ')': if line[0:4].upper() == ')OFF': apl_exit("Bye bye") print('⎕', evaluate(line)) except EOFError: apl_exit(None) def apl_quit (): """ Quit without clean up """ print () sys.exit(0) def apl_exit (message): """ Clean up and quit """ if message is None: print () else: print (message) sys.exit(0) # EOF Add exception handling to the read-evaluate-loop""" The top level of the APL Read-Evaluate-Print loop UNDER DEVELOPMENT This version adds simple APL exception handling """ import sys from apl_exception import APL_Exception as apl_exception # ------------------------------ def evaluate(expression): """ Evaluate an APL expression - dummy version """ return (expression) def read_evaluate_print (prompt): """ Read input, echo input """ try: while True: print(end=prompt) line = input().lstrip() if line: if line[0] == ')': if line[0:4].upper() == ')OFF': apl_exit("Bye bye") try: result = evaluate(line) except apl_exception as e: print(' '*(len(prompt)+len(line)-len(e.line)),end="^\n") result = e.message finally: print('⎕', result) except EOFError: apl_exit(None) def apl_quit (): """ Quit without clean up """ print() sys.exit(0) def apl_exit (message): """ Clean up and quit """ if message is None: print() else: print(message) sys.exit(0) # EOF
<commit_before>""" The top level of the APL Read-Evaluate-Print loop UNDER DEVELOPMENT This version adds a dummy evaluate routine. """ import sys # ------------------------------ def evaluate(expression): """ Evaluate an APL expression - dummy version """ return (expression) def read_evaluate_print (prompt): """ Read input, echo input """ try: while True: print(end=prompt) line = input().lstrip() if line: if line[0] == ')': if line[0:4].upper() == ')OFF': apl_exit("Bye bye") print('⎕', evaluate(line)) except EOFError: apl_exit(None) def apl_quit (): """ Quit without clean up """ print () sys.exit(0) def apl_exit (message): """ Clean up and quit """ if message is None: print () else: print (message) sys.exit(0) # EOF <commit_msg>Add exception handling to the read-evaluate-loop<commit_after>""" The top level of the APL Read-Evaluate-Print loop UNDER DEVELOPMENT This version adds simple APL exception handling """ import sys from apl_exception import APL_Exception as apl_exception # ------------------------------ def evaluate(expression): """ Evaluate an APL expression - dummy version """ return (expression) def read_evaluate_print (prompt): """ Read input, echo input """ try: while True: print(end=prompt) line = input().lstrip() if line: if line[0] == ')': if line[0:4].upper() == ')OFF': apl_exit("Bye bye") try: result = evaluate(line) except apl_exception as e: print(' '*(len(prompt)+len(line)-len(e.line)),end="^\n") result = e.message finally: print('⎕', result) except EOFError: apl_exit(None) def apl_quit (): """ Quit without clean up """ print() sys.exit(0) def apl_exit (message): """ Clean up and quit """ if message is None: print() else: print(message) sys.exit(0) # EOF
574fba0650e4c68b7a31533207c26d8d57ed49c2
run.py
run.py
from sys import argv from examples.connect4.connect4 import Connect4Network if __name__ == '__main__': if argv[1] == 'connect4': Connect4Network().play()
from sys import argv from examples.connect4.connect4 import Connect4Network if __name__ == '__main__': if len(argv) > 1: if argv[1] == 'connect4': Connect4Network().play()
Check length of argv before accessing it
Check length of argv before accessing it
Python
mit
tysonzero/py-ann
from sys import argv from examples.connect4.connect4 import Connect4Network if __name__ == '__main__': if argv[1] == 'connect4': Connect4Network().play() Check length of argv before accessing it
from sys import argv from examples.connect4.connect4 import Connect4Network if __name__ == '__main__': if len(argv) > 1: if argv[1] == 'connect4': Connect4Network().play()
<commit_before>from sys import argv from examples.connect4.connect4 import Connect4Network if __name__ == '__main__': if argv[1] == 'connect4': Connect4Network().play() <commit_msg>Check length of argv before accessing it<commit_after>
from sys import argv from examples.connect4.connect4 import Connect4Network if __name__ == '__main__': if len(argv) > 1: if argv[1] == 'connect4': Connect4Network().play()
from sys import argv from examples.connect4.connect4 import Connect4Network if __name__ == '__main__': if argv[1] == 'connect4': Connect4Network().play() Check length of argv before accessing itfrom sys import argv from examples.connect4.connect4 import Connect4Network if __name__ == '__main__': if len(argv) > 1: if argv[1] == 'connect4': Connect4Network().play()
<commit_before>from sys import argv from examples.connect4.connect4 import Connect4Network if __name__ == '__main__': if argv[1] == 'connect4': Connect4Network().play() <commit_msg>Check length of argv before accessing it<commit_after>from sys import argv from examples.connect4.connect4 import Connect4Network if __name__ == '__main__': if len(argv) > 1: if argv[1] == 'connect4': Connect4Network().play()
a7867806a6bd3abfd6bf2bcac6c490965be000e2
tests/test_completeness.py
tests/test_completeness.py
import unittest as unittest from syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()
import unittest as unittest from jscodegen.syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()
Fix an issue in the tests
Fix an issue in the tests
Python
mit
ksons/jscodegen.py
import unittest as unittest from syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()Fix an issue in the tests
import unittest as unittest from jscodegen.syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()
<commit_before>import unittest as unittest from syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()<commit_msg>Fix an issue in the tests<commit_after>
import unittest as unittest from jscodegen.syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()
import unittest as unittest from syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()Fix an issue in the testsimport unittest as unittest from jscodegen.syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()
<commit_before>import unittest as unittest from syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()<commit_msg>Fix an issue in the tests<commit_after>import unittest as unittest from jscodegen.syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()
eb0a5e5768c2f699cbc1f6395d3c14320d8e730d
template/__init__.py
template/__init__.py
#!/usr/bin/env python # pylint: disable=import-error """Generate files from Jinja2 templates and environment variables.""" from __future__ import ( absolute_import, division, print_function, unicode_literals, ) # pylint: disable=duplicate-code from os import environ from sys import stdin, stdout import argparse from argparse import ArgumentParser from jinja2 import Environment import template.filters __version__ = "0.6.4" def render(template_string): """Render the template.""" env = Environment(autoescape=True) # Add all functions in template.filters as Jinja filters. # pylint: disable=invalid-name for tf in filter(lambda x: not x.startswith("_"), dir(template.filters)): env.filters[tf] = getattr(template.filters, tf) t = env.from_string(template_string) return t.render(environ) def main(): """Main entrypoint.""" parser = ArgumentParser( description="""A CLI tool for generating files from Jinja2 templates and environment variables.""" ) parser.add_argument( "filename", help="Input filename", type=argparse.FileType("r"), nargs="?", ) parser.add_argument( "-o", "--output", help="Output to filename", type=argparse.FileType("w"), ) args = parser.parse_args() infd = args.filename if args.filename else stdin outfd = args.output if args.output else stdout print(render(infd.read()), file=outfd) if __name__ == "__main__": main()
#!/usr/bin/env python # pylint: disable=import-error """A CLI tool for generating files from Jinja2 templates and environment variables.""" from __future__ import ( absolute_import, division, print_function, unicode_literals, ) # pylint: disable=duplicate-code from os import environ from sys import stdin, stdout import argparse from argparse import ArgumentParser from jinja2 import Environment import template.filters __version__ = "0.6.4" def render(template_string): """Render the template.""" env = Environment(autoescape=True) # Add all functions in template.filters as Jinja filters. # pylint: disable=invalid-name for tf in filter(lambda x: not x.startswith("_"), dir(template.filters)): env.filters[tf] = getattr(template.filters, tf) t = env.from_string(template_string) return t.render(environ) def main(): """Main entrypoint.""" parser = ArgumentParser(description=__doc__) parser.add_argument( "filename", help="Input filename", type=argparse.FileType("r"), nargs="?", ) parser.add_argument( "-o", "--output", help="Output to filename", type=argparse.FileType("w"), ) args = parser.parse_args() infd = args.filename if args.filename else stdin outfd = args.output if args.output else stdout print(render(infd.read()), file=outfd) if __name__ == "__main__": main()
Use the docstring as the CLI description.
Use the docstring as the CLI description.
Python
agpl-3.0
adarnimrod/template,adarnimrod/template
#!/usr/bin/env python # pylint: disable=import-error """Generate files from Jinja2 templates and environment variables.""" from __future__ import ( absolute_import, division, print_function, unicode_literals, ) # pylint: disable=duplicate-code from os import environ from sys import stdin, stdout import argparse from argparse import ArgumentParser from jinja2 import Environment import template.filters __version__ = "0.6.4" def render(template_string): """Render the template.""" env = Environment(autoescape=True) # Add all functions in template.filters as Jinja filters. # pylint: disable=invalid-name for tf in filter(lambda x: not x.startswith("_"), dir(template.filters)): env.filters[tf] = getattr(template.filters, tf) t = env.from_string(template_string) return t.render(environ) def main(): """Main entrypoint.""" parser = ArgumentParser( description="""A CLI tool for generating files from Jinja2 templates and environment variables.""" ) parser.add_argument( "filename", help="Input filename", type=argparse.FileType("r"), nargs="?", ) parser.add_argument( "-o", "--output", help="Output to filename", type=argparse.FileType("w"), ) args = parser.parse_args() infd = args.filename if args.filename else stdin outfd = args.output if args.output else stdout print(render(infd.read()), file=outfd) if __name__ == "__main__": main() Use the docstring as the CLI description.
#!/usr/bin/env python # pylint: disable=import-error """A CLI tool for generating files from Jinja2 templates and environment variables.""" from __future__ import ( absolute_import, division, print_function, unicode_literals, ) # pylint: disable=duplicate-code from os import environ from sys import stdin, stdout import argparse from argparse import ArgumentParser from jinja2 import Environment import template.filters __version__ = "0.6.4" def render(template_string): """Render the template.""" env = Environment(autoescape=True) # Add all functions in template.filters as Jinja filters. # pylint: disable=invalid-name for tf in filter(lambda x: not x.startswith("_"), dir(template.filters)): env.filters[tf] = getattr(template.filters, tf) t = env.from_string(template_string) return t.render(environ) def main(): """Main entrypoint.""" parser = ArgumentParser(description=__doc__) parser.add_argument( "filename", help="Input filename", type=argparse.FileType("r"), nargs="?", ) parser.add_argument( "-o", "--output", help="Output to filename", type=argparse.FileType("w"), ) args = parser.parse_args() infd = args.filename if args.filename else stdin outfd = args.output if args.output else stdout print(render(infd.read()), file=outfd) if __name__ == "__main__": main()
<commit_before>#!/usr/bin/env python # pylint: disable=import-error """Generate files from Jinja2 templates and environment variables.""" from __future__ import ( absolute_import, division, print_function, unicode_literals, ) # pylint: disable=duplicate-code from os import environ from sys import stdin, stdout import argparse from argparse import ArgumentParser from jinja2 import Environment import template.filters __version__ = "0.6.4" def render(template_string): """Render the template.""" env = Environment(autoescape=True) # Add all functions in template.filters as Jinja filters. # pylint: disable=invalid-name for tf in filter(lambda x: not x.startswith("_"), dir(template.filters)): env.filters[tf] = getattr(template.filters, tf) t = env.from_string(template_string) return t.render(environ) def main(): """Main entrypoint.""" parser = ArgumentParser( description="""A CLI tool for generating files from Jinja2 templates and environment variables.""" ) parser.add_argument( "filename", help="Input filename", type=argparse.FileType("r"), nargs="?", ) parser.add_argument( "-o", "--output", help="Output to filename", type=argparse.FileType("w"), ) args = parser.parse_args() infd = args.filename if args.filename else stdin outfd = args.output if args.output else stdout print(render(infd.read()), file=outfd) if __name__ == "__main__": main() <commit_msg>Use the docstring as the CLI description.<commit_after>
#!/usr/bin/env python # pylint: disable=import-error """A CLI tool for generating files from Jinja2 templates and environment variables.""" from __future__ import ( absolute_import, division, print_function, unicode_literals, ) # pylint: disable=duplicate-code from os import environ from sys import stdin, stdout import argparse from argparse import ArgumentParser from jinja2 import Environment import template.filters __version__ = "0.6.4" def render(template_string): """Render the template.""" env = Environment(autoescape=True) # Add all functions in template.filters as Jinja filters. # pylint: disable=invalid-name for tf in filter(lambda x: not x.startswith("_"), dir(template.filters)): env.filters[tf] = getattr(template.filters, tf) t = env.from_string(template_string) return t.render(environ) def main(): """Main entrypoint.""" parser = ArgumentParser(description=__doc__) parser.add_argument( "filename", help="Input filename", type=argparse.FileType("r"), nargs="?", ) parser.add_argument( "-o", "--output", help="Output to filename", type=argparse.FileType("w"), ) args = parser.parse_args() infd = args.filename if args.filename else stdin outfd = args.output if args.output else stdout print(render(infd.read()), file=outfd) if __name__ == "__main__": main()
#!/usr/bin/env python # pylint: disable=import-error """Generate files from Jinja2 templates and environment variables.""" from __future__ import ( absolute_import, division, print_function, unicode_literals, ) # pylint: disable=duplicate-code from os import environ from sys import stdin, stdout import argparse from argparse import ArgumentParser from jinja2 import Environment import template.filters __version__ = "0.6.4" def render(template_string): """Render the template.""" env = Environment(autoescape=True) # Add all functions in template.filters as Jinja filters. # pylint: disable=invalid-name for tf in filter(lambda x: not x.startswith("_"), dir(template.filters)): env.filters[tf] = getattr(template.filters, tf) t = env.from_string(template_string) return t.render(environ) def main(): """Main entrypoint.""" parser = ArgumentParser( description="""A CLI tool for generating files from Jinja2 templates and environment variables.""" ) parser.add_argument( "filename", help="Input filename", type=argparse.FileType("r"), nargs="?", ) parser.add_argument( "-o", "--output", help="Output to filename", type=argparse.FileType("w"), ) args = parser.parse_args() infd = args.filename if args.filename else stdin outfd = args.output if args.output else stdout print(render(infd.read()), file=outfd) if __name__ == "__main__": main() Use the docstring as the CLI description.#!/usr/bin/env python # pylint: disable=import-error """A CLI tool for generating files from Jinja2 templates and environment variables.""" from __future__ import ( absolute_import, division, print_function, unicode_literals, ) # pylint: disable=duplicate-code from os import environ from sys import stdin, stdout import argparse from argparse import ArgumentParser from jinja2 import Environment import template.filters __version__ = "0.6.4" def render(template_string): """Render the template.""" env = Environment(autoescape=True) # Add all functions in template.filters as Jinja filters. # pylint: disable=invalid-name for tf in filter(lambda x: not x.startswith("_"), dir(template.filters)): env.filters[tf] = getattr(template.filters, tf) t = env.from_string(template_string) return t.render(environ) def main(): """Main entrypoint.""" parser = ArgumentParser(description=__doc__) parser.add_argument( "filename", help="Input filename", type=argparse.FileType("r"), nargs="?", ) parser.add_argument( "-o", "--output", help="Output to filename", type=argparse.FileType("w"), ) args = parser.parse_args() infd = args.filename if args.filename else stdin outfd = args.output if args.output else stdout print(render(infd.read()), file=outfd) if __name__ == "__main__": main()
<commit_before>#!/usr/bin/env python # pylint: disable=import-error """Generate files from Jinja2 templates and environment variables.""" from __future__ import ( absolute_import, division, print_function, unicode_literals, ) # pylint: disable=duplicate-code from os import environ from sys import stdin, stdout import argparse from argparse import ArgumentParser from jinja2 import Environment import template.filters __version__ = "0.6.4" def render(template_string): """Render the template.""" env = Environment(autoescape=True) # Add all functions in template.filters as Jinja filters. # pylint: disable=invalid-name for tf in filter(lambda x: not x.startswith("_"), dir(template.filters)): env.filters[tf] = getattr(template.filters, tf) t = env.from_string(template_string) return t.render(environ) def main(): """Main entrypoint.""" parser = ArgumentParser( description="""A CLI tool for generating files from Jinja2 templates and environment variables.""" ) parser.add_argument( "filename", help="Input filename", type=argparse.FileType("r"), nargs="?", ) parser.add_argument( "-o", "--output", help="Output to filename", type=argparse.FileType("w"), ) args = parser.parse_args() infd = args.filename if args.filename else stdin outfd = args.output if args.output else stdout print(render(infd.read()), file=outfd) if __name__ == "__main__": main() <commit_msg>Use the docstring as the CLI description.<commit_after>#!/usr/bin/env python # pylint: disable=import-error """A CLI tool for generating files from Jinja2 templates and environment variables.""" from __future__ import ( absolute_import, division, print_function, unicode_literals, ) # pylint: disable=duplicate-code from os import environ from sys import stdin, stdout import argparse from argparse import ArgumentParser from jinja2 import Environment import template.filters __version__ = "0.6.4" def render(template_string): """Render the template.""" env = Environment(autoescape=True) # Add all functions in template.filters as Jinja filters. # pylint: disable=invalid-name for tf in filter(lambda x: not x.startswith("_"), dir(template.filters)): env.filters[tf] = getattr(template.filters, tf) t = env.from_string(template_string) return t.render(environ) def main(): """Main entrypoint.""" parser = ArgumentParser(description=__doc__) parser.add_argument( "filename", help="Input filename", type=argparse.FileType("r"), nargs="?", ) parser.add_argument( "-o", "--output", help="Output to filename", type=argparse.FileType("w"), ) args = parser.parse_args() infd = args.filename if args.filename else stdin outfd = args.output if args.output else stdout print(render(infd.read()), file=outfd) if __name__ == "__main__": main()
3aa7dd8d1247fe9d426049c8b09db0a5b1af9067
prediction/test_tuner.py
prediction/test_tuner.py
from .tuner import Hyperband import unittest class HyperbandTestCase(unittest.TestCase): def test_run(self): observed_ns = [] observed_rs = [] observed_cs = [] def _get(n): observed_ns.append(n) return list(range(n)) def _test(r, c): observed_rs.append(int(r)) observed_cs.append(len(c)) return [r * c for c in c] tuner = Hyperband() tuner.run(_get, _test) expected_ns = [81, 34, 15, 8, 5] expected_rs = [1, 3, 9, 27, 81, 3, 9, 27, 81, 9, 27, 81, 27, 81, 81] expected_cs = [81, 27, 9, 3, 1, 27, 9, 3, 1, 9, 3, 1, 6, 2, 5] self.assertEqual(expected_ns, observed_ns) self.assertEqual(expected_rs, observed_rs) self.assertEqual(expected_rs, observed_rs)
from .tuner import Hyperband import unittest class HyperbandTestCase(unittest.TestCase): def test_run(self): observed_ns = [] observed_rs = [] observed_cs = [] def _get(n): observed_ns.append(n) return list(range(n)) def _test(r, c): observed_rs.append(int(r)) observed_cs.append(len(c)) return [r * c for c in c] tuner = Hyperband() tuner.run(_get, _test) expected_ns = [81, 34, 15, 8, 5] expected_rs = [1, 3, 9, 27, 81, 3, 9, 27, 81, 9, 27, 81, 27, 81, 81] expected_cs = [81, 27, 9, 3, 1, 27, 9, 3, 1, 9, 3, 1, 6, 2, 5] self.assertEqual(expected_ns, observed_ns) self.assertEqual(expected_rs, observed_rs) self.assertEqual(expected_cs, observed_cs)
Fix a typo in Hyperband’s test
Fix a typo in Hyperband’s test
Python
mit
learning-on-chip/google-cluster-prediction
from .tuner import Hyperband import unittest class HyperbandTestCase(unittest.TestCase): def test_run(self): observed_ns = [] observed_rs = [] observed_cs = [] def _get(n): observed_ns.append(n) return list(range(n)) def _test(r, c): observed_rs.append(int(r)) observed_cs.append(len(c)) return [r * c for c in c] tuner = Hyperband() tuner.run(_get, _test) expected_ns = [81, 34, 15, 8, 5] expected_rs = [1, 3, 9, 27, 81, 3, 9, 27, 81, 9, 27, 81, 27, 81, 81] expected_cs = [81, 27, 9, 3, 1, 27, 9, 3, 1, 9, 3, 1, 6, 2, 5] self.assertEqual(expected_ns, observed_ns) self.assertEqual(expected_rs, observed_rs) self.assertEqual(expected_rs, observed_rs) Fix a typo in Hyperband’s test
from .tuner import Hyperband import unittest class HyperbandTestCase(unittest.TestCase): def test_run(self): observed_ns = [] observed_rs = [] observed_cs = [] def _get(n): observed_ns.append(n) return list(range(n)) def _test(r, c): observed_rs.append(int(r)) observed_cs.append(len(c)) return [r * c for c in c] tuner = Hyperband() tuner.run(_get, _test) expected_ns = [81, 34, 15, 8, 5] expected_rs = [1, 3, 9, 27, 81, 3, 9, 27, 81, 9, 27, 81, 27, 81, 81] expected_cs = [81, 27, 9, 3, 1, 27, 9, 3, 1, 9, 3, 1, 6, 2, 5] self.assertEqual(expected_ns, observed_ns) self.assertEqual(expected_rs, observed_rs) self.assertEqual(expected_cs, observed_cs)
<commit_before>from .tuner import Hyperband import unittest class HyperbandTestCase(unittest.TestCase): def test_run(self): observed_ns = [] observed_rs = [] observed_cs = [] def _get(n): observed_ns.append(n) return list(range(n)) def _test(r, c): observed_rs.append(int(r)) observed_cs.append(len(c)) return [r * c for c in c] tuner = Hyperband() tuner.run(_get, _test) expected_ns = [81, 34, 15, 8, 5] expected_rs = [1, 3, 9, 27, 81, 3, 9, 27, 81, 9, 27, 81, 27, 81, 81] expected_cs = [81, 27, 9, 3, 1, 27, 9, 3, 1, 9, 3, 1, 6, 2, 5] self.assertEqual(expected_ns, observed_ns) self.assertEqual(expected_rs, observed_rs) self.assertEqual(expected_rs, observed_rs) <commit_msg>Fix a typo in Hyperband’s test<commit_after>
from .tuner import Hyperband import unittest class HyperbandTestCase(unittest.TestCase): def test_run(self): observed_ns = [] observed_rs = [] observed_cs = [] def _get(n): observed_ns.append(n) return list(range(n)) def _test(r, c): observed_rs.append(int(r)) observed_cs.append(len(c)) return [r * c for c in c] tuner = Hyperband() tuner.run(_get, _test) expected_ns = [81, 34, 15, 8, 5] expected_rs = [1, 3, 9, 27, 81, 3, 9, 27, 81, 9, 27, 81, 27, 81, 81] expected_cs = [81, 27, 9, 3, 1, 27, 9, 3, 1, 9, 3, 1, 6, 2, 5] self.assertEqual(expected_ns, observed_ns) self.assertEqual(expected_rs, observed_rs) self.assertEqual(expected_cs, observed_cs)
from .tuner import Hyperband import unittest class HyperbandTestCase(unittest.TestCase): def test_run(self): observed_ns = [] observed_rs = [] observed_cs = [] def _get(n): observed_ns.append(n) return list(range(n)) def _test(r, c): observed_rs.append(int(r)) observed_cs.append(len(c)) return [r * c for c in c] tuner = Hyperband() tuner.run(_get, _test) expected_ns = [81, 34, 15, 8, 5] expected_rs = [1, 3, 9, 27, 81, 3, 9, 27, 81, 9, 27, 81, 27, 81, 81] expected_cs = [81, 27, 9, 3, 1, 27, 9, 3, 1, 9, 3, 1, 6, 2, 5] self.assertEqual(expected_ns, observed_ns) self.assertEqual(expected_rs, observed_rs) self.assertEqual(expected_rs, observed_rs) Fix a typo in Hyperband’s testfrom .tuner import Hyperband import unittest class HyperbandTestCase(unittest.TestCase): def test_run(self): observed_ns = [] observed_rs = [] observed_cs = [] def _get(n): observed_ns.append(n) return list(range(n)) def _test(r, c): observed_rs.append(int(r)) observed_cs.append(len(c)) return [r * c for c in c] tuner = Hyperband() tuner.run(_get, _test) expected_ns = [81, 34, 15, 8, 5] expected_rs = [1, 3, 9, 27, 81, 3, 9, 27, 81, 9, 27, 81, 27, 81, 81] expected_cs = [81, 27, 9, 3, 1, 27, 9, 3, 1, 9, 3, 1, 6, 2, 5] self.assertEqual(expected_ns, observed_ns) self.assertEqual(expected_rs, observed_rs) self.assertEqual(expected_cs, observed_cs)
<commit_before>from .tuner import Hyperband import unittest class HyperbandTestCase(unittest.TestCase): def test_run(self): observed_ns = [] observed_rs = [] observed_cs = [] def _get(n): observed_ns.append(n) return list(range(n)) def _test(r, c): observed_rs.append(int(r)) observed_cs.append(len(c)) return [r * c for c in c] tuner = Hyperband() tuner.run(_get, _test) expected_ns = [81, 34, 15, 8, 5] expected_rs = [1, 3, 9, 27, 81, 3, 9, 27, 81, 9, 27, 81, 27, 81, 81] expected_cs = [81, 27, 9, 3, 1, 27, 9, 3, 1, 9, 3, 1, 6, 2, 5] self.assertEqual(expected_ns, observed_ns) self.assertEqual(expected_rs, observed_rs) self.assertEqual(expected_rs, observed_rs) <commit_msg>Fix a typo in Hyperband’s test<commit_after>from .tuner import Hyperband import unittest class HyperbandTestCase(unittest.TestCase): def test_run(self): observed_ns = [] observed_rs = [] observed_cs = [] def _get(n): observed_ns.append(n) return list(range(n)) def _test(r, c): observed_rs.append(int(r)) observed_cs.append(len(c)) return [r * c for c in c] tuner = Hyperband() tuner.run(_get, _test) expected_ns = [81, 34, 15, 8, 5] expected_rs = [1, 3, 9, 27, 81, 3, 9, 27, 81, 9, 27, 81, 27, 81, 81] expected_cs = [81, 27, 9, 3, 1, 27, 9, 3, 1, 9, 3, 1, 6, 2, 5] self.assertEqual(expected_ns, observed_ns) self.assertEqual(expected_rs, observed_rs) self.assertEqual(expected_cs, observed_cs)
14ef0cc78c327c67a2acdb68915d427e4babdd4b
run_tests.py
run_tests.py
from os import getcwd from sys import path as ppath ppath.insert(1,getcwd()+'/modules') # TODO: win32 compatibilite (python path) import unittest from lifegame.test.LifeGameTestSuite import LifeGameTestSuite from tests.TestSuite import TestSuite # TODO: Lister les tests ailleurs ? Recuperer les suite de tests de module auto # (rappel: avant on utilise config.config mais il y avait un import croise) runnable = unittest.TestSuite() tests_suites = [TestSuite(), LifeGameTestSuite()] for testsuite in tests_suites: for test_case in testsuite.get_test_cases(): runnable.addTest(unittest.makeSuite(test_case)) runner=unittest.TextTestRunner() runner.run(runnable)
from os import getcwd from sys import path as ppath ppath.insert(1,getcwd()+'/modules') # TODO: win32 compatibilite (python path) import unittest from lifegame.test.LifeGameTestSuite import LifeGameTestSuite from tests.TestSuite import TestSuite # TODO: Lister les tests ailleurs ? Recuperer les suite de tests de module auto # (rappel: avant on utilise config.config mais il y avait un import croise) runnable = unittest.TestSuite() tests_suites = [TestSuite(), LifeGameTestSuite()] for testsuite in tests_suites: for test_case in testsuite.get_test_cases(): runnable.addTest(unittest.makeSuite(test_case)) runner=unittest.TextTestRunner() exit(runner.run(runnable))
Test running: exit with run test status
Test running: exit with run test status
Python
apache-2.0
buxx/synergine
from os import getcwd from sys import path as ppath ppath.insert(1,getcwd()+'/modules') # TODO: win32 compatibilite (python path) import unittest from lifegame.test.LifeGameTestSuite import LifeGameTestSuite from tests.TestSuite import TestSuite # TODO: Lister les tests ailleurs ? Recuperer les suite de tests de module auto # (rappel: avant on utilise config.config mais il y avait un import croise) runnable = unittest.TestSuite() tests_suites = [TestSuite(), LifeGameTestSuite()] for testsuite in tests_suites: for test_case in testsuite.get_test_cases(): runnable.addTest(unittest.makeSuite(test_case)) runner=unittest.TextTestRunner() runner.run(runnable)Test running: exit with run test status
from os import getcwd from sys import path as ppath ppath.insert(1,getcwd()+'/modules') # TODO: win32 compatibilite (python path) import unittest from lifegame.test.LifeGameTestSuite import LifeGameTestSuite from tests.TestSuite import TestSuite # TODO: Lister les tests ailleurs ? Recuperer les suite de tests de module auto # (rappel: avant on utilise config.config mais il y avait un import croise) runnable = unittest.TestSuite() tests_suites = [TestSuite(), LifeGameTestSuite()] for testsuite in tests_suites: for test_case in testsuite.get_test_cases(): runnable.addTest(unittest.makeSuite(test_case)) runner=unittest.TextTestRunner() exit(runner.run(runnable))
<commit_before>from os import getcwd from sys import path as ppath ppath.insert(1,getcwd()+'/modules') # TODO: win32 compatibilite (python path) import unittest from lifegame.test.LifeGameTestSuite import LifeGameTestSuite from tests.TestSuite import TestSuite # TODO: Lister les tests ailleurs ? Recuperer les suite de tests de module auto # (rappel: avant on utilise config.config mais il y avait un import croise) runnable = unittest.TestSuite() tests_suites = [TestSuite(), LifeGameTestSuite()] for testsuite in tests_suites: for test_case in testsuite.get_test_cases(): runnable.addTest(unittest.makeSuite(test_case)) runner=unittest.TextTestRunner() runner.run(runnable)<commit_msg>Test running: exit with run test status<commit_after>
from os import getcwd from sys import path as ppath ppath.insert(1,getcwd()+'/modules') # TODO: win32 compatibilite (python path) import unittest from lifegame.test.LifeGameTestSuite import LifeGameTestSuite from tests.TestSuite import TestSuite # TODO: Lister les tests ailleurs ? Recuperer les suite de tests de module auto # (rappel: avant on utilise config.config mais il y avait un import croise) runnable = unittest.TestSuite() tests_suites = [TestSuite(), LifeGameTestSuite()] for testsuite in tests_suites: for test_case in testsuite.get_test_cases(): runnable.addTest(unittest.makeSuite(test_case)) runner=unittest.TextTestRunner() exit(runner.run(runnable))
from os import getcwd from sys import path as ppath ppath.insert(1,getcwd()+'/modules') # TODO: win32 compatibilite (python path) import unittest from lifegame.test.LifeGameTestSuite import LifeGameTestSuite from tests.TestSuite import TestSuite # TODO: Lister les tests ailleurs ? Recuperer les suite de tests de module auto # (rappel: avant on utilise config.config mais il y avait un import croise) runnable = unittest.TestSuite() tests_suites = [TestSuite(), LifeGameTestSuite()] for testsuite in tests_suites: for test_case in testsuite.get_test_cases(): runnable.addTest(unittest.makeSuite(test_case)) runner=unittest.TextTestRunner() runner.run(runnable)Test running: exit with run test statusfrom os import getcwd from sys import path as ppath ppath.insert(1,getcwd()+'/modules') # TODO: win32 compatibilite (python path) import unittest from lifegame.test.LifeGameTestSuite import LifeGameTestSuite from tests.TestSuite import TestSuite # TODO: Lister les tests ailleurs ? Recuperer les suite de tests de module auto # (rappel: avant on utilise config.config mais il y avait un import croise) runnable = unittest.TestSuite() tests_suites = [TestSuite(), LifeGameTestSuite()] for testsuite in tests_suites: for test_case in testsuite.get_test_cases(): runnable.addTest(unittest.makeSuite(test_case)) runner=unittest.TextTestRunner() exit(runner.run(runnable))
<commit_before>from os import getcwd from sys import path as ppath ppath.insert(1,getcwd()+'/modules') # TODO: win32 compatibilite (python path) import unittest from lifegame.test.LifeGameTestSuite import LifeGameTestSuite from tests.TestSuite import TestSuite # TODO: Lister les tests ailleurs ? Recuperer les suite de tests de module auto # (rappel: avant on utilise config.config mais il y avait un import croise) runnable = unittest.TestSuite() tests_suites = [TestSuite(), LifeGameTestSuite()] for testsuite in tests_suites: for test_case in testsuite.get_test_cases(): runnable.addTest(unittest.makeSuite(test_case)) runner=unittest.TextTestRunner() runner.run(runnable)<commit_msg>Test running: exit with run test status<commit_after>from os import getcwd from sys import path as ppath ppath.insert(1,getcwd()+'/modules') # TODO: win32 compatibilite (python path) import unittest from lifegame.test.LifeGameTestSuite import LifeGameTestSuite from tests.TestSuite import TestSuite # TODO: Lister les tests ailleurs ? Recuperer les suite de tests de module auto # (rappel: avant on utilise config.config mais il y avait un import croise) runnable = unittest.TestSuite() tests_suites = [TestSuite(), LifeGameTestSuite()] for testsuite in tests_suites: for test_case in testsuite.get_test_cases(): runnable.addTest(unittest.makeSuite(test_case)) runner=unittest.TextTestRunner() exit(runner.run(runnable))
ae593fd5de74a123b5064ef2e018b4955dc9e6c9
runserver.py
runserver.py
#!/usr/local/bin/python2.7 import sys from datetime import date from main import app from upload_s3 import set_metadata from flask_frozen import Freezer # cron is called with 3 arguments, should only run in the first week of month cron_condition = len(sys.argv) == 3 and date.today().day > 8 force_update = len(sys.argv) == 2 and sys.argv[1] == 'freeze' if len(sys.argv) > 1: # if runserver is passed an argument if cron_condition or force_update: print "Cron succeeded" #freezer = Freezer(app) #freezer.freeze() #set_metadata() else: app.run()
#!/usr/local/bin/python2.7 import sys from datetime import date from main import app from upload_s3 import set_metadata from flask_frozen import Freezer # cron is called with 3 arguments, should only run in the first week of month cron_condition = len(sys.argv) == 3 and date.today().day < 8 force_update = len(sys.argv) == 2 and sys.argv[1] == 'freeze' if len(sys.argv) > 1: # if runserver is passed an argument if cron_condition or force_update: freezer = Freezer(app) freezer.freeze() set_metadata() else: app.run()
Complete testing, change back to production code
Complete testing, change back to production code
Python
apache-2.0
vprnet/dorothys-list,vprnet/dorothys-list,vprnet/dorothys-list
#!/usr/local/bin/python2.7 import sys from datetime import date from main import app from upload_s3 import set_metadata from flask_frozen import Freezer # cron is called with 3 arguments, should only run in the first week of month cron_condition = len(sys.argv) == 3 and date.today().day > 8 force_update = len(sys.argv) == 2 and sys.argv[1] == 'freeze' if len(sys.argv) > 1: # if runserver is passed an argument if cron_condition or force_update: print "Cron succeeded" #freezer = Freezer(app) #freezer.freeze() #set_metadata() else: app.run() Complete testing, change back to production code
#!/usr/local/bin/python2.7 import sys from datetime import date from main import app from upload_s3 import set_metadata from flask_frozen import Freezer # cron is called with 3 arguments, should only run in the first week of month cron_condition = len(sys.argv) == 3 and date.today().day < 8 force_update = len(sys.argv) == 2 and sys.argv[1] == 'freeze' if len(sys.argv) > 1: # if runserver is passed an argument if cron_condition or force_update: freezer = Freezer(app) freezer.freeze() set_metadata() else: app.run()
<commit_before>#!/usr/local/bin/python2.7 import sys from datetime import date from main import app from upload_s3 import set_metadata from flask_frozen import Freezer # cron is called with 3 arguments, should only run in the first week of month cron_condition = len(sys.argv) == 3 and date.today().day > 8 force_update = len(sys.argv) == 2 and sys.argv[1] == 'freeze' if len(sys.argv) > 1: # if runserver is passed an argument if cron_condition or force_update: print "Cron succeeded" #freezer = Freezer(app) #freezer.freeze() #set_metadata() else: app.run() <commit_msg>Complete testing, change back to production code<commit_after>
#!/usr/local/bin/python2.7 import sys from datetime import date from main import app from upload_s3 import set_metadata from flask_frozen import Freezer # cron is called with 3 arguments, should only run in the first week of month cron_condition = len(sys.argv) == 3 and date.today().day < 8 force_update = len(sys.argv) == 2 and sys.argv[1] == 'freeze' if len(sys.argv) > 1: # if runserver is passed an argument if cron_condition or force_update: freezer = Freezer(app) freezer.freeze() set_metadata() else: app.run()
#!/usr/local/bin/python2.7 import sys from datetime import date from main import app from upload_s3 import set_metadata from flask_frozen import Freezer # cron is called with 3 arguments, should only run in the first week of month cron_condition = len(sys.argv) == 3 and date.today().day > 8 force_update = len(sys.argv) == 2 and sys.argv[1] == 'freeze' if len(sys.argv) > 1: # if runserver is passed an argument if cron_condition or force_update: print "Cron succeeded" #freezer = Freezer(app) #freezer.freeze() #set_metadata() else: app.run() Complete testing, change back to production code#!/usr/local/bin/python2.7 import sys from datetime import date from main import app from upload_s3 import set_metadata from flask_frozen import Freezer # cron is called with 3 arguments, should only run in the first week of month cron_condition = len(sys.argv) == 3 and date.today().day < 8 force_update = len(sys.argv) == 2 and sys.argv[1] == 'freeze' if len(sys.argv) > 1: # if runserver is passed an argument if cron_condition or force_update: freezer = Freezer(app) freezer.freeze() set_metadata() else: app.run()
<commit_before>#!/usr/local/bin/python2.7 import sys from datetime import date from main import app from upload_s3 import set_metadata from flask_frozen import Freezer # cron is called with 3 arguments, should only run in the first week of month cron_condition = len(sys.argv) == 3 and date.today().day > 8 force_update = len(sys.argv) == 2 and sys.argv[1] == 'freeze' if len(sys.argv) > 1: # if runserver is passed an argument if cron_condition or force_update: print "Cron succeeded" #freezer = Freezer(app) #freezer.freeze() #set_metadata() else: app.run() <commit_msg>Complete testing, change back to production code<commit_after>#!/usr/local/bin/python2.7 import sys from datetime import date from main import app from upload_s3 import set_metadata from flask_frozen import Freezer # cron is called with 3 arguments, should only run in the first week of month cron_condition = len(sys.argv) == 3 and date.today().day < 8 force_update = len(sys.argv) == 2 and sys.argv[1] == 'freeze' if len(sys.argv) > 1: # if runserver is passed an argument if cron_condition or force_update: freezer = Freezer(app) freezer.freeze() set_metadata() else: app.run()
3cef3e4774923b81e622f03aee44a933293c6a8d
modelview/migrations/0036_auto_20170322_1622.py
modelview/migrations/0036_auto_20170322_1622.py
# -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2017-03-22 15:22 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('modelview', '0035_auto_20160426_1721'), ] operations = [ migrations.AlterField( model_name='basicfactsheet', name='source_of_funding', field=models.CharField(help_text='What is the main source of funding?', max_length=200, null=True, verbose_name='Source of funding'), ), ]
# -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2017-03-22 15:22 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ #('modelview', '0035_auto_20160426_1721'), ('modelview', '0035_auto_20170724_1801'), ] operations = [ migrations.AlterField( model_name='basicfactsheet', name='source_of_funding', field=models.CharField(help_text='What is the main source of funding?', max_length=200, null=True, verbose_name='Source of funding'), ), ]
Change reference from modelview/0035_auto_20160426 to 0035_auto_20170724
Change reference from modelview/0035_auto_20160426 to 0035_auto_20170724
Python
agpl-3.0
openego/oeplatform,tom-heimbrodt/oeplatform,tom-heimbrodt/oeplatform,openego/oeplatform,tom-heimbrodt/oeplatform,openego/oeplatform,openego/oeplatform
# -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2017-03-22 15:22 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('modelview', '0035_auto_20160426_1721'), ] operations = [ migrations.AlterField( model_name='basicfactsheet', name='source_of_funding', field=models.CharField(help_text='What is the main source of funding?', max_length=200, null=True, verbose_name='Source of funding'), ), ] Change reference from modelview/0035_auto_20160426 to 0035_auto_20170724
# -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2017-03-22 15:22 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ #('modelview', '0035_auto_20160426_1721'), ('modelview', '0035_auto_20170724_1801'), ] operations = [ migrations.AlterField( model_name='basicfactsheet', name='source_of_funding', field=models.CharField(help_text='What is the main source of funding?', max_length=200, null=True, verbose_name='Source of funding'), ), ]
<commit_before># -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2017-03-22 15:22 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('modelview', '0035_auto_20160426_1721'), ] operations = [ migrations.AlterField( model_name='basicfactsheet', name='source_of_funding', field=models.CharField(help_text='What is the main source of funding?', max_length=200, null=True, verbose_name='Source of funding'), ), ] <commit_msg>Change reference from modelview/0035_auto_20160426 to 0035_auto_20170724<commit_after>
# -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2017-03-22 15:22 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ #('modelview', '0035_auto_20160426_1721'), ('modelview', '0035_auto_20170724_1801'), ] operations = [ migrations.AlterField( model_name='basicfactsheet', name='source_of_funding', field=models.CharField(help_text='What is the main source of funding?', max_length=200, null=True, verbose_name='Source of funding'), ), ]
# -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2017-03-22 15:22 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('modelview', '0035_auto_20160426_1721'), ] operations = [ migrations.AlterField( model_name='basicfactsheet', name='source_of_funding', field=models.CharField(help_text='What is the main source of funding?', max_length=200, null=True, verbose_name='Source of funding'), ), ] Change reference from modelview/0035_auto_20160426 to 0035_auto_20170724# -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2017-03-22 15:22 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ #('modelview', '0035_auto_20160426_1721'), ('modelview', '0035_auto_20170724_1801'), ] operations = [ migrations.AlterField( model_name='basicfactsheet', name='source_of_funding', field=models.CharField(help_text='What is the main source of funding?', max_length=200, null=True, verbose_name='Source of funding'), ), ]
<commit_before># -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2017-03-22 15:22 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('modelview', '0035_auto_20160426_1721'), ] operations = [ migrations.AlterField( model_name='basicfactsheet', name='source_of_funding', field=models.CharField(help_text='What is the main source of funding?', max_length=200, null=True, verbose_name='Source of funding'), ), ] <commit_msg>Change reference from modelview/0035_auto_20160426 to 0035_auto_20170724<commit_after># -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2017-03-22 15:22 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ #('modelview', '0035_auto_20160426_1721'), ('modelview', '0035_auto_20170724_1801'), ] operations = [ migrations.AlterField( model_name='basicfactsheet', name='source_of_funding', field=models.CharField(help_text='What is the main source of funding?', max_length=200, null=True, verbose_name='Source of funding'), ), ]
c50d9efe011417162ab6e99e19973dcde6b3ddfa
utils/get_collection_object_count.py
utils/get_collection_object_count.py
#!/usr/bin/env python # -*- coding: utf8 -*- import sys import argparse from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo def main(argv=None): parser = argparse.ArgumentParser( description='Print count of objects for a given collection.') parser.add_argument('path', help="Nuxeo path to collection") parser.add_argument( '--pynuxrc', default='~/.pynuxrc-prod', help="rcfile for use with pynux utils") parser.add_argument( '--components', action='store_true', help="show counts for object components") if argv is None: argv = parser.parse_args() dh = DeepHarvestNuxeo(argv.path, '', pynuxrc=argv.pynuxrc) print "about to fetch objects for path {}".format(dh.path) objects = dh.fetch_objects() object_count = len(objects) print "finished fetching objects. {} found".format(object_count) if not argv.components: return print "about to iterate through objects and get components" component_count = 0 for obj in objects: components = dh.fetch_components(obj) component_count = component_count + len(components) print "finished fetching components. {} found".format(component_count) print "Grand Total: {}".format(object_count + component_count) if __name__ == "__main__": sys.exit(main())
#!/usr/bin/env python # -*- coding: utf8 -*- import sys import argparse from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo def main(argv=None): parser = argparse.ArgumentParser( description='Print count of objects for a given collection.') parser.add_argument('path', help="Nuxeo path to collection") parser.add_argument( '--pynuxrc', default='~/.pynuxrc', help="rcfile for use with pynux utils") parser.add_argument( '--components', action='store_true', help="show counts for object components") if argv is None: argv = parser.parse_args() dh = DeepHarvestNuxeo(argv.path, '', pynuxrc=argv.pynuxrc) print "about to fetch objects for path {}".format(dh.path) objects = dh.fetch_objects() object_count = len(objects) print "finished fetching objects. {} found".format(object_count) uid_set = set() for obj in objects: uid_set.add(obj['uid']) unique = len(uid_set) print "unique uid count: {}".format(unique) if not argv.components: return print "about to iterate through objects and get components" component_count = 0 for obj in objects: components = dh.fetch_components(obj) component_count = component_count + len(components) print "finished fetching components. {} found".format(component_count) print "Grand Total: {}".format(object_count + component_count) if __name__ == "__main__": sys.exit(main())
Change default arg value for pynuxrc
Change default arg value for pynuxrc
Python
bsd-3-clause
barbarahui/nuxeo-calisphere,barbarahui/nuxeo-calisphere
#!/usr/bin/env python # -*- coding: utf8 -*- import sys import argparse from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo def main(argv=None): parser = argparse.ArgumentParser( description='Print count of objects for a given collection.') parser.add_argument('path', help="Nuxeo path to collection") parser.add_argument( '--pynuxrc', default='~/.pynuxrc-prod', help="rcfile for use with pynux utils") parser.add_argument( '--components', action='store_true', help="show counts for object components") if argv is None: argv = parser.parse_args() dh = DeepHarvestNuxeo(argv.path, '', pynuxrc=argv.pynuxrc) print "about to fetch objects for path {}".format(dh.path) objects = dh.fetch_objects() object_count = len(objects) print "finished fetching objects. {} found".format(object_count) if not argv.components: return print "about to iterate through objects and get components" component_count = 0 for obj in objects: components = dh.fetch_components(obj) component_count = component_count + len(components) print "finished fetching components. {} found".format(component_count) print "Grand Total: {}".format(object_count + component_count) if __name__ == "__main__": sys.exit(main()) Change default arg value for pynuxrc
#!/usr/bin/env python # -*- coding: utf8 -*- import sys import argparse from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo def main(argv=None): parser = argparse.ArgumentParser( description='Print count of objects for a given collection.') parser.add_argument('path', help="Nuxeo path to collection") parser.add_argument( '--pynuxrc', default='~/.pynuxrc', help="rcfile for use with pynux utils") parser.add_argument( '--components', action='store_true', help="show counts for object components") if argv is None: argv = parser.parse_args() dh = DeepHarvestNuxeo(argv.path, '', pynuxrc=argv.pynuxrc) print "about to fetch objects for path {}".format(dh.path) objects = dh.fetch_objects() object_count = len(objects) print "finished fetching objects. {} found".format(object_count) uid_set = set() for obj in objects: uid_set.add(obj['uid']) unique = len(uid_set) print "unique uid count: {}".format(unique) if not argv.components: return print "about to iterate through objects and get components" component_count = 0 for obj in objects: components = dh.fetch_components(obj) component_count = component_count + len(components) print "finished fetching components. {} found".format(component_count) print "Grand Total: {}".format(object_count + component_count) if __name__ == "__main__": sys.exit(main())
<commit_before>#!/usr/bin/env python # -*- coding: utf8 -*- import sys import argparse from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo def main(argv=None): parser = argparse.ArgumentParser( description='Print count of objects for a given collection.') parser.add_argument('path', help="Nuxeo path to collection") parser.add_argument( '--pynuxrc', default='~/.pynuxrc-prod', help="rcfile for use with pynux utils") parser.add_argument( '--components', action='store_true', help="show counts for object components") if argv is None: argv = parser.parse_args() dh = DeepHarvestNuxeo(argv.path, '', pynuxrc=argv.pynuxrc) print "about to fetch objects for path {}".format(dh.path) objects = dh.fetch_objects() object_count = len(objects) print "finished fetching objects. {} found".format(object_count) if not argv.components: return print "about to iterate through objects and get components" component_count = 0 for obj in objects: components = dh.fetch_components(obj) component_count = component_count + len(components) print "finished fetching components. {} found".format(component_count) print "Grand Total: {}".format(object_count + component_count) if __name__ == "__main__": sys.exit(main()) <commit_msg>Change default arg value for pynuxrc<commit_after>
#!/usr/bin/env python # -*- coding: utf8 -*- import sys import argparse from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo def main(argv=None): parser = argparse.ArgumentParser( description='Print count of objects for a given collection.') parser.add_argument('path', help="Nuxeo path to collection") parser.add_argument( '--pynuxrc', default='~/.pynuxrc', help="rcfile for use with pynux utils") parser.add_argument( '--components', action='store_true', help="show counts for object components") if argv is None: argv = parser.parse_args() dh = DeepHarvestNuxeo(argv.path, '', pynuxrc=argv.pynuxrc) print "about to fetch objects for path {}".format(dh.path) objects = dh.fetch_objects() object_count = len(objects) print "finished fetching objects. {} found".format(object_count) uid_set = set() for obj in objects: uid_set.add(obj['uid']) unique = len(uid_set) print "unique uid count: {}".format(unique) if not argv.components: return print "about to iterate through objects and get components" component_count = 0 for obj in objects: components = dh.fetch_components(obj) component_count = component_count + len(components) print "finished fetching components. {} found".format(component_count) print "Grand Total: {}".format(object_count + component_count) if __name__ == "__main__": sys.exit(main())
#!/usr/bin/env python # -*- coding: utf8 -*- import sys import argparse from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo def main(argv=None): parser = argparse.ArgumentParser( description='Print count of objects for a given collection.') parser.add_argument('path', help="Nuxeo path to collection") parser.add_argument( '--pynuxrc', default='~/.pynuxrc-prod', help="rcfile for use with pynux utils") parser.add_argument( '--components', action='store_true', help="show counts for object components") if argv is None: argv = parser.parse_args() dh = DeepHarvestNuxeo(argv.path, '', pynuxrc=argv.pynuxrc) print "about to fetch objects for path {}".format(dh.path) objects = dh.fetch_objects() object_count = len(objects) print "finished fetching objects. {} found".format(object_count) if not argv.components: return print "about to iterate through objects and get components" component_count = 0 for obj in objects: components = dh.fetch_components(obj) component_count = component_count + len(components) print "finished fetching components. {} found".format(component_count) print "Grand Total: {}".format(object_count + component_count) if __name__ == "__main__": sys.exit(main()) Change default arg value for pynuxrc#!/usr/bin/env python # -*- coding: utf8 -*- import sys import argparse from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo def main(argv=None): parser = argparse.ArgumentParser( description='Print count of objects for a given collection.') parser.add_argument('path', help="Nuxeo path to collection") parser.add_argument( '--pynuxrc', default='~/.pynuxrc', help="rcfile for use with pynux utils") parser.add_argument( '--components', action='store_true', help="show counts for object components") if argv is None: argv = parser.parse_args() dh = DeepHarvestNuxeo(argv.path, '', pynuxrc=argv.pynuxrc) print "about to fetch objects for path {}".format(dh.path) objects = dh.fetch_objects() object_count = len(objects) print "finished fetching objects. {} found".format(object_count) uid_set = set() for obj in objects: uid_set.add(obj['uid']) unique = len(uid_set) print "unique uid count: {}".format(unique) if not argv.components: return print "about to iterate through objects and get components" component_count = 0 for obj in objects: components = dh.fetch_components(obj) component_count = component_count + len(components) print "finished fetching components. {} found".format(component_count) print "Grand Total: {}".format(object_count + component_count) if __name__ == "__main__": sys.exit(main())
<commit_before>#!/usr/bin/env python # -*- coding: utf8 -*- import sys import argparse from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo def main(argv=None): parser = argparse.ArgumentParser( description='Print count of objects for a given collection.') parser.add_argument('path', help="Nuxeo path to collection") parser.add_argument( '--pynuxrc', default='~/.pynuxrc-prod', help="rcfile for use with pynux utils") parser.add_argument( '--components', action='store_true', help="show counts for object components") if argv is None: argv = parser.parse_args() dh = DeepHarvestNuxeo(argv.path, '', pynuxrc=argv.pynuxrc) print "about to fetch objects for path {}".format(dh.path) objects = dh.fetch_objects() object_count = len(objects) print "finished fetching objects. {} found".format(object_count) if not argv.components: return print "about to iterate through objects and get components" component_count = 0 for obj in objects: components = dh.fetch_components(obj) component_count = component_count + len(components) print "finished fetching components. {} found".format(component_count) print "Grand Total: {}".format(object_count + component_count) if __name__ == "__main__": sys.exit(main()) <commit_msg>Change default arg value for pynuxrc<commit_after>#!/usr/bin/env python # -*- coding: utf8 -*- import sys import argparse from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo def main(argv=None): parser = argparse.ArgumentParser( description='Print count of objects for a given collection.') parser.add_argument('path', help="Nuxeo path to collection") parser.add_argument( '--pynuxrc', default='~/.pynuxrc', help="rcfile for use with pynux utils") parser.add_argument( '--components', action='store_true', help="show counts for object components") if argv is None: argv = parser.parse_args() dh = DeepHarvestNuxeo(argv.path, '', pynuxrc=argv.pynuxrc) print "about to fetch objects for path {}".format(dh.path) objects = dh.fetch_objects() object_count = len(objects) print "finished fetching objects. {} found".format(object_count) uid_set = set() for obj in objects: uid_set.add(obj['uid']) unique = len(uid_set) print "unique uid count: {}".format(unique) if not argv.components: return print "about to iterate through objects and get components" component_count = 0 for obj in objects: components = dh.fetch_components(obj) component_count = component_count + len(components) print "finished fetching components. {} found".format(component_count) print "Grand Total: {}".format(object_count + component_count) if __name__ == "__main__": sys.exit(main())
dad05c2eff78383e3179fd1f4a4502fe553afecd
tests/lib/docker_compose_tools.py
tests/lib/docker_compose_tools.py
# pylint: disable=line-too-long """ Handles docker compose """ from lib.docker_tools import docker_compose from lib.tools import show_status def set_up(pg_version, es_version): """ Start containers """ compose = docker_compose(pg_version, es_version) show_status( "Starting testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) show_status("Stopping and Removing any old containers...") compose("stop") compose("rm", "--force") show_status("Building new images...") compose("build") show_status("Starting new containers...") compose("up", "-d") show_status("Testing environment started") def tear_down(pg_version, es_version): """ Stop containers """ compose = docker_compose(pg_version, es_version) show_status( "Stopping testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) compose("down") show_status("Testing environment stopped")
# pylint: disable=line-too-long """ Handles docker compose """ import sys from sh import ErrorReturnCode from lib.docker_tools import docker_compose from lib.tools import show_status def set_up(pg_version, es_version): """ Start containers """ compose = docker_compose(pg_version, es_version) show_status( "Starting testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) show_status("Stopping and Removing any old containers...") compose("stop") compose("rm", "--force") show_status("Building new images...") try: compose("build") except ErrorReturnCode as exc: print("Failed to build images...") print(exc.stdout.decode("utf-8")) print() print(exc.stderr.decode("utf-8")) sys.exit(1) show_status("Starting new containers...") compose("up", "-d") show_status("Testing environment started") def tear_down(pg_version, es_version): """ Stop containers """ compose = docker_compose(pg_version, es_version) show_status( "Stopping testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) compose("down") show_status("Testing environment stopped")
Print full stdout / stderr when image build fails
Print full stdout / stderr when image build fails
Python
mit
matthewfranglen/postgres-elasticsearch-fdw
# pylint: disable=line-too-long """ Handles docker compose """ from lib.docker_tools import docker_compose from lib.tools import show_status def set_up(pg_version, es_version): """ Start containers """ compose = docker_compose(pg_version, es_version) show_status( "Starting testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) show_status("Stopping and Removing any old containers...") compose("stop") compose("rm", "--force") show_status("Building new images...") compose("build") show_status("Starting new containers...") compose("up", "-d") show_status("Testing environment started") def tear_down(pg_version, es_version): """ Stop containers """ compose = docker_compose(pg_version, es_version) show_status( "Stopping testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) compose("down") show_status("Testing environment stopped") Print full stdout / stderr when image build fails
# pylint: disable=line-too-long """ Handles docker compose """ import sys from sh import ErrorReturnCode from lib.docker_tools import docker_compose from lib.tools import show_status def set_up(pg_version, es_version): """ Start containers """ compose = docker_compose(pg_version, es_version) show_status( "Starting testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) show_status("Stopping and Removing any old containers...") compose("stop") compose("rm", "--force") show_status("Building new images...") try: compose("build") except ErrorReturnCode as exc: print("Failed to build images...") print(exc.stdout.decode("utf-8")) print() print(exc.stderr.decode("utf-8")) sys.exit(1) show_status("Starting new containers...") compose("up", "-d") show_status("Testing environment started") def tear_down(pg_version, es_version): """ Stop containers """ compose = docker_compose(pg_version, es_version) show_status( "Stopping testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) compose("down") show_status("Testing environment stopped")
<commit_before># pylint: disable=line-too-long """ Handles docker compose """ from lib.docker_tools import docker_compose from lib.tools import show_status def set_up(pg_version, es_version): """ Start containers """ compose = docker_compose(pg_version, es_version) show_status( "Starting testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) show_status("Stopping and Removing any old containers...") compose("stop") compose("rm", "--force") show_status("Building new images...") compose("build") show_status("Starting new containers...") compose("up", "-d") show_status("Testing environment started") def tear_down(pg_version, es_version): """ Stop containers """ compose = docker_compose(pg_version, es_version) show_status( "Stopping testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) compose("down") show_status("Testing environment stopped") <commit_msg>Print full stdout / stderr when image build fails<commit_after>
# pylint: disable=line-too-long """ Handles docker compose """ import sys from sh import ErrorReturnCode from lib.docker_tools import docker_compose from lib.tools import show_status def set_up(pg_version, es_version): """ Start containers """ compose = docker_compose(pg_version, es_version) show_status( "Starting testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) show_status("Stopping and Removing any old containers...") compose("stop") compose("rm", "--force") show_status("Building new images...") try: compose("build") except ErrorReturnCode as exc: print("Failed to build images...") print(exc.stdout.decode("utf-8")) print() print(exc.stderr.decode("utf-8")) sys.exit(1) show_status("Starting new containers...") compose("up", "-d") show_status("Testing environment started") def tear_down(pg_version, es_version): """ Stop containers """ compose = docker_compose(pg_version, es_version) show_status( "Stopping testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) compose("down") show_status("Testing environment stopped")
# pylint: disable=line-too-long """ Handles docker compose """ from lib.docker_tools import docker_compose from lib.tools import show_status def set_up(pg_version, es_version): """ Start containers """ compose = docker_compose(pg_version, es_version) show_status( "Starting testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) show_status("Stopping and Removing any old containers...") compose("stop") compose("rm", "--force") show_status("Building new images...") compose("build") show_status("Starting new containers...") compose("up", "-d") show_status("Testing environment started") def tear_down(pg_version, es_version): """ Stop containers """ compose = docker_compose(pg_version, es_version) show_status( "Stopping testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) compose("down") show_status("Testing environment stopped") Print full stdout / stderr when image build fails# pylint: disable=line-too-long """ Handles docker compose """ import sys from sh import ErrorReturnCode from lib.docker_tools import docker_compose from lib.tools import show_status def set_up(pg_version, es_version): """ Start containers """ compose = docker_compose(pg_version, es_version) show_status( "Starting testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) show_status("Stopping and Removing any old containers...") compose("stop") compose("rm", "--force") show_status("Building new images...") try: compose("build") except ErrorReturnCode as exc: print("Failed to build images...") print(exc.stdout.decode("utf-8")) print() print(exc.stderr.decode("utf-8")) sys.exit(1) show_status("Starting new containers...") compose("up", "-d") show_status("Testing environment started") def tear_down(pg_version, es_version): """ Stop containers """ compose = docker_compose(pg_version, es_version) show_status( "Stopping testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) compose("down") show_status("Testing environment stopped")
<commit_before># pylint: disable=line-too-long """ Handles docker compose """ from lib.docker_tools import docker_compose from lib.tools import show_status def set_up(pg_version, es_version): """ Start containers """ compose = docker_compose(pg_version, es_version) show_status( "Starting testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) show_status("Stopping and Removing any old containers...") compose("stop") compose("rm", "--force") show_status("Building new images...") compose("build") show_status("Starting new containers...") compose("up", "-d") show_status("Testing environment started") def tear_down(pg_version, es_version): """ Stop containers """ compose = docker_compose(pg_version, es_version) show_status( "Stopping testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) compose("down") show_status("Testing environment stopped") <commit_msg>Print full stdout / stderr when image build fails<commit_after># pylint: disable=line-too-long """ Handles docker compose """ import sys from sh import ErrorReturnCode from lib.docker_tools import docker_compose from lib.tools import show_status def set_up(pg_version, es_version): """ Start containers """ compose = docker_compose(pg_version, es_version) show_status( "Starting testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) show_status("Stopping and Removing any old containers...") compose("stop") compose("rm", "--force") show_status("Building new images...") try: compose("build") except ErrorReturnCode as exc: print("Failed to build images...") print(exc.stdout.decode("utf-8")) print() print(exc.stderr.decode("utf-8")) sys.exit(1) show_status("Starting new containers...") compose("up", "-d") show_status("Testing environment started") def tear_down(pg_version, es_version): """ Stop containers """ compose = docker_compose(pg_version, es_version) show_status( "Stopping testing environment for PostgreSQL {pg_version} with Elasticsearch {es_version}...".format( pg_version=pg_version, es_version=es_version ) ) compose("down") show_status("Testing environment stopped")
482bcbaeddf3d18c445b780a083d723a85a6dd88
wok/renderers.py
wok/renderers.py
from markdown import markdown import docutils.core from docutils.writers.html4css1 import Writer as rst_html_writer class Renderer(object): extensions = [] @classmethod def render(cls, plain): return plain class Markdown(Renderer): extensions = ['markdown', 'mkd'] @classmethod def render(cls, plain): return markdown(plain, ['def_list', 'footnotes']) class ReStructuredText(Renderer): extensions = ['rst'] @classmethod def render(cls, plain): w = rst_html_writer() return docutils.core.publish_parts(plain, writer=w)['body'] class Plain(Renderer): extensions = 'txt' @classmethod def render(cls, plain): return plain.replace('\n', '<br>') all = [Renderer, Plain, Markdown, ReStructuredText]
from markdown import markdown import docutils.core from docutils.writers.html4css1 import Writer as rst_html_writer class Renderer(object): extensions = [] @classmethod def render(cls, plain): return plain class Markdown(Renderer): extensions = ['markdown', 'mkd'] @classmethod def render(cls, plain): return markdown(plain, ['def_list', 'footnotes', 'codehilite']) class ReStructuredText(Renderer): extensions = ['rst'] @classmethod def render(cls, plain): w = rst_html_writer() return docutils.core.publish_parts(plain, writer=w)['body'] class Plain(Renderer): extensions = 'txt' @classmethod def render(cls, plain): return plain.replace('\n', '<br>') all = [Renderer, Plain, Markdown, ReStructuredText]
Enable syntax highlighting with Markdown.
Enable syntax highlighting with Markdown.
Python
mit
wummel/wok,mythmon/wok,ngokevin/wok,matt-garman/wok,Avaren/wok,algor512/wok,edunham/wok,ngokevin/wok,chrplace/wok,moreati/wok,mythmon/wok,algor512/wok,vaygr/wok,vaygr/wok,edunham/wok,algor512/wok,moreati/wok,abbgrade/wok,edunham/wok,gchriz/wok,abbgrade/wok,matt-garman/wok,chrplace/wok,chrplace/wok,matt-garman/wok,jneves/wok,wummel/wok,moreati/wok,jneves/wok,gchriz/wok,Avaren/wok,vaygr/wok,mythmon/wok,gchriz/wok,jneves/wok,Avaren/wok
from markdown import markdown import docutils.core from docutils.writers.html4css1 import Writer as rst_html_writer class Renderer(object): extensions = [] @classmethod def render(cls, plain): return plain class Markdown(Renderer): extensions = ['markdown', 'mkd'] @classmethod def render(cls, plain): return markdown(plain, ['def_list', 'footnotes']) class ReStructuredText(Renderer): extensions = ['rst'] @classmethod def render(cls, plain): w = rst_html_writer() return docutils.core.publish_parts(plain, writer=w)['body'] class Plain(Renderer): extensions = 'txt' @classmethod def render(cls, plain): return plain.replace('\n', '<br>') all = [Renderer, Plain, Markdown, ReStructuredText] Enable syntax highlighting with Markdown.
from markdown import markdown import docutils.core from docutils.writers.html4css1 import Writer as rst_html_writer class Renderer(object): extensions = [] @classmethod def render(cls, plain): return plain class Markdown(Renderer): extensions = ['markdown', 'mkd'] @classmethod def render(cls, plain): return markdown(plain, ['def_list', 'footnotes', 'codehilite']) class ReStructuredText(Renderer): extensions = ['rst'] @classmethod def render(cls, plain): w = rst_html_writer() return docutils.core.publish_parts(plain, writer=w)['body'] class Plain(Renderer): extensions = 'txt' @classmethod def render(cls, plain): return plain.replace('\n', '<br>') all = [Renderer, Plain, Markdown, ReStructuredText]
<commit_before>from markdown import markdown import docutils.core from docutils.writers.html4css1 import Writer as rst_html_writer class Renderer(object): extensions = [] @classmethod def render(cls, plain): return plain class Markdown(Renderer): extensions = ['markdown', 'mkd'] @classmethod def render(cls, plain): return markdown(plain, ['def_list', 'footnotes']) class ReStructuredText(Renderer): extensions = ['rst'] @classmethod def render(cls, plain): w = rst_html_writer() return docutils.core.publish_parts(plain, writer=w)['body'] class Plain(Renderer): extensions = 'txt' @classmethod def render(cls, plain): return plain.replace('\n', '<br>') all = [Renderer, Plain, Markdown, ReStructuredText] <commit_msg>Enable syntax highlighting with Markdown.<commit_after>
from markdown import markdown import docutils.core from docutils.writers.html4css1 import Writer as rst_html_writer class Renderer(object): extensions = [] @classmethod def render(cls, plain): return plain class Markdown(Renderer): extensions = ['markdown', 'mkd'] @classmethod def render(cls, plain): return markdown(plain, ['def_list', 'footnotes', 'codehilite']) class ReStructuredText(Renderer): extensions = ['rst'] @classmethod def render(cls, plain): w = rst_html_writer() return docutils.core.publish_parts(plain, writer=w)['body'] class Plain(Renderer): extensions = 'txt' @classmethod def render(cls, plain): return plain.replace('\n', '<br>') all = [Renderer, Plain, Markdown, ReStructuredText]
from markdown import markdown import docutils.core from docutils.writers.html4css1 import Writer as rst_html_writer class Renderer(object): extensions = [] @classmethod def render(cls, plain): return plain class Markdown(Renderer): extensions = ['markdown', 'mkd'] @classmethod def render(cls, plain): return markdown(plain, ['def_list', 'footnotes']) class ReStructuredText(Renderer): extensions = ['rst'] @classmethod def render(cls, plain): w = rst_html_writer() return docutils.core.publish_parts(plain, writer=w)['body'] class Plain(Renderer): extensions = 'txt' @classmethod def render(cls, plain): return plain.replace('\n', '<br>') all = [Renderer, Plain, Markdown, ReStructuredText] Enable syntax highlighting with Markdown.from markdown import markdown import docutils.core from docutils.writers.html4css1 import Writer as rst_html_writer class Renderer(object): extensions = [] @classmethod def render(cls, plain): return plain class Markdown(Renderer): extensions = ['markdown', 'mkd'] @classmethod def render(cls, plain): return markdown(plain, ['def_list', 'footnotes', 'codehilite']) class ReStructuredText(Renderer): extensions = ['rst'] @classmethod def render(cls, plain): w = rst_html_writer() return docutils.core.publish_parts(plain, writer=w)['body'] class Plain(Renderer): extensions = 'txt' @classmethod def render(cls, plain): return plain.replace('\n', '<br>') all = [Renderer, Plain, Markdown, ReStructuredText]
<commit_before>from markdown import markdown import docutils.core from docutils.writers.html4css1 import Writer as rst_html_writer class Renderer(object): extensions = [] @classmethod def render(cls, plain): return plain class Markdown(Renderer): extensions = ['markdown', 'mkd'] @classmethod def render(cls, plain): return markdown(plain, ['def_list', 'footnotes']) class ReStructuredText(Renderer): extensions = ['rst'] @classmethod def render(cls, plain): w = rst_html_writer() return docutils.core.publish_parts(plain, writer=w)['body'] class Plain(Renderer): extensions = 'txt' @classmethod def render(cls, plain): return plain.replace('\n', '<br>') all = [Renderer, Plain, Markdown, ReStructuredText] <commit_msg>Enable syntax highlighting with Markdown.<commit_after>from markdown import markdown import docutils.core from docutils.writers.html4css1 import Writer as rst_html_writer class Renderer(object): extensions = [] @classmethod def render(cls, plain): return plain class Markdown(Renderer): extensions = ['markdown', 'mkd'] @classmethod def render(cls, plain): return markdown(plain, ['def_list', 'footnotes', 'codehilite']) class ReStructuredText(Renderer): extensions = ['rst'] @classmethod def render(cls, plain): w = rst_html_writer() return docutils.core.publish_parts(plain, writer=w)['body'] class Plain(Renderer): extensions = 'txt' @classmethod def render(cls, plain): return plain.replace('\n', '<br>') all = [Renderer, Plain, Markdown, ReStructuredText]
7fd5db30d8ce8b6ecc8fdb87f695f484e5e9f38f
jacquard/wsgi.py
jacquard/wsgi.py
""" WSGI application target. This module is designed for use when running the Jacquard server from a WSGI web server such as waitress or gunicorn. `jacquard.wsgi` would be the module to target, picking up the WSGI application from `app`. In this case, the configuration file can be specified through the environment variable `JACQUARD_CONFIG`; if left unspecified, the file 'config.cfg' in the current working directory is assumed. """ from jacquard.config import load_config from jacquard.service import get_wsgi_app from .cli import DEFAULT_CONFIG_FILE_PATH app = get_wsgi_app(load_config(DEFAULT_CONFIG_FILE_PATH))
""" WSGI application target. This module is designed for use when running the Jacquard server from a WSGI web server such as waitress or gunicorn. `jacquard.wsgi` would be the module to target, picking up the WSGI application from `app`. In this case, the configuration file can be specified through the environment variable `JACQUARD_CONFIG`; if left unspecified, the file 'config.cfg' in the current working directory is assumed. """ import os import logging from jacquard.config import load_config from jacquard.service import get_wsgi_app from .cli import DEFAULT_CONFIG_FILE_PATH LOG_LEVEL = os.environ.get('JACQUARD_LOG_LEVEL', 'errors').lower() KNOWN_LOG_LEVELS = { 'debug': logging.DEBUG, 'info': logging.INFO, 'errors': logging.ERROR, } from jacquard.utils import check_keys check_keys((LOG_LEVEL,), KNOWN_LOG_LEVELS, RuntimeError) logging.basicConfig(level=KNOWN_LOG_LEVELS[LOG_LEVEL]) app = get_wsgi_app(load_config(DEFAULT_CONFIG_FILE_PATH))
Add environment variable for configuring logging
Add environment variable for configuring logging
Python
mit
prophile/jacquard,prophile/jacquard
""" WSGI application target. This module is designed for use when running the Jacquard server from a WSGI web server such as waitress or gunicorn. `jacquard.wsgi` would be the module to target, picking up the WSGI application from `app`. In this case, the configuration file can be specified through the environment variable `JACQUARD_CONFIG`; if left unspecified, the file 'config.cfg' in the current working directory is assumed. """ from jacquard.config import load_config from jacquard.service import get_wsgi_app from .cli import DEFAULT_CONFIG_FILE_PATH app = get_wsgi_app(load_config(DEFAULT_CONFIG_FILE_PATH)) Add environment variable for configuring logging
""" WSGI application target. This module is designed for use when running the Jacquard server from a WSGI web server such as waitress or gunicorn. `jacquard.wsgi` would be the module to target, picking up the WSGI application from `app`. In this case, the configuration file can be specified through the environment variable `JACQUARD_CONFIG`; if left unspecified, the file 'config.cfg' in the current working directory is assumed. """ import os import logging from jacquard.config import load_config from jacquard.service import get_wsgi_app from .cli import DEFAULT_CONFIG_FILE_PATH LOG_LEVEL = os.environ.get('JACQUARD_LOG_LEVEL', 'errors').lower() KNOWN_LOG_LEVELS = { 'debug': logging.DEBUG, 'info': logging.INFO, 'errors': logging.ERROR, } from jacquard.utils import check_keys check_keys((LOG_LEVEL,), KNOWN_LOG_LEVELS, RuntimeError) logging.basicConfig(level=KNOWN_LOG_LEVELS[LOG_LEVEL]) app = get_wsgi_app(load_config(DEFAULT_CONFIG_FILE_PATH))
<commit_before>""" WSGI application target. This module is designed for use when running the Jacquard server from a WSGI web server such as waitress or gunicorn. `jacquard.wsgi` would be the module to target, picking up the WSGI application from `app`. In this case, the configuration file can be specified through the environment variable `JACQUARD_CONFIG`; if left unspecified, the file 'config.cfg' in the current working directory is assumed. """ from jacquard.config import load_config from jacquard.service import get_wsgi_app from .cli import DEFAULT_CONFIG_FILE_PATH app = get_wsgi_app(load_config(DEFAULT_CONFIG_FILE_PATH)) <commit_msg>Add environment variable for configuring logging<commit_after>
""" WSGI application target. This module is designed for use when running the Jacquard server from a WSGI web server such as waitress or gunicorn. `jacquard.wsgi` would be the module to target, picking up the WSGI application from `app`. In this case, the configuration file can be specified through the environment variable `JACQUARD_CONFIG`; if left unspecified, the file 'config.cfg' in the current working directory is assumed. """ import os import logging from jacquard.config import load_config from jacquard.service import get_wsgi_app from .cli import DEFAULT_CONFIG_FILE_PATH LOG_LEVEL = os.environ.get('JACQUARD_LOG_LEVEL', 'errors').lower() KNOWN_LOG_LEVELS = { 'debug': logging.DEBUG, 'info': logging.INFO, 'errors': logging.ERROR, } from jacquard.utils import check_keys check_keys((LOG_LEVEL,), KNOWN_LOG_LEVELS, RuntimeError) logging.basicConfig(level=KNOWN_LOG_LEVELS[LOG_LEVEL]) app = get_wsgi_app(load_config(DEFAULT_CONFIG_FILE_PATH))
""" WSGI application target. This module is designed for use when running the Jacquard server from a WSGI web server such as waitress or gunicorn. `jacquard.wsgi` would be the module to target, picking up the WSGI application from `app`. In this case, the configuration file can be specified through the environment variable `JACQUARD_CONFIG`; if left unspecified, the file 'config.cfg' in the current working directory is assumed. """ from jacquard.config import load_config from jacquard.service import get_wsgi_app from .cli import DEFAULT_CONFIG_FILE_PATH app = get_wsgi_app(load_config(DEFAULT_CONFIG_FILE_PATH)) Add environment variable for configuring logging""" WSGI application target. This module is designed for use when running the Jacquard server from a WSGI web server such as waitress or gunicorn. `jacquard.wsgi` would be the module to target, picking up the WSGI application from `app`. In this case, the configuration file can be specified through the environment variable `JACQUARD_CONFIG`; if left unspecified, the file 'config.cfg' in the current working directory is assumed. """ import os import logging from jacquard.config import load_config from jacquard.service import get_wsgi_app from .cli import DEFAULT_CONFIG_FILE_PATH LOG_LEVEL = os.environ.get('JACQUARD_LOG_LEVEL', 'errors').lower() KNOWN_LOG_LEVELS = { 'debug': logging.DEBUG, 'info': logging.INFO, 'errors': logging.ERROR, } from jacquard.utils import check_keys check_keys((LOG_LEVEL,), KNOWN_LOG_LEVELS, RuntimeError) logging.basicConfig(level=KNOWN_LOG_LEVELS[LOG_LEVEL]) app = get_wsgi_app(load_config(DEFAULT_CONFIG_FILE_PATH))
<commit_before>""" WSGI application target. This module is designed for use when running the Jacquard server from a WSGI web server such as waitress or gunicorn. `jacquard.wsgi` would be the module to target, picking up the WSGI application from `app`. In this case, the configuration file can be specified through the environment variable `JACQUARD_CONFIG`; if left unspecified, the file 'config.cfg' in the current working directory is assumed. """ from jacquard.config import load_config from jacquard.service import get_wsgi_app from .cli import DEFAULT_CONFIG_FILE_PATH app = get_wsgi_app(load_config(DEFAULT_CONFIG_FILE_PATH)) <commit_msg>Add environment variable for configuring logging<commit_after>""" WSGI application target. This module is designed for use when running the Jacquard server from a WSGI web server such as waitress or gunicorn. `jacquard.wsgi` would be the module to target, picking up the WSGI application from `app`. In this case, the configuration file can be specified through the environment variable `JACQUARD_CONFIG`; if left unspecified, the file 'config.cfg' in the current working directory is assumed. """ import os import logging from jacquard.config import load_config from jacquard.service import get_wsgi_app from .cli import DEFAULT_CONFIG_FILE_PATH LOG_LEVEL = os.environ.get('JACQUARD_LOG_LEVEL', 'errors').lower() KNOWN_LOG_LEVELS = { 'debug': logging.DEBUG, 'info': logging.INFO, 'errors': logging.ERROR, } from jacquard.utils import check_keys check_keys((LOG_LEVEL,), KNOWN_LOG_LEVELS, RuntimeError) logging.basicConfig(level=KNOWN_LOG_LEVELS[LOG_LEVEL]) app = get_wsgi_app(load_config(DEFAULT_CONFIG_FILE_PATH))
eef8498388c672b25344a3f6fd8c05166e4ed4f6
xea_core/urls.py
xea_core/urls.py
"""xea_core URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.9/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'api/auth/jwt/', include('jwt_knox.urls')), url(r'^api/', include('api.urls')), ]
"""xea_core URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.9/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'api/auth/jwt/', include('jwt_knox.urls', namespace='jwt_knox')), url(r'^api/', include('api.urls')), ]
Add namespace to jwt_knox URLs
Add namespace to jwt_knox URLs
Python
agpl-3.0
gpul-org/xea-core
"""xea_core URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.9/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'api/auth/jwt/', include('jwt_knox.urls')), url(r'^api/', include('api.urls')), ] Add namespace to jwt_knox URLs
"""xea_core URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.9/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'api/auth/jwt/', include('jwt_knox.urls', namespace='jwt_knox')), url(r'^api/', include('api.urls')), ]
<commit_before>"""xea_core URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.9/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'api/auth/jwt/', include('jwt_knox.urls')), url(r'^api/', include('api.urls')), ] <commit_msg>Add namespace to jwt_knox URLs<commit_after>
"""xea_core URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.9/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'api/auth/jwt/', include('jwt_knox.urls', namespace='jwt_knox')), url(r'^api/', include('api.urls')), ]
"""xea_core URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.9/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'api/auth/jwt/', include('jwt_knox.urls')), url(r'^api/', include('api.urls')), ] Add namespace to jwt_knox URLs"""xea_core URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.9/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'api/auth/jwt/', include('jwt_knox.urls', namespace='jwt_knox')), url(r'^api/', include('api.urls')), ]
<commit_before>"""xea_core URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.9/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'api/auth/jwt/', include('jwt_knox.urls')), url(r'^api/', include('api.urls')), ] <commit_msg>Add namespace to jwt_knox URLs<commit_after>"""xea_core URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.9/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'api/auth/jwt/', include('jwt_knox.urls', namespace='jwt_knox')), url(r'^api/', include('api.urls')), ]
6cce924d849f209fb8251acd4c21d65ad0daa24c
plugins/basic_info_plugin.py
plugins/basic_info_plugin.py
import string import textwrap from veryprettytable import VeryPrettyTable from plugins import BasePlugin __author__ = 'peter' class BasicInfoPlugin(BasePlugin): short_description = 'List some basic info about the string in a table' header = 'Basic info:' default = True description = textwrap.dedent('''\ This plugin provides some basic info about the string such as: - Length - Presence of alpha/digits/raw bytes''') key = '--basic' def handle(self): table = VeryPrettyTable() table.field_names = ['String', 'Length', '# Digits', '# Alpha', '# Punct.', '# Control'] for s in self.args['STRING']: table.add_row((s, len(s), sum(x.isdigit() for x in s), sum(x.isalpha() for x in s), sum(x in string.punctuation for x in s), sum(x not in string.printable for x in s))) return str(table) + '\n'
import string import textwrap from veryprettytable import VeryPrettyTable from plugins import BasePlugin from plugins.util import green, red __author__ = 'peter' class BasicInfoPlugin(BasePlugin): short_description = 'List some basic info about the string in a table' header = 'Basic info:' default = True description = textwrap.dedent('''\ This plugin provides some basic info about the string such as: - Length - Presence of alpha/digits/raw bytes''') key = '--basic' def handle(self): table = VeryPrettyTable() table.field_names = ['String', 'Length', '# Digits', '# Alpha', '# Punct.', '# Control', 'Hex?'] for s in self.args['STRING']: table.add_row((s, len(s), sum(x.isdigit() for x in s), sum(x.isalpha() for x in s), sum(x in string.punctuation for x in s), sum(x not in string.printable for x in s), green('✔') if all(x in string.hexdigits for x in s) else red('✗'))) return str(table) + '\n'
Add hex check to basic info
Add hex check to basic info
Python
mit
Sakartu/stringinfo
import string import textwrap from veryprettytable import VeryPrettyTable from plugins import BasePlugin __author__ = 'peter' class BasicInfoPlugin(BasePlugin): short_description = 'List some basic info about the string in a table' header = 'Basic info:' default = True description = textwrap.dedent('''\ This plugin provides some basic info about the string such as: - Length - Presence of alpha/digits/raw bytes''') key = '--basic' def handle(self): table = VeryPrettyTable() table.field_names = ['String', 'Length', '# Digits', '# Alpha', '# Punct.', '# Control'] for s in self.args['STRING']: table.add_row((s, len(s), sum(x.isdigit() for x in s), sum(x.isalpha() for x in s), sum(x in string.punctuation for x in s), sum(x not in string.printable for x in s))) return str(table) + '\n'Add hex check to basic info
import string import textwrap from veryprettytable import VeryPrettyTable from plugins import BasePlugin from plugins.util import green, red __author__ = 'peter' class BasicInfoPlugin(BasePlugin): short_description = 'List some basic info about the string in a table' header = 'Basic info:' default = True description = textwrap.dedent('''\ This plugin provides some basic info about the string such as: - Length - Presence of alpha/digits/raw bytes''') key = '--basic' def handle(self): table = VeryPrettyTable() table.field_names = ['String', 'Length', '# Digits', '# Alpha', '# Punct.', '# Control', 'Hex?'] for s in self.args['STRING']: table.add_row((s, len(s), sum(x.isdigit() for x in s), sum(x.isalpha() for x in s), sum(x in string.punctuation for x in s), sum(x not in string.printable for x in s), green('✔') if all(x in string.hexdigits for x in s) else red('✗'))) return str(table) + '\n'
<commit_before>import string import textwrap from veryprettytable import VeryPrettyTable from plugins import BasePlugin __author__ = 'peter' class BasicInfoPlugin(BasePlugin): short_description = 'List some basic info about the string in a table' header = 'Basic info:' default = True description = textwrap.dedent('''\ This plugin provides some basic info about the string such as: - Length - Presence of alpha/digits/raw bytes''') key = '--basic' def handle(self): table = VeryPrettyTable() table.field_names = ['String', 'Length', '# Digits', '# Alpha', '# Punct.', '# Control'] for s in self.args['STRING']: table.add_row((s, len(s), sum(x.isdigit() for x in s), sum(x.isalpha() for x in s), sum(x in string.punctuation for x in s), sum(x not in string.printable for x in s))) return str(table) + '\n'<commit_msg>Add hex check to basic info<commit_after>
import string import textwrap from veryprettytable import VeryPrettyTable from plugins import BasePlugin from plugins.util import green, red __author__ = 'peter' class BasicInfoPlugin(BasePlugin): short_description = 'List some basic info about the string in a table' header = 'Basic info:' default = True description = textwrap.dedent('''\ This plugin provides some basic info about the string such as: - Length - Presence of alpha/digits/raw bytes''') key = '--basic' def handle(self): table = VeryPrettyTable() table.field_names = ['String', 'Length', '# Digits', '# Alpha', '# Punct.', '# Control', 'Hex?'] for s in self.args['STRING']: table.add_row((s, len(s), sum(x.isdigit() for x in s), sum(x.isalpha() for x in s), sum(x in string.punctuation for x in s), sum(x not in string.printable for x in s), green('✔') if all(x in string.hexdigits for x in s) else red('✗'))) return str(table) + '\n'
import string import textwrap from veryprettytable import VeryPrettyTable from plugins import BasePlugin __author__ = 'peter' class BasicInfoPlugin(BasePlugin): short_description = 'List some basic info about the string in a table' header = 'Basic info:' default = True description = textwrap.dedent('''\ This plugin provides some basic info about the string such as: - Length - Presence of alpha/digits/raw bytes''') key = '--basic' def handle(self): table = VeryPrettyTable() table.field_names = ['String', 'Length', '# Digits', '# Alpha', '# Punct.', '# Control'] for s in self.args['STRING']: table.add_row((s, len(s), sum(x.isdigit() for x in s), sum(x.isalpha() for x in s), sum(x in string.punctuation for x in s), sum(x not in string.printable for x in s))) return str(table) + '\n'Add hex check to basic infoimport string import textwrap from veryprettytable import VeryPrettyTable from plugins import BasePlugin from plugins.util import green, red __author__ = 'peter' class BasicInfoPlugin(BasePlugin): short_description = 'List some basic info about the string in a table' header = 'Basic info:' default = True description = textwrap.dedent('''\ This plugin provides some basic info about the string such as: - Length - Presence of alpha/digits/raw bytes''') key = '--basic' def handle(self): table = VeryPrettyTable() table.field_names = ['String', 'Length', '# Digits', '# Alpha', '# Punct.', '# Control', 'Hex?'] for s in self.args['STRING']: table.add_row((s, len(s), sum(x.isdigit() for x in s), sum(x.isalpha() for x in s), sum(x in string.punctuation for x in s), sum(x not in string.printable for x in s), green('✔') if all(x in string.hexdigits for x in s) else red('✗'))) return str(table) + '\n'
<commit_before>import string import textwrap from veryprettytable import VeryPrettyTable from plugins import BasePlugin __author__ = 'peter' class BasicInfoPlugin(BasePlugin): short_description = 'List some basic info about the string in a table' header = 'Basic info:' default = True description = textwrap.dedent('''\ This plugin provides some basic info about the string such as: - Length - Presence of alpha/digits/raw bytes''') key = '--basic' def handle(self): table = VeryPrettyTable() table.field_names = ['String', 'Length', '# Digits', '# Alpha', '# Punct.', '# Control'] for s in self.args['STRING']: table.add_row((s, len(s), sum(x.isdigit() for x in s), sum(x.isalpha() for x in s), sum(x in string.punctuation for x in s), sum(x not in string.printable for x in s))) return str(table) + '\n'<commit_msg>Add hex check to basic info<commit_after>import string import textwrap from veryprettytable import VeryPrettyTable from plugins import BasePlugin from plugins.util import green, red __author__ = 'peter' class BasicInfoPlugin(BasePlugin): short_description = 'List some basic info about the string in a table' header = 'Basic info:' default = True description = textwrap.dedent('''\ This plugin provides some basic info about the string such as: - Length - Presence of alpha/digits/raw bytes''') key = '--basic' def handle(self): table = VeryPrettyTable() table.field_names = ['String', 'Length', '# Digits', '# Alpha', '# Punct.', '# Control', 'Hex?'] for s in self.args['STRING']: table.add_row((s, len(s), sum(x.isdigit() for x in s), sum(x.isalpha() for x in s), sum(x in string.punctuation for x in s), sum(x not in string.printable for x in s), green('✔') if all(x in string.hexdigits for x in s) else red('✗'))) return str(table) + '\n'
1e82bf0e23775dfc2541dc8c2dec4447c20f2cda
tests/submitsaved.py
tests/submitsaved.py
#!/usr/bin/env python2 import pickle, sys sys.path.append('../') from lewas.parsers import UnitParser, AttrParser, field_rangler import lewas.datastores import lewas.models config = "../config" config = lewas.readConfig(config) lewas.datastores.submitRequest(pickle.load(open(sys.argv[1])), config, False)
#!/usr/bin/env python2 import pickle, sys sys.path.append('../') from lewas.parsers import UnitParser, AttrParser, field_rangler import lewas.datastores import lewas.models config = "../config" config = lewas.readConfig(config) for fn in sys.argv[1:]: lewas.datastores.submitRequest(pickle.load(open(fn)), config, False) print "processed", fn
Add capability to resubmit multiple files
Add capability to resubmit multiple files
Python
mit
LEWASatVT/lewas
#!/usr/bin/env python2 import pickle, sys sys.path.append('../') from lewas.parsers import UnitParser, AttrParser, field_rangler import lewas.datastores import lewas.models config = "../config" config = lewas.readConfig(config) lewas.datastores.submitRequest(pickle.load(open(sys.argv[1])), config, False) Add capability to resubmit multiple files
#!/usr/bin/env python2 import pickle, sys sys.path.append('../') from lewas.parsers import UnitParser, AttrParser, field_rangler import lewas.datastores import lewas.models config = "../config" config = lewas.readConfig(config) for fn in sys.argv[1:]: lewas.datastores.submitRequest(pickle.load(open(fn)), config, False) print "processed", fn
<commit_before>#!/usr/bin/env python2 import pickle, sys sys.path.append('../') from lewas.parsers import UnitParser, AttrParser, field_rangler import lewas.datastores import lewas.models config = "../config" config = lewas.readConfig(config) lewas.datastores.submitRequest(pickle.load(open(sys.argv[1])), config, False) <commit_msg>Add capability to resubmit multiple files<commit_after>
#!/usr/bin/env python2 import pickle, sys sys.path.append('../') from lewas.parsers import UnitParser, AttrParser, field_rangler import lewas.datastores import lewas.models config = "../config" config = lewas.readConfig(config) for fn in sys.argv[1:]: lewas.datastores.submitRequest(pickle.load(open(fn)), config, False) print "processed", fn
#!/usr/bin/env python2 import pickle, sys sys.path.append('../') from lewas.parsers import UnitParser, AttrParser, field_rangler import lewas.datastores import lewas.models config = "../config" config = lewas.readConfig(config) lewas.datastores.submitRequest(pickle.load(open(sys.argv[1])), config, False) Add capability to resubmit multiple files#!/usr/bin/env python2 import pickle, sys sys.path.append('../') from lewas.parsers import UnitParser, AttrParser, field_rangler import lewas.datastores import lewas.models config = "../config" config = lewas.readConfig(config) for fn in sys.argv[1:]: lewas.datastores.submitRequest(pickle.load(open(fn)), config, False) print "processed", fn
<commit_before>#!/usr/bin/env python2 import pickle, sys sys.path.append('../') from lewas.parsers import UnitParser, AttrParser, field_rangler import lewas.datastores import lewas.models config = "../config" config = lewas.readConfig(config) lewas.datastores.submitRequest(pickle.load(open(sys.argv[1])), config, False) <commit_msg>Add capability to resubmit multiple files<commit_after>#!/usr/bin/env python2 import pickle, sys sys.path.append('../') from lewas.parsers import UnitParser, AttrParser, field_rangler import lewas.datastores import lewas.models config = "../config" config = lewas.readConfig(config) for fn in sys.argv[1:]: lewas.datastores.submitRequest(pickle.load(open(fn)), config, False) print "processed", fn
d8a2eb66521f49c535d54cfa8a66965bc0fc8caf
tests/test_client.py
tests/test_client.py
from __future__ import unicode_literals import pytest from bugbuzz import BugBuzzClient @pytest.fixture def bugbuzz_client(base_url='http://localhost'): return BugBuzzClient(base_url) def test_random_access_key(): keys = set() for _ in range(100): client = bugbuzz_client() keys.add(client.aes_key) assert len(keys) == 100
from __future__ import unicode_literals import pytest from Crypto.Cipher import AES from bugbuzz import BugBuzzClient from bugbuzz import pkcs5_unpad @pytest.fixture def bugbuzz_client(base_url='http://localhost'): return BugBuzzClient(base_url) def test_random_access_key(): keys = set() for _ in range(100): client = bugbuzz_client() keys.add(client.aes_key) assert len(keys) == 100 def test_encrypt_decrypt(bugbuzz_client): plaintext = b'super foobar' iv, encrypted = bugbuzz_client.encrypt(plaintext) assert encrypted != plaintext aes = AES.new(bugbuzz_client.aes_key, AES.MODE_CBC, iv) assert pkcs5_unpad(aes.decrypt(encrypted)) == plaintext
Add test for encrypt and decrypt
Add test for encrypt and decrypt
Python
mit
victorlin/bugbuzz-python,victorlin/bugbuzz-python
from __future__ import unicode_literals import pytest from bugbuzz import BugBuzzClient @pytest.fixture def bugbuzz_client(base_url='http://localhost'): return BugBuzzClient(base_url) def test_random_access_key(): keys = set() for _ in range(100): client = bugbuzz_client() keys.add(client.aes_key) assert len(keys) == 100 Add test for encrypt and decrypt
from __future__ import unicode_literals import pytest from Crypto.Cipher import AES from bugbuzz import BugBuzzClient from bugbuzz import pkcs5_unpad @pytest.fixture def bugbuzz_client(base_url='http://localhost'): return BugBuzzClient(base_url) def test_random_access_key(): keys = set() for _ in range(100): client = bugbuzz_client() keys.add(client.aes_key) assert len(keys) == 100 def test_encrypt_decrypt(bugbuzz_client): plaintext = b'super foobar' iv, encrypted = bugbuzz_client.encrypt(plaintext) assert encrypted != plaintext aes = AES.new(bugbuzz_client.aes_key, AES.MODE_CBC, iv) assert pkcs5_unpad(aes.decrypt(encrypted)) == plaintext
<commit_before>from __future__ import unicode_literals import pytest from bugbuzz import BugBuzzClient @pytest.fixture def bugbuzz_client(base_url='http://localhost'): return BugBuzzClient(base_url) def test_random_access_key(): keys = set() for _ in range(100): client = bugbuzz_client() keys.add(client.aes_key) assert len(keys) == 100 <commit_msg>Add test for encrypt and decrypt<commit_after>
from __future__ import unicode_literals import pytest from Crypto.Cipher import AES from bugbuzz import BugBuzzClient from bugbuzz import pkcs5_unpad @pytest.fixture def bugbuzz_client(base_url='http://localhost'): return BugBuzzClient(base_url) def test_random_access_key(): keys = set() for _ in range(100): client = bugbuzz_client() keys.add(client.aes_key) assert len(keys) == 100 def test_encrypt_decrypt(bugbuzz_client): plaintext = b'super foobar' iv, encrypted = bugbuzz_client.encrypt(plaintext) assert encrypted != plaintext aes = AES.new(bugbuzz_client.aes_key, AES.MODE_CBC, iv) assert pkcs5_unpad(aes.decrypt(encrypted)) == plaintext
from __future__ import unicode_literals import pytest from bugbuzz import BugBuzzClient @pytest.fixture def bugbuzz_client(base_url='http://localhost'): return BugBuzzClient(base_url) def test_random_access_key(): keys = set() for _ in range(100): client = bugbuzz_client() keys.add(client.aes_key) assert len(keys) == 100 Add test for encrypt and decryptfrom __future__ import unicode_literals import pytest from Crypto.Cipher import AES from bugbuzz import BugBuzzClient from bugbuzz import pkcs5_unpad @pytest.fixture def bugbuzz_client(base_url='http://localhost'): return BugBuzzClient(base_url) def test_random_access_key(): keys = set() for _ in range(100): client = bugbuzz_client() keys.add(client.aes_key) assert len(keys) == 100 def test_encrypt_decrypt(bugbuzz_client): plaintext = b'super foobar' iv, encrypted = bugbuzz_client.encrypt(plaintext) assert encrypted != plaintext aes = AES.new(bugbuzz_client.aes_key, AES.MODE_CBC, iv) assert pkcs5_unpad(aes.decrypt(encrypted)) == plaintext
<commit_before>from __future__ import unicode_literals import pytest from bugbuzz import BugBuzzClient @pytest.fixture def bugbuzz_client(base_url='http://localhost'): return BugBuzzClient(base_url) def test_random_access_key(): keys = set() for _ in range(100): client = bugbuzz_client() keys.add(client.aes_key) assert len(keys) == 100 <commit_msg>Add test for encrypt and decrypt<commit_after>from __future__ import unicode_literals import pytest from Crypto.Cipher import AES from bugbuzz import BugBuzzClient from bugbuzz import pkcs5_unpad @pytest.fixture def bugbuzz_client(base_url='http://localhost'): return BugBuzzClient(base_url) def test_random_access_key(): keys = set() for _ in range(100): client = bugbuzz_client() keys.add(client.aes_key) assert len(keys) == 100 def test_encrypt_decrypt(bugbuzz_client): plaintext = b'super foobar' iv, encrypted = bugbuzz_client.encrypt(plaintext) assert encrypted != plaintext aes = AES.new(bugbuzz_client.aes_key, AES.MODE_CBC, iv) assert pkcs5_unpad(aes.decrypt(encrypted)) == plaintext
5d44e71da2835ff9cdeb9fa21d0fcd8645c4365a
fjord/urls.py
fjord/urls.py
from django.conf import settings from django.conf.urls.defaults import patterns, include from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch patch() from django.contrib import admin from adminplus import AdminSitePlus admin.site = AdminSitePlus() admin.autodiscover() urlpatterns = patterns('', (r'', include('fjord.analytics.urls')), (r'', include('fjord.base.urls')), (r'', include('fjord.feedback.urls')), # Generate a robots.txt (r'^robots\.txt$', lambda r: HttpResponse( ("User-agent: *\n%s: /" % 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'), mimetype="text/plain" ) ), (r'^browserid/', include('django_browserid.urls')), (r'^admin/', include(admin.site.urls)), ) # In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += staticfiles_urlpatterns()
from django.conf import settings from django.conf.urls.defaults import patterns, include from django.contrib.auth.decorators import login_required from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch patch() from django.contrib import admin from adminplus import AdminSitePlus admin.site = AdminSitePlus() admin.autodiscover() admin.site.login = login_required(admin.site.login) urlpatterns = patterns('', (r'', include('fjord.analytics.urls')), (r'', include('fjord.base.urls')), (r'', include('fjord.feedback.urls')), # Generate a robots.txt (r'^robots\.txt$', lambda r: HttpResponse( ("User-agent: *\n%s: /" % 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'), mimetype="text/plain" ) ), (r'^browserid/', include('django_browserid.urls')), (r'^admin/', include(admin.site.urls)), ) # In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += staticfiles_urlpatterns()
Revert "Nix helpful admin login requirement"
Revert "Nix helpful admin login requirement" This reverts commit 684dc38622d6cbe70879fb900ce5d73146a0cb40. We can put it back in because we're going to stick with LDAP basic auth.
Python
bsd-3-clause
lgp171188/fjord,rlr/fjord,DESHRAJ/fjord,hoosteeno/fjord,lgp171188/fjord,staranjeet/fjord,lgp171188/fjord,mozilla/fjord,Ritsyy/fjord,DESHRAJ/fjord,mozilla/fjord,mozilla/fjord,hoosteeno/fjord,rlr/fjord,mozilla/fjord,hoosteeno/fjord,rlr/fjord,staranjeet/fjord,Ritsyy/fjord,DESHRAJ/fjord,Ritsyy/fjord,hoosteeno/fjord,staranjeet/fjord,staranjeet/fjord,lgp171188/fjord,Ritsyy/fjord,rlr/fjord
from django.conf import settings from django.conf.urls.defaults import patterns, include from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch patch() from django.contrib import admin from adminplus import AdminSitePlus admin.site = AdminSitePlus() admin.autodiscover() urlpatterns = patterns('', (r'', include('fjord.analytics.urls')), (r'', include('fjord.base.urls')), (r'', include('fjord.feedback.urls')), # Generate a robots.txt (r'^robots\.txt$', lambda r: HttpResponse( ("User-agent: *\n%s: /" % 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'), mimetype="text/plain" ) ), (r'^browserid/', include('django_browserid.urls')), (r'^admin/', include(admin.site.urls)), ) # In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += staticfiles_urlpatterns() Revert "Nix helpful admin login requirement" This reverts commit 684dc38622d6cbe70879fb900ce5d73146a0cb40. We can put it back in because we're going to stick with LDAP basic auth.
from django.conf import settings from django.conf.urls.defaults import patterns, include from django.contrib.auth.decorators import login_required from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch patch() from django.contrib import admin from adminplus import AdminSitePlus admin.site = AdminSitePlus() admin.autodiscover() admin.site.login = login_required(admin.site.login) urlpatterns = patterns('', (r'', include('fjord.analytics.urls')), (r'', include('fjord.base.urls')), (r'', include('fjord.feedback.urls')), # Generate a robots.txt (r'^robots\.txt$', lambda r: HttpResponse( ("User-agent: *\n%s: /" % 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'), mimetype="text/plain" ) ), (r'^browserid/', include('django_browserid.urls')), (r'^admin/', include(admin.site.urls)), ) # In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += staticfiles_urlpatterns()
<commit_before>from django.conf import settings from django.conf.urls.defaults import patterns, include from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch patch() from django.contrib import admin from adminplus import AdminSitePlus admin.site = AdminSitePlus() admin.autodiscover() urlpatterns = patterns('', (r'', include('fjord.analytics.urls')), (r'', include('fjord.base.urls')), (r'', include('fjord.feedback.urls')), # Generate a robots.txt (r'^robots\.txt$', lambda r: HttpResponse( ("User-agent: *\n%s: /" % 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'), mimetype="text/plain" ) ), (r'^browserid/', include('django_browserid.urls')), (r'^admin/', include(admin.site.urls)), ) # In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += staticfiles_urlpatterns() <commit_msg>Revert "Nix helpful admin login requirement" This reverts commit 684dc38622d6cbe70879fb900ce5d73146a0cb40. We can put it back in because we're going to stick with LDAP basic auth.<commit_after>
from django.conf import settings from django.conf.urls.defaults import patterns, include from django.contrib.auth.decorators import login_required from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch patch() from django.contrib import admin from adminplus import AdminSitePlus admin.site = AdminSitePlus() admin.autodiscover() admin.site.login = login_required(admin.site.login) urlpatterns = patterns('', (r'', include('fjord.analytics.urls')), (r'', include('fjord.base.urls')), (r'', include('fjord.feedback.urls')), # Generate a robots.txt (r'^robots\.txt$', lambda r: HttpResponse( ("User-agent: *\n%s: /" % 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'), mimetype="text/plain" ) ), (r'^browserid/', include('django_browserid.urls')), (r'^admin/', include(admin.site.urls)), ) # In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += staticfiles_urlpatterns()
from django.conf import settings from django.conf.urls.defaults import patterns, include from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch patch() from django.contrib import admin from adminplus import AdminSitePlus admin.site = AdminSitePlus() admin.autodiscover() urlpatterns = patterns('', (r'', include('fjord.analytics.urls')), (r'', include('fjord.base.urls')), (r'', include('fjord.feedback.urls')), # Generate a robots.txt (r'^robots\.txt$', lambda r: HttpResponse( ("User-agent: *\n%s: /" % 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'), mimetype="text/plain" ) ), (r'^browserid/', include('django_browserid.urls')), (r'^admin/', include(admin.site.urls)), ) # In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += staticfiles_urlpatterns() Revert "Nix helpful admin login requirement" This reverts commit 684dc38622d6cbe70879fb900ce5d73146a0cb40. We can put it back in because we're going to stick with LDAP basic auth.from django.conf import settings from django.conf.urls.defaults import patterns, include from django.contrib.auth.decorators import login_required from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch patch() from django.contrib import admin from adminplus import AdminSitePlus admin.site = AdminSitePlus() admin.autodiscover() admin.site.login = login_required(admin.site.login) urlpatterns = patterns('', (r'', include('fjord.analytics.urls')), (r'', include('fjord.base.urls')), (r'', include('fjord.feedback.urls')), # Generate a robots.txt (r'^robots\.txt$', lambda r: HttpResponse( ("User-agent: *\n%s: /" % 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'), mimetype="text/plain" ) ), (r'^browserid/', include('django_browserid.urls')), (r'^admin/', include(admin.site.urls)), ) # In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += staticfiles_urlpatterns()
<commit_before>from django.conf import settings from django.conf.urls.defaults import patterns, include from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch patch() from django.contrib import admin from adminplus import AdminSitePlus admin.site = AdminSitePlus() admin.autodiscover() urlpatterns = patterns('', (r'', include('fjord.analytics.urls')), (r'', include('fjord.base.urls')), (r'', include('fjord.feedback.urls')), # Generate a robots.txt (r'^robots\.txt$', lambda r: HttpResponse( ("User-agent: *\n%s: /" % 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'), mimetype="text/plain" ) ), (r'^browserid/', include('django_browserid.urls')), (r'^admin/', include(admin.site.urls)), ) # In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += staticfiles_urlpatterns() <commit_msg>Revert "Nix helpful admin login requirement" This reverts commit 684dc38622d6cbe70879fb900ce5d73146a0cb40. We can put it back in because we're going to stick with LDAP basic auth.<commit_after>from django.conf import settings from django.conf.urls.defaults import patterns, include from django.contrib.auth.decorators import login_required from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch patch() from django.contrib import admin from adminplus import AdminSitePlus admin.site = AdminSitePlus() admin.autodiscover() admin.site.login = login_required(admin.site.login) urlpatterns = patterns('', (r'', include('fjord.analytics.urls')), (r'', include('fjord.base.urls')), (r'', include('fjord.feedback.urls')), # Generate a robots.txt (r'^robots\.txt$', lambda r: HttpResponse( ("User-agent: *\n%s: /" % 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'), mimetype="text/plain" ) ), (r'^browserid/', include('django_browserid.urls')), (r'^admin/', include(admin.site.urls)), ) # In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += staticfiles_urlpatterns()
904644ba540fbf6dcd47fb5a006d03a64d299fb2
src/yoi/pages.py
src/yoi/pages.py
from flask import request from flaskext.genshi import render_response from yoi.app import app @app.route('/') def get_index(): return render_response('index.html') @app.route('/home') def get_home(): return render_response('home.html') @app.route('/journal') def get_journal(): return render_response('journal.html') @app.route('/new-entry') def get_new_entry(): return render_response('new-entry.html')
from flask import request from flaskext.genshi import render_response from yoi.app import app @app.route('/') def index(): return render_response('index.html') @app.route('/home') def home(): return render_response('home.html') @app.route('/journal') def journal(): return render_response('journal.html') @app.route('/new-entry') def new_entry(): return render_response('new-entry.html')
Remove `get_` prefix on page functions.
Remove `get_` prefix on page functions.
Python
mit
doptio/you-owe-it,doptio/you-owe-it,doptio/you-owe-it,doptio/you-owe-it
from flask import request from flaskext.genshi import render_response from yoi.app import app @app.route('/') def get_index(): return render_response('index.html') @app.route('/home') def get_home(): return render_response('home.html') @app.route('/journal') def get_journal(): return render_response('journal.html') @app.route('/new-entry') def get_new_entry(): return render_response('new-entry.html') Remove `get_` prefix on page functions.
from flask import request from flaskext.genshi import render_response from yoi.app import app @app.route('/') def index(): return render_response('index.html') @app.route('/home') def home(): return render_response('home.html') @app.route('/journal') def journal(): return render_response('journal.html') @app.route('/new-entry') def new_entry(): return render_response('new-entry.html')
<commit_before>from flask import request from flaskext.genshi import render_response from yoi.app import app @app.route('/') def get_index(): return render_response('index.html') @app.route('/home') def get_home(): return render_response('home.html') @app.route('/journal') def get_journal(): return render_response('journal.html') @app.route('/new-entry') def get_new_entry(): return render_response('new-entry.html') <commit_msg>Remove `get_` prefix on page functions.<commit_after>
from flask import request from flaskext.genshi import render_response from yoi.app import app @app.route('/') def index(): return render_response('index.html') @app.route('/home') def home(): return render_response('home.html') @app.route('/journal') def journal(): return render_response('journal.html') @app.route('/new-entry') def new_entry(): return render_response('new-entry.html')
from flask import request from flaskext.genshi import render_response from yoi.app import app @app.route('/') def get_index(): return render_response('index.html') @app.route('/home') def get_home(): return render_response('home.html') @app.route('/journal') def get_journal(): return render_response('journal.html') @app.route('/new-entry') def get_new_entry(): return render_response('new-entry.html') Remove `get_` prefix on page functions.from flask import request from flaskext.genshi import render_response from yoi.app import app @app.route('/') def index(): return render_response('index.html') @app.route('/home') def home(): return render_response('home.html') @app.route('/journal') def journal(): return render_response('journal.html') @app.route('/new-entry') def new_entry(): return render_response('new-entry.html')
<commit_before>from flask import request from flaskext.genshi import render_response from yoi.app import app @app.route('/') def get_index(): return render_response('index.html') @app.route('/home') def get_home(): return render_response('home.html') @app.route('/journal') def get_journal(): return render_response('journal.html') @app.route('/new-entry') def get_new_entry(): return render_response('new-entry.html') <commit_msg>Remove `get_` prefix on page functions.<commit_after>from flask import request from flaskext.genshi import render_response from yoi.app import app @app.route('/') def index(): return render_response('index.html') @app.route('/home') def home(): return render_response('home.html') @app.route('/journal') def journal(): return render_response('journal.html') @app.route('/new-entry') def new_entry(): return render_response('new-entry.html')
8365945ef62b8f9cd37022302e8ee6299716720d
masterfirefoxos/base/helpers.py
masterfirefoxos/base/helpers.py
from feincms.templatetags.feincms_tags import feincms_render_region from jingo import register from jinja2 import Markup @register.function def render_region(feincms_page, region, request): return Markup(feincms_render_region(None, feincms_page, region, request))
from django.contrib.staticfiles.templatetags.staticfiles import static as static_helper from feincms.templatetags.feincms_tags import feincms_render_region from jingo import register from jinja2 import Markup static = register.function(static_helper) @register.function def render_region(feincms_page, region, request): return Markup(feincms_render_region(None, feincms_page, region, request))
Add static helper for jinja2
Add static helper for jinja2
Python
mpl-2.0
craigcook/masterfirefoxos,glogiotatidis/masterfirefoxos,mozilla/masterfirefoxos,craigcook/masterfirefoxos,enng0227/masterfirefoxos,glogiotatidis/masterfirefoxos,mozilla/masterfirefoxos,liu21st/masterfirefoxos,enng0227/masterfirefoxos,enng0227/masterfirefoxos,mozilla/masterfirefoxos,glogiotatidis/masterfirefoxos,glogiotatidis/masterfirefoxos,liu21st/masterfirefoxos,liu21st/masterfirefoxos,enng0227/masterfirefoxos,craigcook/masterfirefoxos,liu21st/masterfirefoxos,mozilla/masterfirefoxos,craigcook/masterfirefoxos
from feincms.templatetags.feincms_tags import feincms_render_region from jingo import register from jinja2 import Markup @register.function def render_region(feincms_page, region, request): return Markup(feincms_render_region(None, feincms_page, region, request)) Add static helper for jinja2
from django.contrib.staticfiles.templatetags.staticfiles import static as static_helper from feincms.templatetags.feincms_tags import feincms_render_region from jingo import register from jinja2 import Markup static = register.function(static_helper) @register.function def render_region(feincms_page, region, request): return Markup(feincms_render_region(None, feincms_page, region, request))
<commit_before>from feincms.templatetags.feincms_tags import feincms_render_region from jingo import register from jinja2 import Markup @register.function def render_region(feincms_page, region, request): return Markup(feincms_render_region(None, feincms_page, region, request)) <commit_msg>Add static helper for jinja2<commit_after>
from django.contrib.staticfiles.templatetags.staticfiles import static as static_helper from feincms.templatetags.feincms_tags import feincms_render_region from jingo import register from jinja2 import Markup static = register.function(static_helper) @register.function def render_region(feincms_page, region, request): return Markup(feincms_render_region(None, feincms_page, region, request))
from feincms.templatetags.feincms_tags import feincms_render_region from jingo import register from jinja2 import Markup @register.function def render_region(feincms_page, region, request): return Markup(feincms_render_region(None, feincms_page, region, request)) Add static helper for jinja2from django.contrib.staticfiles.templatetags.staticfiles import static as static_helper from feincms.templatetags.feincms_tags import feincms_render_region from jingo import register from jinja2 import Markup static = register.function(static_helper) @register.function def render_region(feincms_page, region, request): return Markup(feincms_render_region(None, feincms_page, region, request))
<commit_before>from feincms.templatetags.feincms_tags import feincms_render_region from jingo import register from jinja2 import Markup @register.function def render_region(feincms_page, region, request): return Markup(feincms_render_region(None, feincms_page, region, request)) <commit_msg>Add static helper for jinja2<commit_after>from django.contrib.staticfiles.templatetags.staticfiles import static as static_helper from feincms.templatetags.feincms_tags import feincms_render_region from jingo import register from jinja2 import Markup static = register.function(static_helper) @register.function def render_region(feincms_page, region, request): return Markup(feincms_render_region(None, feincms_page, region, request))
538a48b0ec6589d4d8fcf1c4253898a0fd3367e3
hello.py
hello.py
# Based on "Flask is Fun" demo from Flask documentation. import os from flask import Flask app = Flask(__name__) @app.route("/") def hello(): return "Hello World!" if __name__ == "__main__": app.config.update( SERVER_NAME='127.0.0.1:%d' % int(os.environ.get('PORT', 5000)), ) app.run()
# Based on "Flask is Fun" demo from Flask documentation. import os from flask import Flask app = Flask(__name__) @app.route("/") def hello(): return "Hello World!" if __name__ == "__main__": app.config.update( SERVER_NAME='127.0.0.1:%d' % int(os.environ.get('PORT', 5000)), ) app.run(host='0.0.0.0')
Add external access for demo process
Add external access for demo process Avoid using Lynx web browser on a Raspberry pi for reaching localhost on console :-)
Python
mit
smartmob-project/strawboss-demo,smartmob-project/strawboss-demo
# Based on "Flask is Fun" demo from Flask documentation. import os from flask import Flask app = Flask(__name__) @app.route("/") def hello(): return "Hello World!" if __name__ == "__main__": app.config.update( SERVER_NAME='127.0.0.1:%d' % int(os.environ.get('PORT', 5000)), ) app.run() Add external access for demo process Avoid using Lynx web browser on a Raspberry pi for reaching localhost on console :-)
# Based on "Flask is Fun" demo from Flask documentation. import os from flask import Flask app = Flask(__name__) @app.route("/") def hello(): return "Hello World!" if __name__ == "__main__": app.config.update( SERVER_NAME='127.0.0.1:%d' % int(os.environ.get('PORT', 5000)), ) app.run(host='0.0.0.0')
<commit_before># Based on "Flask is Fun" demo from Flask documentation. import os from flask import Flask app = Flask(__name__) @app.route("/") def hello(): return "Hello World!" if __name__ == "__main__": app.config.update( SERVER_NAME='127.0.0.1:%d' % int(os.environ.get('PORT', 5000)), ) app.run() <commit_msg>Add external access for demo process Avoid using Lynx web browser on a Raspberry pi for reaching localhost on console :-)<commit_after>
# Based on "Flask is Fun" demo from Flask documentation. import os from flask import Flask app = Flask(__name__) @app.route("/") def hello(): return "Hello World!" if __name__ == "__main__": app.config.update( SERVER_NAME='127.0.0.1:%d' % int(os.environ.get('PORT', 5000)), ) app.run(host='0.0.0.0')
# Based on "Flask is Fun" demo from Flask documentation. import os from flask import Flask app = Flask(__name__) @app.route("/") def hello(): return "Hello World!" if __name__ == "__main__": app.config.update( SERVER_NAME='127.0.0.1:%d' % int(os.environ.get('PORT', 5000)), ) app.run() Add external access for demo process Avoid using Lynx web browser on a Raspberry pi for reaching localhost on console :-)# Based on "Flask is Fun" demo from Flask documentation. import os from flask import Flask app = Flask(__name__) @app.route("/") def hello(): return "Hello World!" if __name__ == "__main__": app.config.update( SERVER_NAME='127.0.0.1:%d' % int(os.environ.get('PORT', 5000)), ) app.run(host='0.0.0.0')
<commit_before># Based on "Flask is Fun" demo from Flask documentation. import os from flask import Flask app = Flask(__name__) @app.route("/") def hello(): return "Hello World!" if __name__ == "__main__": app.config.update( SERVER_NAME='127.0.0.1:%d' % int(os.environ.get('PORT', 5000)), ) app.run() <commit_msg>Add external access for demo process Avoid using Lynx web browser on a Raspberry pi for reaching localhost on console :-)<commit_after># Based on "Flask is Fun" demo from Flask documentation. import os from flask import Flask app = Flask(__name__) @app.route("/") def hello(): return "Hello World!" if __name__ == "__main__": app.config.update( SERVER_NAME='127.0.0.1:%d' % int(os.environ.get('PORT', 5000)), ) app.run(host='0.0.0.0')
1598c699dc6bdf5d6edd700b70e11df207412dcd
hackernews.py
hackernews.py
import requests class HackerNews(): def __init__(self): self.url = 'https://hacker-news.firebaseio.com/v0/{uri}' def request(self, method, uri): url = self.url.format(uri=uri) return requests.request(method, url) def item(self, item_id): r = self.request('GET', 'item/{item_id}.json'.format(item_id=item_id)) return r.json() def user(self, user_id): r = self.request('GET', 'user/{user_id}.json'.format(user_id=user_id)) return r.json() def top_stories(self): r = self.request('GET', 'topstories.json') return r.json() def max_item(self): r = self.request('GET', 'maxitem.json') return r.json() def updates(self): r = self.request('GET', 'updates.json') return r.json()
from datetime import datetime import requests class HackerNews(): def __init__(self, timeout=5): self.url = 'https://hacker-news.firebaseio.com/v0/{uri}' self.timeout = timeout def request(self, method, uri): url = self.url.format(uri=uri) return requests.request(method, url, timeout=self.timeout) def item(self, item_id): r = self.request('GET', 'item/{item_id}.json'.format(item_id=item_id)) item = r.json() item['time'] = datetime.fromtimestamp(item['time']) return item def user(self, user_id): r = self.request('GET', 'user/{user_id}.json'.format(user_id=user_id)) user = r.json() user['created'] = datetime.fromtimestamp(user['created']) return user def top_stories(self): r = self.request('GET', 'topstories.json') return r.json() def max_item(self): r = self.request('GET', 'maxitem.json') return r.json() def updates(self): r = self.request('GET', 'updates.json') return r.json()
Convert timestamps to native datetime objects (breaking change)
Convert timestamps to native datetime objects (breaking change)
Python
mit
abrinsmead/hackernews-python
import requests class HackerNews(): def __init__(self): self.url = 'https://hacker-news.firebaseio.com/v0/{uri}' def request(self, method, uri): url = self.url.format(uri=uri) return requests.request(method, url) def item(self, item_id): r = self.request('GET', 'item/{item_id}.json'.format(item_id=item_id)) return r.json() def user(self, user_id): r = self.request('GET', 'user/{user_id}.json'.format(user_id=user_id)) return r.json() def top_stories(self): r = self.request('GET', 'topstories.json') return r.json() def max_item(self): r = self.request('GET', 'maxitem.json') return r.json() def updates(self): r = self.request('GET', 'updates.json') return r.json() Convert timestamps to native datetime objects (breaking change)
from datetime import datetime import requests class HackerNews(): def __init__(self, timeout=5): self.url = 'https://hacker-news.firebaseio.com/v0/{uri}' self.timeout = timeout def request(self, method, uri): url = self.url.format(uri=uri) return requests.request(method, url, timeout=self.timeout) def item(self, item_id): r = self.request('GET', 'item/{item_id}.json'.format(item_id=item_id)) item = r.json() item['time'] = datetime.fromtimestamp(item['time']) return item def user(self, user_id): r = self.request('GET', 'user/{user_id}.json'.format(user_id=user_id)) user = r.json() user['created'] = datetime.fromtimestamp(user['created']) return user def top_stories(self): r = self.request('GET', 'topstories.json') return r.json() def max_item(self): r = self.request('GET', 'maxitem.json') return r.json() def updates(self): r = self.request('GET', 'updates.json') return r.json()
<commit_before>import requests class HackerNews(): def __init__(self): self.url = 'https://hacker-news.firebaseio.com/v0/{uri}' def request(self, method, uri): url = self.url.format(uri=uri) return requests.request(method, url) def item(self, item_id): r = self.request('GET', 'item/{item_id}.json'.format(item_id=item_id)) return r.json() def user(self, user_id): r = self.request('GET', 'user/{user_id}.json'.format(user_id=user_id)) return r.json() def top_stories(self): r = self.request('GET', 'topstories.json') return r.json() def max_item(self): r = self.request('GET', 'maxitem.json') return r.json() def updates(self): r = self.request('GET', 'updates.json') return r.json() <commit_msg>Convert timestamps to native datetime objects (breaking change)<commit_after>
from datetime import datetime import requests class HackerNews(): def __init__(self, timeout=5): self.url = 'https://hacker-news.firebaseio.com/v0/{uri}' self.timeout = timeout def request(self, method, uri): url = self.url.format(uri=uri) return requests.request(method, url, timeout=self.timeout) def item(self, item_id): r = self.request('GET', 'item/{item_id}.json'.format(item_id=item_id)) item = r.json() item['time'] = datetime.fromtimestamp(item['time']) return item def user(self, user_id): r = self.request('GET', 'user/{user_id}.json'.format(user_id=user_id)) user = r.json() user['created'] = datetime.fromtimestamp(user['created']) return user def top_stories(self): r = self.request('GET', 'topstories.json') return r.json() def max_item(self): r = self.request('GET', 'maxitem.json') return r.json() def updates(self): r = self.request('GET', 'updates.json') return r.json()
import requests class HackerNews(): def __init__(self): self.url = 'https://hacker-news.firebaseio.com/v0/{uri}' def request(self, method, uri): url = self.url.format(uri=uri) return requests.request(method, url) def item(self, item_id): r = self.request('GET', 'item/{item_id}.json'.format(item_id=item_id)) return r.json() def user(self, user_id): r = self.request('GET', 'user/{user_id}.json'.format(user_id=user_id)) return r.json() def top_stories(self): r = self.request('GET', 'topstories.json') return r.json() def max_item(self): r = self.request('GET', 'maxitem.json') return r.json() def updates(self): r = self.request('GET', 'updates.json') return r.json() Convert timestamps to native datetime objects (breaking change)from datetime import datetime import requests class HackerNews(): def __init__(self, timeout=5): self.url = 'https://hacker-news.firebaseio.com/v0/{uri}' self.timeout = timeout def request(self, method, uri): url = self.url.format(uri=uri) return requests.request(method, url, timeout=self.timeout) def item(self, item_id): r = self.request('GET', 'item/{item_id}.json'.format(item_id=item_id)) item = r.json() item['time'] = datetime.fromtimestamp(item['time']) return item def user(self, user_id): r = self.request('GET', 'user/{user_id}.json'.format(user_id=user_id)) user = r.json() user['created'] = datetime.fromtimestamp(user['created']) return user def top_stories(self): r = self.request('GET', 'topstories.json') return r.json() def max_item(self): r = self.request('GET', 'maxitem.json') return r.json() def updates(self): r = self.request('GET', 'updates.json') return r.json()
<commit_before>import requests class HackerNews(): def __init__(self): self.url = 'https://hacker-news.firebaseio.com/v0/{uri}' def request(self, method, uri): url = self.url.format(uri=uri) return requests.request(method, url) def item(self, item_id): r = self.request('GET', 'item/{item_id}.json'.format(item_id=item_id)) return r.json() def user(self, user_id): r = self.request('GET', 'user/{user_id}.json'.format(user_id=user_id)) return r.json() def top_stories(self): r = self.request('GET', 'topstories.json') return r.json() def max_item(self): r = self.request('GET', 'maxitem.json') return r.json() def updates(self): r = self.request('GET', 'updates.json') return r.json() <commit_msg>Convert timestamps to native datetime objects (breaking change)<commit_after>from datetime import datetime import requests class HackerNews(): def __init__(self, timeout=5): self.url = 'https://hacker-news.firebaseio.com/v0/{uri}' self.timeout = timeout def request(self, method, uri): url = self.url.format(uri=uri) return requests.request(method, url, timeout=self.timeout) def item(self, item_id): r = self.request('GET', 'item/{item_id}.json'.format(item_id=item_id)) item = r.json() item['time'] = datetime.fromtimestamp(item['time']) return item def user(self, user_id): r = self.request('GET', 'user/{user_id}.json'.format(user_id=user_id)) user = r.json() user['created'] = datetime.fromtimestamp(user['created']) return user def top_stories(self): r = self.request('GET', 'topstories.json') return r.json() def max_item(self): r = self.request('GET', 'maxitem.json') return r.json() def updates(self): r = self.request('GET', 'updates.json') return r.json()
8b7a7e09a53d23b37266c67995756cf1dadf520d
conanfile.py
conanfile.py
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.1.2" class CMakeIncludeGuardConan(ConanFile): name = "cmake-include-guard" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) requires = ("cmake-module-common/master@smspillaz/cmake-module-common", ) generators = "cmake" url = "http://github.com/polysquare/cmake-include-guard" licence = "MIT" def source(self): zip_name = "cmake-include-guard.zip" download("https://github.com/polysquare/" "cmake-include-guard/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cmake-include-guard", src="cmake-include-guard-" + VERSION, keep_path=True)
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.1.2" class CMakeIncludeGuardConan(ConanFile): name = "cmake-include-guard" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" url = "http://github.com/polysquare/cmake-include-guard" licence = "MIT" options = { "dev": [True, False] } default_options = "dev=False" def requirements(self): if self.options.dev: self.requires("cmake-module-common/master@smspillaz/cmake-module-common") def source(self): zip_name = "cmake-include-guard.zip" download("https://github.com/polysquare/" "cmake-include-guard/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cmake-include-guard", src="cmake-include-guard-" + VERSION, keep_path=True)
Make cmake-module-common a dev-only requirement
conan: Make cmake-module-common a dev-only requirement
Python
mit
polysquare/cmake-include-guard
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.1.2" class CMakeIncludeGuardConan(ConanFile): name = "cmake-include-guard" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) requires = ("cmake-module-common/master@smspillaz/cmake-module-common", ) generators = "cmake" url = "http://github.com/polysquare/cmake-include-guard" licence = "MIT" def source(self): zip_name = "cmake-include-guard.zip" download("https://github.com/polysquare/" "cmake-include-guard/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cmake-include-guard", src="cmake-include-guard-" + VERSION, keep_path=True) conan: Make cmake-module-common a dev-only requirement
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.1.2" class CMakeIncludeGuardConan(ConanFile): name = "cmake-include-guard" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" url = "http://github.com/polysquare/cmake-include-guard" licence = "MIT" options = { "dev": [True, False] } default_options = "dev=False" def requirements(self): if self.options.dev: self.requires("cmake-module-common/master@smspillaz/cmake-module-common") def source(self): zip_name = "cmake-include-guard.zip" download("https://github.com/polysquare/" "cmake-include-guard/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cmake-include-guard", src="cmake-include-guard-" + VERSION, keep_path=True)
<commit_before>from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.1.2" class CMakeIncludeGuardConan(ConanFile): name = "cmake-include-guard" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) requires = ("cmake-module-common/master@smspillaz/cmake-module-common", ) generators = "cmake" url = "http://github.com/polysquare/cmake-include-guard" licence = "MIT" def source(self): zip_name = "cmake-include-guard.zip" download("https://github.com/polysquare/" "cmake-include-guard/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cmake-include-guard", src="cmake-include-guard-" + VERSION, keep_path=True) <commit_msg>conan: Make cmake-module-common a dev-only requirement<commit_after>
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.1.2" class CMakeIncludeGuardConan(ConanFile): name = "cmake-include-guard" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" url = "http://github.com/polysquare/cmake-include-guard" licence = "MIT" options = { "dev": [True, False] } default_options = "dev=False" def requirements(self): if self.options.dev: self.requires("cmake-module-common/master@smspillaz/cmake-module-common") def source(self): zip_name = "cmake-include-guard.zip" download("https://github.com/polysquare/" "cmake-include-guard/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cmake-include-guard", src="cmake-include-guard-" + VERSION, keep_path=True)
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.1.2" class CMakeIncludeGuardConan(ConanFile): name = "cmake-include-guard" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) requires = ("cmake-module-common/master@smspillaz/cmake-module-common", ) generators = "cmake" url = "http://github.com/polysquare/cmake-include-guard" licence = "MIT" def source(self): zip_name = "cmake-include-guard.zip" download("https://github.com/polysquare/" "cmake-include-guard/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cmake-include-guard", src="cmake-include-guard-" + VERSION, keep_path=True) conan: Make cmake-module-common a dev-only requirementfrom conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.1.2" class CMakeIncludeGuardConan(ConanFile): name = "cmake-include-guard" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" url = "http://github.com/polysquare/cmake-include-guard" licence = "MIT" options = { "dev": [True, False] } default_options = "dev=False" def requirements(self): if self.options.dev: self.requires("cmake-module-common/master@smspillaz/cmake-module-common") def source(self): zip_name = "cmake-include-guard.zip" download("https://github.com/polysquare/" "cmake-include-guard/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cmake-include-guard", src="cmake-include-guard-" + VERSION, keep_path=True)
<commit_before>from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.1.2" class CMakeIncludeGuardConan(ConanFile): name = "cmake-include-guard" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) requires = ("cmake-module-common/master@smspillaz/cmake-module-common", ) generators = "cmake" url = "http://github.com/polysquare/cmake-include-guard" licence = "MIT" def source(self): zip_name = "cmake-include-guard.zip" download("https://github.com/polysquare/" "cmake-include-guard/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cmake-include-guard", src="cmake-include-guard-" + VERSION, keep_path=True) <commit_msg>conan: Make cmake-module-common a dev-only requirement<commit_after>from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.1.2" class CMakeIncludeGuardConan(ConanFile): name = "cmake-include-guard" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" url = "http://github.com/polysquare/cmake-include-guard" licence = "MIT" options = { "dev": [True, False] } default_options = "dev=False" def requirements(self): if self.options.dev: self.requires("cmake-module-common/master@smspillaz/cmake-module-common") def source(self): zip_name = "cmake-include-guard.zip" download("https://github.com/polysquare/" "cmake-include-guard/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cmake-include-guard", src="cmake-include-guard-" + VERSION, keep_path=True)
70686bc62b85d36894b6e7909ceaa25d122ffe95
users/serializers.py
users/serializers.py
from rest_framework import serializers from django.contrib.auth.models import User, Group class GroupSerializer(serializers.ModelSerializer): class Meta: model = Group fields = ('name',) class UserSerializer(serializers.ModelSerializer): groups_complete = GroupSerializer(source='groups', read_only=True, many=True) class Meta: model = User fields = ('id', 'username', 'password', 'first_name', 'last_name', 'groups', 'groups_complete') extra_kwargs = { 'username': {'required': True}, 'password': {'required': True} }
from rest_framework import serializers from django.contrib.auth.models import User, Group class GroupSerializer(serializers.ModelSerializer): class Meta: model = Group fields = ('name',) class UserSerializer(serializers.ModelSerializer): class Meta: model = User fields = ('id', 'username', 'password', 'first_name', 'last_name', 'groups') extra_kwargs = { 'username': {'required': True}, 'password': {'required': False} }
Remove nested serializer for groups, make password optional in serializer for user edition
Remove nested serializer for groups, make password optional in serializer for user edition
Python
mit
fernandolobato/balarco,fernandolobato/balarco,fernandolobato/balarco
from rest_framework import serializers from django.contrib.auth.models import User, Group class GroupSerializer(serializers.ModelSerializer): class Meta: model = Group fields = ('name',) class UserSerializer(serializers.ModelSerializer): groups_complete = GroupSerializer(source='groups', read_only=True, many=True) class Meta: model = User fields = ('id', 'username', 'password', 'first_name', 'last_name', 'groups', 'groups_complete') extra_kwargs = { 'username': {'required': True}, 'password': {'required': True} } Remove nested serializer for groups, make password optional in serializer for user edition
from rest_framework import serializers from django.contrib.auth.models import User, Group class GroupSerializer(serializers.ModelSerializer): class Meta: model = Group fields = ('name',) class UserSerializer(serializers.ModelSerializer): class Meta: model = User fields = ('id', 'username', 'password', 'first_name', 'last_name', 'groups') extra_kwargs = { 'username': {'required': True}, 'password': {'required': False} }
<commit_before>from rest_framework import serializers from django.contrib.auth.models import User, Group class GroupSerializer(serializers.ModelSerializer): class Meta: model = Group fields = ('name',) class UserSerializer(serializers.ModelSerializer): groups_complete = GroupSerializer(source='groups', read_only=True, many=True) class Meta: model = User fields = ('id', 'username', 'password', 'first_name', 'last_name', 'groups', 'groups_complete') extra_kwargs = { 'username': {'required': True}, 'password': {'required': True} } <commit_msg>Remove nested serializer for groups, make password optional in serializer for user edition<commit_after>
from rest_framework import serializers from django.contrib.auth.models import User, Group class GroupSerializer(serializers.ModelSerializer): class Meta: model = Group fields = ('name',) class UserSerializer(serializers.ModelSerializer): class Meta: model = User fields = ('id', 'username', 'password', 'first_name', 'last_name', 'groups') extra_kwargs = { 'username': {'required': True}, 'password': {'required': False} }
from rest_framework import serializers from django.contrib.auth.models import User, Group class GroupSerializer(serializers.ModelSerializer): class Meta: model = Group fields = ('name',) class UserSerializer(serializers.ModelSerializer): groups_complete = GroupSerializer(source='groups', read_only=True, many=True) class Meta: model = User fields = ('id', 'username', 'password', 'first_name', 'last_name', 'groups', 'groups_complete') extra_kwargs = { 'username': {'required': True}, 'password': {'required': True} } Remove nested serializer for groups, make password optional in serializer for user editionfrom rest_framework import serializers from django.contrib.auth.models import User, Group class GroupSerializer(serializers.ModelSerializer): class Meta: model = Group fields = ('name',) class UserSerializer(serializers.ModelSerializer): class Meta: model = User fields = ('id', 'username', 'password', 'first_name', 'last_name', 'groups') extra_kwargs = { 'username': {'required': True}, 'password': {'required': False} }
<commit_before>from rest_framework import serializers from django.contrib.auth.models import User, Group class GroupSerializer(serializers.ModelSerializer): class Meta: model = Group fields = ('name',) class UserSerializer(serializers.ModelSerializer): groups_complete = GroupSerializer(source='groups', read_only=True, many=True) class Meta: model = User fields = ('id', 'username', 'password', 'first_name', 'last_name', 'groups', 'groups_complete') extra_kwargs = { 'username': {'required': True}, 'password': {'required': True} } <commit_msg>Remove nested serializer for groups, make password optional in serializer for user edition<commit_after>from rest_framework import serializers from django.contrib.auth.models import User, Group class GroupSerializer(serializers.ModelSerializer): class Meta: model = Group fields = ('name',) class UserSerializer(serializers.ModelSerializer): class Meta: model = User fields = ('id', 'username', 'password', 'first_name', 'last_name', 'groups') extra_kwargs = { 'username': {'required': True}, 'password': {'required': False} }
7d266a2e0f9d636272f60636d79703166b6377d2
tests/test_exceptions.py
tests/test_exceptions.py
# Tests for SecretStorage # Author: Dmitry Shachnev, 2013 # License: BSD # Various exception tests import unittest import secretstorage from secretstorage.exceptions import ItemNotFoundException class ExceptionsTest(unittest.TestCase): """A test case that ensures that all SecretStorage exceptions are raised correctly.""" @classmethod def setUpClass(cls): cls.bus = secretstorage.dbus_init(main_loop=False) cls.collection = secretstorage.Collection(cls.bus) def test_double_deleting(self): item = self.collection.create_item('MyItem', {'application': 'secretstorage-test'}, b'pa$$word') item.delete() self.assertRaises(ItemNotFoundException, item.delete) def test_non_existing_item(self): self.assertRaises(ItemNotFoundException, secretstorage.Item, self.bus, '/not/existing/path') def test_non_existing_collection(self): self.assertRaises(ItemNotFoundException, secretstorage.get_collection_by_alias, self.bus, 'non-existing-alias') if __name__ == '__main__': unittest.main()
# Tests for SecretStorage # Author: Dmitry Shachnev, 2013 # License: BSD # Various exception tests import unittest import secretstorage from secretstorage.exceptions import ItemNotFoundException class ExceptionsTest(unittest.TestCase): """A test case that ensures that all SecretStorage exceptions are raised correctly.""" @classmethod def setUpClass(cls): cls.bus = secretstorage.dbus_init(main_loop=False) cls.collection = secretstorage.Collection(cls.bus) def test_double_deleting(self): item = self.collection.create_item('MyItem', {'application': 'secretstorage-test'}, b'pa$$word') item.delete() self.assertRaises(ItemNotFoundException, item.delete) def test_non_existing_item(self): self.assertRaises(ItemNotFoundException, secretstorage.Item, self.bus, '/not/existing/path') def test_non_existing_collection(self): self.assertRaises(ItemNotFoundException, secretstorage.get_collection_by_alias, self.bus, 'non-existing-alias') if __name__ == '__main__': unittest.main()
Remove unwanted whitespace in tests
Remove unwanted whitespace in tests
Python
bsd-3-clause
mitya57/secretstorage
# Tests for SecretStorage # Author: Dmitry Shachnev, 2013 # License: BSD # Various exception tests import unittest import secretstorage from secretstorage.exceptions import ItemNotFoundException class ExceptionsTest(unittest.TestCase): """A test case that ensures that all SecretStorage exceptions are raised correctly.""" @classmethod def setUpClass(cls): cls.bus = secretstorage.dbus_init(main_loop=False) cls.collection = secretstorage.Collection(cls.bus) def test_double_deleting(self): item = self.collection.create_item('MyItem', {'application': 'secretstorage-test'}, b'pa$$word') item.delete() self.assertRaises(ItemNotFoundException, item.delete) def test_non_existing_item(self): self.assertRaises(ItemNotFoundException, secretstorage.Item, self.bus, '/not/existing/path') def test_non_existing_collection(self): self.assertRaises(ItemNotFoundException, secretstorage.get_collection_by_alias, self.bus, 'non-existing-alias') if __name__ == '__main__': unittest.main() Remove unwanted whitespace in tests
# Tests for SecretStorage # Author: Dmitry Shachnev, 2013 # License: BSD # Various exception tests import unittest import secretstorage from secretstorage.exceptions import ItemNotFoundException class ExceptionsTest(unittest.TestCase): """A test case that ensures that all SecretStorage exceptions are raised correctly.""" @classmethod def setUpClass(cls): cls.bus = secretstorage.dbus_init(main_loop=False) cls.collection = secretstorage.Collection(cls.bus) def test_double_deleting(self): item = self.collection.create_item('MyItem', {'application': 'secretstorage-test'}, b'pa$$word') item.delete() self.assertRaises(ItemNotFoundException, item.delete) def test_non_existing_item(self): self.assertRaises(ItemNotFoundException, secretstorage.Item, self.bus, '/not/existing/path') def test_non_existing_collection(self): self.assertRaises(ItemNotFoundException, secretstorage.get_collection_by_alias, self.bus, 'non-existing-alias') if __name__ == '__main__': unittest.main()
<commit_before># Tests for SecretStorage # Author: Dmitry Shachnev, 2013 # License: BSD # Various exception tests import unittest import secretstorage from secretstorage.exceptions import ItemNotFoundException class ExceptionsTest(unittest.TestCase): """A test case that ensures that all SecretStorage exceptions are raised correctly.""" @classmethod def setUpClass(cls): cls.bus = secretstorage.dbus_init(main_loop=False) cls.collection = secretstorage.Collection(cls.bus) def test_double_deleting(self): item = self.collection.create_item('MyItem', {'application': 'secretstorage-test'}, b'pa$$word') item.delete() self.assertRaises(ItemNotFoundException, item.delete) def test_non_existing_item(self): self.assertRaises(ItemNotFoundException, secretstorage.Item, self.bus, '/not/existing/path') def test_non_existing_collection(self): self.assertRaises(ItemNotFoundException, secretstorage.get_collection_by_alias, self.bus, 'non-existing-alias') if __name__ == '__main__': unittest.main() <commit_msg>Remove unwanted whitespace in tests<commit_after>
# Tests for SecretStorage # Author: Dmitry Shachnev, 2013 # License: BSD # Various exception tests import unittest import secretstorage from secretstorage.exceptions import ItemNotFoundException class ExceptionsTest(unittest.TestCase): """A test case that ensures that all SecretStorage exceptions are raised correctly.""" @classmethod def setUpClass(cls): cls.bus = secretstorage.dbus_init(main_loop=False) cls.collection = secretstorage.Collection(cls.bus) def test_double_deleting(self): item = self.collection.create_item('MyItem', {'application': 'secretstorage-test'}, b'pa$$word') item.delete() self.assertRaises(ItemNotFoundException, item.delete) def test_non_existing_item(self): self.assertRaises(ItemNotFoundException, secretstorage.Item, self.bus, '/not/existing/path') def test_non_existing_collection(self): self.assertRaises(ItemNotFoundException, secretstorage.get_collection_by_alias, self.bus, 'non-existing-alias') if __name__ == '__main__': unittest.main()
# Tests for SecretStorage # Author: Dmitry Shachnev, 2013 # License: BSD # Various exception tests import unittest import secretstorage from secretstorage.exceptions import ItemNotFoundException class ExceptionsTest(unittest.TestCase): """A test case that ensures that all SecretStorage exceptions are raised correctly.""" @classmethod def setUpClass(cls): cls.bus = secretstorage.dbus_init(main_loop=False) cls.collection = secretstorage.Collection(cls.bus) def test_double_deleting(self): item = self.collection.create_item('MyItem', {'application': 'secretstorage-test'}, b'pa$$word') item.delete() self.assertRaises(ItemNotFoundException, item.delete) def test_non_existing_item(self): self.assertRaises(ItemNotFoundException, secretstorage.Item, self.bus, '/not/existing/path') def test_non_existing_collection(self): self.assertRaises(ItemNotFoundException, secretstorage.get_collection_by_alias, self.bus, 'non-existing-alias') if __name__ == '__main__': unittest.main() Remove unwanted whitespace in tests# Tests for SecretStorage # Author: Dmitry Shachnev, 2013 # License: BSD # Various exception tests import unittest import secretstorage from secretstorage.exceptions import ItemNotFoundException class ExceptionsTest(unittest.TestCase): """A test case that ensures that all SecretStorage exceptions are raised correctly.""" @classmethod def setUpClass(cls): cls.bus = secretstorage.dbus_init(main_loop=False) cls.collection = secretstorage.Collection(cls.bus) def test_double_deleting(self): item = self.collection.create_item('MyItem', {'application': 'secretstorage-test'}, b'pa$$word') item.delete() self.assertRaises(ItemNotFoundException, item.delete) def test_non_existing_item(self): self.assertRaises(ItemNotFoundException, secretstorage.Item, self.bus, '/not/existing/path') def test_non_existing_collection(self): self.assertRaises(ItemNotFoundException, secretstorage.get_collection_by_alias, self.bus, 'non-existing-alias') if __name__ == '__main__': unittest.main()
<commit_before># Tests for SecretStorage # Author: Dmitry Shachnev, 2013 # License: BSD # Various exception tests import unittest import secretstorage from secretstorage.exceptions import ItemNotFoundException class ExceptionsTest(unittest.TestCase): """A test case that ensures that all SecretStorage exceptions are raised correctly.""" @classmethod def setUpClass(cls): cls.bus = secretstorage.dbus_init(main_loop=False) cls.collection = secretstorage.Collection(cls.bus) def test_double_deleting(self): item = self.collection.create_item('MyItem', {'application': 'secretstorage-test'}, b'pa$$word') item.delete() self.assertRaises(ItemNotFoundException, item.delete) def test_non_existing_item(self): self.assertRaises(ItemNotFoundException, secretstorage.Item, self.bus, '/not/existing/path') def test_non_existing_collection(self): self.assertRaises(ItemNotFoundException, secretstorage.get_collection_by_alias, self.bus, 'non-existing-alias') if __name__ == '__main__': unittest.main() <commit_msg>Remove unwanted whitespace in tests<commit_after># Tests for SecretStorage # Author: Dmitry Shachnev, 2013 # License: BSD # Various exception tests import unittest import secretstorage from secretstorage.exceptions import ItemNotFoundException class ExceptionsTest(unittest.TestCase): """A test case that ensures that all SecretStorage exceptions are raised correctly.""" @classmethod def setUpClass(cls): cls.bus = secretstorage.dbus_init(main_loop=False) cls.collection = secretstorage.Collection(cls.bus) def test_double_deleting(self): item = self.collection.create_item('MyItem', {'application': 'secretstorage-test'}, b'pa$$word') item.delete() self.assertRaises(ItemNotFoundException, item.delete) def test_non_existing_item(self): self.assertRaises(ItemNotFoundException, secretstorage.Item, self.bus, '/not/existing/path') def test_non_existing_collection(self): self.assertRaises(ItemNotFoundException, secretstorage.get_collection_by_alias, self.bus, 'non-existing-alias') if __name__ == '__main__': unittest.main()
c86ccf75fdf49115697548bb046bc83f348a1aba
tests/test_exceptions.py
tests/test_exceptions.py
from framewirc import exceptions class MissingAttributesTest: def test_message(self): attrs = ['some', 'attrs'] expected = 'Required attribute(s) missing: {}'.format(attrs) exception = exceptions.MissingAttributes(attrs) assert str(exception) == expected
from framewirc import exceptions def test_message(): attrs = ['some', 'attrs'] expected = 'Required attribute(s) missing: {}'.format(attrs) exception = exceptions.MissingAttributes(attrs) assert str(exception) == expected
Fix test that wasn't running
Fix test that wasn't running
Python
bsd-2-clause
meshy/framewirc
from framewirc import exceptions class MissingAttributesTest: def test_message(self): attrs = ['some', 'attrs'] expected = 'Required attribute(s) missing: {}'.format(attrs) exception = exceptions.MissingAttributes(attrs) assert str(exception) == expected Fix test that wasn't running
from framewirc import exceptions def test_message(): attrs = ['some', 'attrs'] expected = 'Required attribute(s) missing: {}'.format(attrs) exception = exceptions.MissingAttributes(attrs) assert str(exception) == expected
<commit_before>from framewirc import exceptions class MissingAttributesTest: def test_message(self): attrs = ['some', 'attrs'] expected = 'Required attribute(s) missing: {}'.format(attrs) exception = exceptions.MissingAttributes(attrs) assert str(exception) == expected <commit_msg>Fix test that wasn't running<commit_after>
from framewirc import exceptions def test_message(): attrs = ['some', 'attrs'] expected = 'Required attribute(s) missing: {}'.format(attrs) exception = exceptions.MissingAttributes(attrs) assert str(exception) == expected
from framewirc import exceptions class MissingAttributesTest: def test_message(self): attrs = ['some', 'attrs'] expected = 'Required attribute(s) missing: {}'.format(attrs) exception = exceptions.MissingAttributes(attrs) assert str(exception) == expected Fix test that wasn't runningfrom framewirc import exceptions def test_message(): attrs = ['some', 'attrs'] expected = 'Required attribute(s) missing: {}'.format(attrs) exception = exceptions.MissingAttributes(attrs) assert str(exception) == expected
<commit_before>from framewirc import exceptions class MissingAttributesTest: def test_message(self): attrs = ['some', 'attrs'] expected = 'Required attribute(s) missing: {}'.format(attrs) exception = exceptions.MissingAttributes(attrs) assert str(exception) == expected <commit_msg>Fix test that wasn't running<commit_after>from framewirc import exceptions def test_message(): attrs = ['some', 'attrs'] expected = 'Required attribute(s) missing: {}'.format(attrs) exception = exceptions.MissingAttributes(attrs) assert str(exception) == expected
4b926ab5de00bc4885021fff0b2ac3679703707a
nova/db/sqlalchemy/migrate_repo/versions/034_change_instance_id_in_migrations.py
nova/db/sqlalchemy/migrate_repo/versions/034_change_instance_id_in_migrations.py
# vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License.from sqlalchemy import * from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id)
# vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License.from sqlalchemy import * from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) if migrate_engine.name == "mysql": migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \ "`migrations_ibfk_1`;") migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id)
Drop FK before dropping instance_id column.
Drop FK before dropping instance_id column.
Python
apache-2.0
klmitch/nova,hanlind/nova,aristanetworks/arista-ovs-nova,fajoy/nova,dawnpower/nova,gooddata/openstack-nova,joker946/nova,shail2810/nova,orbitfp7/nova,jeffrey4l/nova,zaina/nova,tianweizhang/nova,ruslanloman/nova,mgagne/nova,yosshy/nova,cloudbase/nova,paulmathews/nova,JioCloud/nova_test_latest,savi-dev/nova,russellb/nova,nikesh-mahalka/nova,BeyondTheClouds/nova,scripnichenko/nova,leilihh/nova,gooddata/openstack-nova,vmturbo/nova,JianyuWang/nova,badock/nova,tudorvio/nova,CloudServer/nova,TieWei/nova,sileht/deb-openstack-nova,MountainWei/nova,fnordahl/nova,mmnelemane/nova,psiwczak/openstack,bclau/nova,Stavitsky/nova,NeCTAR-RC/nova,viggates/nova,jianghuaw/nova,aristanetworks/arista-ovs-nova,CiscoSystems/nova,yosshy/nova,cloudbase/nova-virtualbox,affo/nova,dstroppa/openstack-smartos-nova-grizzly,akash1808/nova,yrobla/nova,ruslanloman/nova,openstack/nova,vladikr/nova_drafts,maoy/zknova,bigswitch/nova,apporc/nova,barnsnake351/nova,dims/nova,salv-orlando/MyRepo,spring-week-topos/nova-week,devendermishrajio/nova_test_latest,edulramirez/nova,saleemjaveds/https-github.com-openstack-nova,Triv90/Nova,dims/nova,paulmathews/nova,houshengbo/nova_vmware_compute_driver,angdraug/nova,vmturbo/nova,watonyweng/nova,usc-isi/nova,gooddata/openstack-nova,sridevikoushik31/openstack,NoBodyCam/TftpPxeBootBareMetal,usc-isi/extra-specs,SUSE-Cloud/nova,mikalstill/nova,belmiromoreira/nova,Triv90/Nova,vmturbo/nova,cyx1231st/nova,petrutlucian94/nova,shahar-stratoscale/nova,rahulunair/nova,MountainWei/nova,gspilio/nova,berrange/nova,gspilio/nova,felixma/nova,nikesh-mahalka/nova,redhat-openstack/nova,badock/nova,noironetworks/nova,joker946/nova,alexandrucoman/vbox-nova-driver,shahar-stratoscale/nova,isyippee/nova,yrobla/nova,salv-orlando/MyRepo,petrutlucian94/nova_dev,rajalokan/nova,maelnor/nova,russellb/nova,isyippee/nova,rahulunair/nova,houshengbo/nova_vmware_compute_driver,tealover/nova,whitepages/nova,alvarolopez/nova,TwinkleChawla/nova,maoy/zknova,mahak/nova,usc-isi/nova,cyx1231st/nova,iuliat/nova,tianweizhang/nova,eayunstack/nova,ewindisch/nova,JioCloud/nova,DirectXMan12/nova-hacking,virtualopensystems/nova,BeyondTheClouds/nova,leilihh/nova,petrutlucian94/nova_dev,rajalokan/nova,luogangyi/bcec-nova,tudorvio/nova,raildo/nova,shootstar/novatest,fajoy/nova,sridevikoushik31/openstack,maelnor/nova,tanglei528/nova,tanglei528/nova,ewindisch/nova,j-carpentier/nova,josephsuh/extra-specs,Brocade-OpenSource/OpenStack-DNRM-Nova,eneabio/nova,rrader/nova-docker-plugin,LoHChina/nova,mahak/nova,TwinkleChawla/nova,Juniper/nova,imsplitbit/nova,mikalstill/nova,scripnichenko/nova,akash1808/nova_test_latest,OpenAcademy-OpenStack/nova-scheduler,CiscoSystems/nova,rajalokan/nova,bgxavier/nova,NoBodyCam/TftpPxeBootBareMetal,saleemjaveds/https-github.com-openstack-nova,alexandrucoman/vbox-nova-driver,noironetworks/nova,NewpTone/stacklab-nova,zzicewind/nova,tealover/nova,phenoxim/nova,yrobla/nova,Francis-Liu/animated-broccoli,devoid/nova,mgagne/nova,DirectXMan12/nova-hacking,Metaswitch/calico-nova,sridevikoushik31/nova,KarimAllah/nova,berrange/nova,qwefi/nova,fnordahl/nova,blueboxgroup/nova,CCI-MOC/nova,kimjaejoong/nova,klmitch/nova,CCI-MOC/nova,spring-week-topos/nova-week,edulramirez/nova,akash1808/nova_test_latest,angdraug/nova,josephsuh/extra-specs,Yuriy-Leonov/nova,affo/nova,double12gzh/nova,iuliat/nova,projectcalico/calico-nova,vmturbo/nova,redhat-openstack/nova,Yuriy-Leonov/nova,josephsuh/extra-specs,gspilio/nova,savi-dev/nova,Triv90/Nova,mandeepdhami/nova,rahulunair/nova,devendermishrajio/nova,zhimin711/nova,savi-dev/nova,DirectXMan12/nova-hacking,mandeepdhami/nova,Juniper/nova,dstroppa/openstack-smartos-nova-grizzly,double12gzh/nova,Juniper/nova,projectcalico/calico-nova,takeshineshiro/nova,blueboxgroup/nova,devendermishrajio/nova_test_latest,adelina-t/nova,JianyuWang/nova,eonpatapon/nova,j-carpentier/nova,dawnpower/nova,bigswitch/nova,tangfeixiong/nova,SUSE-Cloud/nova,silenceli/nova,devendermishrajio/nova,sebrandon1/nova,plumgrid/plumgrid-nova,CEG-FYP-OpenStack/scheduler,Metaswitch/calico-nova,Yusuke1987/openstack_template,JioCloud/nova_test_latest,sridevikoushik31/openstack,citrix-openstack-build/nova,ntt-sic/nova,ted-gould/nova,sebrandon1/nova,klmitch/nova,fajoy/nova,salv-orlando/MyRepo,sacharya/nova,Tehsmash/nova,NeCTAR-RC/nova,mikalstill/nova,phenoxim/nova,Stavitsky/nova,dstroppa/openstack-smartos-nova-grizzly,eneabio/nova,shail2810/nova,alvarolopez/nova,eneabio/nova,tangfeixiong/nova,akash1808/nova,usc-isi/extra-specs,hanlind/nova,klmitch/nova,rickerc/nova_audit,bgxavier/nova,usc-isi/extra-specs,usc-isi/nova,psiwczak/openstack,varunarya10/nova_test_latest,thomasem/nova,viggates/nova,JioCloud/nova,yatinkumbhare/openstack-nova,bclau/nova,barnsnake351/nova,takeshineshiro/nova,alaski/nova,openstack/nova,Francis-Liu/animated-broccoli,alaski/nova,leilihh/novaha,ted-gould/nova,felixma/nova,leilihh/novaha,CEG-FYP-OpenStack/scheduler,ntt-sic/nova,jianghuaw/nova,paulmathews/nova,CloudServer/nova,sridevikoushik31/nova,rrader/nova-docker-plugin,varunarya10/nova_test_latest,LoHChina/nova,russellb/nova,maheshp/novatest,qwefi/nova,silenceli/nova,eonpatapon/nova,sacharya/nova,thomasem/nova,psiwczak/openstack,NoBodyCam/TftpPxeBootBareMetal,rickerc/nova_audit,KarimAllah/nova,sileht/deb-openstack-nova,eharney/nova,shootstar/novatest,cloudbase/nova-virtualbox,rajalokan/nova,Tehsmash/nova,virtualopensystems/nova,cloudbau/nova,houshengbo/nova_vmware_compute_driver,maheshp/novatest,Brocade-OpenSource/OpenStack-DNRM-Nova,kimjaejoong/nova,eayunstack/nova,gooddata/openstack-nova,maheshp/novatest,cernops/nova,raildo/nova,sridevikoushik31/nova,mahak/nova,Juniper/nova,cloudbau/nova,maoy/zknova,sridevikoushik31/nova,jianghuaw/nova,luogangyi/bcec-nova,sebrandon1/nova,plumgrid/plumgrid-nova,jeffrey4l/nova,imsplitbit/nova,NewpTone/stacklab-nova,Yusuke1987/openstack_template,eharney/nova,orbitfp7/nova,zzicewind/nova,openstack/nova,jianghuaw/nova,apporc/nova,zhimin711/nova,vladikr/nova_drafts,cloudbase/nova,BeyondTheClouds/nova,cernops/nova,watonyweng/nova,mmnelemane/nova,OpenAcademy-OpenStack/nova-scheduler,aristanetworks/arista-ovs-nova,adelina-t/nova,citrix-openstack-build/nova,sileht/deb-openstack-nova,cloudbase/nova,NewpTone/stacklab-nova,zaina/nova,KarimAllah/nova,petrutlucian94/nova,yatinkumbhare/openstack-nova,TieWei/nova,whitepages/nova,hanlind/nova,devoid/nova,belmiromoreira/nova,cernops/nova
# vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License.from sqlalchemy import * from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id) Drop FK before dropping instance_id column.
# vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License.from sqlalchemy import * from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) if migrate_engine.name == "mysql": migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \ "`migrations_ibfk_1`;") migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id)
<commit_before># vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License.from sqlalchemy import * from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id) <commit_msg>Drop FK before dropping instance_id column.<commit_after>
# vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License.from sqlalchemy import * from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) if migrate_engine.name == "mysql": migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \ "`migrations_ibfk_1`;") migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id)
# vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License.from sqlalchemy import * from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id) Drop FK before dropping instance_id column.# vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License.from sqlalchemy import * from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) if migrate_engine.name == "mysql": migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \ "`migrations_ibfk_1`;") migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id)
<commit_before># vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License.from sqlalchemy import * from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id) <commit_msg>Drop FK before dropping instance_id column.<commit_after># vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License.from sqlalchemy import * from sqlalchemy import Column, Integer, String, MetaData, Table meta = MetaData() # # Tables to alter # # instance_id = Column('instance_id', Integer()) instance_uuid = Column('instance_uuid', String(255)) def upgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.create_column(instance_uuid) if migrate_engine.name == "mysql": migrate_engine.execute("ALTER TABLE migrations DROP FOREIGN KEY " \ "`migrations_ibfk_1`;") migrations.c.instance_id.drop() def downgrade(migrate_engine): meta.bind = migrate_engine migrations = Table('migrations', meta, autoload=True) migrations.c.instance_uuid.drop() migrations.create_column(instance_id)
cd9c6f96f68391cf333b68e1ea28e513e869aca3
troposphere/detective.py
troposphere/detective.py
# Copyright (c) 2020, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from . import AWSObject class Graph(AWSObject): resource_type = "AWS::Detective::Graph" props = { } class MemberInvitation(AWSObject): resource_type = "AWS::Detective::MemberInvitation" props = { 'GraphArn': (basestring, True), 'MemberEmailAddress': (basestring, True), 'MemberId': (basestring, True), 'Message': (basestring, False), }
# Copyright (c) 2020, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from . import AWSObject from .validators import boolean class Graph(AWSObject): resource_type = "AWS::Detective::Graph" props = { } class MemberInvitation(AWSObject): resource_type = "AWS::Detective::MemberInvitation" props = { 'DisableEmailNotification': (boolean, False), 'GraphArn': (basestring, True), 'MemberEmailAddress': (basestring, True), 'MemberId': (basestring, True), 'Message': (basestring, False), }
Update Detective per 2021-03-15 changes
Update Detective per 2021-03-15 changes
Python
bsd-2-clause
cloudtools/troposphere,cloudtools/troposphere
# Copyright (c) 2020, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from . import AWSObject class Graph(AWSObject): resource_type = "AWS::Detective::Graph" props = { } class MemberInvitation(AWSObject): resource_type = "AWS::Detective::MemberInvitation" props = { 'GraphArn': (basestring, True), 'MemberEmailAddress': (basestring, True), 'MemberId': (basestring, True), 'Message': (basestring, False), } Update Detective per 2021-03-15 changes
# Copyright (c) 2020, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from . import AWSObject from .validators import boolean class Graph(AWSObject): resource_type = "AWS::Detective::Graph" props = { } class MemberInvitation(AWSObject): resource_type = "AWS::Detective::MemberInvitation" props = { 'DisableEmailNotification': (boolean, False), 'GraphArn': (basestring, True), 'MemberEmailAddress': (basestring, True), 'MemberId': (basestring, True), 'Message': (basestring, False), }
<commit_before># Copyright (c) 2020, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from . import AWSObject class Graph(AWSObject): resource_type = "AWS::Detective::Graph" props = { } class MemberInvitation(AWSObject): resource_type = "AWS::Detective::MemberInvitation" props = { 'GraphArn': (basestring, True), 'MemberEmailAddress': (basestring, True), 'MemberId': (basestring, True), 'Message': (basestring, False), } <commit_msg>Update Detective per 2021-03-15 changes<commit_after>
# Copyright (c) 2020, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from . import AWSObject from .validators import boolean class Graph(AWSObject): resource_type = "AWS::Detective::Graph" props = { } class MemberInvitation(AWSObject): resource_type = "AWS::Detective::MemberInvitation" props = { 'DisableEmailNotification': (boolean, False), 'GraphArn': (basestring, True), 'MemberEmailAddress': (basestring, True), 'MemberId': (basestring, True), 'Message': (basestring, False), }
# Copyright (c) 2020, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from . import AWSObject class Graph(AWSObject): resource_type = "AWS::Detective::Graph" props = { } class MemberInvitation(AWSObject): resource_type = "AWS::Detective::MemberInvitation" props = { 'GraphArn': (basestring, True), 'MemberEmailAddress': (basestring, True), 'MemberId': (basestring, True), 'Message': (basestring, False), } Update Detective per 2021-03-15 changes# Copyright (c) 2020, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from . import AWSObject from .validators import boolean class Graph(AWSObject): resource_type = "AWS::Detective::Graph" props = { } class MemberInvitation(AWSObject): resource_type = "AWS::Detective::MemberInvitation" props = { 'DisableEmailNotification': (boolean, False), 'GraphArn': (basestring, True), 'MemberEmailAddress': (basestring, True), 'MemberId': (basestring, True), 'Message': (basestring, False), }
<commit_before># Copyright (c) 2020, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from . import AWSObject class Graph(AWSObject): resource_type = "AWS::Detective::Graph" props = { } class MemberInvitation(AWSObject): resource_type = "AWS::Detective::MemberInvitation" props = { 'GraphArn': (basestring, True), 'MemberEmailAddress': (basestring, True), 'MemberId': (basestring, True), 'Message': (basestring, False), } <commit_msg>Update Detective per 2021-03-15 changes<commit_after># Copyright (c) 2020, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from . import AWSObject from .validators import boolean class Graph(AWSObject): resource_type = "AWS::Detective::Graph" props = { } class MemberInvitation(AWSObject): resource_type = "AWS::Detective::MemberInvitation" props = { 'DisableEmailNotification': (boolean, False), 'GraphArn': (basestring, True), 'MemberEmailAddress': (basestring, True), 'MemberId': (basestring, True), 'Message': (basestring, False), }
08a1bfc233dc44f40ee0b8a5950eda19bca1d6a6
quran_tafseer/serializers.py
quran_tafseer/serializers.py
from django.urls import reverse from rest_framework import serializers from .models import Tafseer, TafseerText class TafseerSerializer(serializers.ModelSerializer): class Meta: model = Tafseer fields = ['id', 'name'] class TafseerTextSerializer(serializers.ModelSerializer): tafseer_id = serializers.IntegerField(source='tafseer.id') tafseer_name = serializers.CharField(source='tafseer.name') ayah_url = serializers.SerializerMethodField() def get_ayah_url(self, obj): return reverse('ayah-detail', kwargs={'number': obj.ayah.number, 'sura_num': obj.ayah.sura.pk}) class Meta: model = TafseerText fields = ['tafseer_id', 'tafseer_name', 'ayah_url', 'ayah', 'text']
from django.urls import reverse from rest_framework import serializers from .models import Tafseer, TafseerText class TafseerSerializer(serializers.ModelSerializer): class Meta: model = Tafseer fields = ['id', 'name'] class TafseerTextSerializer(serializers.ModelSerializer): tafseer_id = serializers.IntegerField(source='tafseer.id') tafseer_name = serializers.CharField(source='tafseer.name') ayah_url = serializers.SerializerMethodField() ayah_number = serializers.IntegerField(source='ayah') def get_ayah_url(self, obj): return reverse('ayah-detail', kwargs={'number': obj.ayah.number, 'sura_num': obj.ayah.sura.pk}) class Meta: model = TafseerText fields = ['tafseer_id', 'tafseer_name', 'ayah_url', 'ayah_number', 'text']
Change serializer field name from ayah to ayah_number
Change serializer field name from ayah to ayah_number
Python
mit
EmadMokhtar/tafseer_api
from django.urls import reverse from rest_framework import serializers from .models import Tafseer, TafseerText class TafseerSerializer(serializers.ModelSerializer): class Meta: model = Tafseer fields = ['id', 'name'] class TafseerTextSerializer(serializers.ModelSerializer): tafseer_id = serializers.IntegerField(source='tafseer.id') tafseer_name = serializers.CharField(source='tafseer.name') ayah_url = serializers.SerializerMethodField() def get_ayah_url(self, obj): return reverse('ayah-detail', kwargs={'number': obj.ayah.number, 'sura_num': obj.ayah.sura.pk}) class Meta: model = TafseerText fields = ['tafseer_id', 'tafseer_name', 'ayah_url', 'ayah', 'text'] Change serializer field name from ayah to ayah_number
from django.urls import reverse from rest_framework import serializers from .models import Tafseer, TafseerText class TafseerSerializer(serializers.ModelSerializer): class Meta: model = Tafseer fields = ['id', 'name'] class TafseerTextSerializer(serializers.ModelSerializer): tafseer_id = serializers.IntegerField(source='tafseer.id') tafseer_name = serializers.CharField(source='tafseer.name') ayah_url = serializers.SerializerMethodField() ayah_number = serializers.IntegerField(source='ayah') def get_ayah_url(self, obj): return reverse('ayah-detail', kwargs={'number': obj.ayah.number, 'sura_num': obj.ayah.sura.pk}) class Meta: model = TafseerText fields = ['tafseer_id', 'tafseer_name', 'ayah_url', 'ayah_number', 'text']
<commit_before>from django.urls import reverse from rest_framework import serializers from .models import Tafseer, TafseerText class TafseerSerializer(serializers.ModelSerializer): class Meta: model = Tafseer fields = ['id', 'name'] class TafseerTextSerializer(serializers.ModelSerializer): tafseer_id = serializers.IntegerField(source='tafseer.id') tafseer_name = serializers.CharField(source='tafseer.name') ayah_url = serializers.SerializerMethodField() def get_ayah_url(self, obj): return reverse('ayah-detail', kwargs={'number': obj.ayah.number, 'sura_num': obj.ayah.sura.pk}) class Meta: model = TafseerText fields = ['tafseer_id', 'tafseer_name', 'ayah_url', 'ayah', 'text'] <commit_msg>Change serializer field name from ayah to ayah_number<commit_after>
from django.urls import reverse from rest_framework import serializers from .models import Tafseer, TafseerText class TafseerSerializer(serializers.ModelSerializer): class Meta: model = Tafseer fields = ['id', 'name'] class TafseerTextSerializer(serializers.ModelSerializer): tafseer_id = serializers.IntegerField(source='tafseer.id') tafseer_name = serializers.CharField(source='tafseer.name') ayah_url = serializers.SerializerMethodField() ayah_number = serializers.IntegerField(source='ayah') def get_ayah_url(self, obj): return reverse('ayah-detail', kwargs={'number': obj.ayah.number, 'sura_num': obj.ayah.sura.pk}) class Meta: model = TafseerText fields = ['tafseer_id', 'tafseer_name', 'ayah_url', 'ayah_number', 'text']
from django.urls import reverse from rest_framework import serializers from .models import Tafseer, TafseerText class TafseerSerializer(serializers.ModelSerializer): class Meta: model = Tafseer fields = ['id', 'name'] class TafseerTextSerializer(serializers.ModelSerializer): tafseer_id = serializers.IntegerField(source='tafseer.id') tafseer_name = serializers.CharField(source='tafseer.name') ayah_url = serializers.SerializerMethodField() def get_ayah_url(self, obj): return reverse('ayah-detail', kwargs={'number': obj.ayah.number, 'sura_num': obj.ayah.sura.pk}) class Meta: model = TafseerText fields = ['tafseer_id', 'tafseer_name', 'ayah_url', 'ayah', 'text'] Change serializer field name from ayah to ayah_numberfrom django.urls import reverse from rest_framework import serializers from .models import Tafseer, TafseerText class TafseerSerializer(serializers.ModelSerializer): class Meta: model = Tafseer fields = ['id', 'name'] class TafseerTextSerializer(serializers.ModelSerializer): tafseer_id = serializers.IntegerField(source='tafseer.id') tafseer_name = serializers.CharField(source='tafseer.name') ayah_url = serializers.SerializerMethodField() ayah_number = serializers.IntegerField(source='ayah') def get_ayah_url(self, obj): return reverse('ayah-detail', kwargs={'number': obj.ayah.number, 'sura_num': obj.ayah.sura.pk}) class Meta: model = TafseerText fields = ['tafseer_id', 'tafseer_name', 'ayah_url', 'ayah_number', 'text']
<commit_before>from django.urls import reverse from rest_framework import serializers from .models import Tafseer, TafseerText class TafseerSerializer(serializers.ModelSerializer): class Meta: model = Tafseer fields = ['id', 'name'] class TafseerTextSerializer(serializers.ModelSerializer): tafseer_id = serializers.IntegerField(source='tafseer.id') tafseer_name = serializers.CharField(source='tafseer.name') ayah_url = serializers.SerializerMethodField() def get_ayah_url(self, obj): return reverse('ayah-detail', kwargs={'number': obj.ayah.number, 'sura_num': obj.ayah.sura.pk}) class Meta: model = TafseerText fields = ['tafseer_id', 'tafseer_name', 'ayah_url', 'ayah', 'text'] <commit_msg>Change serializer field name from ayah to ayah_number<commit_after>from django.urls import reverse from rest_framework import serializers from .models import Tafseer, TafseerText class TafseerSerializer(serializers.ModelSerializer): class Meta: model = Tafseer fields = ['id', 'name'] class TafseerTextSerializer(serializers.ModelSerializer): tafseer_id = serializers.IntegerField(source='tafseer.id') tafseer_name = serializers.CharField(source='tafseer.name') ayah_url = serializers.SerializerMethodField() ayah_number = serializers.IntegerField(source='ayah') def get_ayah_url(self, obj): return reverse('ayah-detail', kwargs={'number': obj.ayah.number, 'sura_num': obj.ayah.sura.pk}) class Meta: model = TafseerText fields = ['tafseer_id', 'tafseer_name', 'ayah_url', 'ayah_number', 'text']
efb636d392dab44fcc110b8d17933476ba90996d
Donut/__init__.py
Donut/__init__.py
import flask import sqlalchemy from Donut import config, constants from Donut.modules import example app = flask.Flask(__name__) app.debug = False app.secret_key = config.SECRET_KEY # Maximum file upload size, in bytes. app.config['MAX_CONTENT_LENGTH'] = constants.MAX_CONTENT_LENGTH # Load blueprint modules app.register_blueprint(example.blueprint, url_prefix='/example') # Create database engine object. # TODO##DatabaseWork: We currently don't have a database set up, so we can't # reference sqlalchemy yet. However, it serves as a good example implementation. # engine = sqlalchemy.create_engine(config.DB_URI, convert_unicode=True) @app.before_request def before_request(): """Logic executed before request is processed.""" # TODO#DatabaseWork uncomment this line # flask.g.db = engine.connect() @app.teardown_request def teardown_request(exception): """Logic executed after every request is finished.""" # TODO#DatabaseWork uncomment these lines # if flask.g.db != None: # flask.g.db.close() # After initialization, import the routes. from Donut import routes
import flask import sqlalchemy import os from Donut import constants from Donut.modules import example app = flask.Flask(__name__) app.debug = False # Get app config, if we're not testing on travis. if 'TRAVIS' not in os.environ: app.config.from_object('Donut.config') # Maximum file upload size, in bytes. app.config['MAX_CONTENT_LENGTH'] = constants.MAX_CONTENT_LENGTH # Load blueprint modules app.register_blueprint(example.blueprint, url_prefix='/example') # Create database engine object. # TODO##DatabaseWork: We currently don't have a database set up, so we can't # reference sqlalchemy yet. However, it serves as a good example implementation. # engine = sqlalchemy.create_engine(app.config['DB_URI'], convert_unicode=True) @app.before_request def before_request(): """Logic executed before request is processed.""" # TODO#DatabaseWork uncomment this line # flask.g.db = engine.connect() @app.teardown_request def teardown_request(exception): """Logic executed after every request is finished.""" # TODO#DatabaseWork uncomment these lines # if flask.g.db != None: # flask.g.db.close() # After initialization, import the routes. from Donut import routes
Change how we get config settings
Change how we get config settings - Flask recommends using `app.config.from_object` to get config settings. - Don't get config settings if we're using travis, since it won't have the settings from the repo.
Python
mit
ASCIT/donut,ASCIT/donut-python,ASCIT/donut,ASCIT/donut-python,ASCIT/donut
import flask import sqlalchemy from Donut import config, constants from Donut.modules import example app = flask.Flask(__name__) app.debug = False app.secret_key = config.SECRET_KEY # Maximum file upload size, in bytes. app.config['MAX_CONTENT_LENGTH'] = constants.MAX_CONTENT_LENGTH # Load blueprint modules app.register_blueprint(example.blueprint, url_prefix='/example') # Create database engine object. # TODO##DatabaseWork: We currently don't have a database set up, so we can't # reference sqlalchemy yet. However, it serves as a good example implementation. # engine = sqlalchemy.create_engine(config.DB_URI, convert_unicode=True) @app.before_request def before_request(): """Logic executed before request is processed.""" # TODO#DatabaseWork uncomment this line # flask.g.db = engine.connect() @app.teardown_request def teardown_request(exception): """Logic executed after every request is finished.""" # TODO#DatabaseWork uncomment these lines # if flask.g.db != None: # flask.g.db.close() # After initialization, import the routes. from Donut import routes Change how we get config settings - Flask recommends using `app.config.from_object` to get config settings. - Don't get config settings if we're using travis, since it won't have the settings from the repo.
import flask import sqlalchemy import os from Donut import constants from Donut.modules import example app = flask.Flask(__name__) app.debug = False # Get app config, if we're not testing on travis. if 'TRAVIS' not in os.environ: app.config.from_object('Donut.config') # Maximum file upload size, in bytes. app.config['MAX_CONTENT_LENGTH'] = constants.MAX_CONTENT_LENGTH # Load blueprint modules app.register_blueprint(example.blueprint, url_prefix='/example') # Create database engine object. # TODO##DatabaseWork: We currently don't have a database set up, so we can't # reference sqlalchemy yet. However, it serves as a good example implementation. # engine = sqlalchemy.create_engine(app.config['DB_URI'], convert_unicode=True) @app.before_request def before_request(): """Logic executed before request is processed.""" # TODO#DatabaseWork uncomment this line # flask.g.db = engine.connect() @app.teardown_request def teardown_request(exception): """Logic executed after every request is finished.""" # TODO#DatabaseWork uncomment these lines # if flask.g.db != None: # flask.g.db.close() # After initialization, import the routes. from Donut import routes
<commit_before>import flask import sqlalchemy from Donut import config, constants from Donut.modules import example app = flask.Flask(__name__) app.debug = False app.secret_key = config.SECRET_KEY # Maximum file upload size, in bytes. app.config['MAX_CONTENT_LENGTH'] = constants.MAX_CONTENT_LENGTH # Load blueprint modules app.register_blueprint(example.blueprint, url_prefix='/example') # Create database engine object. # TODO##DatabaseWork: We currently don't have a database set up, so we can't # reference sqlalchemy yet. However, it serves as a good example implementation. # engine = sqlalchemy.create_engine(config.DB_URI, convert_unicode=True) @app.before_request def before_request(): """Logic executed before request is processed.""" # TODO#DatabaseWork uncomment this line # flask.g.db = engine.connect() @app.teardown_request def teardown_request(exception): """Logic executed after every request is finished.""" # TODO#DatabaseWork uncomment these lines # if flask.g.db != None: # flask.g.db.close() # After initialization, import the routes. from Donut import routes <commit_msg>Change how we get config settings - Flask recommends using `app.config.from_object` to get config settings. - Don't get config settings if we're using travis, since it won't have the settings from the repo.<commit_after>
import flask import sqlalchemy import os from Donut import constants from Donut.modules import example app = flask.Flask(__name__) app.debug = False # Get app config, if we're not testing on travis. if 'TRAVIS' not in os.environ: app.config.from_object('Donut.config') # Maximum file upload size, in bytes. app.config['MAX_CONTENT_LENGTH'] = constants.MAX_CONTENT_LENGTH # Load blueprint modules app.register_blueprint(example.blueprint, url_prefix='/example') # Create database engine object. # TODO##DatabaseWork: We currently don't have a database set up, so we can't # reference sqlalchemy yet. However, it serves as a good example implementation. # engine = sqlalchemy.create_engine(app.config['DB_URI'], convert_unicode=True) @app.before_request def before_request(): """Logic executed before request is processed.""" # TODO#DatabaseWork uncomment this line # flask.g.db = engine.connect() @app.teardown_request def teardown_request(exception): """Logic executed after every request is finished.""" # TODO#DatabaseWork uncomment these lines # if flask.g.db != None: # flask.g.db.close() # After initialization, import the routes. from Donut import routes
import flask import sqlalchemy from Donut import config, constants from Donut.modules import example app = flask.Flask(__name__) app.debug = False app.secret_key = config.SECRET_KEY # Maximum file upload size, in bytes. app.config['MAX_CONTENT_LENGTH'] = constants.MAX_CONTENT_LENGTH # Load blueprint modules app.register_blueprint(example.blueprint, url_prefix='/example') # Create database engine object. # TODO##DatabaseWork: We currently don't have a database set up, so we can't # reference sqlalchemy yet. However, it serves as a good example implementation. # engine = sqlalchemy.create_engine(config.DB_URI, convert_unicode=True) @app.before_request def before_request(): """Logic executed before request is processed.""" # TODO#DatabaseWork uncomment this line # flask.g.db = engine.connect() @app.teardown_request def teardown_request(exception): """Logic executed after every request is finished.""" # TODO#DatabaseWork uncomment these lines # if flask.g.db != None: # flask.g.db.close() # After initialization, import the routes. from Donut import routes Change how we get config settings - Flask recommends using `app.config.from_object` to get config settings. - Don't get config settings if we're using travis, since it won't have the settings from the repo.import flask import sqlalchemy import os from Donut import constants from Donut.modules import example app = flask.Flask(__name__) app.debug = False # Get app config, if we're not testing on travis. if 'TRAVIS' not in os.environ: app.config.from_object('Donut.config') # Maximum file upload size, in bytes. app.config['MAX_CONTENT_LENGTH'] = constants.MAX_CONTENT_LENGTH # Load blueprint modules app.register_blueprint(example.blueprint, url_prefix='/example') # Create database engine object. # TODO##DatabaseWork: We currently don't have a database set up, so we can't # reference sqlalchemy yet. However, it serves as a good example implementation. # engine = sqlalchemy.create_engine(app.config['DB_URI'], convert_unicode=True) @app.before_request def before_request(): """Logic executed before request is processed.""" # TODO#DatabaseWork uncomment this line # flask.g.db = engine.connect() @app.teardown_request def teardown_request(exception): """Logic executed after every request is finished.""" # TODO#DatabaseWork uncomment these lines # if flask.g.db != None: # flask.g.db.close() # After initialization, import the routes. from Donut import routes
<commit_before>import flask import sqlalchemy from Donut import config, constants from Donut.modules import example app = flask.Flask(__name__) app.debug = False app.secret_key = config.SECRET_KEY # Maximum file upload size, in bytes. app.config['MAX_CONTENT_LENGTH'] = constants.MAX_CONTENT_LENGTH # Load blueprint modules app.register_blueprint(example.blueprint, url_prefix='/example') # Create database engine object. # TODO##DatabaseWork: We currently don't have a database set up, so we can't # reference sqlalchemy yet. However, it serves as a good example implementation. # engine = sqlalchemy.create_engine(config.DB_URI, convert_unicode=True) @app.before_request def before_request(): """Logic executed before request is processed.""" # TODO#DatabaseWork uncomment this line # flask.g.db = engine.connect() @app.teardown_request def teardown_request(exception): """Logic executed after every request is finished.""" # TODO#DatabaseWork uncomment these lines # if flask.g.db != None: # flask.g.db.close() # After initialization, import the routes. from Donut import routes <commit_msg>Change how we get config settings - Flask recommends using `app.config.from_object` to get config settings. - Don't get config settings if we're using travis, since it won't have the settings from the repo.<commit_after>import flask import sqlalchemy import os from Donut import constants from Donut.modules import example app = flask.Flask(__name__) app.debug = False # Get app config, if we're not testing on travis. if 'TRAVIS' not in os.environ: app.config.from_object('Donut.config') # Maximum file upload size, in bytes. app.config['MAX_CONTENT_LENGTH'] = constants.MAX_CONTENT_LENGTH # Load blueprint modules app.register_blueprint(example.blueprint, url_prefix='/example') # Create database engine object. # TODO##DatabaseWork: We currently don't have a database set up, so we can't # reference sqlalchemy yet. However, it serves as a good example implementation. # engine = sqlalchemy.create_engine(app.config['DB_URI'], convert_unicode=True) @app.before_request def before_request(): """Logic executed before request is processed.""" # TODO#DatabaseWork uncomment this line # flask.g.db = engine.connect() @app.teardown_request def teardown_request(exception): """Logic executed after every request is finished.""" # TODO#DatabaseWork uncomment these lines # if flask.g.db != None: # flask.g.db.close() # After initialization, import the routes. from Donut import routes
31c79697db0d5c973cff9b845ed28845695ecb02
website/addons/twofactor/views.py
website/addons/twofactor/views.py
# -*- coding: utf-8 -*- import httplib as http from framework import request from framework.auth.decorators import must_be_logged_in from framework.exceptions import HTTPError from website.project.decorators import must_have_addon @must_be_logged_in @must_have_addon('twofactor', 'user') def user_settings(user_addon, *args, **kwargs): code = request.json.get('code') if code is None: raise HTTPError(code=http.BAD_REQUEST) if user_addon.verify_code(code): user_addon.is_confirmed = True raise HTTPError(http.FORBIDDEN, data=dict( message_short='Forbidden', message_long='The two-factor verification code you provided is invalid.' ))
# -*- coding: utf-8 -*- import httplib as http from framework import request from framework.auth.decorators import must_be_logged_in from framework.exceptions import HTTPError from website.project.decorators import must_have_addon @must_be_logged_in @must_have_addon('twofactor', 'user') def user_settings(user_addon, *args, **kwargs): code = request.json.get('code') if code is None: raise HTTPError(code=http.BAD_REQUEST) if user_addon.verify_code(code): user_addon.is_confirmed = True user_addon.save() return {'message': 'Successfully verified two-factor authentication.'}, http.OK raise HTTPError(http.FORBIDDEN, data=dict( message_short='Forbidden', message_long='The two-factor verification code you provided is invalid.' ))
Fix response when user successfully confirms 2fa
Fix response when user successfully confirms 2fa
Python
apache-2.0
CenterForOpenScience/osf.io,revanthkolli/osf.io,jnayak1/osf.io,mluo613/osf.io,caneruguz/osf.io,alexschiller/osf.io,adlius/osf.io,himanshuo/osf.io,CenterForOpenScience/osf.io,samanehsan/osf.io,jeffreyliu3230/osf.io,cslzchen/osf.io,cwisecarver/osf.io,billyhunt/osf.io,SSJohns/osf.io,cosenal/osf.io,GageGaskins/osf.io,kushG/osf.io,zamattiac/osf.io,njantrania/osf.io,petermalcolm/osf.io,jinluyuan/osf.io,bdyetton/prettychart,monikagrabowska/osf.io,rdhyee/osf.io,monikagrabowska/osf.io,chennan47/osf.io,haoyuchen1992/osf.io,cwisecarver/osf.io,ZobairAlijan/osf.io,RomanZWang/osf.io,TomHeatwole/osf.io,wearpants/osf.io,binoculars/osf.io,jolene-esposito/osf.io,billyhunt/osf.io,reinaH/osf.io,kushG/osf.io,GageGaskins/osf.io,kwierman/osf.io,leb2dg/osf.io,felliott/osf.io,revanthkolli/osf.io,Johnetordoff/osf.io,caseyrygt/osf.io,Nesiehr/osf.io,acshi/osf.io,barbour-em/osf.io,doublebits/osf.io,acshi/osf.io,zkraime/osf.io,TomBaxter/osf.io,kushG/osf.io,baylee-d/osf.io,adlius/osf.io,cosenal/osf.io,jolene-esposito/osf.io,kushG/osf.io,njantrania/osf.io,DanielSBrown/osf.io,sbt9uc/osf.io,adlius/osf.io,MerlinZhang/osf.io,caseyrygt/osf.io,zachjanicki/osf.io,SSJohns/osf.io,TomBaxter/osf.io,cldershem/osf.io,reinaH/osf.io,revanthkolli/osf.io,zachjanicki/osf.io,sloria/osf.io,felliott/osf.io,hmoco/osf.io,hmoco/osf.io,brianjgeiger/osf.io,HarryRybacki/osf.io,danielneis/osf.io,emetsger/osf.io,chennan47/osf.io,icereval/osf.io,RomanZWang/osf.io,jeffreyliu3230/osf.io,kch8qx/osf.io,chennan47/osf.io,binoculars/osf.io,KAsante95/osf.io,amyshi188/osf.io,brianjgeiger/osf.io,ticklemepierce/osf.io,cwisecarver/osf.io,AndrewSallans/osf.io,mluo613/osf.io,alexschiller/osf.io,fabianvf/osf.io,alexschiller/osf.io,mluke93/osf.io,abought/osf.io,MerlinZhang/osf.io,zachjanicki/osf.io,kch8qx/osf.io,MerlinZhang/osf.io,dplorimer/osf,cosenal/osf.io,Nesiehr/osf.io,mluo613/osf.io,aaxelb/osf.io,asanfilippo7/osf.io,arpitar/osf.io,cwisecarver/osf.io,lamdnhan/osf.io,arpitar/osf.io,ckc6cz/osf.io,zamattiac/osf.io,Johnetordoff/osf.io,wearpants/osf.io,alexschiller/osf.io,TomBaxter/osf.io,icereval/osf.io,doublebits/osf.io,crcresearch/osf.io,HarryRybacki/osf.io,KAsante95/osf.io,haoyuchen1992/osf.io,caseyrygt/osf.io,haoyuchen1992/osf.io,HalcyonChimera/osf.io,mattclark/osf.io,dplorimer/osf,GaryKriebel/osf.io,doublebits/osf.io,samchrisinger/osf.io,fabianvf/osf.io,wearpants/osf.io,jinluyuan/osf.io,chrisseto/osf.io,samanehsan/osf.io,mluke93/osf.io,leb2dg/osf.io,GaryKriebel/osf.io,laurenrevere/osf.io,kch8qx/osf.io,ticklemepierce/osf.io,cslzchen/osf.io,ckc6cz/osf.io,pattisdr/osf.io,brandonPurvis/osf.io,leb2dg/osf.io,brandonPurvis/osf.io,laurenrevere/osf.io,sloria/osf.io,GageGaskins/osf.io,TomHeatwole/osf.io,arpitar/osf.io,aaxelb/osf.io,KAsante95/osf.io,amyshi188/osf.io,HarryRybacki/osf.io,brandonPurvis/osf.io,Nesiehr/osf.io,cosenal/osf.io,jinluyuan/osf.io,zamattiac/osf.io,abought/osf.io,KAsante95/osf.io,rdhyee/osf.io,samchrisinger/osf.io,TomHeatwole/osf.io,acshi/osf.io,sbt9uc/osf.io,chrisseto/osf.io,felliott/osf.io,Johnetordoff/osf.io,DanielSBrown/osf.io,emetsger/osf.io,caseyrygt/osf.io,caseyrollins/osf.io,TomHeatwole/osf.io,lamdnhan/osf.io,lyndsysimon/osf.io,crcresearch/osf.io,zkraime/osf.io,mluo613/osf.io,brandonPurvis/osf.io,ckc6cz/osf.io,erinspace/osf.io,jnayak1/osf.io,HarryRybacki/osf.io,mluke93/osf.io,felliott/osf.io,mfraezz/osf.io,billyhunt/osf.io,brianjgeiger/osf.io,saradbowman/osf.io,chrisseto/osf.io,Nesiehr/osf.io,SSJohns/osf.io,dplorimer/osf,jmcarp/osf.io,cslzchen/osf.io,crcresearch/osf.io,samchrisinger/osf.io,revanthkolli/osf.io,bdyetton/prettychart,abought/osf.io,erinspace/osf.io,caseyrollins/osf.io,GageGaskins/osf.io,lyndsysimon/osf.io,cslzchen/osf.io,reinaH/osf.io,wearpants/osf.io,zkraime/osf.io,lamdnhan/osf.io,jeffreyliu3230/osf.io,RomanZWang/osf.io,lyndsysimon/osf.io,caneruguz/osf.io,GageGaskins/osf.io,mfraezz/osf.io,pattisdr/osf.io,baylee-d/osf.io,Ghalko/osf.io,GaryKriebel/osf.io,chrisseto/osf.io,Ghalko/osf.io,mattclark/osf.io,HalcyonChimera/osf.io,billyhunt/osf.io,acshi/osf.io,cldershem/osf.io,rdhyee/osf.io,alexschiller/osf.io,asanfilippo7/osf.io,arpitar/osf.io,baylee-d/osf.io,samanehsan/osf.io,rdhyee/osf.io,acshi/osf.io,mfraezz/osf.io,njantrania/osf.io,monikagrabowska/osf.io,fabianvf/osf.io,mattclark/osf.io,kch8qx/osf.io,ticklemepierce/osf.io,saradbowman/osf.io,GaryKriebel/osf.io,fabianvf/osf.io,abought/osf.io,kwierman/osf.io,aaxelb/osf.io,MerlinZhang/osf.io,barbour-em/osf.io,jinluyuan/osf.io,monikagrabowska/osf.io,sbt9uc/osf.io,SSJohns/osf.io,KAsante95/osf.io,dplorimer/osf,himanshuo/osf.io,jnayak1/osf.io,barbour-em/osf.io,cldershem/osf.io,sloria/osf.io,HalcyonChimera/osf.io,sbt9uc/osf.io,caneruguz/osf.io,Johnetordoff/osf.io,Ghalko/osf.io,ZobairAlijan/osf.io,RomanZWang/osf.io,RomanZWang/osf.io,leb2dg/osf.io,mfraezz/osf.io,jmcarp/osf.io,AndrewSallans/osf.io,DanielSBrown/osf.io,pattisdr/osf.io,petermalcolm/osf.io,reinaH/osf.io,erinspace/osf.io,mluo613/osf.io,zachjanicki/osf.io,jolene-esposito/osf.io,laurenrevere/osf.io,amyshi188/osf.io,CenterForOpenScience/osf.io,binoculars/osf.io,CenterForOpenScience/osf.io,ckc6cz/osf.io,hmoco/osf.io,zkraime/osf.io,doublebits/osf.io,kwierman/osf.io,jeffreyliu3230/osf.io,himanshuo/osf.io,jmcarp/osf.io,emetsger/osf.io,himanshuo/osf.io,lamdnhan/osf.io,icereval/osf.io,asanfilippo7/osf.io,doublebits/osf.io,bdyetton/prettychart,njantrania/osf.io,hmoco/osf.io,caneruguz/osf.io,haoyuchen1992/osf.io,lyndsysimon/osf.io,asanfilippo7/osf.io,aaxelb/osf.io,samanehsan/osf.io,cldershem/osf.io,ZobairAlijan/osf.io,jmcarp/osf.io,adlius/osf.io,kch8qx/osf.io,emetsger/osf.io,danielneis/osf.io,mluke93/osf.io,samchrisinger/osf.io,caseyrollins/osf.io,amyshi188/osf.io,ticklemepierce/osf.io,jolene-esposito/osf.io,HalcyonChimera/osf.io,danielneis/osf.io,kwierman/osf.io,DanielSBrown/osf.io,brianjgeiger/osf.io,danielneis/osf.io,petermalcolm/osf.io,brandonPurvis/osf.io,barbour-em/osf.io,Ghalko/osf.io,petermalcolm/osf.io,ZobairAlijan/osf.io,monikagrabowska/osf.io,jnayak1/osf.io,billyhunt/osf.io,bdyetton/prettychart,zamattiac/osf.io
# -*- coding: utf-8 -*- import httplib as http from framework import request from framework.auth.decorators import must_be_logged_in from framework.exceptions import HTTPError from website.project.decorators import must_have_addon @must_be_logged_in @must_have_addon('twofactor', 'user') def user_settings(user_addon, *args, **kwargs): code = request.json.get('code') if code is None: raise HTTPError(code=http.BAD_REQUEST) if user_addon.verify_code(code): user_addon.is_confirmed = True raise HTTPError(http.FORBIDDEN, data=dict( message_short='Forbidden', message_long='The two-factor verification code you provided is invalid.' )) Fix response when user successfully confirms 2fa
# -*- coding: utf-8 -*- import httplib as http from framework import request from framework.auth.decorators import must_be_logged_in from framework.exceptions import HTTPError from website.project.decorators import must_have_addon @must_be_logged_in @must_have_addon('twofactor', 'user') def user_settings(user_addon, *args, **kwargs): code = request.json.get('code') if code is None: raise HTTPError(code=http.BAD_REQUEST) if user_addon.verify_code(code): user_addon.is_confirmed = True user_addon.save() return {'message': 'Successfully verified two-factor authentication.'}, http.OK raise HTTPError(http.FORBIDDEN, data=dict( message_short='Forbidden', message_long='The two-factor verification code you provided is invalid.' ))
<commit_before># -*- coding: utf-8 -*- import httplib as http from framework import request from framework.auth.decorators import must_be_logged_in from framework.exceptions import HTTPError from website.project.decorators import must_have_addon @must_be_logged_in @must_have_addon('twofactor', 'user') def user_settings(user_addon, *args, **kwargs): code = request.json.get('code') if code is None: raise HTTPError(code=http.BAD_REQUEST) if user_addon.verify_code(code): user_addon.is_confirmed = True raise HTTPError(http.FORBIDDEN, data=dict( message_short='Forbidden', message_long='The two-factor verification code you provided is invalid.' )) <commit_msg>Fix response when user successfully confirms 2fa<commit_after>
# -*- coding: utf-8 -*- import httplib as http from framework import request from framework.auth.decorators import must_be_logged_in from framework.exceptions import HTTPError from website.project.decorators import must_have_addon @must_be_logged_in @must_have_addon('twofactor', 'user') def user_settings(user_addon, *args, **kwargs): code = request.json.get('code') if code is None: raise HTTPError(code=http.BAD_REQUEST) if user_addon.verify_code(code): user_addon.is_confirmed = True user_addon.save() return {'message': 'Successfully verified two-factor authentication.'}, http.OK raise HTTPError(http.FORBIDDEN, data=dict( message_short='Forbidden', message_long='The two-factor verification code you provided is invalid.' ))
# -*- coding: utf-8 -*- import httplib as http from framework import request from framework.auth.decorators import must_be_logged_in from framework.exceptions import HTTPError from website.project.decorators import must_have_addon @must_be_logged_in @must_have_addon('twofactor', 'user') def user_settings(user_addon, *args, **kwargs): code = request.json.get('code') if code is None: raise HTTPError(code=http.BAD_REQUEST) if user_addon.verify_code(code): user_addon.is_confirmed = True raise HTTPError(http.FORBIDDEN, data=dict( message_short='Forbidden', message_long='The two-factor verification code you provided is invalid.' )) Fix response when user successfully confirms 2fa# -*- coding: utf-8 -*- import httplib as http from framework import request from framework.auth.decorators import must_be_logged_in from framework.exceptions import HTTPError from website.project.decorators import must_have_addon @must_be_logged_in @must_have_addon('twofactor', 'user') def user_settings(user_addon, *args, **kwargs): code = request.json.get('code') if code is None: raise HTTPError(code=http.BAD_REQUEST) if user_addon.verify_code(code): user_addon.is_confirmed = True user_addon.save() return {'message': 'Successfully verified two-factor authentication.'}, http.OK raise HTTPError(http.FORBIDDEN, data=dict( message_short='Forbidden', message_long='The two-factor verification code you provided is invalid.' ))
<commit_before># -*- coding: utf-8 -*- import httplib as http from framework import request from framework.auth.decorators import must_be_logged_in from framework.exceptions import HTTPError from website.project.decorators import must_have_addon @must_be_logged_in @must_have_addon('twofactor', 'user') def user_settings(user_addon, *args, **kwargs): code = request.json.get('code') if code is None: raise HTTPError(code=http.BAD_REQUEST) if user_addon.verify_code(code): user_addon.is_confirmed = True raise HTTPError(http.FORBIDDEN, data=dict( message_short='Forbidden', message_long='The two-factor verification code you provided is invalid.' )) <commit_msg>Fix response when user successfully confirms 2fa<commit_after># -*- coding: utf-8 -*- import httplib as http from framework import request from framework.auth.decorators import must_be_logged_in from framework.exceptions import HTTPError from website.project.decorators import must_have_addon @must_be_logged_in @must_have_addon('twofactor', 'user') def user_settings(user_addon, *args, **kwargs): code = request.json.get('code') if code is None: raise HTTPError(code=http.BAD_REQUEST) if user_addon.verify_code(code): user_addon.is_confirmed = True user_addon.save() return {'message': 'Successfully verified two-factor authentication.'}, http.OK raise HTTPError(http.FORBIDDEN, data=dict( message_short='Forbidden', message_long='The two-factor verification code you provided is invalid.' ))
8ce2cbff7063b97e249dd87dd2acb8d83bdbf509
examples/py/fetch-ohlcv-sequentially.py
examples/py/fetch-ohlcv-sequentially.py
# -*- coding: utf-8 -*- import os import sys import time # ----------------------------------------------------------------------------- root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root) # ----------------------------------------------------------------------------- import ccxt # noqa: E402 # ----------------------------------------------------------------------------- # common constants msec = 1000 minute = 60 * msec # ----------------------------------------------------------------------------- kraken = ccxt.kraken() # ----------------------------------------------------------------------------- from_datetime = '2017-09-01 00:00:00' from_timestamp = kraken.parse8601(from_datetime) # ----------------------------------------------------------------------------- now = kraken.milliseconds() # ----------------------------------------------------------------------------- while from_timestamp < now: print('Fetching candles starting from', kraken.iso8601(from_timestamp)) ohlcvs = kraken.fetch_ohlcv('BTC/USD', '1m', from_timestamp) # don't hit the rateLimit or you will be banned time.sleep(kraken.rateLimit / msec) # Kraken returns 720 candles for 1m timeframe at once from_timestamp += len(ohlcvs) * minute
# -*- coding: utf-8 -*- import os import sys import time # ----------------------------------------------------------------------------- root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root) # ----------------------------------------------------------------------------- import ccxt # noqa: E402 # ----------------------------------------------------------------------------- # common constants msec = 1000 minute = 60 * msec # ----------------------------------------------------------------------------- kraken = ccxt.kraken() # ----------------------------------------------------------------------------- from_datetime = '2017-09-01 00:00:00' from_timestamp = kraken.parse8601(from_datetime) # ----------------------------------------------------------------------------- now = kraken.milliseconds() # ----------------------------------------------------------------------------- data = [] while from_timestamp < now: print('Fetching candles starting from', kraken.iso8601(from_timestamp)) ohlcvs = kraken.fetch_ohlcv('BTC/USD', '1m', from_timestamp) # don't hit the rateLimit or you will be banned time.sleep(3 * kraken.rateLimit / msec) # Kraken returns 720 candles for 1m timeframe at once from_timestamp += len(ohlcvs) * minute data += ohlcvs
Add delay 3 times the rate limit to avoid API errors
Add delay 3 times the rate limit to avoid API errors
Python
mit
tritoanst/ccxt,tritoanst/ccxt,ccxt/ccxt,ccxt/ccxt,tritoanst/ccxt,ccxt/ccxt,ccxt/ccxt,ccxt/ccxt,tritoanst/ccxt
# -*- coding: utf-8 -*- import os import sys import time # ----------------------------------------------------------------------------- root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root) # ----------------------------------------------------------------------------- import ccxt # noqa: E402 # ----------------------------------------------------------------------------- # common constants msec = 1000 minute = 60 * msec # ----------------------------------------------------------------------------- kraken = ccxt.kraken() # ----------------------------------------------------------------------------- from_datetime = '2017-09-01 00:00:00' from_timestamp = kraken.parse8601(from_datetime) # ----------------------------------------------------------------------------- now = kraken.milliseconds() # ----------------------------------------------------------------------------- while from_timestamp < now: print('Fetching candles starting from', kraken.iso8601(from_timestamp)) ohlcvs = kraken.fetch_ohlcv('BTC/USD', '1m', from_timestamp) # don't hit the rateLimit or you will be banned time.sleep(kraken.rateLimit / msec) # Kraken returns 720 candles for 1m timeframe at once from_timestamp += len(ohlcvs) * minute Add delay 3 times the rate limit to avoid API errors
# -*- coding: utf-8 -*- import os import sys import time # ----------------------------------------------------------------------------- root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root) # ----------------------------------------------------------------------------- import ccxt # noqa: E402 # ----------------------------------------------------------------------------- # common constants msec = 1000 minute = 60 * msec # ----------------------------------------------------------------------------- kraken = ccxt.kraken() # ----------------------------------------------------------------------------- from_datetime = '2017-09-01 00:00:00' from_timestamp = kraken.parse8601(from_datetime) # ----------------------------------------------------------------------------- now = kraken.milliseconds() # ----------------------------------------------------------------------------- data = [] while from_timestamp < now: print('Fetching candles starting from', kraken.iso8601(from_timestamp)) ohlcvs = kraken.fetch_ohlcv('BTC/USD', '1m', from_timestamp) # don't hit the rateLimit or you will be banned time.sleep(3 * kraken.rateLimit / msec) # Kraken returns 720 candles for 1m timeframe at once from_timestamp += len(ohlcvs) * minute data += ohlcvs
<commit_before># -*- coding: utf-8 -*- import os import sys import time # ----------------------------------------------------------------------------- root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root) # ----------------------------------------------------------------------------- import ccxt # noqa: E402 # ----------------------------------------------------------------------------- # common constants msec = 1000 minute = 60 * msec # ----------------------------------------------------------------------------- kraken = ccxt.kraken() # ----------------------------------------------------------------------------- from_datetime = '2017-09-01 00:00:00' from_timestamp = kraken.parse8601(from_datetime) # ----------------------------------------------------------------------------- now = kraken.milliseconds() # ----------------------------------------------------------------------------- while from_timestamp < now: print('Fetching candles starting from', kraken.iso8601(from_timestamp)) ohlcvs = kraken.fetch_ohlcv('BTC/USD', '1m', from_timestamp) # don't hit the rateLimit or you will be banned time.sleep(kraken.rateLimit / msec) # Kraken returns 720 candles for 1m timeframe at once from_timestamp += len(ohlcvs) * minute <commit_msg>Add delay 3 times the rate limit to avoid API errors<commit_after>
# -*- coding: utf-8 -*- import os import sys import time # ----------------------------------------------------------------------------- root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root) # ----------------------------------------------------------------------------- import ccxt # noqa: E402 # ----------------------------------------------------------------------------- # common constants msec = 1000 minute = 60 * msec # ----------------------------------------------------------------------------- kraken = ccxt.kraken() # ----------------------------------------------------------------------------- from_datetime = '2017-09-01 00:00:00' from_timestamp = kraken.parse8601(from_datetime) # ----------------------------------------------------------------------------- now = kraken.milliseconds() # ----------------------------------------------------------------------------- data = [] while from_timestamp < now: print('Fetching candles starting from', kraken.iso8601(from_timestamp)) ohlcvs = kraken.fetch_ohlcv('BTC/USD', '1m', from_timestamp) # don't hit the rateLimit or you will be banned time.sleep(3 * kraken.rateLimit / msec) # Kraken returns 720 candles for 1m timeframe at once from_timestamp += len(ohlcvs) * minute data += ohlcvs
# -*- coding: utf-8 -*- import os import sys import time # ----------------------------------------------------------------------------- root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root) # ----------------------------------------------------------------------------- import ccxt # noqa: E402 # ----------------------------------------------------------------------------- # common constants msec = 1000 minute = 60 * msec # ----------------------------------------------------------------------------- kraken = ccxt.kraken() # ----------------------------------------------------------------------------- from_datetime = '2017-09-01 00:00:00' from_timestamp = kraken.parse8601(from_datetime) # ----------------------------------------------------------------------------- now = kraken.milliseconds() # ----------------------------------------------------------------------------- while from_timestamp < now: print('Fetching candles starting from', kraken.iso8601(from_timestamp)) ohlcvs = kraken.fetch_ohlcv('BTC/USD', '1m', from_timestamp) # don't hit the rateLimit or you will be banned time.sleep(kraken.rateLimit / msec) # Kraken returns 720 candles for 1m timeframe at once from_timestamp += len(ohlcvs) * minute Add delay 3 times the rate limit to avoid API errors# -*- coding: utf-8 -*- import os import sys import time # ----------------------------------------------------------------------------- root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root) # ----------------------------------------------------------------------------- import ccxt # noqa: E402 # ----------------------------------------------------------------------------- # common constants msec = 1000 minute = 60 * msec # ----------------------------------------------------------------------------- kraken = ccxt.kraken() # ----------------------------------------------------------------------------- from_datetime = '2017-09-01 00:00:00' from_timestamp = kraken.parse8601(from_datetime) # ----------------------------------------------------------------------------- now = kraken.milliseconds() # ----------------------------------------------------------------------------- data = [] while from_timestamp < now: print('Fetching candles starting from', kraken.iso8601(from_timestamp)) ohlcvs = kraken.fetch_ohlcv('BTC/USD', '1m', from_timestamp) # don't hit the rateLimit or you will be banned time.sleep(3 * kraken.rateLimit / msec) # Kraken returns 720 candles for 1m timeframe at once from_timestamp += len(ohlcvs) * minute data += ohlcvs
<commit_before># -*- coding: utf-8 -*- import os import sys import time # ----------------------------------------------------------------------------- root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root) # ----------------------------------------------------------------------------- import ccxt # noqa: E402 # ----------------------------------------------------------------------------- # common constants msec = 1000 minute = 60 * msec # ----------------------------------------------------------------------------- kraken = ccxt.kraken() # ----------------------------------------------------------------------------- from_datetime = '2017-09-01 00:00:00' from_timestamp = kraken.parse8601(from_datetime) # ----------------------------------------------------------------------------- now = kraken.milliseconds() # ----------------------------------------------------------------------------- while from_timestamp < now: print('Fetching candles starting from', kraken.iso8601(from_timestamp)) ohlcvs = kraken.fetch_ohlcv('BTC/USD', '1m', from_timestamp) # don't hit the rateLimit or you will be banned time.sleep(kraken.rateLimit / msec) # Kraken returns 720 candles for 1m timeframe at once from_timestamp += len(ohlcvs) * minute <commit_msg>Add delay 3 times the rate limit to avoid API errors<commit_after># -*- coding: utf-8 -*- import os import sys import time # ----------------------------------------------------------------------------- root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root) # ----------------------------------------------------------------------------- import ccxt # noqa: E402 # ----------------------------------------------------------------------------- # common constants msec = 1000 minute = 60 * msec # ----------------------------------------------------------------------------- kraken = ccxt.kraken() # ----------------------------------------------------------------------------- from_datetime = '2017-09-01 00:00:00' from_timestamp = kraken.parse8601(from_datetime) # ----------------------------------------------------------------------------- now = kraken.milliseconds() # ----------------------------------------------------------------------------- data = [] while from_timestamp < now: print('Fetching candles starting from', kraken.iso8601(from_timestamp)) ohlcvs = kraken.fetch_ohlcv('BTC/USD', '1m', from_timestamp) # don't hit the rateLimit or you will be banned time.sleep(3 * kraken.rateLimit / msec) # Kraken returns 720 candles for 1m timeframe at once from_timestamp += len(ohlcvs) * minute data += ohlcvs
6e6c5bb9f02a4f9d380ee20216f710a6de0b0583
fenum.py
fenum.py
#!/bin/env python import os import sys def main(): if len(sys.argv) == 1 or sys.argv[1].lower() == "-h" or sys.argv[1].lower() == "--help": print("Syntax: fenum.py [files...]") print("\tEnumerate the given files (starting at 1) in the same order as they are passed to the script.") return for k,v in enumerate(sys.argv[1:], 1): path, name = os.path.split(v if not v.endswith("/") else v[:-1]) if path: path += "/" try: fname = "{}{} - {}".format(path, str(k).zfill(len(str(len(sys.argv) - 1))), name) print("\"{}\" -> \"{}\"".format(v, fname)) os.rename(v, fname) except Exception as e: print(str(e)) main()
#!/bin/env python import os import sys import math def main(): if len(sys.argv) == 1 or sys.argv[1].lower() == "-h" or sys.argv[1].lower() == "--help": print("Syntax: fenum.py [files...]") print("\tEnumerate the given files (starting at 1) in the same order as they are passed to the script.") return for k,v in enumerate(sys.argv[1:], 1): path, name = os.path.split(v if not v.endswith("/") else v[:-1]) if path: path += "/" try: fname = "{}{} - {}".format( path, str(k).zfill(int(math.log10(len(sys.argv) - 1)) + 1), name) print("\"{}\" -> \"{}\"".format(v, fname)) os.rename(v, fname) except Exception as e: print(str(e)) main()
Use log10 to get the amount of digits
Use log10 to get the amount of digits
Python
mit
mphe/scripts,mphe/scripts,mall0c/scripts,mall0c/scripts
#!/bin/env python import os import sys def main(): if len(sys.argv) == 1 or sys.argv[1].lower() == "-h" or sys.argv[1].lower() == "--help": print("Syntax: fenum.py [files...]") print("\tEnumerate the given files (starting at 1) in the same order as they are passed to the script.") return for k,v in enumerate(sys.argv[1:], 1): path, name = os.path.split(v if not v.endswith("/") else v[:-1]) if path: path += "/" try: fname = "{}{} - {}".format(path, str(k).zfill(len(str(len(sys.argv) - 1))), name) print("\"{}\" -> \"{}\"".format(v, fname)) os.rename(v, fname) except Exception as e: print(str(e)) main() Use log10 to get the amount of digits
#!/bin/env python import os import sys import math def main(): if len(sys.argv) == 1 or sys.argv[1].lower() == "-h" or sys.argv[1].lower() == "--help": print("Syntax: fenum.py [files...]") print("\tEnumerate the given files (starting at 1) in the same order as they are passed to the script.") return for k,v in enumerate(sys.argv[1:], 1): path, name = os.path.split(v if not v.endswith("/") else v[:-1]) if path: path += "/" try: fname = "{}{} - {}".format( path, str(k).zfill(int(math.log10(len(sys.argv) - 1)) + 1), name) print("\"{}\" -> \"{}\"".format(v, fname)) os.rename(v, fname) except Exception as e: print(str(e)) main()
<commit_before>#!/bin/env python import os import sys def main(): if len(sys.argv) == 1 or sys.argv[1].lower() == "-h" or sys.argv[1].lower() == "--help": print("Syntax: fenum.py [files...]") print("\tEnumerate the given files (starting at 1) in the same order as they are passed to the script.") return for k,v in enumerate(sys.argv[1:], 1): path, name = os.path.split(v if not v.endswith("/") else v[:-1]) if path: path += "/" try: fname = "{}{} - {}".format(path, str(k).zfill(len(str(len(sys.argv) - 1))), name) print("\"{}\" -> \"{}\"".format(v, fname)) os.rename(v, fname) except Exception as e: print(str(e)) main() <commit_msg>Use log10 to get the amount of digits<commit_after>
#!/bin/env python import os import sys import math def main(): if len(sys.argv) == 1 or sys.argv[1].lower() == "-h" or sys.argv[1].lower() == "--help": print("Syntax: fenum.py [files...]") print("\tEnumerate the given files (starting at 1) in the same order as they are passed to the script.") return for k,v in enumerate(sys.argv[1:], 1): path, name = os.path.split(v if not v.endswith("/") else v[:-1]) if path: path += "/" try: fname = "{}{} - {}".format( path, str(k).zfill(int(math.log10(len(sys.argv) - 1)) + 1), name) print("\"{}\" -> \"{}\"".format(v, fname)) os.rename(v, fname) except Exception as e: print(str(e)) main()
#!/bin/env python import os import sys def main(): if len(sys.argv) == 1 or sys.argv[1].lower() == "-h" or sys.argv[1].lower() == "--help": print("Syntax: fenum.py [files...]") print("\tEnumerate the given files (starting at 1) in the same order as they are passed to the script.") return for k,v in enumerate(sys.argv[1:], 1): path, name = os.path.split(v if not v.endswith("/") else v[:-1]) if path: path += "/" try: fname = "{}{} - {}".format(path, str(k).zfill(len(str(len(sys.argv) - 1))), name) print("\"{}\" -> \"{}\"".format(v, fname)) os.rename(v, fname) except Exception as e: print(str(e)) main() Use log10 to get the amount of digits#!/bin/env python import os import sys import math def main(): if len(sys.argv) == 1 or sys.argv[1].lower() == "-h" or sys.argv[1].lower() == "--help": print("Syntax: fenum.py [files...]") print("\tEnumerate the given files (starting at 1) in the same order as they are passed to the script.") return for k,v in enumerate(sys.argv[1:], 1): path, name = os.path.split(v if not v.endswith("/") else v[:-1]) if path: path += "/" try: fname = "{}{} - {}".format( path, str(k).zfill(int(math.log10(len(sys.argv) - 1)) + 1), name) print("\"{}\" -> \"{}\"".format(v, fname)) os.rename(v, fname) except Exception as e: print(str(e)) main()
<commit_before>#!/bin/env python import os import sys def main(): if len(sys.argv) == 1 or sys.argv[1].lower() == "-h" or sys.argv[1].lower() == "--help": print("Syntax: fenum.py [files...]") print("\tEnumerate the given files (starting at 1) in the same order as they are passed to the script.") return for k,v in enumerate(sys.argv[1:], 1): path, name = os.path.split(v if not v.endswith("/") else v[:-1]) if path: path += "/" try: fname = "{}{} - {}".format(path, str(k).zfill(len(str(len(sys.argv) - 1))), name) print("\"{}\" -> \"{}\"".format(v, fname)) os.rename(v, fname) except Exception as e: print(str(e)) main() <commit_msg>Use log10 to get the amount of digits<commit_after>#!/bin/env python import os import sys import math def main(): if len(sys.argv) == 1 or sys.argv[1].lower() == "-h" or sys.argv[1].lower() == "--help": print("Syntax: fenum.py [files...]") print("\tEnumerate the given files (starting at 1) in the same order as they are passed to the script.") return for k,v in enumerate(sys.argv[1:], 1): path, name = os.path.split(v if not v.endswith("/") else v[:-1]) if path: path += "/" try: fname = "{}{} - {}".format( path, str(k).zfill(int(math.log10(len(sys.argv) - 1)) + 1), name) print("\"{}\" -> \"{}\"".format(v, fname)) os.rename(v, fname) except Exception as e: print(str(e)) main()
1c58cdadb93180557275682fa34433de2b736445
forms.py
forms.py
from flask.ext.wtf import Form from wtforms import SelectField, BooleanField, IntegerField, TextField, \ validators # TODO add vailidation class TeamForm(Form): number = IntegerField("Number", [validators.Required(), validators.NumberRange(min=1, max=99999)]) name = TextField("Name", [validators.Required(), validators.Length(min=1, max=50)]) affiliation = TextField("Affiliation", [validators.Length(min=1, max=200)]) city = TextField("City", [validators.Length(min=1, max=50)]) state = TextField("State", [validators.Length(min=2, max=2)]) class ScoreForm(Form): team_id = SelectField(u'Team', coerce=int) tree_branch_is_closer = BooleanField(default=False) tree_branch_is_intact = BooleanField(default=False) cargo_plane_location = SelectField(choices=[('0', 'None'), ('1', 'Yellow only'), ('2', 'Light blue')])
from flask.ext.wtf import Form from wtforms import SelectField, BooleanField, IntegerField, TextField, \ validators class TeamForm(Form): number = IntegerField("Number", [validators.Required(), validators.NumberRange(min=1, max=99999)]) name = TextField("Name", [validators.Required(), validators.Length(min=1, max=50)]) affiliation = TextField("Affiliation", [validators.Length(min=1, max=200)]) city = TextField("City", [validators.Length(min=1, max=50)]) state = TextField("State", [validators.Length(min=2, max=2)]) # TODO add validation class ScoreForm(Form): team_id = SelectField(u'Team', coerce=int) tree_branch_is_closer = BooleanField(default=False) tree_branch_is_intact = BooleanField(default=False) cargo_plane_location = SelectField(choices=[('0', 'None'), ('1', 'Yellow only'), ('2', 'Light blue')])
Move TODO tag to correct class
Move TODO tag to correct class
Python
mit
rtfoley/scorepy,rtfoley/scorepy,rtfoley/scorepy
from flask.ext.wtf import Form from wtforms import SelectField, BooleanField, IntegerField, TextField, \ validators # TODO add vailidation class TeamForm(Form): number = IntegerField("Number", [validators.Required(), validators.NumberRange(min=1, max=99999)]) name = TextField("Name", [validators.Required(), validators.Length(min=1, max=50)]) affiliation = TextField("Affiliation", [validators.Length(min=1, max=200)]) city = TextField("City", [validators.Length(min=1, max=50)]) state = TextField("State", [validators.Length(min=2, max=2)]) class ScoreForm(Form): team_id = SelectField(u'Team', coerce=int) tree_branch_is_closer = BooleanField(default=False) tree_branch_is_intact = BooleanField(default=False) cargo_plane_location = SelectField(choices=[('0', 'None'), ('1', 'Yellow only'), ('2', 'Light blue')]) Move TODO tag to correct class
from flask.ext.wtf import Form from wtforms import SelectField, BooleanField, IntegerField, TextField, \ validators class TeamForm(Form): number = IntegerField("Number", [validators.Required(), validators.NumberRange(min=1, max=99999)]) name = TextField("Name", [validators.Required(), validators.Length(min=1, max=50)]) affiliation = TextField("Affiliation", [validators.Length(min=1, max=200)]) city = TextField("City", [validators.Length(min=1, max=50)]) state = TextField("State", [validators.Length(min=2, max=2)]) # TODO add validation class ScoreForm(Form): team_id = SelectField(u'Team', coerce=int) tree_branch_is_closer = BooleanField(default=False) tree_branch_is_intact = BooleanField(default=False) cargo_plane_location = SelectField(choices=[('0', 'None'), ('1', 'Yellow only'), ('2', 'Light blue')])
<commit_before>from flask.ext.wtf import Form from wtforms import SelectField, BooleanField, IntegerField, TextField, \ validators # TODO add vailidation class TeamForm(Form): number = IntegerField("Number", [validators.Required(), validators.NumberRange(min=1, max=99999)]) name = TextField("Name", [validators.Required(), validators.Length(min=1, max=50)]) affiliation = TextField("Affiliation", [validators.Length(min=1, max=200)]) city = TextField("City", [validators.Length(min=1, max=50)]) state = TextField("State", [validators.Length(min=2, max=2)]) class ScoreForm(Form): team_id = SelectField(u'Team', coerce=int) tree_branch_is_closer = BooleanField(default=False) tree_branch_is_intact = BooleanField(default=False) cargo_plane_location = SelectField(choices=[('0', 'None'), ('1', 'Yellow only'), ('2', 'Light blue')]) <commit_msg>Move TODO tag to correct class<commit_after>
from flask.ext.wtf import Form from wtforms import SelectField, BooleanField, IntegerField, TextField, \ validators class TeamForm(Form): number = IntegerField("Number", [validators.Required(), validators.NumberRange(min=1, max=99999)]) name = TextField("Name", [validators.Required(), validators.Length(min=1, max=50)]) affiliation = TextField("Affiliation", [validators.Length(min=1, max=200)]) city = TextField("City", [validators.Length(min=1, max=50)]) state = TextField("State", [validators.Length(min=2, max=2)]) # TODO add validation class ScoreForm(Form): team_id = SelectField(u'Team', coerce=int) tree_branch_is_closer = BooleanField(default=False) tree_branch_is_intact = BooleanField(default=False) cargo_plane_location = SelectField(choices=[('0', 'None'), ('1', 'Yellow only'), ('2', 'Light blue')])
from flask.ext.wtf import Form from wtforms import SelectField, BooleanField, IntegerField, TextField, \ validators # TODO add vailidation class TeamForm(Form): number = IntegerField("Number", [validators.Required(), validators.NumberRange(min=1, max=99999)]) name = TextField("Name", [validators.Required(), validators.Length(min=1, max=50)]) affiliation = TextField("Affiliation", [validators.Length(min=1, max=200)]) city = TextField("City", [validators.Length(min=1, max=50)]) state = TextField("State", [validators.Length(min=2, max=2)]) class ScoreForm(Form): team_id = SelectField(u'Team', coerce=int) tree_branch_is_closer = BooleanField(default=False) tree_branch_is_intact = BooleanField(default=False) cargo_plane_location = SelectField(choices=[('0', 'None'), ('1', 'Yellow only'), ('2', 'Light blue')]) Move TODO tag to correct classfrom flask.ext.wtf import Form from wtforms import SelectField, BooleanField, IntegerField, TextField, \ validators class TeamForm(Form): number = IntegerField("Number", [validators.Required(), validators.NumberRange(min=1, max=99999)]) name = TextField("Name", [validators.Required(), validators.Length(min=1, max=50)]) affiliation = TextField("Affiliation", [validators.Length(min=1, max=200)]) city = TextField("City", [validators.Length(min=1, max=50)]) state = TextField("State", [validators.Length(min=2, max=2)]) # TODO add validation class ScoreForm(Form): team_id = SelectField(u'Team', coerce=int) tree_branch_is_closer = BooleanField(default=False) tree_branch_is_intact = BooleanField(default=False) cargo_plane_location = SelectField(choices=[('0', 'None'), ('1', 'Yellow only'), ('2', 'Light blue')])
<commit_before>from flask.ext.wtf import Form from wtforms import SelectField, BooleanField, IntegerField, TextField, \ validators # TODO add vailidation class TeamForm(Form): number = IntegerField("Number", [validators.Required(), validators.NumberRange(min=1, max=99999)]) name = TextField("Name", [validators.Required(), validators.Length(min=1, max=50)]) affiliation = TextField("Affiliation", [validators.Length(min=1, max=200)]) city = TextField("City", [validators.Length(min=1, max=50)]) state = TextField("State", [validators.Length(min=2, max=2)]) class ScoreForm(Form): team_id = SelectField(u'Team', coerce=int) tree_branch_is_closer = BooleanField(default=False) tree_branch_is_intact = BooleanField(default=False) cargo_plane_location = SelectField(choices=[('0', 'None'), ('1', 'Yellow only'), ('2', 'Light blue')]) <commit_msg>Move TODO tag to correct class<commit_after>from flask.ext.wtf import Form from wtforms import SelectField, BooleanField, IntegerField, TextField, \ validators class TeamForm(Form): number = IntegerField("Number", [validators.Required(), validators.NumberRange(min=1, max=99999)]) name = TextField("Name", [validators.Required(), validators.Length(min=1, max=50)]) affiliation = TextField("Affiliation", [validators.Length(min=1, max=200)]) city = TextField("City", [validators.Length(min=1, max=50)]) state = TextField("State", [validators.Length(min=2, max=2)]) # TODO add validation class ScoreForm(Form): team_id = SelectField(u'Team', coerce=int) tree_branch_is_closer = BooleanField(default=False) tree_branch_is_intact = BooleanField(default=False) cargo_plane_location = SelectField(choices=[('0', 'None'), ('1', 'Yellow only'), ('2', 'Light blue')])
c1008646ad8eac1de30d0c1bfd95caebd66f7fa1
pax/plugins/peak_processing/ClassifyPeaks.py
pax/plugins/peak_processing/ClassifyPeaks.py
from pax import plugin, units class AdHocClassification(plugin.TransformPlugin): def transform_event(self, event): for peak in event.peaks: # Don't work on noise and lone_hit if peak.type in ('unknown', 'lone_hit'): continue if peak.range_90p_area < 150 * units.ns: peak.type = 's1' elif peak.range_90p_area > 200 * units.ns: if peak.area > 5: peak.type = 's2' else: peak.type = 'coincidence' return event
from pax import plugin, units class AdHocClassification(plugin.TransformPlugin): def transform_event(self, event): for peak in event.peaks: # Don't work on noise and lone_hit if peak.type in ('noise', 'lone_hit'): continue if peak.range_90p_area < 150 * units.ns: peak.type = 's1' elif peak.range_90p_area > 200 * units.ns: if peak.area > 5: peak.type = 's2' else: peak.type = 'coincidence' return event
Fix goof in classification fix
Fix goof in classification fix
Python
bsd-3-clause
XENON1T/pax,XENON1T/pax
from pax import plugin, units class AdHocClassification(plugin.TransformPlugin): def transform_event(self, event): for peak in event.peaks: # Don't work on noise and lone_hit if peak.type in ('unknown', 'lone_hit'): continue if peak.range_90p_area < 150 * units.ns: peak.type = 's1' elif peak.range_90p_area > 200 * units.ns: if peak.area > 5: peak.type = 's2' else: peak.type = 'coincidence' return event Fix goof in classification fix
from pax import plugin, units class AdHocClassification(plugin.TransformPlugin): def transform_event(self, event): for peak in event.peaks: # Don't work on noise and lone_hit if peak.type in ('noise', 'lone_hit'): continue if peak.range_90p_area < 150 * units.ns: peak.type = 's1' elif peak.range_90p_area > 200 * units.ns: if peak.area > 5: peak.type = 's2' else: peak.type = 'coincidence' return event
<commit_before>from pax import plugin, units class AdHocClassification(plugin.TransformPlugin): def transform_event(self, event): for peak in event.peaks: # Don't work on noise and lone_hit if peak.type in ('unknown', 'lone_hit'): continue if peak.range_90p_area < 150 * units.ns: peak.type = 's1' elif peak.range_90p_area > 200 * units.ns: if peak.area > 5: peak.type = 's2' else: peak.type = 'coincidence' return event <commit_msg>Fix goof in classification fix<commit_after>
from pax import plugin, units class AdHocClassification(plugin.TransformPlugin): def transform_event(self, event): for peak in event.peaks: # Don't work on noise and lone_hit if peak.type in ('noise', 'lone_hit'): continue if peak.range_90p_area < 150 * units.ns: peak.type = 's1' elif peak.range_90p_area > 200 * units.ns: if peak.area > 5: peak.type = 's2' else: peak.type = 'coincidence' return event
from pax import plugin, units class AdHocClassification(plugin.TransformPlugin): def transform_event(self, event): for peak in event.peaks: # Don't work on noise and lone_hit if peak.type in ('unknown', 'lone_hit'): continue if peak.range_90p_area < 150 * units.ns: peak.type = 's1' elif peak.range_90p_area > 200 * units.ns: if peak.area > 5: peak.type = 's2' else: peak.type = 'coincidence' return event Fix goof in classification fixfrom pax import plugin, units class AdHocClassification(plugin.TransformPlugin): def transform_event(self, event): for peak in event.peaks: # Don't work on noise and lone_hit if peak.type in ('noise', 'lone_hit'): continue if peak.range_90p_area < 150 * units.ns: peak.type = 's1' elif peak.range_90p_area > 200 * units.ns: if peak.area > 5: peak.type = 's2' else: peak.type = 'coincidence' return event
<commit_before>from pax import plugin, units class AdHocClassification(plugin.TransformPlugin): def transform_event(self, event): for peak in event.peaks: # Don't work on noise and lone_hit if peak.type in ('unknown', 'lone_hit'): continue if peak.range_90p_area < 150 * units.ns: peak.type = 's1' elif peak.range_90p_area > 200 * units.ns: if peak.area > 5: peak.type = 's2' else: peak.type = 'coincidence' return event <commit_msg>Fix goof in classification fix<commit_after>from pax import plugin, units class AdHocClassification(plugin.TransformPlugin): def transform_event(self, event): for peak in event.peaks: # Don't work on noise and lone_hit if peak.type in ('noise', 'lone_hit'): continue if peak.range_90p_area < 150 * units.ns: peak.type = 's1' elif peak.range_90p_area > 200 * units.ns: if peak.area > 5: peak.type = 's2' else: peak.type = 'coincidence' return event
9e42514ac030cc2fa3aab89addfa48cf0342f991
kremlin/utils.py
kremlin/utils.py
""" # # #### ##### # # ##### # # # # # # # # # ## ## # # # ## # # ### #### #### # # # # # # # # ##### # # # # # # # # ## # # # # # # ##### # # # # # # # # Kremlin Magical Everything System Glasnost Image Board and Boredom Inhibitor """ import os from PIL import Image def mkthumb(fp, h=128, w=128): """ Create a thumbnail for an image fp filesystem path to the full size image h height (default is 128) w width (default is 128) The thumbnail will be unceremoniously dumped in the same directory with 'thumbnail' between the file name and extension. """ size = (h, w) f, ext = os.path.splitext(fp) im = Image.open(fp) im.thumbnail(size, Image.ANTIALIAS) im.save('.thumbnail'.join([f, ext]))
""" # # #### ##### # # ##### # # # # # # # # # ## ## # # # ## # # ### #### #### # # # # # # # # ##### # # # # # # # # ## # # # # # # ##### # # # # # # # # Kremlin Magical Everything System Glasnost Image Board and Boredom Inhibitor """ import os from PIL import Image def mkthumb(fp, h=128, w=128): """ Create a thumbnail for an image fp filesystem path to the full size image h height (default is 128) w width (default is 128) The thumbnail will be unceremoniously dumped in the same directory with 'thumbnail' between the file name and extension. """ size = (h, w) f, ext = os.path.splitext(fp) with Image.open(fp) as im: im.thumbnail(size, Image.ANTIALIAS) im.save('.thumbnail'.join([f, ext]))
Use context manager for PIL Open
Use context manager for PIL Open This is currently kind of buggy with Python3. A harmless warning is emitted since the handle is still closed late inside PIL/Pillow.
Python
bsd-2-clause
glasnost/kremlin,glasnost/kremlin,glasnost/kremlin
""" # # #### ##### # # ##### # # # # # # # # # ## ## # # # ## # # ### #### #### # # # # # # # # ##### # # # # # # # # ## # # # # # # ##### # # # # # # # # Kremlin Magical Everything System Glasnost Image Board and Boredom Inhibitor """ import os from PIL import Image def mkthumb(fp, h=128, w=128): """ Create a thumbnail for an image fp filesystem path to the full size image h height (default is 128) w width (default is 128) The thumbnail will be unceremoniously dumped in the same directory with 'thumbnail' between the file name and extension. """ size = (h, w) f, ext = os.path.splitext(fp) im = Image.open(fp) im.thumbnail(size, Image.ANTIALIAS) im.save('.thumbnail'.join([f, ext])) Use context manager for PIL Open This is currently kind of buggy with Python3. A harmless warning is emitted since the handle is still closed late inside PIL/Pillow.
""" # # #### ##### # # ##### # # # # # # # # # ## ## # # # ## # # ### #### #### # # # # # # # # ##### # # # # # # # # ## # # # # # # ##### # # # # # # # # Kremlin Magical Everything System Glasnost Image Board and Boredom Inhibitor """ import os from PIL import Image def mkthumb(fp, h=128, w=128): """ Create a thumbnail for an image fp filesystem path to the full size image h height (default is 128) w width (default is 128) The thumbnail will be unceremoniously dumped in the same directory with 'thumbnail' between the file name and extension. """ size = (h, w) f, ext = os.path.splitext(fp) with Image.open(fp) as im: im.thumbnail(size, Image.ANTIALIAS) im.save('.thumbnail'.join([f, ext]))
<commit_before>""" # # #### ##### # # ##### # # # # # # # # # ## ## # # # ## # # ### #### #### # # # # # # # # ##### # # # # # # # # ## # # # # # # ##### # # # # # # # # Kremlin Magical Everything System Glasnost Image Board and Boredom Inhibitor """ import os from PIL import Image def mkthumb(fp, h=128, w=128): """ Create a thumbnail for an image fp filesystem path to the full size image h height (default is 128) w width (default is 128) The thumbnail will be unceremoniously dumped in the same directory with 'thumbnail' between the file name and extension. """ size = (h, w) f, ext = os.path.splitext(fp) im = Image.open(fp) im.thumbnail(size, Image.ANTIALIAS) im.save('.thumbnail'.join([f, ext])) <commit_msg>Use context manager for PIL Open This is currently kind of buggy with Python3. A harmless warning is emitted since the handle is still closed late inside PIL/Pillow.<commit_after>
""" # # #### ##### # # ##### # # # # # # # # # ## ## # # # ## # # ### #### #### # # # # # # # # ##### # # # # # # # # ## # # # # # # ##### # # # # # # # # Kremlin Magical Everything System Glasnost Image Board and Boredom Inhibitor """ import os from PIL import Image def mkthumb(fp, h=128, w=128): """ Create a thumbnail for an image fp filesystem path to the full size image h height (default is 128) w width (default is 128) The thumbnail will be unceremoniously dumped in the same directory with 'thumbnail' between the file name and extension. """ size = (h, w) f, ext = os.path.splitext(fp) with Image.open(fp) as im: im.thumbnail(size, Image.ANTIALIAS) im.save('.thumbnail'.join([f, ext]))
""" # # #### ##### # # ##### # # # # # # # # # ## ## # # # ## # # ### #### #### # # # # # # # # ##### # # # # # # # # ## # # # # # # ##### # # # # # # # # Kremlin Magical Everything System Glasnost Image Board and Boredom Inhibitor """ import os from PIL import Image def mkthumb(fp, h=128, w=128): """ Create a thumbnail for an image fp filesystem path to the full size image h height (default is 128) w width (default is 128) The thumbnail will be unceremoniously dumped in the same directory with 'thumbnail' between the file name and extension. """ size = (h, w) f, ext = os.path.splitext(fp) im = Image.open(fp) im.thumbnail(size, Image.ANTIALIAS) im.save('.thumbnail'.join([f, ext])) Use context manager for PIL Open This is currently kind of buggy with Python3. A harmless warning is emitted since the handle is still closed late inside PIL/Pillow.""" # # #### ##### # # ##### # # # # # # # # # ## ## # # # ## # # ### #### #### # # # # # # # # ##### # # # # # # # # ## # # # # # # ##### # # # # # # # # Kremlin Magical Everything System Glasnost Image Board and Boredom Inhibitor """ import os from PIL import Image def mkthumb(fp, h=128, w=128): """ Create a thumbnail for an image fp filesystem path to the full size image h height (default is 128) w width (default is 128) The thumbnail will be unceremoniously dumped in the same directory with 'thumbnail' between the file name and extension. """ size = (h, w) f, ext = os.path.splitext(fp) with Image.open(fp) as im: im.thumbnail(size, Image.ANTIALIAS) im.save('.thumbnail'.join([f, ext]))
<commit_before>""" # # #### ##### # # ##### # # # # # # # # # ## ## # # # ## # # ### #### #### # # # # # # # # ##### # # # # # # # # ## # # # # # # ##### # # # # # # # # Kremlin Magical Everything System Glasnost Image Board and Boredom Inhibitor """ import os from PIL import Image def mkthumb(fp, h=128, w=128): """ Create a thumbnail for an image fp filesystem path to the full size image h height (default is 128) w width (default is 128) The thumbnail will be unceremoniously dumped in the same directory with 'thumbnail' between the file name and extension. """ size = (h, w) f, ext = os.path.splitext(fp) im = Image.open(fp) im.thumbnail(size, Image.ANTIALIAS) im.save('.thumbnail'.join([f, ext])) <commit_msg>Use context manager for PIL Open This is currently kind of buggy with Python3. A harmless warning is emitted since the handle is still closed late inside PIL/Pillow.<commit_after>""" # # #### ##### # # ##### # # # # # # # # # ## ## # # # ## # # ### #### #### # # # # # # # # ##### # # # # # # # # ## # # # # # # ##### # # # # # # # # Kremlin Magical Everything System Glasnost Image Board and Boredom Inhibitor """ import os from PIL import Image def mkthumb(fp, h=128, w=128): """ Create a thumbnail for an image fp filesystem path to the full size image h height (default is 128) w width (default is 128) The thumbnail will be unceremoniously dumped in the same directory with 'thumbnail' between the file name and extension. """ size = (h, w) f, ext = os.path.splitext(fp) with Image.open(fp) as im: im.thumbnail(size, Image.ANTIALIAS) im.save('.thumbnail'.join([f, ext]))
12d2e8033c46f06f5442cba40a7c2c673196ad1c
akanda/horizon/routers/views.py
akanda/horizon/routers/views.py
from django.utils.translation import ugettext_lazy as _ # noqa from horizon import exceptions from openstack_dashboard import api def get_interfaces_data(self): try: router_id = self.kwargs['router_id'] router = api.quantum.router_get(self.request, router_id) ports = [api.quantum.Port(p) for p in router.ports] except Exception: ports = [] msg = _( 'Port list can not be retrieved for router ID %s' % self.kwargs.get('router_id') ) exceptions.handle(self.request, msg) for p in ports: p.set_id_as_name_if_empty() return ports
from django.utils.translation import ugettext_lazy as _ # noqa from horizon import exceptions from openstack_dashboard import api def get_interfaces_data(self): try: router_id = self.kwargs['router_id'] router = api.quantum.router_get(self.request, router_id) # Note(rods): Right now we are listing, for both normal and # admin users, all the ports on the user's networks # the router is associated with. We may want in the # future show the ports on the mgt and the external # networks for the admin users. ports = [api.quantum.Port(p) for p in router.ports if p['device_owner'] == 'network:router_interface'] except Exception: ports = [] msg = _( 'Port list can not be retrieved for router ID %s' % self.kwargs.get('router_id') ) exceptions.handle(self.request, msg) for p in ports: p.set_id_as_name_if_empty() return ports
Fix the router's interfaces listing view to show only the interfaces on the user's networks filtering out interfaces on the mgt and public networks.
Fix the router's interfaces listing view to show only the interfaces on the user's networks filtering out interfaces on the mgt and public networks. DHC-1512 Change-Id: I9b68b75d5e8325c4c70090fa500a417e23b1836f Signed-off-by: Rosario Di Somma <73b2fe5f91895aea2b4d0e8942a5edf9f18fa897@dreamhost.com>
Python
apache-2.0
dreamhost/akanda-horizon,dreamhost/akanda-horizon
from django.utils.translation import ugettext_lazy as _ # noqa from horizon import exceptions from openstack_dashboard import api def get_interfaces_data(self): try: router_id = self.kwargs['router_id'] router = api.quantum.router_get(self.request, router_id) ports = [api.quantum.Port(p) for p in router.ports] except Exception: ports = [] msg = _( 'Port list can not be retrieved for router ID %s' % self.kwargs.get('router_id') ) exceptions.handle(self.request, msg) for p in ports: p.set_id_as_name_if_empty() return ports Fix the router's interfaces listing view to show only the interfaces on the user's networks filtering out interfaces on the mgt and public networks. DHC-1512 Change-Id: I9b68b75d5e8325c4c70090fa500a417e23b1836f Signed-off-by: Rosario Di Somma <73b2fe5f91895aea2b4d0e8942a5edf9f18fa897@dreamhost.com>
from django.utils.translation import ugettext_lazy as _ # noqa from horizon import exceptions from openstack_dashboard import api def get_interfaces_data(self): try: router_id = self.kwargs['router_id'] router = api.quantum.router_get(self.request, router_id) # Note(rods): Right now we are listing, for both normal and # admin users, all the ports on the user's networks # the router is associated with. We may want in the # future show the ports on the mgt and the external # networks for the admin users. ports = [api.quantum.Port(p) for p in router.ports if p['device_owner'] == 'network:router_interface'] except Exception: ports = [] msg = _( 'Port list can not be retrieved for router ID %s' % self.kwargs.get('router_id') ) exceptions.handle(self.request, msg) for p in ports: p.set_id_as_name_if_empty() return ports
<commit_before>from django.utils.translation import ugettext_lazy as _ # noqa from horizon import exceptions from openstack_dashboard import api def get_interfaces_data(self): try: router_id = self.kwargs['router_id'] router = api.quantum.router_get(self.request, router_id) ports = [api.quantum.Port(p) for p in router.ports] except Exception: ports = [] msg = _( 'Port list can not be retrieved for router ID %s' % self.kwargs.get('router_id') ) exceptions.handle(self.request, msg) for p in ports: p.set_id_as_name_if_empty() return ports <commit_msg>Fix the router's interfaces listing view to show only the interfaces on the user's networks filtering out interfaces on the mgt and public networks. DHC-1512 Change-Id: I9b68b75d5e8325c4c70090fa500a417e23b1836f Signed-off-by: Rosario Di Somma <73b2fe5f91895aea2b4d0e8942a5edf9f18fa897@dreamhost.com><commit_after>
from django.utils.translation import ugettext_lazy as _ # noqa from horizon import exceptions from openstack_dashboard import api def get_interfaces_data(self): try: router_id = self.kwargs['router_id'] router = api.quantum.router_get(self.request, router_id) # Note(rods): Right now we are listing, for both normal and # admin users, all the ports on the user's networks # the router is associated with. We may want in the # future show the ports on the mgt and the external # networks for the admin users. ports = [api.quantum.Port(p) for p in router.ports if p['device_owner'] == 'network:router_interface'] except Exception: ports = [] msg = _( 'Port list can not be retrieved for router ID %s' % self.kwargs.get('router_id') ) exceptions.handle(self.request, msg) for p in ports: p.set_id_as_name_if_empty() return ports
from django.utils.translation import ugettext_lazy as _ # noqa from horizon import exceptions from openstack_dashboard import api def get_interfaces_data(self): try: router_id = self.kwargs['router_id'] router = api.quantum.router_get(self.request, router_id) ports = [api.quantum.Port(p) for p in router.ports] except Exception: ports = [] msg = _( 'Port list can not be retrieved for router ID %s' % self.kwargs.get('router_id') ) exceptions.handle(self.request, msg) for p in ports: p.set_id_as_name_if_empty() return ports Fix the router's interfaces listing view to show only the interfaces on the user's networks filtering out interfaces on the mgt and public networks. DHC-1512 Change-Id: I9b68b75d5e8325c4c70090fa500a417e23b1836f Signed-off-by: Rosario Di Somma <73b2fe5f91895aea2b4d0e8942a5edf9f18fa897@dreamhost.com>from django.utils.translation import ugettext_lazy as _ # noqa from horizon import exceptions from openstack_dashboard import api def get_interfaces_data(self): try: router_id = self.kwargs['router_id'] router = api.quantum.router_get(self.request, router_id) # Note(rods): Right now we are listing, for both normal and # admin users, all the ports on the user's networks # the router is associated with. We may want in the # future show the ports on the mgt and the external # networks for the admin users. ports = [api.quantum.Port(p) for p in router.ports if p['device_owner'] == 'network:router_interface'] except Exception: ports = [] msg = _( 'Port list can not be retrieved for router ID %s' % self.kwargs.get('router_id') ) exceptions.handle(self.request, msg) for p in ports: p.set_id_as_name_if_empty() return ports
<commit_before>from django.utils.translation import ugettext_lazy as _ # noqa from horizon import exceptions from openstack_dashboard import api def get_interfaces_data(self): try: router_id = self.kwargs['router_id'] router = api.quantum.router_get(self.request, router_id) ports = [api.quantum.Port(p) for p in router.ports] except Exception: ports = [] msg = _( 'Port list can not be retrieved for router ID %s' % self.kwargs.get('router_id') ) exceptions.handle(self.request, msg) for p in ports: p.set_id_as_name_if_empty() return ports <commit_msg>Fix the router's interfaces listing view to show only the interfaces on the user's networks filtering out interfaces on the mgt and public networks. DHC-1512 Change-Id: I9b68b75d5e8325c4c70090fa500a417e23b1836f Signed-off-by: Rosario Di Somma <73b2fe5f91895aea2b4d0e8942a5edf9f18fa897@dreamhost.com><commit_after>from django.utils.translation import ugettext_lazy as _ # noqa from horizon import exceptions from openstack_dashboard import api def get_interfaces_data(self): try: router_id = self.kwargs['router_id'] router = api.quantum.router_get(self.request, router_id) # Note(rods): Right now we are listing, for both normal and # admin users, all the ports on the user's networks # the router is associated with. We may want in the # future show the ports on the mgt and the external # networks for the admin users. ports = [api.quantum.Port(p) for p in router.ports if p['device_owner'] == 'network:router_interface'] except Exception: ports = [] msg = _( 'Port list can not be retrieved for router ID %s' % self.kwargs.get('router_id') ) exceptions.handle(self.request, msg) for p in ports: p.set_id_as_name_if_empty() return ports
a2530b9cd2baf70591e377b6eb7d5104493989a1
test/conftest.py
test/conftest.py
def pytest_addoption(parser): parser.addoption("--domain", action="append", default=[], help="list of stringinputs to pass to test functions") def pytest_generate_tests(metafunc): if 'domain' in metafunc.fixturenames: metafunc.parametrize("domain", metafunc.config.option.domain)
def pytest_addoption(parser): parser.addoption("--domain", action="append", default=[], help="list of stringinputs to pass to test functions") parser.addoption("--url", action="append", default=[], help="list of stringinputs to pass to test functions") def pytest_generate_tests(metafunc): if 'domain' in metafunc.fixturenames: metafunc.parametrize("domain", metafunc.config.option.domain) if 'url' in metafunc.fixturenames: metafunc.parametrize("url", metafunc.config.option.url)
Add test parameter for site url
Add test parameter for site url
Python
mit
HIIT/mediacollection
def pytest_addoption(parser): parser.addoption("--domain", action="append", default=[], help="list of stringinputs to pass to test functions") def pytest_generate_tests(metafunc): if 'domain' in metafunc.fixturenames: metafunc.parametrize("domain", metafunc.config.option.domain) Add test parameter for site url
def pytest_addoption(parser): parser.addoption("--domain", action="append", default=[], help="list of stringinputs to pass to test functions") parser.addoption("--url", action="append", default=[], help="list of stringinputs to pass to test functions") def pytest_generate_tests(metafunc): if 'domain' in metafunc.fixturenames: metafunc.parametrize("domain", metafunc.config.option.domain) if 'url' in metafunc.fixturenames: metafunc.parametrize("url", metafunc.config.option.url)
<commit_before>def pytest_addoption(parser): parser.addoption("--domain", action="append", default=[], help="list of stringinputs to pass to test functions") def pytest_generate_tests(metafunc): if 'domain' in metafunc.fixturenames: metafunc.parametrize("domain", metafunc.config.option.domain) <commit_msg>Add test parameter for site url<commit_after>
def pytest_addoption(parser): parser.addoption("--domain", action="append", default=[], help="list of stringinputs to pass to test functions") parser.addoption("--url", action="append", default=[], help="list of stringinputs to pass to test functions") def pytest_generate_tests(metafunc): if 'domain' in metafunc.fixturenames: metafunc.parametrize("domain", metafunc.config.option.domain) if 'url' in metafunc.fixturenames: metafunc.parametrize("url", metafunc.config.option.url)
def pytest_addoption(parser): parser.addoption("--domain", action="append", default=[], help="list of stringinputs to pass to test functions") def pytest_generate_tests(metafunc): if 'domain' in metafunc.fixturenames: metafunc.parametrize("domain", metafunc.config.option.domain) Add test parameter for site urldef pytest_addoption(parser): parser.addoption("--domain", action="append", default=[], help="list of stringinputs to pass to test functions") parser.addoption("--url", action="append", default=[], help="list of stringinputs to pass to test functions") def pytest_generate_tests(metafunc): if 'domain' in metafunc.fixturenames: metafunc.parametrize("domain", metafunc.config.option.domain) if 'url' in metafunc.fixturenames: metafunc.parametrize("url", metafunc.config.option.url)
<commit_before>def pytest_addoption(parser): parser.addoption("--domain", action="append", default=[], help="list of stringinputs to pass to test functions") def pytest_generate_tests(metafunc): if 'domain' in metafunc.fixturenames: metafunc.parametrize("domain", metafunc.config.option.domain) <commit_msg>Add test parameter for site url<commit_after>def pytest_addoption(parser): parser.addoption("--domain", action="append", default=[], help="list of stringinputs to pass to test functions") parser.addoption("--url", action="append", default=[], help="list of stringinputs to pass to test functions") def pytest_generate_tests(metafunc): if 'domain' in metafunc.fixturenames: metafunc.parametrize("domain", metafunc.config.option.domain) if 'url' in metafunc.fixturenames: metafunc.parametrize("url", metafunc.config.option.url)
e0af42d5431c42ee5e12fb228978ac8ce9c62807
server/modules/persistent.py
server/modules/persistent.py
#!/usr/bin/python3 import argparse import os import icon_lib parser = argparse.ArgumentParser(description='iconograph persistent') parser.add_argument( '--chroot-path', dest='chroot_path', action='store', required=True) FLAGS = parser.parse_args() def main(): module = icon_lib.IconModule(FLAGS.chroot_path) os.mkdir(os.path.join(FLAGS.chroot_path, 'persistent')) tool_path = os.path.join(FLAGS.chroot_path, 'icon', 'persistent') os.makedirs(tool_path, exist_ok=True) script = os.path.join(tool_path, 'startup.sh') with open(script, 'w') as fh: os.chmod(fh.fileno(), 0o755) fh.write("""\ #!/bin/bash set -ex e2fsck -a /persistent mount -o data=journal,noatime,sync LABEL=PERSISTENT /persistent """) with module.ServiceFile('persistent.service') as fh: fh.write(""" [Unit] Description=Mount /persistent DefaultDependencies=no Conflicts=shutdown.target After=systemd-remount-fs.service Before=sysinit.target [Service] Type=oneshot RemainAfterExit=yes ExecStart=/icon/persistent/startup.sh [Install] WantedBy=sysinit.target """) module.EnableService('persistent.service') if __name__ == '__main__': main()
#!/usr/bin/python3 import argparse import os import icon_lib parser = argparse.ArgumentParser(description='iconograph persistent') parser.add_argument( '--chroot-path', dest='chroot_path', action='store', required=True) FLAGS = parser.parse_args() def main(): module = icon_lib.IconModule(FLAGS.chroot_path) os.mkdir(os.path.join(FLAGS.chroot_path, 'persistent')) tool_path = os.path.join(FLAGS.chroot_path, 'icon', 'persistent') os.makedirs(tool_path, exist_ok=True) script = os.path.join(tool_path, 'startup.sh') with open(script, 'w') as fh: os.chmod(fh.fileno(), 0o755) fh.write("""\ #!/bin/bash set -ex e2fsck -a /persistent mount -o noatime LABEL=PERSISTENT /persistent """) with module.ServiceFile('persistent.service') as fh: fh.write(""" [Unit] Description=Mount /persistent DefaultDependencies=no Conflicts=shutdown.target After=systemd-remount-fs.service Before=sysinit.target [Service] Type=oneshot RemainAfterExit=yes ExecStart=/icon/persistent/startup.sh [Install] WantedBy=sysinit.target """) module.EnableService('persistent.service') if __name__ == '__main__': main()
Remove flags that drastically slow down writes
Remove flags that drastically slow down writes
Python
apache-2.0
robot-tools/iconograph,robot-tools/iconograph,robot-tools/iconograph,robot-tools/iconograph
#!/usr/bin/python3 import argparse import os import icon_lib parser = argparse.ArgumentParser(description='iconograph persistent') parser.add_argument( '--chroot-path', dest='chroot_path', action='store', required=True) FLAGS = parser.parse_args() def main(): module = icon_lib.IconModule(FLAGS.chroot_path) os.mkdir(os.path.join(FLAGS.chroot_path, 'persistent')) tool_path = os.path.join(FLAGS.chroot_path, 'icon', 'persistent') os.makedirs(tool_path, exist_ok=True) script = os.path.join(tool_path, 'startup.sh') with open(script, 'w') as fh: os.chmod(fh.fileno(), 0o755) fh.write("""\ #!/bin/bash set -ex e2fsck -a /persistent mount -o data=journal,noatime,sync LABEL=PERSISTENT /persistent """) with module.ServiceFile('persistent.service') as fh: fh.write(""" [Unit] Description=Mount /persistent DefaultDependencies=no Conflicts=shutdown.target After=systemd-remount-fs.service Before=sysinit.target [Service] Type=oneshot RemainAfterExit=yes ExecStart=/icon/persistent/startup.sh [Install] WantedBy=sysinit.target """) module.EnableService('persistent.service') if __name__ == '__main__': main() Remove flags that drastically slow down writes
#!/usr/bin/python3 import argparse import os import icon_lib parser = argparse.ArgumentParser(description='iconograph persistent') parser.add_argument( '--chroot-path', dest='chroot_path', action='store', required=True) FLAGS = parser.parse_args() def main(): module = icon_lib.IconModule(FLAGS.chroot_path) os.mkdir(os.path.join(FLAGS.chroot_path, 'persistent')) tool_path = os.path.join(FLAGS.chroot_path, 'icon', 'persistent') os.makedirs(tool_path, exist_ok=True) script = os.path.join(tool_path, 'startup.sh') with open(script, 'w') as fh: os.chmod(fh.fileno(), 0o755) fh.write("""\ #!/bin/bash set -ex e2fsck -a /persistent mount -o noatime LABEL=PERSISTENT /persistent """) with module.ServiceFile('persistent.service') as fh: fh.write(""" [Unit] Description=Mount /persistent DefaultDependencies=no Conflicts=shutdown.target After=systemd-remount-fs.service Before=sysinit.target [Service] Type=oneshot RemainAfterExit=yes ExecStart=/icon/persistent/startup.sh [Install] WantedBy=sysinit.target """) module.EnableService('persistent.service') if __name__ == '__main__': main()
<commit_before>#!/usr/bin/python3 import argparse import os import icon_lib parser = argparse.ArgumentParser(description='iconograph persistent') parser.add_argument( '--chroot-path', dest='chroot_path', action='store', required=True) FLAGS = parser.parse_args() def main(): module = icon_lib.IconModule(FLAGS.chroot_path) os.mkdir(os.path.join(FLAGS.chroot_path, 'persistent')) tool_path = os.path.join(FLAGS.chroot_path, 'icon', 'persistent') os.makedirs(tool_path, exist_ok=True) script = os.path.join(tool_path, 'startup.sh') with open(script, 'w') as fh: os.chmod(fh.fileno(), 0o755) fh.write("""\ #!/bin/bash set -ex e2fsck -a /persistent mount -o data=journal,noatime,sync LABEL=PERSISTENT /persistent """) with module.ServiceFile('persistent.service') as fh: fh.write(""" [Unit] Description=Mount /persistent DefaultDependencies=no Conflicts=shutdown.target After=systemd-remount-fs.service Before=sysinit.target [Service] Type=oneshot RemainAfterExit=yes ExecStart=/icon/persistent/startup.sh [Install] WantedBy=sysinit.target """) module.EnableService('persistent.service') if __name__ == '__main__': main() <commit_msg>Remove flags that drastically slow down writes<commit_after>
#!/usr/bin/python3 import argparse import os import icon_lib parser = argparse.ArgumentParser(description='iconograph persistent') parser.add_argument( '--chroot-path', dest='chroot_path', action='store', required=True) FLAGS = parser.parse_args() def main(): module = icon_lib.IconModule(FLAGS.chroot_path) os.mkdir(os.path.join(FLAGS.chroot_path, 'persistent')) tool_path = os.path.join(FLAGS.chroot_path, 'icon', 'persistent') os.makedirs(tool_path, exist_ok=True) script = os.path.join(tool_path, 'startup.sh') with open(script, 'w') as fh: os.chmod(fh.fileno(), 0o755) fh.write("""\ #!/bin/bash set -ex e2fsck -a /persistent mount -o noatime LABEL=PERSISTENT /persistent """) with module.ServiceFile('persistent.service') as fh: fh.write(""" [Unit] Description=Mount /persistent DefaultDependencies=no Conflicts=shutdown.target After=systemd-remount-fs.service Before=sysinit.target [Service] Type=oneshot RemainAfterExit=yes ExecStart=/icon/persistent/startup.sh [Install] WantedBy=sysinit.target """) module.EnableService('persistent.service') if __name__ == '__main__': main()
#!/usr/bin/python3 import argparse import os import icon_lib parser = argparse.ArgumentParser(description='iconograph persistent') parser.add_argument( '--chroot-path', dest='chroot_path', action='store', required=True) FLAGS = parser.parse_args() def main(): module = icon_lib.IconModule(FLAGS.chroot_path) os.mkdir(os.path.join(FLAGS.chroot_path, 'persistent')) tool_path = os.path.join(FLAGS.chroot_path, 'icon', 'persistent') os.makedirs(tool_path, exist_ok=True) script = os.path.join(tool_path, 'startup.sh') with open(script, 'w') as fh: os.chmod(fh.fileno(), 0o755) fh.write("""\ #!/bin/bash set -ex e2fsck -a /persistent mount -o data=journal,noatime,sync LABEL=PERSISTENT /persistent """) with module.ServiceFile('persistent.service') as fh: fh.write(""" [Unit] Description=Mount /persistent DefaultDependencies=no Conflicts=shutdown.target After=systemd-remount-fs.service Before=sysinit.target [Service] Type=oneshot RemainAfterExit=yes ExecStart=/icon/persistent/startup.sh [Install] WantedBy=sysinit.target """) module.EnableService('persistent.service') if __name__ == '__main__': main() Remove flags that drastically slow down writes#!/usr/bin/python3 import argparse import os import icon_lib parser = argparse.ArgumentParser(description='iconograph persistent') parser.add_argument( '--chroot-path', dest='chroot_path', action='store', required=True) FLAGS = parser.parse_args() def main(): module = icon_lib.IconModule(FLAGS.chroot_path) os.mkdir(os.path.join(FLAGS.chroot_path, 'persistent')) tool_path = os.path.join(FLAGS.chroot_path, 'icon', 'persistent') os.makedirs(tool_path, exist_ok=True) script = os.path.join(tool_path, 'startup.sh') with open(script, 'w') as fh: os.chmod(fh.fileno(), 0o755) fh.write("""\ #!/bin/bash set -ex e2fsck -a /persistent mount -o noatime LABEL=PERSISTENT /persistent """) with module.ServiceFile('persistent.service') as fh: fh.write(""" [Unit] Description=Mount /persistent DefaultDependencies=no Conflicts=shutdown.target After=systemd-remount-fs.service Before=sysinit.target [Service] Type=oneshot RemainAfterExit=yes ExecStart=/icon/persistent/startup.sh [Install] WantedBy=sysinit.target """) module.EnableService('persistent.service') if __name__ == '__main__': main()
<commit_before>#!/usr/bin/python3 import argparse import os import icon_lib parser = argparse.ArgumentParser(description='iconograph persistent') parser.add_argument( '--chroot-path', dest='chroot_path', action='store', required=True) FLAGS = parser.parse_args() def main(): module = icon_lib.IconModule(FLAGS.chroot_path) os.mkdir(os.path.join(FLAGS.chroot_path, 'persistent')) tool_path = os.path.join(FLAGS.chroot_path, 'icon', 'persistent') os.makedirs(tool_path, exist_ok=True) script = os.path.join(tool_path, 'startup.sh') with open(script, 'w') as fh: os.chmod(fh.fileno(), 0o755) fh.write("""\ #!/bin/bash set -ex e2fsck -a /persistent mount -o data=journal,noatime,sync LABEL=PERSISTENT /persistent """) with module.ServiceFile('persistent.service') as fh: fh.write(""" [Unit] Description=Mount /persistent DefaultDependencies=no Conflicts=shutdown.target After=systemd-remount-fs.service Before=sysinit.target [Service] Type=oneshot RemainAfterExit=yes ExecStart=/icon/persistent/startup.sh [Install] WantedBy=sysinit.target """) module.EnableService('persistent.service') if __name__ == '__main__': main() <commit_msg>Remove flags that drastically slow down writes<commit_after>#!/usr/bin/python3 import argparse import os import icon_lib parser = argparse.ArgumentParser(description='iconograph persistent') parser.add_argument( '--chroot-path', dest='chroot_path', action='store', required=True) FLAGS = parser.parse_args() def main(): module = icon_lib.IconModule(FLAGS.chroot_path) os.mkdir(os.path.join(FLAGS.chroot_path, 'persistent')) tool_path = os.path.join(FLAGS.chroot_path, 'icon', 'persistent') os.makedirs(tool_path, exist_ok=True) script = os.path.join(tool_path, 'startup.sh') with open(script, 'w') as fh: os.chmod(fh.fileno(), 0o755) fh.write("""\ #!/bin/bash set -ex e2fsck -a /persistent mount -o noatime LABEL=PERSISTENT /persistent """) with module.ServiceFile('persistent.service') as fh: fh.write(""" [Unit] Description=Mount /persistent DefaultDependencies=no Conflicts=shutdown.target After=systemd-remount-fs.service Before=sysinit.target [Service] Type=oneshot RemainAfterExit=yes ExecStart=/icon/persistent/startup.sh [Install] WantedBy=sysinit.target """) module.EnableService('persistent.service') if __name__ == '__main__': main()
ec2092c683f721e32a2d1d9792f296e140d6ba45
paperwork_parser/exceptions.py
paperwork_parser/exceptions.py
class InvalidPDFError(ValueError): pass class FieldParseError(Exception): pass
class InvalidPDFError(ValueError): pass class UnknownVariantError(Exception): pass class FieldParseError(Exception): pass
Add new exception type for unknown variants
Add new exception type for unknown variants
Python
mit
loanzen/zen_document_parser
class InvalidPDFError(ValueError): pass class FieldParseError(Exception): pass Add new exception type for unknown variants
class InvalidPDFError(ValueError): pass class UnknownVariantError(Exception): pass class FieldParseError(Exception): pass
<commit_before> class InvalidPDFError(ValueError): pass class FieldParseError(Exception): pass <commit_msg>Add new exception type for unknown variants<commit_after>
class InvalidPDFError(ValueError): pass class UnknownVariantError(Exception): pass class FieldParseError(Exception): pass
class InvalidPDFError(ValueError): pass class FieldParseError(Exception): pass Add new exception type for unknown variants class InvalidPDFError(ValueError): pass class UnknownVariantError(Exception): pass class FieldParseError(Exception): pass
<commit_before> class InvalidPDFError(ValueError): pass class FieldParseError(Exception): pass <commit_msg>Add new exception type for unknown variants<commit_after> class InvalidPDFError(ValueError): pass class UnknownVariantError(Exception): pass class FieldParseError(Exception): pass
0b49114a6b0830fa0b05d32803ae52526b8e48ca
gnsq/backofftimer.py
gnsq/backofftimer.py
from random import randint class BackoffTimer(object): def __init__(self, ratio=1, max_interval=None, min_interval=None): self.c = 0 self.ratio = ratio self.max_interval = max_interval self.min_interval = min_interval def is_reset(self): return self.c == 0 def reset(self): self.c = 0 return self def success(self): self.c = max(self.c - 1, 0) return self def failure(self): self.c += 1 return self def get_interval(self): k = pow(2, self.c) - 1 interval = randint(0, k) * self.ratio if self.max_interval is not None: interval = min(interval, self.max_interval) if self.min_interval is not None: interval = max(interval, self.min_interval) return interval
import random class BackoffTimer(object): def __init__(self, ratio=1, max_interval=None, min_interval=None): self.c = 0 self.ratio = ratio self.max_interval = max_interval self.min_interval = min_interval def is_reset(self): return self.c == 0 def reset(self): self.c = 0 return self def success(self): self.c = max(self.c - 1, 0) return self def failure(self): self.c += 1 return self def get_interval(self): k = pow(2, self.c) - 1 interval = random.random() * k * self.ratio if self.max_interval is not None: interval = min(interval, self.max_interval) if self.min_interval is not None: interval = max(interval, self.min_interval) return interval
Return float for interval instead of int.
Return float for interval instead of int.
Python
bsd-3-clause
hiringsolved/gnsq,wtolson/gnsq,wtolson/gnsq
from random import randint class BackoffTimer(object): def __init__(self, ratio=1, max_interval=None, min_interval=None): self.c = 0 self.ratio = ratio self.max_interval = max_interval self.min_interval = min_interval def is_reset(self): return self.c == 0 def reset(self): self.c = 0 return self def success(self): self.c = max(self.c - 1, 0) return self def failure(self): self.c += 1 return self def get_interval(self): k = pow(2, self.c) - 1 interval = randint(0, k) * self.ratio if self.max_interval is not None: interval = min(interval, self.max_interval) if self.min_interval is not None: interval = max(interval, self.min_interval) return interval Return float for interval instead of int.
import random class BackoffTimer(object): def __init__(self, ratio=1, max_interval=None, min_interval=None): self.c = 0 self.ratio = ratio self.max_interval = max_interval self.min_interval = min_interval def is_reset(self): return self.c == 0 def reset(self): self.c = 0 return self def success(self): self.c = max(self.c - 1, 0) return self def failure(self): self.c += 1 return self def get_interval(self): k = pow(2, self.c) - 1 interval = random.random() * k * self.ratio if self.max_interval is not None: interval = min(interval, self.max_interval) if self.min_interval is not None: interval = max(interval, self.min_interval) return interval
<commit_before>from random import randint class BackoffTimer(object): def __init__(self, ratio=1, max_interval=None, min_interval=None): self.c = 0 self.ratio = ratio self.max_interval = max_interval self.min_interval = min_interval def is_reset(self): return self.c == 0 def reset(self): self.c = 0 return self def success(self): self.c = max(self.c - 1, 0) return self def failure(self): self.c += 1 return self def get_interval(self): k = pow(2, self.c) - 1 interval = randint(0, k) * self.ratio if self.max_interval is not None: interval = min(interval, self.max_interval) if self.min_interval is not None: interval = max(interval, self.min_interval) return interval <commit_msg>Return float for interval instead of int.<commit_after>
import random class BackoffTimer(object): def __init__(self, ratio=1, max_interval=None, min_interval=None): self.c = 0 self.ratio = ratio self.max_interval = max_interval self.min_interval = min_interval def is_reset(self): return self.c == 0 def reset(self): self.c = 0 return self def success(self): self.c = max(self.c - 1, 0) return self def failure(self): self.c += 1 return self def get_interval(self): k = pow(2, self.c) - 1 interval = random.random() * k * self.ratio if self.max_interval is not None: interval = min(interval, self.max_interval) if self.min_interval is not None: interval = max(interval, self.min_interval) return interval
from random import randint class BackoffTimer(object): def __init__(self, ratio=1, max_interval=None, min_interval=None): self.c = 0 self.ratio = ratio self.max_interval = max_interval self.min_interval = min_interval def is_reset(self): return self.c == 0 def reset(self): self.c = 0 return self def success(self): self.c = max(self.c - 1, 0) return self def failure(self): self.c += 1 return self def get_interval(self): k = pow(2, self.c) - 1 interval = randint(0, k) * self.ratio if self.max_interval is not None: interval = min(interval, self.max_interval) if self.min_interval is not None: interval = max(interval, self.min_interval) return interval Return float for interval instead of int.import random class BackoffTimer(object): def __init__(self, ratio=1, max_interval=None, min_interval=None): self.c = 0 self.ratio = ratio self.max_interval = max_interval self.min_interval = min_interval def is_reset(self): return self.c == 0 def reset(self): self.c = 0 return self def success(self): self.c = max(self.c - 1, 0) return self def failure(self): self.c += 1 return self def get_interval(self): k = pow(2, self.c) - 1 interval = random.random() * k * self.ratio if self.max_interval is not None: interval = min(interval, self.max_interval) if self.min_interval is not None: interval = max(interval, self.min_interval) return interval
<commit_before>from random import randint class BackoffTimer(object): def __init__(self, ratio=1, max_interval=None, min_interval=None): self.c = 0 self.ratio = ratio self.max_interval = max_interval self.min_interval = min_interval def is_reset(self): return self.c == 0 def reset(self): self.c = 0 return self def success(self): self.c = max(self.c - 1, 0) return self def failure(self): self.c += 1 return self def get_interval(self): k = pow(2, self.c) - 1 interval = randint(0, k) * self.ratio if self.max_interval is not None: interval = min(interval, self.max_interval) if self.min_interval is not None: interval = max(interval, self.min_interval) return interval <commit_msg>Return float for interval instead of int.<commit_after>import random class BackoffTimer(object): def __init__(self, ratio=1, max_interval=None, min_interval=None): self.c = 0 self.ratio = ratio self.max_interval = max_interval self.min_interval = min_interval def is_reset(self): return self.c == 0 def reset(self): self.c = 0 return self def success(self): self.c = max(self.c - 1, 0) return self def failure(self): self.c += 1 return self def get_interval(self): k = pow(2, self.c) - 1 interval = random.random() * k * self.ratio if self.max_interval is not None: interval = min(interval, self.max_interval) if self.min_interval is not None: interval = max(interval, self.min_interval) return interval
1d74b003818e260ae1f453cb26f1c9efc29e8ba2
scripts/run_unit_test.py
scripts/run_unit_test.py
#!/usr/bin/env python import serial import os import sys import time # Make and flash the unit test FILE_LOCATION = os.path.dirname(os.path.abspath(__file__)) os.chdir(FILE_LOCATION + "/../") print os.system("make flash_unit_test") # Ask the user to reset the board raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...") # Open a serial port time.sleep(1) print 'Connecting to /dev/serial/by-id/usb-eecs567_final_project-if00' ser = serial.Serial("/dev/serial/by-id/usb-eecs567_final_project-if00", 115200) # time.sleep(1) # Send data to start USB OTG print 'Write start' ser.write("start") print 'Run test' # Read until we see the finished text result = '' try: while True: num_chars = ser.inWaiting() if num_chars: new = '' try: new = ser.read(num_chars) except: print '\nFailed to read' sys.stdout.write(new) result += new if result.find("Finished") != -1: break finally: # Print the result so the user can see and close the serial port #print result ser.close()
#!/usr/bin/env python import serial import os import sys import time # Make and flash the unit test FILE_LOCATION = os.path.dirname(os.path.abspath(__file__)) os.chdir(FILE_LOCATION + "/../") print os.system("make flash_unit_test") # Ask the user to reset the board raw_input("""\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...""") # Open a serial port time.sleep(1) print 'Connecting to /dev/serial/by-id/usb-eecs567_final_project-if00' ser = serial.Serial("/dev/serial/by-id/usb-eecs567_final_project-if00", 115200) # time.sleep(1) # Send data to start USB OTG print 'Write start' ser.write("start") print 'Run test' # Read until we see the finished text result = '' try: while True: num_chars = ser.inWaiting() if num_chars: new = '' try: new = ser.read(num_chars) except: print '\nFailed to read' sys.stdout.write(new) result += new if result.find("Finished") != -1: break finally: # Close the serial port ser.close()
Bring scripts folder up to pep8 standards
PEP8: Bring scripts folder up to pep8 standards
Python
mit
fnivek/eecs567-final-project,fnivek/eecs567-final-project,fnivek/eecs567-final-project,fnivek/eecs567-final-project,fnivek/eecs567-final-project
#!/usr/bin/env python import serial import os import sys import time # Make and flash the unit test FILE_LOCATION = os.path.dirname(os.path.abspath(__file__)) os.chdir(FILE_LOCATION + "/../") print os.system("make flash_unit_test") # Ask the user to reset the board raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...") # Open a serial port time.sleep(1) print 'Connecting to /dev/serial/by-id/usb-eecs567_final_project-if00' ser = serial.Serial("/dev/serial/by-id/usb-eecs567_final_project-if00", 115200) # time.sleep(1) # Send data to start USB OTG print 'Write start' ser.write("start") print 'Run test' # Read until we see the finished text result = '' try: while True: num_chars = ser.inWaiting() if num_chars: new = '' try: new = ser.read(num_chars) except: print '\nFailed to read' sys.stdout.write(new) result += new if result.find("Finished") != -1: break finally: # Print the result so the user can see and close the serial port #print result ser.close()PEP8: Bring scripts folder up to pep8 standards
#!/usr/bin/env python import serial import os import sys import time # Make and flash the unit test FILE_LOCATION = os.path.dirname(os.path.abspath(__file__)) os.chdir(FILE_LOCATION + "/../") print os.system("make flash_unit_test") # Ask the user to reset the board raw_input("""\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...""") # Open a serial port time.sleep(1) print 'Connecting to /dev/serial/by-id/usb-eecs567_final_project-if00' ser = serial.Serial("/dev/serial/by-id/usb-eecs567_final_project-if00", 115200) # time.sleep(1) # Send data to start USB OTG print 'Write start' ser.write("start") print 'Run test' # Read until we see the finished text result = '' try: while True: num_chars = ser.inWaiting() if num_chars: new = '' try: new = ser.read(num_chars) except: print '\nFailed to read' sys.stdout.write(new) result += new if result.find("Finished") != -1: break finally: # Close the serial port ser.close()
<commit_before>#!/usr/bin/env python import serial import os import sys import time # Make and flash the unit test FILE_LOCATION = os.path.dirname(os.path.abspath(__file__)) os.chdir(FILE_LOCATION + "/../") print os.system("make flash_unit_test") # Ask the user to reset the board raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...") # Open a serial port time.sleep(1) print 'Connecting to /dev/serial/by-id/usb-eecs567_final_project-if00' ser = serial.Serial("/dev/serial/by-id/usb-eecs567_final_project-if00", 115200) # time.sleep(1) # Send data to start USB OTG print 'Write start' ser.write("start") print 'Run test' # Read until we see the finished text result = '' try: while True: num_chars = ser.inWaiting() if num_chars: new = '' try: new = ser.read(num_chars) except: print '\nFailed to read' sys.stdout.write(new) result += new if result.find("Finished") != -1: break finally: # Print the result so the user can see and close the serial port #print result ser.close()<commit_msg>PEP8: Bring scripts folder up to pep8 standards<commit_after>
#!/usr/bin/env python import serial import os import sys import time # Make and flash the unit test FILE_LOCATION = os.path.dirname(os.path.abspath(__file__)) os.chdir(FILE_LOCATION + "/../") print os.system("make flash_unit_test") # Ask the user to reset the board raw_input("""\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...""") # Open a serial port time.sleep(1) print 'Connecting to /dev/serial/by-id/usb-eecs567_final_project-if00' ser = serial.Serial("/dev/serial/by-id/usb-eecs567_final_project-if00", 115200) # time.sleep(1) # Send data to start USB OTG print 'Write start' ser.write("start") print 'Run test' # Read until we see the finished text result = '' try: while True: num_chars = ser.inWaiting() if num_chars: new = '' try: new = ser.read(num_chars) except: print '\nFailed to read' sys.stdout.write(new) result += new if result.find("Finished") != -1: break finally: # Close the serial port ser.close()
#!/usr/bin/env python import serial import os import sys import time # Make and flash the unit test FILE_LOCATION = os.path.dirname(os.path.abspath(__file__)) os.chdir(FILE_LOCATION + "/../") print os.system("make flash_unit_test") # Ask the user to reset the board raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...") # Open a serial port time.sleep(1) print 'Connecting to /dev/serial/by-id/usb-eecs567_final_project-if00' ser = serial.Serial("/dev/serial/by-id/usb-eecs567_final_project-if00", 115200) # time.sleep(1) # Send data to start USB OTG print 'Write start' ser.write("start") print 'Run test' # Read until we see the finished text result = '' try: while True: num_chars = ser.inWaiting() if num_chars: new = '' try: new = ser.read(num_chars) except: print '\nFailed to read' sys.stdout.write(new) result += new if result.find("Finished") != -1: break finally: # Print the result so the user can see and close the serial port #print result ser.close()PEP8: Bring scripts folder up to pep8 standards#!/usr/bin/env python import serial import os import sys import time # Make and flash the unit test FILE_LOCATION = os.path.dirname(os.path.abspath(__file__)) os.chdir(FILE_LOCATION + "/../") print os.system("make flash_unit_test") # Ask the user to reset the board raw_input("""\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...""") # Open a serial port time.sleep(1) print 'Connecting to /dev/serial/by-id/usb-eecs567_final_project-if00' ser = serial.Serial("/dev/serial/by-id/usb-eecs567_final_project-if00", 115200) # time.sleep(1) # Send data to start USB OTG print 'Write start' ser.write("start") print 'Run test' # Read until we see the finished text result = '' try: while True: num_chars = ser.inWaiting() if num_chars: new = '' try: new = ser.read(num_chars) except: print '\nFailed to read' sys.stdout.write(new) result += new if result.find("Finished") != -1: break finally: # Close the serial port ser.close()
<commit_before>#!/usr/bin/env python import serial import os import sys import time # Make and flash the unit test FILE_LOCATION = os.path.dirname(os.path.abspath(__file__)) os.chdir(FILE_LOCATION + "/../") print os.system("make flash_unit_test") # Ask the user to reset the board raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...") # Open a serial port time.sleep(1) print 'Connecting to /dev/serial/by-id/usb-eecs567_final_project-if00' ser = serial.Serial("/dev/serial/by-id/usb-eecs567_final_project-if00", 115200) # time.sleep(1) # Send data to start USB OTG print 'Write start' ser.write("start") print 'Run test' # Read until we see the finished text result = '' try: while True: num_chars = ser.inWaiting() if num_chars: new = '' try: new = ser.read(num_chars) except: print '\nFailed to read' sys.stdout.write(new) result += new if result.find("Finished") != -1: break finally: # Print the result so the user can see and close the serial port #print result ser.close()<commit_msg>PEP8: Bring scripts folder up to pep8 standards<commit_after>#!/usr/bin/env python import serial import os import sys import time # Make and flash the unit test FILE_LOCATION = os.path.dirname(os.path.abspath(__file__)) os.chdir(FILE_LOCATION + "/../") print os.system("make flash_unit_test") # Ask the user to reset the board raw_input("""\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...""") # Open a serial port time.sleep(1) print 'Connecting to /dev/serial/by-id/usb-eecs567_final_project-if00' ser = serial.Serial("/dev/serial/by-id/usb-eecs567_final_project-if00", 115200) # time.sleep(1) # Send data to start USB OTG print 'Write start' ser.write("start") print 'Run test' # Read until we see the finished text result = '' try: while True: num_chars = ser.inWaiting() if num_chars: new = '' try: new = ser.read(num_chars) except: print '\nFailed to read' sys.stdout.write(new) result += new if result.find("Finished") != -1: break finally: # Close the serial port ser.close()
515855105be9d26ad1a272177a64604e0fa547a5
wait_for_agent_update.py
wait_for_agent_update.py
#!/usr/bin/env python __metaclass__ = type from jujupy import ( check_wordpress, Environment, format_listing, until_timeout, ) from collections import defaultdict import sys def agent_update(environment, version): env = Environment(environment) for ignored in until_timeout(30): versions = defaultdict(list) status = env.get_status() for item_name, item in env.agent_items(status): versions[item.get('agent-version', 'unknown')].append(item_name) if versions.keys() == [version]: break print format_listing(versions, version) sys.stdout.flush() else: raise Exception('Some versions did not update.') def main(): try: agent_update(sys.argv[1], sys.argv[2]) except Exception as e: print e sys.exit(1) if __name__ == '__main__': main()
#!/usr/bin/env python __metaclass__ = type from jujupy import ( check_wordpress, Environment, format_listing, until_timeout, ) from collections import defaultdict import sys def agent_update(environment, version): env = Environment(environment) for ignored in until_timeout(300): versions = defaultdict(list) status = env.get_status() for item_name, item in env.agent_items(status): versions[item.get('agent-version', 'unknown')].append(item_name) if versions.keys() == [version]: break print format_listing(versions, version) sys.stdout.flush() else: raise Exception('Some versions did not update.') def main(): try: agent_update(sys.argv[1], sys.argv[2]) except Exception as e: print e sys.exit(1) if __name__ == '__main__': main()
Increase timeout for agent update to 5 minutes.
Increase timeout for agent update to 5 minutes.
Python
agpl-3.0
mjs/juju,mjs/juju,mjs/juju,mjs/juju,mjs/juju,mjs/juju,mjs/juju
#!/usr/bin/env python __metaclass__ = type from jujupy import ( check_wordpress, Environment, format_listing, until_timeout, ) from collections import defaultdict import sys def agent_update(environment, version): env = Environment(environment) for ignored in until_timeout(30): versions = defaultdict(list) status = env.get_status() for item_name, item in env.agent_items(status): versions[item.get('agent-version', 'unknown')].append(item_name) if versions.keys() == [version]: break print format_listing(versions, version) sys.stdout.flush() else: raise Exception('Some versions did not update.') def main(): try: agent_update(sys.argv[1], sys.argv[2]) except Exception as e: print e sys.exit(1) if __name__ == '__main__': main() Increase timeout for agent update to 5 minutes.
#!/usr/bin/env python __metaclass__ = type from jujupy import ( check_wordpress, Environment, format_listing, until_timeout, ) from collections import defaultdict import sys def agent_update(environment, version): env = Environment(environment) for ignored in until_timeout(300): versions = defaultdict(list) status = env.get_status() for item_name, item in env.agent_items(status): versions[item.get('agent-version', 'unknown')].append(item_name) if versions.keys() == [version]: break print format_listing(versions, version) sys.stdout.flush() else: raise Exception('Some versions did not update.') def main(): try: agent_update(sys.argv[1], sys.argv[2]) except Exception as e: print e sys.exit(1) if __name__ == '__main__': main()
<commit_before>#!/usr/bin/env python __metaclass__ = type from jujupy import ( check_wordpress, Environment, format_listing, until_timeout, ) from collections import defaultdict import sys def agent_update(environment, version): env = Environment(environment) for ignored in until_timeout(30): versions = defaultdict(list) status = env.get_status() for item_name, item in env.agent_items(status): versions[item.get('agent-version', 'unknown')].append(item_name) if versions.keys() == [version]: break print format_listing(versions, version) sys.stdout.flush() else: raise Exception('Some versions did not update.') def main(): try: agent_update(sys.argv[1], sys.argv[2]) except Exception as e: print e sys.exit(1) if __name__ == '__main__': main() <commit_msg>Increase timeout for agent update to 5 minutes.<commit_after>
#!/usr/bin/env python __metaclass__ = type from jujupy import ( check_wordpress, Environment, format_listing, until_timeout, ) from collections import defaultdict import sys def agent_update(environment, version): env = Environment(environment) for ignored in until_timeout(300): versions = defaultdict(list) status = env.get_status() for item_name, item in env.agent_items(status): versions[item.get('agent-version', 'unknown')].append(item_name) if versions.keys() == [version]: break print format_listing(versions, version) sys.stdout.flush() else: raise Exception('Some versions did not update.') def main(): try: agent_update(sys.argv[1], sys.argv[2]) except Exception as e: print e sys.exit(1) if __name__ == '__main__': main()
#!/usr/bin/env python __metaclass__ = type from jujupy import ( check_wordpress, Environment, format_listing, until_timeout, ) from collections import defaultdict import sys def agent_update(environment, version): env = Environment(environment) for ignored in until_timeout(30): versions = defaultdict(list) status = env.get_status() for item_name, item in env.agent_items(status): versions[item.get('agent-version', 'unknown')].append(item_name) if versions.keys() == [version]: break print format_listing(versions, version) sys.stdout.flush() else: raise Exception('Some versions did not update.') def main(): try: agent_update(sys.argv[1], sys.argv[2]) except Exception as e: print e sys.exit(1) if __name__ == '__main__': main() Increase timeout for agent update to 5 minutes.#!/usr/bin/env python __metaclass__ = type from jujupy import ( check_wordpress, Environment, format_listing, until_timeout, ) from collections import defaultdict import sys def agent_update(environment, version): env = Environment(environment) for ignored in until_timeout(300): versions = defaultdict(list) status = env.get_status() for item_name, item in env.agent_items(status): versions[item.get('agent-version', 'unknown')].append(item_name) if versions.keys() == [version]: break print format_listing(versions, version) sys.stdout.flush() else: raise Exception('Some versions did not update.') def main(): try: agent_update(sys.argv[1], sys.argv[2]) except Exception as e: print e sys.exit(1) if __name__ == '__main__': main()
<commit_before>#!/usr/bin/env python __metaclass__ = type from jujupy import ( check_wordpress, Environment, format_listing, until_timeout, ) from collections import defaultdict import sys def agent_update(environment, version): env = Environment(environment) for ignored in until_timeout(30): versions = defaultdict(list) status = env.get_status() for item_name, item in env.agent_items(status): versions[item.get('agent-version', 'unknown')].append(item_name) if versions.keys() == [version]: break print format_listing(versions, version) sys.stdout.flush() else: raise Exception('Some versions did not update.') def main(): try: agent_update(sys.argv[1], sys.argv[2]) except Exception as e: print e sys.exit(1) if __name__ == '__main__': main() <commit_msg>Increase timeout for agent update to 5 minutes.<commit_after>#!/usr/bin/env python __metaclass__ = type from jujupy import ( check_wordpress, Environment, format_listing, until_timeout, ) from collections import defaultdict import sys def agent_update(environment, version): env = Environment(environment) for ignored in until_timeout(300): versions = defaultdict(list) status = env.get_status() for item_name, item in env.agent_items(status): versions[item.get('agent-version', 'unknown')].append(item_name) if versions.keys() == [version]: break print format_listing(versions, version) sys.stdout.flush() else: raise Exception('Some versions did not update.') def main(): try: agent_update(sys.argv[1], sys.argv[2]) except Exception as e: print e sys.exit(1) if __name__ == '__main__': main()
d498a9846567e4986ba2a2541b2b4e4719c2c83f
keras/__init__.py
keras/__init__.py
from __future__ import absolute_import from . import activations from . import applications from . import backend from . import datasets from . import engine from . import layers from . import preprocessing from . import utils from . import wrappers from . import callbacks from . import constraints from . import initializers from . import metrics from . import models from . import losses from . import optimizers from . import regularizers __version__ = '2.0.3'
from __future__ import absolute_import from . import activations from . import applications from . import backend from . import datasets from . import engine from . import layers from . import preprocessing from . import utils from . import wrappers from . import callbacks from . import constraints from . import initializers from . import metrics from . import models from . import losses from . import optimizers from . import regularizers # Importable from root because it's technically not a layer from .layers import Input __version__ = '2.0.3'
Make Input importable from root
Make Input importable from root
Python
apache-2.0
keras-team/keras,keras-team/keras
from __future__ import absolute_import from . import activations from . import applications from . import backend from . import datasets from . import engine from . import layers from . import preprocessing from . import utils from . import wrappers from . import callbacks from . import constraints from . import initializers from . import metrics from . import models from . import losses from . import optimizers from . import regularizers __version__ = '2.0.3' Make Input importable from root
from __future__ import absolute_import from . import activations from . import applications from . import backend from . import datasets from . import engine from . import layers from . import preprocessing from . import utils from . import wrappers from . import callbacks from . import constraints from . import initializers from . import metrics from . import models from . import losses from . import optimizers from . import regularizers # Importable from root because it's technically not a layer from .layers import Input __version__ = '2.0.3'
<commit_before>from __future__ import absolute_import from . import activations from . import applications from . import backend from . import datasets from . import engine from . import layers from . import preprocessing from . import utils from . import wrappers from . import callbacks from . import constraints from . import initializers from . import metrics from . import models from . import losses from . import optimizers from . import regularizers __version__ = '2.0.3' <commit_msg>Make Input importable from root<commit_after>
from __future__ import absolute_import from . import activations from . import applications from . import backend from . import datasets from . import engine from . import layers from . import preprocessing from . import utils from . import wrappers from . import callbacks from . import constraints from . import initializers from . import metrics from . import models from . import losses from . import optimizers from . import regularizers # Importable from root because it's technically not a layer from .layers import Input __version__ = '2.0.3'
from __future__ import absolute_import from . import activations from . import applications from . import backend from . import datasets from . import engine from . import layers from . import preprocessing from . import utils from . import wrappers from . import callbacks from . import constraints from . import initializers from . import metrics from . import models from . import losses from . import optimizers from . import regularizers __version__ = '2.0.3' Make Input importable from rootfrom __future__ import absolute_import from . import activations from . import applications from . import backend from . import datasets from . import engine from . import layers from . import preprocessing from . import utils from . import wrappers from . import callbacks from . import constraints from . import initializers from . import metrics from . import models from . import losses from . import optimizers from . import regularizers # Importable from root because it's technically not a layer from .layers import Input __version__ = '2.0.3'
<commit_before>from __future__ import absolute_import from . import activations from . import applications from . import backend from . import datasets from . import engine from . import layers from . import preprocessing from . import utils from . import wrappers from . import callbacks from . import constraints from . import initializers from . import metrics from . import models from . import losses from . import optimizers from . import regularizers __version__ = '2.0.3' <commit_msg>Make Input importable from root<commit_after>from __future__ import absolute_import from . import activations from . import applications from . import backend from . import datasets from . import engine from . import layers from . import preprocessing from . import utils from . import wrappers from . import callbacks from . import constraints from . import initializers from . import metrics from . import models from . import losses from . import optimizers from . import regularizers # Importable from root because it's technically not a layer from .layers import Input __version__ = '2.0.3'
727b94078ab15d04a65adbd57fe6962949bda97e
openacademy/model/openacademy_course.py
openacademy/model/openacademy_course.py
from openerp import models, fields, api class Course(models.Model): ''' This class creates a model for courses ''' _name = 'openacademey.course' name = fields.Char(string='Title', required=True) description = fields.Text(string='Description')
from openerp import models, fields, api class Course(models.Model): ''' This class creates a model for courses ''' _name = 'openacademy.course' name = fields.Char(string='Title', required=True) description = fields.Text(string='Description')
Fix typo in Course table name
Fix typo in Course table name
Python
mit
tebanep/odoo_training_addons
from openerp import models, fields, api class Course(models.Model): ''' This class creates a model for courses ''' _name = 'openacademey.course' name = fields.Char(string='Title', required=True) description = fields.Text(string='Description') Fix typo in Course table name
from openerp import models, fields, api class Course(models.Model): ''' This class creates a model for courses ''' _name = 'openacademy.course' name = fields.Char(string='Title', required=True) description = fields.Text(string='Description')
<commit_before>from openerp import models, fields, api class Course(models.Model): ''' This class creates a model for courses ''' _name = 'openacademey.course' name = fields.Char(string='Title', required=True) description = fields.Text(string='Description') <commit_msg>Fix typo in Course table name<commit_after>
from openerp import models, fields, api class Course(models.Model): ''' This class creates a model for courses ''' _name = 'openacademy.course' name = fields.Char(string='Title', required=True) description = fields.Text(string='Description')
from openerp import models, fields, api class Course(models.Model): ''' This class creates a model for courses ''' _name = 'openacademey.course' name = fields.Char(string='Title', required=True) description = fields.Text(string='Description') Fix typo in Course table namefrom openerp import models, fields, api class Course(models.Model): ''' This class creates a model for courses ''' _name = 'openacademy.course' name = fields.Char(string='Title', required=True) description = fields.Text(string='Description')
<commit_before>from openerp import models, fields, api class Course(models.Model): ''' This class creates a model for courses ''' _name = 'openacademey.course' name = fields.Char(string='Title', required=True) description = fields.Text(string='Description') <commit_msg>Fix typo in Course table name<commit_after>from openerp import models, fields, api class Course(models.Model): ''' This class creates a model for courses ''' _name = 'openacademy.course' name = fields.Char(string='Title', required=True) description = fields.Text(string='Description')
88e99caf6c426cdee602157f9aee120ecf822cad
docs/conf.py
docs/conf.py
import sys from os.path import dirname, abspath sys.path.insert(0, dirname(dirname(abspath(__file__)))) from django.conf import settings settings.configure() project = 'django-slack' version = '' release = '' copyright = '2014, 2015 Chris Lamb' author = 'lamby' extensions = ['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'] html_title = "%s documentation" % project html_static_path = [] master_doc = 'index' exclude_trees = ['_build'] templates_path = ['_templates'] latex_documents = [ ('index', '%s.tex' % project, html_title, author, 'manual', True), ] intersphinx_mapping = {'http://docs.python.org/': None}
import sys from os.path import dirname, abspath sys.path.insert(0, dirname(dirname(abspath(__file__)))) from django.conf import settings settings.configure() project = 'django-slack' version = '' release = '' copyright = '2014, 2015 Chris Lamb' author = 'lamby' extensions = ['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'] html_title = "%s documentation" % project html_static_path = [] html_theme = "sphinx_rtd_theme" master_doc = 'index' exclude_trees = ['_build'] templates_path = ['_templates'] latex_documents = [ ('index', '%s.tex' % project, html_title, author, 'manual', True), ] intersphinx_mapping = {'http://docs.python.org/': None}
Set readthedocs theme in docs build
Set readthedocs theme in docs build
Python
bsd-3-clause
lamby/django-slack
import sys from os.path import dirname, abspath sys.path.insert(0, dirname(dirname(abspath(__file__)))) from django.conf import settings settings.configure() project = 'django-slack' version = '' release = '' copyright = '2014, 2015 Chris Lamb' author = 'lamby' extensions = ['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'] html_title = "%s documentation" % project html_static_path = [] master_doc = 'index' exclude_trees = ['_build'] templates_path = ['_templates'] latex_documents = [ ('index', '%s.tex' % project, html_title, author, 'manual', True), ] intersphinx_mapping = {'http://docs.python.org/': None} Set readthedocs theme in docs build
import sys from os.path import dirname, abspath sys.path.insert(0, dirname(dirname(abspath(__file__)))) from django.conf import settings settings.configure() project = 'django-slack' version = '' release = '' copyright = '2014, 2015 Chris Lamb' author = 'lamby' extensions = ['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'] html_title = "%s documentation" % project html_static_path = [] html_theme = "sphinx_rtd_theme" master_doc = 'index' exclude_trees = ['_build'] templates_path = ['_templates'] latex_documents = [ ('index', '%s.tex' % project, html_title, author, 'manual', True), ] intersphinx_mapping = {'http://docs.python.org/': None}
<commit_before>import sys from os.path import dirname, abspath sys.path.insert(0, dirname(dirname(abspath(__file__)))) from django.conf import settings settings.configure() project = 'django-slack' version = '' release = '' copyright = '2014, 2015 Chris Lamb' author = 'lamby' extensions = ['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'] html_title = "%s documentation" % project html_static_path = [] master_doc = 'index' exclude_trees = ['_build'] templates_path = ['_templates'] latex_documents = [ ('index', '%s.tex' % project, html_title, author, 'manual', True), ] intersphinx_mapping = {'http://docs.python.org/': None} <commit_msg>Set readthedocs theme in docs build<commit_after>
import sys from os.path import dirname, abspath sys.path.insert(0, dirname(dirname(abspath(__file__)))) from django.conf import settings settings.configure() project = 'django-slack' version = '' release = '' copyright = '2014, 2015 Chris Lamb' author = 'lamby' extensions = ['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'] html_title = "%s documentation" % project html_static_path = [] html_theme = "sphinx_rtd_theme" master_doc = 'index' exclude_trees = ['_build'] templates_path = ['_templates'] latex_documents = [ ('index', '%s.tex' % project, html_title, author, 'manual', True), ] intersphinx_mapping = {'http://docs.python.org/': None}
import sys from os.path import dirname, abspath sys.path.insert(0, dirname(dirname(abspath(__file__)))) from django.conf import settings settings.configure() project = 'django-slack' version = '' release = '' copyright = '2014, 2015 Chris Lamb' author = 'lamby' extensions = ['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'] html_title = "%s documentation" % project html_static_path = [] master_doc = 'index' exclude_trees = ['_build'] templates_path = ['_templates'] latex_documents = [ ('index', '%s.tex' % project, html_title, author, 'manual', True), ] intersphinx_mapping = {'http://docs.python.org/': None} Set readthedocs theme in docs buildimport sys from os.path import dirname, abspath sys.path.insert(0, dirname(dirname(abspath(__file__)))) from django.conf import settings settings.configure() project = 'django-slack' version = '' release = '' copyright = '2014, 2015 Chris Lamb' author = 'lamby' extensions = ['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'] html_title = "%s documentation" % project html_static_path = [] html_theme = "sphinx_rtd_theme" master_doc = 'index' exclude_trees = ['_build'] templates_path = ['_templates'] latex_documents = [ ('index', '%s.tex' % project, html_title, author, 'manual', True), ] intersphinx_mapping = {'http://docs.python.org/': None}
<commit_before>import sys from os.path import dirname, abspath sys.path.insert(0, dirname(dirname(abspath(__file__)))) from django.conf import settings settings.configure() project = 'django-slack' version = '' release = '' copyright = '2014, 2015 Chris Lamb' author = 'lamby' extensions = ['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'] html_title = "%s documentation" % project html_static_path = [] master_doc = 'index' exclude_trees = ['_build'] templates_path = ['_templates'] latex_documents = [ ('index', '%s.tex' % project, html_title, author, 'manual', True), ] intersphinx_mapping = {'http://docs.python.org/': None} <commit_msg>Set readthedocs theme in docs build<commit_after>import sys from os.path import dirname, abspath sys.path.insert(0, dirname(dirname(abspath(__file__)))) from django.conf import settings settings.configure() project = 'django-slack' version = '' release = '' copyright = '2014, 2015 Chris Lamb' author = 'lamby' extensions = ['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'] html_title = "%s documentation" % project html_static_path = [] html_theme = "sphinx_rtd_theme" master_doc = 'index' exclude_trees = ['_build'] templates_path = ['_templates'] latex_documents = [ ('index', '%s.tex' % project, html_title, author, 'manual', True), ] intersphinx_mapping = {'http://docs.python.org/': None}
8e20b56c4b91f673429697232926115db46e0c2d
spanky/commands/cmd_users.py
spanky/commands/cmd_users.py
import click from spanky.cli import pass_context from spanky.lib.users import UserInit @click.command('users', short_help='creates users base on /etc/spanky/users') @pass_context def cli(ctx): config = ctx.config.load('users.yml')() user_init = UserInit(config) user_init.build()
import sys import click from spanky.cli import pass_context from spanky.lib.users import UserInit @click.command('users', short_help='creates users base on /etc/spanky/users') @pass_context def cli(ctx): try: config = ctx.config.load('users.yml')() except IOError: # no config lets bail click.echo('No users to install') sys.exit(1) user_init = UserInit(config) user_init.build()
Return a 1 when we don't have user to install.
Return a 1 when we don't have user to install. TODO: Make better return code
Python
bsd-3-clause
pglbutt/spanky,pglbutt/spanky,pglbutt/spanky
import click from spanky.cli import pass_context from spanky.lib.users import UserInit @click.command('users', short_help='creates users base on /etc/spanky/users') @pass_context def cli(ctx): config = ctx.config.load('users.yml')() user_init = UserInit(config) user_init.build() Return a 1 when we don't have user to install. TODO: Make better return code
import sys import click from spanky.cli import pass_context from spanky.lib.users import UserInit @click.command('users', short_help='creates users base on /etc/spanky/users') @pass_context def cli(ctx): try: config = ctx.config.load('users.yml')() except IOError: # no config lets bail click.echo('No users to install') sys.exit(1) user_init = UserInit(config) user_init.build()
<commit_before>import click from spanky.cli import pass_context from spanky.lib.users import UserInit @click.command('users', short_help='creates users base on /etc/spanky/users') @pass_context def cli(ctx): config = ctx.config.load('users.yml')() user_init = UserInit(config) user_init.build() <commit_msg>Return a 1 when we don't have user to install. TODO: Make better return code<commit_after>
import sys import click from spanky.cli import pass_context from spanky.lib.users import UserInit @click.command('users', short_help='creates users base on /etc/spanky/users') @pass_context def cli(ctx): try: config = ctx.config.load('users.yml')() except IOError: # no config lets bail click.echo('No users to install') sys.exit(1) user_init = UserInit(config) user_init.build()
import click from spanky.cli import pass_context from spanky.lib.users import UserInit @click.command('users', short_help='creates users base on /etc/spanky/users') @pass_context def cli(ctx): config = ctx.config.load('users.yml')() user_init = UserInit(config) user_init.build() Return a 1 when we don't have user to install. TODO: Make better return codeimport sys import click from spanky.cli import pass_context from spanky.lib.users import UserInit @click.command('users', short_help='creates users base on /etc/spanky/users') @pass_context def cli(ctx): try: config = ctx.config.load('users.yml')() except IOError: # no config lets bail click.echo('No users to install') sys.exit(1) user_init = UserInit(config) user_init.build()
<commit_before>import click from spanky.cli import pass_context from spanky.lib.users import UserInit @click.command('users', short_help='creates users base on /etc/spanky/users') @pass_context def cli(ctx): config = ctx.config.load('users.yml')() user_init = UserInit(config) user_init.build() <commit_msg>Return a 1 when we don't have user to install. TODO: Make better return code<commit_after>import sys import click from spanky.cli import pass_context from spanky.lib.users import UserInit @click.command('users', short_help='creates users base on /etc/spanky/users') @pass_context def cli(ctx): try: config = ctx.config.load('users.yml')() except IOError: # no config lets bail click.echo('No users to install') sys.exit(1) user_init = UserInit(config) user_init.build()
9970faa24d9f27817c7e9c9e88869bb415d6926c
main/__init__.py
main/__init__.py
from flask import Flask from config import config from . import extensions, modules def create_app(config_name): config_obj = config[config_name]() app = Flask(__name__, static_url_path='/static') # Initializes configuration values. app.config.from_object(config_obj) # Configure SSL if the current platform supports it. if not app.debug and not app.testing and not app.config.get('SSL_DISABLE'): from flask_sslify import SSLify SSLify(app) # Initializes Flask extensions. extensions.init_app(app) # Initializes modules. modules.init_app(app) return app
from flask import Flask, redirect, request from config import config from . import extensions, modules def create_app(config_name): config_obj = config[config_name]() app = Flask(__name__, static_url_path='/static') # Initializes configuration values. app.config.from_object(config_obj) # Configure SSL if the current platform supports it. if not app.debug and not app.testing and not app.config.get('SSL_DISABLE'): from flask_sslify import SSLify SSLify(app) @app.before_request def redirect_www(): """ Redirects www requests to non-www. """ if request.host.startswith('www.'): new_host = request.host[4:] return redirect(f"{request.scheme}://{new_host}/", code=301) # Initializes Flask extensions. extensions.init_app(app) # Initializes modules. modules.init_app(app) return app
Add before_request callback to handle www redirects
Add before_request callback to handle www redirects
Python
mit
ellmetha/morganaubert-resume,ellmetha/morganaubert-resume,ellmetha/morganaubert-resume,ellmetha/morganaubert-resume
from flask import Flask from config import config from . import extensions, modules def create_app(config_name): config_obj = config[config_name]() app = Flask(__name__, static_url_path='/static') # Initializes configuration values. app.config.from_object(config_obj) # Configure SSL if the current platform supports it. if not app.debug and not app.testing and not app.config.get('SSL_DISABLE'): from flask_sslify import SSLify SSLify(app) # Initializes Flask extensions. extensions.init_app(app) # Initializes modules. modules.init_app(app) return app Add before_request callback to handle www redirects
from flask import Flask, redirect, request from config import config from . import extensions, modules def create_app(config_name): config_obj = config[config_name]() app = Flask(__name__, static_url_path='/static') # Initializes configuration values. app.config.from_object(config_obj) # Configure SSL if the current platform supports it. if not app.debug and not app.testing and not app.config.get('SSL_DISABLE'): from flask_sslify import SSLify SSLify(app) @app.before_request def redirect_www(): """ Redirects www requests to non-www. """ if request.host.startswith('www.'): new_host = request.host[4:] return redirect(f"{request.scheme}://{new_host}/", code=301) # Initializes Flask extensions. extensions.init_app(app) # Initializes modules. modules.init_app(app) return app
<commit_before>from flask import Flask from config import config from . import extensions, modules def create_app(config_name): config_obj = config[config_name]() app = Flask(__name__, static_url_path='/static') # Initializes configuration values. app.config.from_object(config_obj) # Configure SSL if the current platform supports it. if not app.debug and not app.testing and not app.config.get('SSL_DISABLE'): from flask_sslify import SSLify SSLify(app) # Initializes Flask extensions. extensions.init_app(app) # Initializes modules. modules.init_app(app) return app <commit_msg>Add before_request callback to handle www redirects<commit_after>
from flask import Flask, redirect, request from config import config from . import extensions, modules def create_app(config_name): config_obj = config[config_name]() app = Flask(__name__, static_url_path='/static') # Initializes configuration values. app.config.from_object(config_obj) # Configure SSL if the current platform supports it. if not app.debug and not app.testing and not app.config.get('SSL_DISABLE'): from flask_sslify import SSLify SSLify(app) @app.before_request def redirect_www(): """ Redirects www requests to non-www. """ if request.host.startswith('www.'): new_host = request.host[4:] return redirect(f"{request.scheme}://{new_host}/", code=301) # Initializes Flask extensions. extensions.init_app(app) # Initializes modules. modules.init_app(app) return app
from flask import Flask from config import config from . import extensions, modules def create_app(config_name): config_obj = config[config_name]() app = Flask(__name__, static_url_path='/static') # Initializes configuration values. app.config.from_object(config_obj) # Configure SSL if the current platform supports it. if not app.debug and not app.testing and not app.config.get('SSL_DISABLE'): from flask_sslify import SSLify SSLify(app) # Initializes Flask extensions. extensions.init_app(app) # Initializes modules. modules.init_app(app) return app Add before_request callback to handle www redirectsfrom flask import Flask, redirect, request from config import config from . import extensions, modules def create_app(config_name): config_obj = config[config_name]() app = Flask(__name__, static_url_path='/static') # Initializes configuration values. app.config.from_object(config_obj) # Configure SSL if the current platform supports it. if not app.debug and not app.testing and not app.config.get('SSL_DISABLE'): from flask_sslify import SSLify SSLify(app) @app.before_request def redirect_www(): """ Redirects www requests to non-www. """ if request.host.startswith('www.'): new_host = request.host[4:] return redirect(f"{request.scheme}://{new_host}/", code=301) # Initializes Flask extensions. extensions.init_app(app) # Initializes modules. modules.init_app(app) return app
<commit_before>from flask import Flask from config import config from . import extensions, modules def create_app(config_name): config_obj = config[config_name]() app = Flask(__name__, static_url_path='/static') # Initializes configuration values. app.config.from_object(config_obj) # Configure SSL if the current platform supports it. if not app.debug and not app.testing and not app.config.get('SSL_DISABLE'): from flask_sslify import SSLify SSLify(app) # Initializes Flask extensions. extensions.init_app(app) # Initializes modules. modules.init_app(app) return app <commit_msg>Add before_request callback to handle www redirects<commit_after>from flask import Flask, redirect, request from config import config from . import extensions, modules def create_app(config_name): config_obj = config[config_name]() app = Flask(__name__, static_url_path='/static') # Initializes configuration values. app.config.from_object(config_obj) # Configure SSL if the current platform supports it. if not app.debug and not app.testing and not app.config.get('SSL_DISABLE'): from flask_sslify import SSLify SSLify(app) @app.before_request def redirect_www(): """ Redirects www requests to non-www. """ if request.host.startswith('www.'): new_host = request.host[4:] return redirect(f"{request.scheme}://{new_host}/", code=301) # Initializes Flask extensions. extensions.init_app(app) # Initializes modules. modules.init_app(app) return app
d0e139d286b18c9dcdc8c46161c4ebdf0f0f8d96
examples/cooperative_binding.py
examples/cooperative_binding.py
import sys import os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), '..')) from crnpy.crn import CRN, from_react_file __author__ = "Elisa Tonello" __copyright__ = "Copyright (c) 2016, Elisa Tonello" __license__ = "BSD" __version__ = "0.0.1" # Cooperative binding print "Creating model..." crn = from_react_file("data/reactions/cooperative_binding") crn.inspect(True) print print("Remove ps1, ps2 and ps3 by qss") crn.remove(qss = ['ps1', 'ps2', 'ps3'], debug = True) for s, f in crn.removed_species: print(s + " = " + str(f)) crn.inspect(True)
import sys import os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), '..')) from crnpy.crn import CRN, from_react_file __author__ = "Elisa Tonello" __copyright__ = "Copyright (c) 2016, Elisa Tonello" __license__ = "BSD" __version__ = "0.0.1" # Cooperative binding print "Creating model..." crn = from_react_file("data/reactions/cooperative_binding") crn.inspect(True) print("") print("Remove ps1, ps2 and ps3 by qssa") crn.remove(qss = ['ps1', 'ps2', 'ps3']) for s, f in crn.removed_species: print(s + " = " + str(f)) crn.inspect(True)
Remove debug and adjusted print.
Remove debug and adjusted print.
Python
bsd-3-clause
etonello/crnpy
import sys import os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), '..')) from crnpy.crn import CRN, from_react_file __author__ = "Elisa Tonello" __copyright__ = "Copyright (c) 2016, Elisa Tonello" __license__ = "BSD" __version__ = "0.0.1" # Cooperative binding print "Creating model..." crn = from_react_file("data/reactions/cooperative_binding") crn.inspect(True) print print("Remove ps1, ps2 and ps3 by qss") crn.remove(qss = ['ps1', 'ps2', 'ps3'], debug = True) for s, f in crn.removed_species: print(s + " = " + str(f)) crn.inspect(True) Remove debug and adjusted print.
import sys import os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), '..')) from crnpy.crn import CRN, from_react_file __author__ = "Elisa Tonello" __copyright__ = "Copyright (c) 2016, Elisa Tonello" __license__ = "BSD" __version__ = "0.0.1" # Cooperative binding print "Creating model..." crn = from_react_file("data/reactions/cooperative_binding") crn.inspect(True) print("") print("Remove ps1, ps2 and ps3 by qssa") crn.remove(qss = ['ps1', 'ps2', 'ps3']) for s, f in crn.removed_species: print(s + " = " + str(f)) crn.inspect(True)
<commit_before>import sys import os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), '..')) from crnpy.crn import CRN, from_react_file __author__ = "Elisa Tonello" __copyright__ = "Copyright (c) 2016, Elisa Tonello" __license__ = "BSD" __version__ = "0.0.1" # Cooperative binding print "Creating model..." crn = from_react_file("data/reactions/cooperative_binding") crn.inspect(True) print print("Remove ps1, ps2 and ps3 by qss") crn.remove(qss = ['ps1', 'ps2', 'ps3'], debug = True) for s, f in crn.removed_species: print(s + " = " + str(f)) crn.inspect(True) <commit_msg>Remove debug and adjusted print.<commit_after>
import sys import os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), '..')) from crnpy.crn import CRN, from_react_file __author__ = "Elisa Tonello" __copyright__ = "Copyright (c) 2016, Elisa Tonello" __license__ = "BSD" __version__ = "0.0.1" # Cooperative binding print "Creating model..." crn = from_react_file("data/reactions/cooperative_binding") crn.inspect(True) print("") print("Remove ps1, ps2 and ps3 by qssa") crn.remove(qss = ['ps1', 'ps2', 'ps3']) for s, f in crn.removed_species: print(s + " = " + str(f)) crn.inspect(True)
import sys import os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), '..')) from crnpy.crn import CRN, from_react_file __author__ = "Elisa Tonello" __copyright__ = "Copyright (c) 2016, Elisa Tonello" __license__ = "BSD" __version__ = "0.0.1" # Cooperative binding print "Creating model..." crn = from_react_file("data/reactions/cooperative_binding") crn.inspect(True) print print("Remove ps1, ps2 and ps3 by qss") crn.remove(qss = ['ps1', 'ps2', 'ps3'], debug = True) for s, f in crn.removed_species: print(s + " = " + str(f)) crn.inspect(True) Remove debug and adjusted print.import sys import os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), '..')) from crnpy.crn import CRN, from_react_file __author__ = "Elisa Tonello" __copyright__ = "Copyright (c) 2016, Elisa Tonello" __license__ = "BSD" __version__ = "0.0.1" # Cooperative binding print "Creating model..." crn = from_react_file("data/reactions/cooperative_binding") crn.inspect(True) print("") print("Remove ps1, ps2 and ps3 by qssa") crn.remove(qss = ['ps1', 'ps2', 'ps3']) for s, f in crn.removed_species: print(s + " = " + str(f)) crn.inspect(True)
<commit_before>import sys import os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), '..')) from crnpy.crn import CRN, from_react_file __author__ = "Elisa Tonello" __copyright__ = "Copyright (c) 2016, Elisa Tonello" __license__ = "BSD" __version__ = "0.0.1" # Cooperative binding print "Creating model..." crn = from_react_file("data/reactions/cooperative_binding") crn.inspect(True) print print("Remove ps1, ps2 and ps3 by qss") crn.remove(qss = ['ps1', 'ps2', 'ps3'], debug = True) for s, f in crn.removed_species: print(s + " = " + str(f)) crn.inspect(True) <commit_msg>Remove debug and adjusted print.<commit_after>import sys import os sys.path.insert(0, os.path.join( os.path.dirname(os.path.realpath(__file__)), '..')) from crnpy.crn import CRN, from_react_file __author__ = "Elisa Tonello" __copyright__ = "Copyright (c) 2016, Elisa Tonello" __license__ = "BSD" __version__ = "0.0.1" # Cooperative binding print "Creating model..." crn = from_react_file("data/reactions/cooperative_binding") crn.inspect(True) print("") print("Remove ps1, ps2 and ps3 by qssa") crn.remove(qss = ['ps1', 'ps2', 'ps3']) for s, f in crn.removed_species: print(s + " = " + str(f)) crn.inspect(True)
67d3193683d2215fdd660bdc086801fe761c7db7
src/views.py
src/views.py
from flask import render_template from app import app @app.route('/') def index(): return render_template('index.html', active='index') @app.route('/contact/') def contact(): return render_template('contact.html', active='contact') @app.context_processor def utility_processor(): def page_title(title=None): return "{} | {}".format(title, app.config['SITE_TITLE']) if title \ else app.config['SITE_TITLE'] def post_source(path): return '{}{}{}'.format(app.config['POST_SOURCE_ROOT'], path, app.config['FLATPAGES_EXTENSION']) return dict(page_title=page_title, post_source=post_source) @app.template_filter('date') def date_filter(date): return date.strftime('%B %-d, %Y')
import os from flask import render_template from flask import send_from_directory from app import app @app.route('/') def index(): return render_template('index.html', active='index') @app.route('/contact/') def contact(): return render_template('contact.html', active='contact') @app.route('/favicon.ico') def favicon(): return send_from_directory(os.path.join(app.root_path, 'static'), 'favicon.ico', mimetype='image/vnd.microsoft.icon') @app.context_processor def utility_processor(): def page_title(title=None): return "{} | {}".format(title, app.config['SITE_TITLE']) if title \ else app.config['SITE_TITLE'] def post_source(path): return '{}{}{}'.format(app.config['POST_SOURCE_ROOT'], path, app.config['FLATPAGES_EXTENSION']) return dict(page_title=page_title, post_source=post_source) @app.template_filter('date') def date_filter(date): return date.strftime('%B %-d, %Y')
Make the favicon available at /favicon.ico
Make the favicon available at /favicon.ico
Python
mit
matachi/MaTachi.github.io,matachi/MaTachi.github.io
from flask import render_template from app import app @app.route('/') def index(): return render_template('index.html', active='index') @app.route('/contact/') def contact(): return render_template('contact.html', active='contact') @app.context_processor def utility_processor(): def page_title(title=None): return "{} | {}".format(title, app.config['SITE_TITLE']) if title \ else app.config['SITE_TITLE'] def post_source(path): return '{}{}{}'.format(app.config['POST_SOURCE_ROOT'], path, app.config['FLATPAGES_EXTENSION']) return dict(page_title=page_title, post_source=post_source) @app.template_filter('date') def date_filter(date): return date.strftime('%B %-d, %Y') Make the favicon available at /favicon.ico
import os from flask import render_template from flask import send_from_directory from app import app @app.route('/') def index(): return render_template('index.html', active='index') @app.route('/contact/') def contact(): return render_template('contact.html', active='contact') @app.route('/favicon.ico') def favicon(): return send_from_directory(os.path.join(app.root_path, 'static'), 'favicon.ico', mimetype='image/vnd.microsoft.icon') @app.context_processor def utility_processor(): def page_title(title=None): return "{} | {}".format(title, app.config['SITE_TITLE']) if title \ else app.config['SITE_TITLE'] def post_source(path): return '{}{}{}'.format(app.config['POST_SOURCE_ROOT'], path, app.config['FLATPAGES_EXTENSION']) return dict(page_title=page_title, post_source=post_source) @app.template_filter('date') def date_filter(date): return date.strftime('%B %-d, %Y')
<commit_before>from flask import render_template from app import app @app.route('/') def index(): return render_template('index.html', active='index') @app.route('/contact/') def contact(): return render_template('contact.html', active='contact') @app.context_processor def utility_processor(): def page_title(title=None): return "{} | {}".format(title, app.config['SITE_TITLE']) if title \ else app.config['SITE_TITLE'] def post_source(path): return '{}{}{}'.format(app.config['POST_SOURCE_ROOT'], path, app.config['FLATPAGES_EXTENSION']) return dict(page_title=page_title, post_source=post_source) @app.template_filter('date') def date_filter(date): return date.strftime('%B %-d, %Y') <commit_msg>Make the favicon available at /favicon.ico<commit_after>
import os from flask import render_template from flask import send_from_directory from app import app @app.route('/') def index(): return render_template('index.html', active='index') @app.route('/contact/') def contact(): return render_template('contact.html', active='contact') @app.route('/favicon.ico') def favicon(): return send_from_directory(os.path.join(app.root_path, 'static'), 'favicon.ico', mimetype='image/vnd.microsoft.icon') @app.context_processor def utility_processor(): def page_title(title=None): return "{} | {}".format(title, app.config['SITE_TITLE']) if title \ else app.config['SITE_TITLE'] def post_source(path): return '{}{}{}'.format(app.config['POST_SOURCE_ROOT'], path, app.config['FLATPAGES_EXTENSION']) return dict(page_title=page_title, post_source=post_source) @app.template_filter('date') def date_filter(date): return date.strftime('%B %-d, %Y')
from flask import render_template from app import app @app.route('/') def index(): return render_template('index.html', active='index') @app.route('/contact/') def contact(): return render_template('contact.html', active='contact') @app.context_processor def utility_processor(): def page_title(title=None): return "{} | {}".format(title, app.config['SITE_TITLE']) if title \ else app.config['SITE_TITLE'] def post_source(path): return '{}{}{}'.format(app.config['POST_SOURCE_ROOT'], path, app.config['FLATPAGES_EXTENSION']) return dict(page_title=page_title, post_source=post_source) @app.template_filter('date') def date_filter(date): return date.strftime('%B %-d, %Y') Make the favicon available at /favicon.icoimport os from flask import render_template from flask import send_from_directory from app import app @app.route('/') def index(): return render_template('index.html', active='index') @app.route('/contact/') def contact(): return render_template('contact.html', active='contact') @app.route('/favicon.ico') def favicon(): return send_from_directory(os.path.join(app.root_path, 'static'), 'favicon.ico', mimetype='image/vnd.microsoft.icon') @app.context_processor def utility_processor(): def page_title(title=None): return "{} | {}".format(title, app.config['SITE_TITLE']) if title \ else app.config['SITE_TITLE'] def post_source(path): return '{}{}{}'.format(app.config['POST_SOURCE_ROOT'], path, app.config['FLATPAGES_EXTENSION']) return dict(page_title=page_title, post_source=post_source) @app.template_filter('date') def date_filter(date): return date.strftime('%B %-d, %Y')
<commit_before>from flask import render_template from app import app @app.route('/') def index(): return render_template('index.html', active='index') @app.route('/contact/') def contact(): return render_template('contact.html', active='contact') @app.context_processor def utility_processor(): def page_title(title=None): return "{} | {}".format(title, app.config['SITE_TITLE']) if title \ else app.config['SITE_TITLE'] def post_source(path): return '{}{}{}'.format(app.config['POST_SOURCE_ROOT'], path, app.config['FLATPAGES_EXTENSION']) return dict(page_title=page_title, post_source=post_source) @app.template_filter('date') def date_filter(date): return date.strftime('%B %-d, %Y') <commit_msg>Make the favicon available at /favicon.ico<commit_after>import os from flask import render_template from flask import send_from_directory from app import app @app.route('/') def index(): return render_template('index.html', active='index') @app.route('/contact/') def contact(): return render_template('contact.html', active='contact') @app.route('/favicon.ico') def favicon(): return send_from_directory(os.path.join(app.root_path, 'static'), 'favicon.ico', mimetype='image/vnd.microsoft.icon') @app.context_processor def utility_processor(): def page_title(title=None): return "{} | {}".format(title, app.config['SITE_TITLE']) if title \ else app.config['SITE_TITLE'] def post_source(path): return '{}{}{}'.format(app.config['POST_SOURCE_ROOT'], path, app.config['FLATPAGES_EXTENSION']) return dict(page_title=page_title, post_source=post_source) @app.template_filter('date') def date_filter(date): return date.strftime('%B %-d, %Y')
0cda8950a661a0d994d7b5824af949ec1e40f584
setup.py
setup.py
from distutils.core import setup setup(name="zutil", version='0.1.4', description="Utilities used for generating zCFD control dictionaries", author="Zenotech", author_email="support@zenotech.com", url="https://zcfd.zenotech.com/", packages=["zutil", "zutil.post", "zutil.analysis", "zutil.plot"], install_requires=[ 'mpi4py', 'ipython<6.0', 'Fabric', 'ipywidgets', 'matplotlib', 'numpy', 'pandas', 'PyYAML' ], )
from distutils.core import setup setup(name="zutil", version='0.1.5', description="Utilities used for generating zCFD control dictionaries", author="Zenotech", author_email="support@zenotech.com", url="https://zcfd.zenotech.com/", packages=["zutil", "zutil.post", "zutil.analysis", "zutil.plot"], install_requires=[ 'ipython<6.0', 'Fabric', 'ipywidgets', 'matplotlib', 'numpy', 'pandas', 'PyYAML' ], extras_require={ "mpi": ["mpi4py"] } )
Make mpi an optional dependency
Make mpi an optional dependency
Python
mit
zCFD/zutil
from distutils.core import setup setup(name="zutil", version='0.1.4', description="Utilities used for generating zCFD control dictionaries", author="Zenotech", author_email="support@zenotech.com", url="https://zcfd.zenotech.com/", packages=["zutil", "zutil.post", "zutil.analysis", "zutil.plot"], install_requires=[ 'mpi4py', 'ipython<6.0', 'Fabric', 'ipywidgets', 'matplotlib', 'numpy', 'pandas', 'PyYAML' ], )Make mpi an optional dependency
from distutils.core import setup setup(name="zutil", version='0.1.5', description="Utilities used for generating zCFD control dictionaries", author="Zenotech", author_email="support@zenotech.com", url="https://zcfd.zenotech.com/", packages=["zutil", "zutil.post", "zutil.analysis", "zutil.plot"], install_requires=[ 'ipython<6.0', 'Fabric', 'ipywidgets', 'matplotlib', 'numpy', 'pandas', 'PyYAML' ], extras_require={ "mpi": ["mpi4py"] } )
<commit_before>from distutils.core import setup setup(name="zutil", version='0.1.4', description="Utilities used for generating zCFD control dictionaries", author="Zenotech", author_email="support@zenotech.com", url="https://zcfd.zenotech.com/", packages=["zutil", "zutil.post", "zutil.analysis", "zutil.plot"], install_requires=[ 'mpi4py', 'ipython<6.0', 'Fabric', 'ipywidgets', 'matplotlib', 'numpy', 'pandas', 'PyYAML' ], )<commit_msg>Make mpi an optional dependency<commit_after>
from distutils.core import setup setup(name="zutil", version='0.1.5', description="Utilities used for generating zCFD control dictionaries", author="Zenotech", author_email="support@zenotech.com", url="https://zcfd.zenotech.com/", packages=["zutil", "zutil.post", "zutil.analysis", "zutil.plot"], install_requires=[ 'ipython<6.0', 'Fabric', 'ipywidgets', 'matplotlib', 'numpy', 'pandas', 'PyYAML' ], extras_require={ "mpi": ["mpi4py"] } )
from distutils.core import setup setup(name="zutil", version='0.1.4', description="Utilities used for generating zCFD control dictionaries", author="Zenotech", author_email="support@zenotech.com", url="https://zcfd.zenotech.com/", packages=["zutil", "zutil.post", "zutil.analysis", "zutil.plot"], install_requires=[ 'mpi4py', 'ipython<6.0', 'Fabric', 'ipywidgets', 'matplotlib', 'numpy', 'pandas', 'PyYAML' ], )Make mpi an optional dependencyfrom distutils.core import setup setup(name="zutil", version='0.1.5', description="Utilities used for generating zCFD control dictionaries", author="Zenotech", author_email="support@zenotech.com", url="https://zcfd.zenotech.com/", packages=["zutil", "zutil.post", "zutil.analysis", "zutil.plot"], install_requires=[ 'ipython<6.0', 'Fabric', 'ipywidgets', 'matplotlib', 'numpy', 'pandas', 'PyYAML' ], extras_require={ "mpi": ["mpi4py"] } )
<commit_before>from distutils.core import setup setup(name="zutil", version='0.1.4', description="Utilities used for generating zCFD control dictionaries", author="Zenotech", author_email="support@zenotech.com", url="https://zcfd.zenotech.com/", packages=["zutil", "zutil.post", "zutil.analysis", "zutil.plot"], install_requires=[ 'mpi4py', 'ipython<6.0', 'Fabric', 'ipywidgets', 'matplotlib', 'numpy', 'pandas', 'PyYAML' ], )<commit_msg>Make mpi an optional dependency<commit_after>from distutils.core import setup setup(name="zutil", version='0.1.5', description="Utilities used for generating zCFD control dictionaries", author="Zenotech", author_email="support@zenotech.com", url="https://zcfd.zenotech.com/", packages=["zutil", "zutil.post", "zutil.analysis", "zutil.plot"], install_requires=[ 'ipython<6.0', 'Fabric', 'ipywidgets', 'matplotlib', 'numpy', 'pandas', 'PyYAML' ], extras_require={ "mpi": ["mpi4py"] } )
334ccd245997373c4ddc21c03657339c4ca20192
setup.py
setup.py
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys try: from setuptools import setup except ImportError: from distutils.core import setup settings = dict() # Publish if sys.argv[-1] == 'publish': os.system('python setup.py sdist upload') sys.exit() settings.update( name='whenpy', version='0.1.0', description='Friendly Dates and Times', long_description=open('README.rst').read(), author='Andy Dirnberger', author_email='dirn@dirnonline.com', url='https://github.com/dirn/when.py', packages=['when'], install_requires=['pytz'], license=open('LICENSE').read(), classifiers=( 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ), ) setup(**settings)
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys try: from setuptools import setup except ImportError: from distutils.core import setup settings = dict() # Publish if sys.argv[-1] == 'publish': os.system('python setup.py sdist upload') sys.exit() settings.update( name='whenpy', version='0.1.0', description='Friendly Dates and Times', long_description=open('README.rst').read(), author='Andy Dirnberger', author_email='dirn@dirnonline.com', url='https://github.com/dirn/when.py', packages=['when'], package_data={'': ['LICENSE']}, include_package_data=True, install_requires=['pytz'], license=open('LICENSE').read(), classifiers=( 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ), ) setup(**settings)
Add LICENSE to package data
Add LICENSE to package data The LICENSE file isn't included with the version found on PyPI. Including it in the `package_data` argument passed to `setup` should fix this.
Python
bsd-3-clause
dirn/When.py
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys try: from setuptools import setup except ImportError: from distutils.core import setup settings = dict() # Publish if sys.argv[-1] == 'publish': os.system('python setup.py sdist upload') sys.exit() settings.update( name='whenpy', version='0.1.0', description='Friendly Dates and Times', long_description=open('README.rst').read(), author='Andy Dirnberger', author_email='dirn@dirnonline.com', url='https://github.com/dirn/when.py', packages=['when'], install_requires=['pytz'], license=open('LICENSE').read(), classifiers=( 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ), ) setup(**settings) Add LICENSE to package data The LICENSE file isn't included with the version found on PyPI. Including it in the `package_data` argument passed to `setup` should fix this.
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys try: from setuptools import setup except ImportError: from distutils.core import setup settings = dict() # Publish if sys.argv[-1] == 'publish': os.system('python setup.py sdist upload') sys.exit() settings.update( name='whenpy', version='0.1.0', description='Friendly Dates and Times', long_description=open('README.rst').read(), author='Andy Dirnberger', author_email='dirn@dirnonline.com', url='https://github.com/dirn/when.py', packages=['when'], package_data={'': ['LICENSE']}, include_package_data=True, install_requires=['pytz'], license=open('LICENSE').read(), classifiers=( 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ), ) setup(**settings)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys try: from setuptools import setup except ImportError: from distutils.core import setup settings = dict() # Publish if sys.argv[-1] == 'publish': os.system('python setup.py sdist upload') sys.exit() settings.update( name='whenpy', version='0.1.0', description='Friendly Dates and Times', long_description=open('README.rst').read(), author='Andy Dirnberger', author_email='dirn@dirnonline.com', url='https://github.com/dirn/when.py', packages=['when'], install_requires=['pytz'], license=open('LICENSE').read(), classifiers=( 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ), ) setup(**settings) <commit_msg>Add LICENSE to package data The LICENSE file isn't included with the version found on PyPI. Including it in the `package_data` argument passed to `setup` should fix this.<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys try: from setuptools import setup except ImportError: from distutils.core import setup settings = dict() # Publish if sys.argv[-1] == 'publish': os.system('python setup.py sdist upload') sys.exit() settings.update( name='whenpy', version='0.1.0', description='Friendly Dates and Times', long_description=open('README.rst').read(), author='Andy Dirnberger', author_email='dirn@dirnonline.com', url='https://github.com/dirn/when.py', packages=['when'], package_data={'': ['LICENSE']}, include_package_data=True, install_requires=['pytz'], license=open('LICENSE').read(), classifiers=( 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ), ) setup(**settings)
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys try: from setuptools import setup except ImportError: from distutils.core import setup settings = dict() # Publish if sys.argv[-1] == 'publish': os.system('python setup.py sdist upload') sys.exit() settings.update( name='whenpy', version='0.1.0', description='Friendly Dates and Times', long_description=open('README.rst').read(), author='Andy Dirnberger', author_email='dirn@dirnonline.com', url='https://github.com/dirn/when.py', packages=['when'], install_requires=['pytz'], license=open('LICENSE').read(), classifiers=( 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ), ) setup(**settings) Add LICENSE to package data The LICENSE file isn't included with the version found on PyPI. Including it in the `package_data` argument passed to `setup` should fix this.#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys try: from setuptools import setup except ImportError: from distutils.core import setup settings = dict() # Publish if sys.argv[-1] == 'publish': os.system('python setup.py sdist upload') sys.exit() settings.update( name='whenpy', version='0.1.0', description='Friendly Dates and Times', long_description=open('README.rst').read(), author='Andy Dirnberger', author_email='dirn@dirnonline.com', url='https://github.com/dirn/when.py', packages=['when'], package_data={'': ['LICENSE']}, include_package_data=True, install_requires=['pytz'], license=open('LICENSE').read(), classifiers=( 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ), ) setup(**settings)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys try: from setuptools import setup except ImportError: from distutils.core import setup settings = dict() # Publish if sys.argv[-1] == 'publish': os.system('python setup.py sdist upload') sys.exit() settings.update( name='whenpy', version='0.1.0', description='Friendly Dates and Times', long_description=open('README.rst').read(), author='Andy Dirnberger', author_email='dirn@dirnonline.com', url='https://github.com/dirn/when.py', packages=['when'], install_requires=['pytz'], license=open('LICENSE').read(), classifiers=( 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ), ) setup(**settings) <commit_msg>Add LICENSE to package data The LICENSE file isn't included with the version found on PyPI. Including it in the `package_data` argument passed to `setup` should fix this.<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys try: from setuptools import setup except ImportError: from distutils.core import setup settings = dict() # Publish if sys.argv[-1] == 'publish': os.system('python setup.py sdist upload') sys.exit() settings.update( name='whenpy', version='0.1.0', description='Friendly Dates and Times', long_description=open('README.rst').read(), author='Andy Dirnberger', author_email='dirn@dirnonline.com', url='https://github.com/dirn/when.py', packages=['when'], package_data={'': ['LICENSE']}, include_package_data=True, install_requires=['pytz'], license=open('LICENSE').read(), classifiers=( 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ), ) setup(**settings)
55851fc7fadd4b6a0241e63e6b708f9b5d4e253d
setup.py
setup.py
from setuptools import setup, find_packages setup( name="django-service-rating-store", version="0.3.0", url='https://github.com/praekelt/django-service-rating-store', license='BSD', description=( "Django app that allows storage and visualisation of Service Rating data posted via REST API"), long_description=open('README.rst', 'r').read(), author='Praekelt Foundation', author_email='devops@praekeltfoundation.org', packages=find_packages(), include_package_data=True, install_requires=[ 'Django', 'django-tastypie', 'South', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: System :: Networking', ], )
from setuptools import setup, find_packages setup( name="django-service-rating-store", version="0.3.0a", url='https://github.com/praekelt/django-service-rating-store', license='BSD', description=( "Django app that allows storage and visualisation of Service Rating data posted via REST API"), long_description=open('README.rst', 'r').read(), author='Praekelt Foundation', author_email='devops@praekeltfoundation.org', packages=find_packages(), include_package_data=True, install_requires=[ 'Django', 'django-tastypie', 'South', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: System :: Networking', ], )
Bump to 0.3.0a post release
Bump to 0.3.0a post release
Python
bsd-3-clause
praekelt/django-service-rating-store
from setuptools import setup, find_packages setup( name="django-service-rating-store", version="0.3.0", url='https://github.com/praekelt/django-service-rating-store', license='BSD', description=( "Django app that allows storage and visualisation of Service Rating data posted via REST API"), long_description=open('README.rst', 'r').read(), author='Praekelt Foundation', author_email='devops@praekeltfoundation.org', packages=find_packages(), include_package_data=True, install_requires=[ 'Django', 'django-tastypie', 'South', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: System :: Networking', ], ) Bump to 0.3.0a post release
from setuptools import setup, find_packages setup( name="django-service-rating-store", version="0.3.0a", url='https://github.com/praekelt/django-service-rating-store', license='BSD', description=( "Django app that allows storage and visualisation of Service Rating data posted via REST API"), long_description=open('README.rst', 'r').read(), author='Praekelt Foundation', author_email='devops@praekeltfoundation.org', packages=find_packages(), include_package_data=True, install_requires=[ 'Django', 'django-tastypie', 'South', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: System :: Networking', ], )
<commit_before>from setuptools import setup, find_packages setup( name="django-service-rating-store", version="0.3.0", url='https://github.com/praekelt/django-service-rating-store', license='BSD', description=( "Django app that allows storage and visualisation of Service Rating data posted via REST API"), long_description=open('README.rst', 'r').read(), author='Praekelt Foundation', author_email='devops@praekeltfoundation.org', packages=find_packages(), include_package_data=True, install_requires=[ 'Django', 'django-tastypie', 'South', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: System :: Networking', ], ) <commit_msg>Bump to 0.3.0a post release<commit_after>
from setuptools import setup, find_packages setup( name="django-service-rating-store", version="0.3.0a", url='https://github.com/praekelt/django-service-rating-store', license='BSD', description=( "Django app that allows storage and visualisation of Service Rating data posted via REST API"), long_description=open('README.rst', 'r').read(), author='Praekelt Foundation', author_email='devops@praekeltfoundation.org', packages=find_packages(), include_package_data=True, install_requires=[ 'Django', 'django-tastypie', 'South', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: System :: Networking', ], )
from setuptools import setup, find_packages setup( name="django-service-rating-store", version="0.3.0", url='https://github.com/praekelt/django-service-rating-store', license='BSD', description=( "Django app that allows storage and visualisation of Service Rating data posted via REST API"), long_description=open('README.rst', 'r').read(), author='Praekelt Foundation', author_email='devops@praekeltfoundation.org', packages=find_packages(), include_package_data=True, install_requires=[ 'Django', 'django-tastypie', 'South', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: System :: Networking', ], ) Bump to 0.3.0a post releasefrom setuptools import setup, find_packages setup( name="django-service-rating-store", version="0.3.0a", url='https://github.com/praekelt/django-service-rating-store', license='BSD', description=( "Django app that allows storage and visualisation of Service Rating data posted via REST API"), long_description=open('README.rst', 'r').read(), author='Praekelt Foundation', author_email='devops@praekeltfoundation.org', packages=find_packages(), include_package_data=True, install_requires=[ 'Django', 'django-tastypie', 'South', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: System :: Networking', ], )
<commit_before>from setuptools import setup, find_packages setup( name="django-service-rating-store", version="0.3.0", url='https://github.com/praekelt/django-service-rating-store', license='BSD', description=( "Django app that allows storage and visualisation of Service Rating data posted via REST API"), long_description=open('README.rst', 'r').read(), author='Praekelt Foundation', author_email='devops@praekeltfoundation.org', packages=find_packages(), include_package_data=True, install_requires=[ 'Django', 'django-tastypie', 'South', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: System :: Networking', ], ) <commit_msg>Bump to 0.3.0a post release<commit_after>from setuptools import setup, find_packages setup( name="django-service-rating-store", version="0.3.0a", url='https://github.com/praekelt/django-service-rating-store', license='BSD', description=( "Django app that allows storage and visualisation of Service Rating data posted via REST API"), long_description=open('README.rst', 'r').read(), author='Praekelt Foundation', author_email='devops@praekeltfoundation.org', packages=find_packages(), include_package_data=True, install_requires=[ 'Django', 'django-tastypie', 'South', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: System :: Networking', ], )
845ed2a406baf333cd289f789aed0089b3ec3be9
setup.py
setup.py
# -*- coding: utf-8 -*- import os from setuptools import setup def read(fname): try: return open(os.path.join(os.path.dirname(__file__), fname)).read() except: return '' setup( name='todoist-python', version='7.0', packages=['todoist', 'todoist.managers'], author='Doist Team', author_email='info@todoist.com', license='BSD', description='todoist-python - The official Todoist Python API library', long_description = read('README.md'), install_requires=[ 'requests', ], # see here for complete list of classifiers # http://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=( 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', ), )
# -*- coding: utf-8 -*- import os from setuptools import setup def read(fname): try: return open(os.path.join(os.path.dirname(__file__), fname)).read() except: return '' setup( name='todoist-python', version='7.0.1', packages=['todoist', 'todoist.managers'], author='Doist Team', author_email='info@todoist.com', license='BSD', description='todoist-python - The official Todoist Python API library', long_description = read('README.md'), install_requires=[ 'requests', ], # see here for complete list of classifiers # http://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=( 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', ), )
Update the PyPI version to 7.0.1.
Update the PyPI version to 7.0.1.
Python
mit
Doist/todoist-python
# -*- coding: utf-8 -*- import os from setuptools import setup def read(fname): try: return open(os.path.join(os.path.dirname(__file__), fname)).read() except: return '' setup( name='todoist-python', version='7.0', packages=['todoist', 'todoist.managers'], author='Doist Team', author_email='info@todoist.com', license='BSD', description='todoist-python - The official Todoist Python API library', long_description = read('README.md'), install_requires=[ 'requests', ], # see here for complete list of classifiers # http://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=( 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', ), ) Update the PyPI version to 7.0.1.
# -*- coding: utf-8 -*- import os from setuptools import setup def read(fname): try: return open(os.path.join(os.path.dirname(__file__), fname)).read() except: return '' setup( name='todoist-python', version='7.0.1', packages=['todoist', 'todoist.managers'], author='Doist Team', author_email='info@todoist.com', license='BSD', description='todoist-python - The official Todoist Python API library', long_description = read('README.md'), install_requires=[ 'requests', ], # see here for complete list of classifiers # http://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=( 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', ), )
<commit_before># -*- coding: utf-8 -*- import os from setuptools import setup def read(fname): try: return open(os.path.join(os.path.dirname(__file__), fname)).read() except: return '' setup( name='todoist-python', version='7.0', packages=['todoist', 'todoist.managers'], author='Doist Team', author_email='info@todoist.com', license='BSD', description='todoist-python - The official Todoist Python API library', long_description = read('README.md'), install_requires=[ 'requests', ], # see here for complete list of classifiers # http://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=( 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', ), ) <commit_msg>Update the PyPI version to 7.0.1.<commit_after>
# -*- coding: utf-8 -*- import os from setuptools import setup def read(fname): try: return open(os.path.join(os.path.dirname(__file__), fname)).read() except: return '' setup( name='todoist-python', version='7.0.1', packages=['todoist', 'todoist.managers'], author='Doist Team', author_email='info@todoist.com', license='BSD', description='todoist-python - The official Todoist Python API library', long_description = read('README.md'), install_requires=[ 'requests', ], # see here for complete list of classifiers # http://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=( 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', ), )
# -*- coding: utf-8 -*- import os from setuptools import setup def read(fname): try: return open(os.path.join(os.path.dirname(__file__), fname)).read() except: return '' setup( name='todoist-python', version='7.0', packages=['todoist', 'todoist.managers'], author='Doist Team', author_email='info@todoist.com', license='BSD', description='todoist-python - The official Todoist Python API library', long_description = read('README.md'), install_requires=[ 'requests', ], # see here for complete list of classifiers # http://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=( 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', ), ) Update the PyPI version to 7.0.1.# -*- coding: utf-8 -*- import os from setuptools import setup def read(fname): try: return open(os.path.join(os.path.dirname(__file__), fname)).read() except: return '' setup( name='todoist-python', version='7.0.1', packages=['todoist', 'todoist.managers'], author='Doist Team', author_email='info@todoist.com', license='BSD', description='todoist-python - The official Todoist Python API library', long_description = read('README.md'), install_requires=[ 'requests', ], # see here for complete list of classifiers # http://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=( 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', ), )
<commit_before># -*- coding: utf-8 -*- import os from setuptools import setup def read(fname): try: return open(os.path.join(os.path.dirname(__file__), fname)).read() except: return '' setup( name='todoist-python', version='7.0', packages=['todoist', 'todoist.managers'], author='Doist Team', author_email='info@todoist.com', license='BSD', description='todoist-python - The official Todoist Python API library', long_description = read('README.md'), install_requires=[ 'requests', ], # see here for complete list of classifiers # http://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=( 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', ), ) <commit_msg>Update the PyPI version to 7.0.1.<commit_after># -*- coding: utf-8 -*- import os from setuptools import setup def read(fname): try: return open(os.path.join(os.path.dirname(__file__), fname)).read() except: return '' setup( name='todoist-python', version='7.0.1', packages=['todoist', 'todoist.managers'], author='Doist Team', author_email='info@todoist.com', license='BSD', description='todoist-python - The official Todoist Python API library', long_description = read('README.md'), install_requires=[ 'requests', ], # see here for complete list of classifiers # http://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=( 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', ), )
e00f59a20c3efa6a8db307f482d54a99f5b1f643
setup.py
setup.py
import os import sys from setuptools import setup # Utility function to read the README file. def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() # The packages we depend on dependencies = [ "Flask==0.9", "MySQL-python==1.2.4c1" ] # If old Python, then we need simplejson if sys.version_info < (2,6): dependencies += ["simplejson>=2.6.2"] setup( name = "pegasus-metrics", version = "0.1", author = "Gideon Juve", author_email = "gideon@isi.edu", description = "Anonymous usage metrics collection and reporting for Pegasus", long_description = read("README.md"), license = "Apache2", url = "https://github.com/pegasus-isi/pegasus-metrics", classifiers = [ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: Apache Software License", ], packages = ["pegasus","pegasus.metrics", "pegasus.metrics"], include_package_data = True, zip_safe = False, scripts = ["bin/pegasus-metrics-server"], install_requires = dependencies )
import os import sys from setuptools import setup # Utility function to read the README file. def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() # The packages we depend on dependencies = [ "Flask==0.9", "MySQL-python==1.2.4c1" ] # If old Python, then we need simplejson if sys.version_info < (2,6): dependencies += ["simplejson>=2.6.2"] setup( name = "pegasus-metrics", version = "0.1", author = "Gideon Juve", author_email = "gideon@isi.edu", description = "Anonymous usage metrics collection and reporting for Pegasus", long_description = read("README.md"), license = "Apache2", url = "https://github.com/pegasus-isi/pegasus-metrics", classifiers = [ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: Apache Software License", ], packages = ["pegasus","pegasus.metrics"], package_data = {"pegasus.metrics" : ["templates/*", "static/*"] }, include_package_data = True, zip_safe = False, scripts = ["bin/pegasus-metrics-server"], install_requires = dependencies )
Include templates and static data in packaging
Include templates and static data in packaging
Python
apache-2.0
pegasus-isi/pegasus-metrics,pegasus-isi/pegasus-metrics,pegasus-isi/pegasus-metrics
import os import sys from setuptools import setup # Utility function to read the README file. def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() # The packages we depend on dependencies = [ "Flask==0.9", "MySQL-python==1.2.4c1" ] # If old Python, then we need simplejson if sys.version_info < (2,6): dependencies += ["simplejson>=2.6.2"] setup( name = "pegasus-metrics", version = "0.1", author = "Gideon Juve", author_email = "gideon@isi.edu", description = "Anonymous usage metrics collection and reporting for Pegasus", long_description = read("README.md"), license = "Apache2", url = "https://github.com/pegasus-isi/pegasus-metrics", classifiers = [ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: Apache Software License", ], packages = ["pegasus","pegasus.metrics", "pegasus.metrics"], include_package_data = True, zip_safe = False, scripts = ["bin/pegasus-metrics-server"], install_requires = dependencies ) Include templates and static data in packaging
import os import sys from setuptools import setup # Utility function to read the README file. def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() # The packages we depend on dependencies = [ "Flask==0.9", "MySQL-python==1.2.4c1" ] # If old Python, then we need simplejson if sys.version_info < (2,6): dependencies += ["simplejson>=2.6.2"] setup( name = "pegasus-metrics", version = "0.1", author = "Gideon Juve", author_email = "gideon@isi.edu", description = "Anonymous usage metrics collection and reporting for Pegasus", long_description = read("README.md"), license = "Apache2", url = "https://github.com/pegasus-isi/pegasus-metrics", classifiers = [ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: Apache Software License", ], packages = ["pegasus","pegasus.metrics"], package_data = {"pegasus.metrics" : ["templates/*", "static/*"] }, include_package_data = True, zip_safe = False, scripts = ["bin/pegasus-metrics-server"], install_requires = dependencies )
<commit_before>import os import sys from setuptools import setup # Utility function to read the README file. def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() # The packages we depend on dependencies = [ "Flask==0.9", "MySQL-python==1.2.4c1" ] # If old Python, then we need simplejson if sys.version_info < (2,6): dependencies += ["simplejson>=2.6.2"] setup( name = "pegasus-metrics", version = "0.1", author = "Gideon Juve", author_email = "gideon@isi.edu", description = "Anonymous usage metrics collection and reporting for Pegasus", long_description = read("README.md"), license = "Apache2", url = "https://github.com/pegasus-isi/pegasus-metrics", classifiers = [ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: Apache Software License", ], packages = ["pegasus","pegasus.metrics", "pegasus.metrics"], include_package_data = True, zip_safe = False, scripts = ["bin/pegasus-metrics-server"], install_requires = dependencies ) <commit_msg>Include templates and static data in packaging<commit_after>
import os import sys from setuptools import setup # Utility function to read the README file. def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() # The packages we depend on dependencies = [ "Flask==0.9", "MySQL-python==1.2.4c1" ] # If old Python, then we need simplejson if sys.version_info < (2,6): dependencies += ["simplejson>=2.6.2"] setup( name = "pegasus-metrics", version = "0.1", author = "Gideon Juve", author_email = "gideon@isi.edu", description = "Anonymous usage metrics collection and reporting for Pegasus", long_description = read("README.md"), license = "Apache2", url = "https://github.com/pegasus-isi/pegasus-metrics", classifiers = [ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: Apache Software License", ], packages = ["pegasus","pegasus.metrics"], package_data = {"pegasus.metrics" : ["templates/*", "static/*"] }, include_package_data = True, zip_safe = False, scripts = ["bin/pegasus-metrics-server"], install_requires = dependencies )
import os import sys from setuptools import setup # Utility function to read the README file. def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() # The packages we depend on dependencies = [ "Flask==0.9", "MySQL-python==1.2.4c1" ] # If old Python, then we need simplejson if sys.version_info < (2,6): dependencies += ["simplejson>=2.6.2"] setup( name = "pegasus-metrics", version = "0.1", author = "Gideon Juve", author_email = "gideon@isi.edu", description = "Anonymous usage metrics collection and reporting for Pegasus", long_description = read("README.md"), license = "Apache2", url = "https://github.com/pegasus-isi/pegasus-metrics", classifiers = [ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: Apache Software License", ], packages = ["pegasus","pegasus.metrics", "pegasus.metrics"], include_package_data = True, zip_safe = False, scripts = ["bin/pegasus-metrics-server"], install_requires = dependencies ) Include templates and static data in packagingimport os import sys from setuptools import setup # Utility function to read the README file. def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() # The packages we depend on dependencies = [ "Flask==0.9", "MySQL-python==1.2.4c1" ] # If old Python, then we need simplejson if sys.version_info < (2,6): dependencies += ["simplejson>=2.6.2"] setup( name = "pegasus-metrics", version = "0.1", author = "Gideon Juve", author_email = "gideon@isi.edu", description = "Anonymous usage metrics collection and reporting for Pegasus", long_description = read("README.md"), license = "Apache2", url = "https://github.com/pegasus-isi/pegasus-metrics", classifiers = [ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: Apache Software License", ], packages = ["pegasus","pegasus.metrics"], package_data = {"pegasus.metrics" : ["templates/*", "static/*"] }, include_package_data = True, zip_safe = False, scripts = ["bin/pegasus-metrics-server"], install_requires = dependencies )
<commit_before>import os import sys from setuptools import setup # Utility function to read the README file. def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() # The packages we depend on dependencies = [ "Flask==0.9", "MySQL-python==1.2.4c1" ] # If old Python, then we need simplejson if sys.version_info < (2,6): dependencies += ["simplejson>=2.6.2"] setup( name = "pegasus-metrics", version = "0.1", author = "Gideon Juve", author_email = "gideon@isi.edu", description = "Anonymous usage metrics collection and reporting for Pegasus", long_description = read("README.md"), license = "Apache2", url = "https://github.com/pegasus-isi/pegasus-metrics", classifiers = [ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: Apache Software License", ], packages = ["pegasus","pegasus.metrics", "pegasus.metrics"], include_package_data = True, zip_safe = False, scripts = ["bin/pegasus-metrics-server"], install_requires = dependencies ) <commit_msg>Include templates and static data in packaging<commit_after>import os import sys from setuptools import setup # Utility function to read the README file. def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() # The packages we depend on dependencies = [ "Flask==0.9", "MySQL-python==1.2.4c1" ] # If old Python, then we need simplejson if sys.version_info < (2,6): dependencies += ["simplejson>=2.6.2"] setup( name = "pegasus-metrics", version = "0.1", author = "Gideon Juve", author_email = "gideon@isi.edu", description = "Anonymous usage metrics collection and reporting for Pegasus", long_description = read("README.md"), license = "Apache2", url = "https://github.com/pegasus-isi/pegasus-metrics", classifiers = [ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: Apache Software License", ], packages = ["pegasus","pegasus.metrics"], package_data = {"pegasus.metrics" : ["templates/*", "static/*"] }, include_package_data = True, zip_safe = False, scripts = ["bin/pegasus-metrics-server"], install_requires = dependencies )
c3ad61a642e0e1149ae7ea1e8bde2efa6bbc9263
setup.py
setup.py
from distutils.core import setup from setuptools import find_packages with open('README.md') as fp: long_description = fp.read() setup( name='sendwithus', version='1.6.6', author='sendwithus', author_email='us@sendwithus.com', packages=find_packages(), scripts=[], url='https://github.com/sendwithus/sendwithus_python', license='LICENSE.txt', description='Python API client for sendwithus.com', long_description=long_description, test_suite="sendwithus.test", install_requires=[ "requests >= 1.1.0", "six >= 1.9.0" ], classifiers=[ "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "License :: OSI Approved :: Apache Software License", "Development Status :: 5 - Production/Stable", "Topic :: Communications :: Email" ] )
from distutils.core import setup from setuptools import find_packages with open('README.md') as fp: long_description = fp.read() setup( name='sendwithus', version='1.6.6', author='sendwithus', author_email='us@sendwithus.com', packages=find_packages(), scripts=[], url='https://github.com/sendwithus/sendwithus_python', license='LICENSE.txt', description='Python API client for sendwithus.com', long_description=long_description, test_suite="sendwithus.test", install_requires=[ "requests >= 1.1.0", "six >= 1.9.0" ], classifiers=[ "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "License :: OSI Approved :: Apache Software License", "Development Status :: 5 - Production/Stable", "Topic :: Communications :: Email" ] )
Update the classifier for all the tested versions of python
Update the classifier for all the tested versions of python
Python
apache-2.0
mefyl/sendwithus_python,sendwithus/sendwithus_python
from distutils.core import setup from setuptools import find_packages with open('README.md') as fp: long_description = fp.read() setup( name='sendwithus', version='1.6.6', author='sendwithus', author_email='us@sendwithus.com', packages=find_packages(), scripts=[], url='https://github.com/sendwithus/sendwithus_python', license='LICENSE.txt', description='Python API client for sendwithus.com', long_description=long_description, test_suite="sendwithus.test", install_requires=[ "requests >= 1.1.0", "six >= 1.9.0" ], classifiers=[ "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "License :: OSI Approved :: Apache Software License", "Development Status :: 5 - Production/Stable", "Topic :: Communications :: Email" ] ) Update the classifier for all the tested versions of python
from distutils.core import setup from setuptools import find_packages with open('README.md') as fp: long_description = fp.read() setup( name='sendwithus', version='1.6.6', author='sendwithus', author_email='us@sendwithus.com', packages=find_packages(), scripts=[], url='https://github.com/sendwithus/sendwithus_python', license='LICENSE.txt', description='Python API client for sendwithus.com', long_description=long_description, test_suite="sendwithus.test", install_requires=[ "requests >= 1.1.0", "six >= 1.9.0" ], classifiers=[ "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "License :: OSI Approved :: Apache Software License", "Development Status :: 5 - Production/Stable", "Topic :: Communications :: Email" ] )
<commit_before>from distutils.core import setup from setuptools import find_packages with open('README.md') as fp: long_description = fp.read() setup( name='sendwithus', version='1.6.6', author='sendwithus', author_email='us@sendwithus.com', packages=find_packages(), scripts=[], url='https://github.com/sendwithus/sendwithus_python', license='LICENSE.txt', description='Python API client for sendwithus.com', long_description=long_description, test_suite="sendwithus.test", install_requires=[ "requests >= 1.1.0", "six >= 1.9.0" ], classifiers=[ "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "License :: OSI Approved :: Apache Software License", "Development Status :: 5 - Production/Stable", "Topic :: Communications :: Email" ] ) <commit_msg>Update the classifier for all the tested versions of python<commit_after>
from distutils.core import setup from setuptools import find_packages with open('README.md') as fp: long_description = fp.read() setup( name='sendwithus', version='1.6.6', author='sendwithus', author_email='us@sendwithus.com', packages=find_packages(), scripts=[], url='https://github.com/sendwithus/sendwithus_python', license='LICENSE.txt', description='Python API client for sendwithus.com', long_description=long_description, test_suite="sendwithus.test", install_requires=[ "requests >= 1.1.0", "six >= 1.9.0" ], classifiers=[ "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "License :: OSI Approved :: Apache Software License", "Development Status :: 5 - Production/Stable", "Topic :: Communications :: Email" ] )
from distutils.core import setup from setuptools import find_packages with open('README.md') as fp: long_description = fp.read() setup( name='sendwithus', version='1.6.6', author='sendwithus', author_email='us@sendwithus.com', packages=find_packages(), scripts=[], url='https://github.com/sendwithus/sendwithus_python', license='LICENSE.txt', description='Python API client for sendwithus.com', long_description=long_description, test_suite="sendwithus.test", install_requires=[ "requests >= 1.1.0", "six >= 1.9.0" ], classifiers=[ "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "License :: OSI Approved :: Apache Software License", "Development Status :: 5 - Production/Stable", "Topic :: Communications :: Email" ] ) Update the classifier for all the tested versions of pythonfrom distutils.core import setup from setuptools import find_packages with open('README.md') as fp: long_description = fp.read() setup( name='sendwithus', version='1.6.6', author='sendwithus', author_email='us@sendwithus.com', packages=find_packages(), scripts=[], url='https://github.com/sendwithus/sendwithus_python', license='LICENSE.txt', description='Python API client for sendwithus.com', long_description=long_description, test_suite="sendwithus.test", install_requires=[ "requests >= 1.1.0", "six >= 1.9.0" ], classifiers=[ "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "License :: OSI Approved :: Apache Software License", "Development Status :: 5 - Production/Stable", "Topic :: Communications :: Email" ] )
<commit_before>from distutils.core import setup from setuptools import find_packages with open('README.md') as fp: long_description = fp.read() setup( name='sendwithus', version='1.6.6', author='sendwithus', author_email='us@sendwithus.com', packages=find_packages(), scripts=[], url='https://github.com/sendwithus/sendwithus_python', license='LICENSE.txt', description='Python API client for sendwithus.com', long_description=long_description, test_suite="sendwithus.test", install_requires=[ "requests >= 1.1.0", "six >= 1.9.0" ], classifiers=[ "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "License :: OSI Approved :: Apache Software License", "Development Status :: 5 - Production/Stable", "Topic :: Communications :: Email" ] ) <commit_msg>Update the classifier for all the tested versions of python<commit_after>from distutils.core import setup from setuptools import find_packages with open('README.md') as fp: long_description = fp.read() setup( name='sendwithus', version='1.6.6', author='sendwithus', author_email='us@sendwithus.com', packages=find_packages(), scripts=[], url='https://github.com/sendwithus/sendwithus_python', license='LICENSE.txt', description='Python API client for sendwithus.com', long_description=long_description, test_suite="sendwithus.test", install_requires=[ "requests >= 1.1.0", "six >= 1.9.0" ], classifiers=[ "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "License :: OSI Approved :: Apache Software License", "Development Status :: 5 - Production/Stable", "Topic :: Communications :: Email" ] )
9f50db5179886770178a07a1167b40f3d17ccbd9
setup.py
setup.py
# -*- coding: utf-8 -*- from setuptools import setup, find_packages classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Topic :: Software Development", "Topic :: Software Development :: Version Control", "Topic :: Text Processing :: Filters", ] setup( name='diff-highlight', version='0.1.0', description='pretty diff highlighter; emphasis changed words in diff', long_description=open("README.rst").read(), classifiers=classifiers, keywords=['mercurial', 'git', 'diff', 'highlight'], author='Takeshi Komiya', author_email='i.tkomiya at gmail.com', url='http://blockdiag.com/', download_url='http://pypi.python.org/pypi/diff-highlight', license='Apache License 2.0', py_modules=['diff_highlight'], packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, entry_points=""" [console_scripts] diff-highlight = highlights.command:highlight_main """ )
# -*- coding: utf-8 -*- from setuptools import setup, find_packages classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Topic :: Software Development", "Topic :: Software Development :: Version Control", "Topic :: Text Processing :: Filters", ] setup( name='diff-highlight', version='0.1.0', description='pretty diff highlighter; emphasis changed words in diff', long_description=open("README.rst").read(), classifiers=classifiers, keywords=['mercurial', 'git', 'diff', 'highlight'], author='Takeshi Komiya', author_email='i.tkomiya at gmail.com', url='http://blockdiag.com/', download_url='http://pypi.python.org/pypi/diff-highlight', license='Apache License 2.0', py_modules=['diff_highlight'], packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, entry_points=""" [console_scripts] diff-highlight = highlights.command:highlight_main """ )
Add py24 and py25 to classifiers
Add py24 and py25 to classifiers
Python
apache-2.0
tk0miya/diff-highlight
# -*- coding: utf-8 -*- from setuptools import setup, find_packages classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Topic :: Software Development", "Topic :: Software Development :: Version Control", "Topic :: Text Processing :: Filters", ] setup( name='diff-highlight', version='0.1.0', description='pretty diff highlighter; emphasis changed words in diff', long_description=open("README.rst").read(), classifiers=classifiers, keywords=['mercurial', 'git', 'diff', 'highlight'], author='Takeshi Komiya', author_email='i.tkomiya at gmail.com', url='http://blockdiag.com/', download_url='http://pypi.python.org/pypi/diff-highlight', license='Apache License 2.0', py_modules=['diff_highlight'], packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, entry_points=""" [console_scripts] diff-highlight = highlights.command:highlight_main """ ) Add py24 and py25 to classifiers
# -*- coding: utf-8 -*- from setuptools import setup, find_packages classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Topic :: Software Development", "Topic :: Software Development :: Version Control", "Topic :: Text Processing :: Filters", ] setup( name='diff-highlight', version='0.1.0', description='pretty diff highlighter; emphasis changed words in diff', long_description=open("README.rst").read(), classifiers=classifiers, keywords=['mercurial', 'git', 'diff', 'highlight'], author='Takeshi Komiya', author_email='i.tkomiya at gmail.com', url='http://blockdiag.com/', download_url='http://pypi.python.org/pypi/diff-highlight', license='Apache License 2.0', py_modules=['diff_highlight'], packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, entry_points=""" [console_scripts] diff-highlight = highlights.command:highlight_main """ )
<commit_before># -*- coding: utf-8 -*- from setuptools import setup, find_packages classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Topic :: Software Development", "Topic :: Software Development :: Version Control", "Topic :: Text Processing :: Filters", ] setup( name='diff-highlight', version='0.1.0', description='pretty diff highlighter; emphasis changed words in diff', long_description=open("README.rst").read(), classifiers=classifiers, keywords=['mercurial', 'git', 'diff', 'highlight'], author='Takeshi Komiya', author_email='i.tkomiya at gmail.com', url='http://blockdiag.com/', download_url='http://pypi.python.org/pypi/diff-highlight', license='Apache License 2.0', py_modules=['diff_highlight'], packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, entry_points=""" [console_scripts] diff-highlight = highlights.command:highlight_main """ ) <commit_msg>Add py24 and py25 to classifiers<commit_after>
# -*- coding: utf-8 -*- from setuptools import setup, find_packages classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Topic :: Software Development", "Topic :: Software Development :: Version Control", "Topic :: Text Processing :: Filters", ] setup( name='diff-highlight', version='0.1.0', description='pretty diff highlighter; emphasis changed words in diff', long_description=open("README.rst").read(), classifiers=classifiers, keywords=['mercurial', 'git', 'diff', 'highlight'], author='Takeshi Komiya', author_email='i.tkomiya at gmail.com', url='http://blockdiag.com/', download_url='http://pypi.python.org/pypi/diff-highlight', license='Apache License 2.0', py_modules=['diff_highlight'], packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, entry_points=""" [console_scripts] diff-highlight = highlights.command:highlight_main """ )
# -*- coding: utf-8 -*- from setuptools import setup, find_packages classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Topic :: Software Development", "Topic :: Software Development :: Version Control", "Topic :: Text Processing :: Filters", ] setup( name='diff-highlight', version='0.1.0', description='pretty diff highlighter; emphasis changed words in diff', long_description=open("README.rst").read(), classifiers=classifiers, keywords=['mercurial', 'git', 'diff', 'highlight'], author='Takeshi Komiya', author_email='i.tkomiya at gmail.com', url='http://blockdiag.com/', download_url='http://pypi.python.org/pypi/diff-highlight', license='Apache License 2.0', py_modules=['diff_highlight'], packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, entry_points=""" [console_scripts] diff-highlight = highlights.command:highlight_main """ ) Add py24 and py25 to classifiers# -*- coding: utf-8 -*- from setuptools import setup, find_packages classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Topic :: Software Development", "Topic :: Software Development :: Version Control", "Topic :: Text Processing :: Filters", ] setup( name='diff-highlight', version='0.1.0', description='pretty diff highlighter; emphasis changed words in diff', long_description=open("README.rst").read(), classifiers=classifiers, keywords=['mercurial', 'git', 'diff', 'highlight'], author='Takeshi Komiya', author_email='i.tkomiya at gmail.com', url='http://blockdiag.com/', download_url='http://pypi.python.org/pypi/diff-highlight', license='Apache License 2.0', py_modules=['diff_highlight'], packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, entry_points=""" [console_scripts] diff-highlight = highlights.command:highlight_main """ )
<commit_before># -*- coding: utf-8 -*- from setuptools import setup, find_packages classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Topic :: Software Development", "Topic :: Software Development :: Version Control", "Topic :: Text Processing :: Filters", ] setup( name='diff-highlight', version='0.1.0', description='pretty diff highlighter; emphasis changed words in diff', long_description=open("README.rst").read(), classifiers=classifiers, keywords=['mercurial', 'git', 'diff', 'highlight'], author='Takeshi Komiya', author_email='i.tkomiya at gmail.com', url='http://blockdiag.com/', download_url='http://pypi.python.org/pypi/diff-highlight', license='Apache License 2.0', py_modules=['diff_highlight'], packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, entry_points=""" [console_scripts] diff-highlight = highlights.command:highlight_main """ ) <commit_msg>Add py24 and py25 to classifiers<commit_after># -*- coding: utf-8 -*- from setuptools import setup, find_packages classifiers = [ "Development Status :: 4 - Beta", "Intended Audience :: System Administrators", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 2.4", "Programming Language :: Python :: 2.5", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Topic :: Software Development", "Topic :: Software Development :: Version Control", "Topic :: Text Processing :: Filters", ] setup( name='diff-highlight', version='0.1.0', description='pretty diff highlighter; emphasis changed words in diff', long_description=open("README.rst").read(), classifiers=classifiers, keywords=['mercurial', 'git', 'diff', 'highlight'], author='Takeshi Komiya', author_email='i.tkomiya at gmail.com', url='http://blockdiag.com/', download_url='http://pypi.python.org/pypi/diff-highlight', license='Apache License 2.0', py_modules=['diff_highlight'], packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, entry_points=""" [console_scripts] diff-highlight = highlights.command:highlight_main """ )
20c0466c951864da74b250c0eeb907ad1dc209b9
setup.py
setup.py
import subprocess import sys from setuptools import Command, setup class RunTests(Command): user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): errno = subprocess.call([sys.executable, '-m', 'unittest', 'parserutils.tests.tests']) raise SystemExit(errno) with open('README.md') as readme: long_description = readme.read() setup( name='parserutils', description='A collection of performant parsing utilities', long_description=long_description, long_description_content_type='text/markdown', keywords='parser,parsing,utils,utilities,collections,dates,elements,numbers,strings,url,xml', version='1.2.1', packages=[ 'parserutils', 'parserutils.tests' ], install_requires=[ 'defusedxml>=0.4.1', 'python-dateutil>=2.4.2', 'six>=1.9.0' ], tests_require=['mock'], url='https://github.com/consbio/parserutils', license='BSD', cmdclass={'test': RunTests} )
import subprocess import sys from setuptools import Command, setup class RunTests(Command): user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): errno = subprocess.call([sys.executable, '-m', 'unittest', 'parserutils.tests.tests']) raise SystemExit(errno) with open('README.md') as readme: long_description = readme.read() setup( name='parserutils', description='A collection of performant parsing utilities', long_description=long_description, long_description_content_type='text/markdown', keywords='parser,parsing,utils,utilities,collections,dates,elements,numbers,strings,url,xml', version='1.2.2', packages=[ 'parserutils', 'parserutils.tests' ], install_requires=[ 'defusedxml>=0.4.1', 'python-dateutil>=2.4.2', 'six>=1.9.0' ], tests_require=['mock'], url='https://github.com/consbio/parserutils', license='BSD', cmdclass={'test': RunTests} )
Increment minor version once again
Increment minor version once again
Python
bsd-3-clause
consbio/parserutils
import subprocess import sys from setuptools import Command, setup class RunTests(Command): user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): errno = subprocess.call([sys.executable, '-m', 'unittest', 'parserutils.tests.tests']) raise SystemExit(errno) with open('README.md') as readme: long_description = readme.read() setup( name='parserutils', description='A collection of performant parsing utilities', long_description=long_description, long_description_content_type='text/markdown', keywords='parser,parsing,utils,utilities,collections,dates,elements,numbers,strings,url,xml', version='1.2.1', packages=[ 'parserutils', 'parserutils.tests' ], install_requires=[ 'defusedxml>=0.4.1', 'python-dateutil>=2.4.2', 'six>=1.9.0' ], tests_require=['mock'], url='https://github.com/consbio/parserutils', license='BSD', cmdclass={'test': RunTests} ) Increment minor version once again
import subprocess import sys from setuptools import Command, setup class RunTests(Command): user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): errno = subprocess.call([sys.executable, '-m', 'unittest', 'parserutils.tests.tests']) raise SystemExit(errno) with open('README.md') as readme: long_description = readme.read() setup( name='parserutils', description='A collection of performant parsing utilities', long_description=long_description, long_description_content_type='text/markdown', keywords='parser,parsing,utils,utilities,collections,dates,elements,numbers,strings,url,xml', version='1.2.2', packages=[ 'parserutils', 'parserutils.tests' ], install_requires=[ 'defusedxml>=0.4.1', 'python-dateutil>=2.4.2', 'six>=1.9.0' ], tests_require=['mock'], url='https://github.com/consbio/parserutils', license='BSD', cmdclass={'test': RunTests} )
<commit_before>import subprocess import sys from setuptools import Command, setup class RunTests(Command): user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): errno = subprocess.call([sys.executable, '-m', 'unittest', 'parserutils.tests.tests']) raise SystemExit(errno) with open('README.md') as readme: long_description = readme.read() setup( name='parserutils', description='A collection of performant parsing utilities', long_description=long_description, long_description_content_type='text/markdown', keywords='parser,parsing,utils,utilities,collections,dates,elements,numbers,strings,url,xml', version='1.2.1', packages=[ 'parserutils', 'parserutils.tests' ], install_requires=[ 'defusedxml>=0.4.1', 'python-dateutil>=2.4.2', 'six>=1.9.0' ], tests_require=['mock'], url='https://github.com/consbio/parserutils', license='BSD', cmdclass={'test': RunTests} ) <commit_msg>Increment minor version once again<commit_after>
import subprocess import sys from setuptools import Command, setup class RunTests(Command): user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): errno = subprocess.call([sys.executable, '-m', 'unittest', 'parserutils.tests.tests']) raise SystemExit(errno) with open('README.md') as readme: long_description = readme.read() setup( name='parserutils', description='A collection of performant parsing utilities', long_description=long_description, long_description_content_type='text/markdown', keywords='parser,parsing,utils,utilities,collections,dates,elements,numbers,strings,url,xml', version='1.2.2', packages=[ 'parserutils', 'parserutils.tests' ], install_requires=[ 'defusedxml>=0.4.1', 'python-dateutil>=2.4.2', 'six>=1.9.0' ], tests_require=['mock'], url='https://github.com/consbio/parserutils', license='BSD', cmdclass={'test': RunTests} )
import subprocess import sys from setuptools import Command, setup class RunTests(Command): user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): errno = subprocess.call([sys.executable, '-m', 'unittest', 'parserutils.tests.tests']) raise SystemExit(errno) with open('README.md') as readme: long_description = readme.read() setup( name='parserutils', description='A collection of performant parsing utilities', long_description=long_description, long_description_content_type='text/markdown', keywords='parser,parsing,utils,utilities,collections,dates,elements,numbers,strings,url,xml', version='1.2.1', packages=[ 'parserutils', 'parserutils.tests' ], install_requires=[ 'defusedxml>=0.4.1', 'python-dateutil>=2.4.2', 'six>=1.9.0' ], tests_require=['mock'], url='https://github.com/consbio/parserutils', license='BSD', cmdclass={'test': RunTests} ) Increment minor version once againimport subprocess import sys from setuptools import Command, setup class RunTests(Command): user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): errno = subprocess.call([sys.executable, '-m', 'unittest', 'parserutils.tests.tests']) raise SystemExit(errno) with open('README.md') as readme: long_description = readme.read() setup( name='parserutils', description='A collection of performant parsing utilities', long_description=long_description, long_description_content_type='text/markdown', keywords='parser,parsing,utils,utilities,collections,dates,elements,numbers,strings,url,xml', version='1.2.2', packages=[ 'parserutils', 'parserutils.tests' ], install_requires=[ 'defusedxml>=0.4.1', 'python-dateutil>=2.4.2', 'six>=1.9.0' ], tests_require=['mock'], url='https://github.com/consbio/parserutils', license='BSD', cmdclass={'test': RunTests} )
<commit_before>import subprocess import sys from setuptools import Command, setup class RunTests(Command): user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): errno = subprocess.call([sys.executable, '-m', 'unittest', 'parserutils.tests.tests']) raise SystemExit(errno) with open('README.md') as readme: long_description = readme.read() setup( name='parserutils', description='A collection of performant parsing utilities', long_description=long_description, long_description_content_type='text/markdown', keywords='parser,parsing,utils,utilities,collections,dates,elements,numbers,strings,url,xml', version='1.2.1', packages=[ 'parserutils', 'parserutils.tests' ], install_requires=[ 'defusedxml>=0.4.1', 'python-dateutil>=2.4.2', 'six>=1.9.0' ], tests_require=['mock'], url='https://github.com/consbio/parserutils', license='BSD', cmdclass={'test': RunTests} ) <commit_msg>Increment minor version once again<commit_after>import subprocess import sys from setuptools import Command, setup class RunTests(Command): user_options = [] def initialize_options(self): pass def finalize_options(self): pass def run(self): errno = subprocess.call([sys.executable, '-m', 'unittest', 'parserutils.tests.tests']) raise SystemExit(errno) with open('README.md') as readme: long_description = readme.read() setup( name='parserutils', description='A collection of performant parsing utilities', long_description=long_description, long_description_content_type='text/markdown', keywords='parser,parsing,utils,utilities,collections,dates,elements,numbers,strings,url,xml', version='1.2.2', packages=[ 'parserutils', 'parserutils.tests' ], install_requires=[ 'defusedxml>=0.4.1', 'python-dateutil>=2.4.2', 'six>=1.9.0' ], tests_require=['mock'], url='https://github.com/consbio/parserutils', license='BSD', cmdclass={'test': RunTests} )
ea9847a31eb4441415f1c11ddf57056f206fc375
reverse-engineering/reveng.py
reverse-engineering/reveng.py
from os.path import dirname from puresnmp.pdu import PDU from puresnmp.test import readbytes_multiple from puresnmp.x690.types import pop_tlv HERE = dirname(__file__) for row in readbytes_multiple('authpriv.hex', HERE): print(row) pdu, _ = pop_tlv(row) print(pdu.pretty())
from os.path import dirname HERE = dirname(__file__) from puresnmp.pdu import PDU from puresnmp.test import readbytes_multiple from puresnmp.x690.types import pop_tlv for row in readbytes_multiple("authpriv.hex", HERE): print(row) pdu, _ = pop_tlv(row) print(pdu.pretty())
Add some files for reverse-engineering
Add some files for reverse-engineering
Python
mit
exhuma/puresnmp,exhuma/puresnmp
from os.path import dirname from puresnmp.pdu import PDU from puresnmp.test import readbytes_multiple from puresnmp.x690.types import pop_tlv HERE = dirname(__file__) for row in readbytes_multiple('authpriv.hex', HERE): print(row) pdu, _ = pop_tlv(row) print(pdu.pretty()) Add some files for reverse-engineering
from os.path import dirname HERE = dirname(__file__) from puresnmp.pdu import PDU from puresnmp.test import readbytes_multiple from puresnmp.x690.types import pop_tlv for row in readbytes_multiple("authpriv.hex", HERE): print(row) pdu, _ = pop_tlv(row) print(pdu.pretty())
<commit_before>from os.path import dirname from puresnmp.pdu import PDU from puresnmp.test import readbytes_multiple from puresnmp.x690.types import pop_tlv HERE = dirname(__file__) for row in readbytes_multiple('authpriv.hex', HERE): print(row) pdu, _ = pop_tlv(row) print(pdu.pretty()) <commit_msg>Add some files for reverse-engineering<commit_after>
from os.path import dirname HERE = dirname(__file__) from puresnmp.pdu import PDU from puresnmp.test import readbytes_multiple from puresnmp.x690.types import pop_tlv for row in readbytes_multiple("authpriv.hex", HERE): print(row) pdu, _ = pop_tlv(row) print(pdu.pretty())
from os.path import dirname from puresnmp.pdu import PDU from puresnmp.test import readbytes_multiple from puresnmp.x690.types import pop_tlv HERE = dirname(__file__) for row in readbytes_multiple('authpriv.hex', HERE): print(row) pdu, _ = pop_tlv(row) print(pdu.pretty()) Add some files for reverse-engineeringfrom os.path import dirname HERE = dirname(__file__) from puresnmp.pdu import PDU from puresnmp.test import readbytes_multiple from puresnmp.x690.types import pop_tlv for row in readbytes_multiple("authpriv.hex", HERE): print(row) pdu, _ = pop_tlv(row) print(pdu.pretty())
<commit_before>from os.path import dirname from puresnmp.pdu import PDU from puresnmp.test import readbytes_multiple from puresnmp.x690.types import pop_tlv HERE = dirname(__file__) for row in readbytes_multiple('authpriv.hex', HERE): print(row) pdu, _ = pop_tlv(row) print(pdu.pretty()) <commit_msg>Add some files for reverse-engineering<commit_after>from os.path import dirname HERE = dirname(__file__) from puresnmp.pdu import PDU from puresnmp.test import readbytes_multiple from puresnmp.x690.types import pop_tlv for row in readbytes_multiple("authpriv.hex", HERE): print(row) pdu, _ = pop_tlv(row) print(pdu.pretty())
e893a860f4a8ad9682f400507948ee20fce1c328
healthcheck/contrib/django/status_endpoint/views.py
healthcheck/contrib/django/status_endpoint/views.py
import json from django.conf import settings from django.views.decorators.http import require_http_methods from django.http import HttpResponse, HttpResponseServerError from healthcheck.healthcheck import ( DjangoDBsHealthCheck, FilesDontExistHealthCheck, HealthChecker) @require_http_methods(['GET']) def status(request): checks = [] if getattr(settings, 'STATUS_CHECK_DBS', True): checks.append(DjangoDBsHealthCheck()) files_to_check = getattr( settings, 'STATUS_CHECK_FILES') if files_to_check: checks.append( FilesDontExistHealthCheck( files_to_check, check_id="quiesce file doesn't exist")) ok, details = HealthChecker(checks)() if not ok: return HttpResponseServerError((json.dumps(details))) return HttpResponse(json.dumps(details))
import json from django.conf import settings from django.views.decorators.http import require_http_methods from django.http import HttpResponse from healthcheck.healthcheck import ( DjangoDBsHealthCheck, FilesDontExistHealthCheck, HealthChecker) class JsonResponse(HttpResponse): def __init__(self, data, **kwargs): kwargs.setdefault('content_type', 'application/json') data = json.dumps(data) super(JsonResponse, self).__init__(content=data, **kwargs) class JsonResponseServerError(JsonResponse): status_code = 500 @require_http_methods(['GET']) def status(request): checks = [] if getattr(settings, 'STATUS_CHECK_DBS', True): checks.append(DjangoDBsHealthCheck()) files_to_check = getattr(settings, 'STATUS_CHECK_FILES') if files_to_check: checks.append(FilesDontExistHealthCheck( files_to_check, check_id="quiesce file doesn't exist")) ok, details = HealthChecker(checks)() if not ok: return JsonResponseServerError(json.dumps(details)) return JsonResponse(details)
Fix content_type for JSON responses
Fix content_type for JSON responses
Python
mit
yola/healthcheck
import json from django.conf import settings from django.views.decorators.http import require_http_methods from django.http import HttpResponse, HttpResponseServerError from healthcheck.healthcheck import ( DjangoDBsHealthCheck, FilesDontExistHealthCheck, HealthChecker) @require_http_methods(['GET']) def status(request): checks = [] if getattr(settings, 'STATUS_CHECK_DBS', True): checks.append(DjangoDBsHealthCheck()) files_to_check = getattr( settings, 'STATUS_CHECK_FILES') if files_to_check: checks.append( FilesDontExistHealthCheck( files_to_check, check_id="quiesce file doesn't exist")) ok, details = HealthChecker(checks)() if not ok: return HttpResponseServerError((json.dumps(details))) return HttpResponse(json.dumps(details)) Fix content_type for JSON responses
import json from django.conf import settings from django.views.decorators.http import require_http_methods from django.http import HttpResponse from healthcheck.healthcheck import ( DjangoDBsHealthCheck, FilesDontExistHealthCheck, HealthChecker) class JsonResponse(HttpResponse): def __init__(self, data, **kwargs): kwargs.setdefault('content_type', 'application/json') data = json.dumps(data) super(JsonResponse, self).__init__(content=data, **kwargs) class JsonResponseServerError(JsonResponse): status_code = 500 @require_http_methods(['GET']) def status(request): checks = [] if getattr(settings, 'STATUS_CHECK_DBS', True): checks.append(DjangoDBsHealthCheck()) files_to_check = getattr(settings, 'STATUS_CHECK_FILES') if files_to_check: checks.append(FilesDontExistHealthCheck( files_to_check, check_id="quiesce file doesn't exist")) ok, details = HealthChecker(checks)() if not ok: return JsonResponseServerError(json.dumps(details)) return JsonResponse(details)
<commit_before>import json from django.conf import settings from django.views.decorators.http import require_http_methods from django.http import HttpResponse, HttpResponseServerError from healthcheck.healthcheck import ( DjangoDBsHealthCheck, FilesDontExistHealthCheck, HealthChecker) @require_http_methods(['GET']) def status(request): checks = [] if getattr(settings, 'STATUS_CHECK_DBS', True): checks.append(DjangoDBsHealthCheck()) files_to_check = getattr( settings, 'STATUS_CHECK_FILES') if files_to_check: checks.append( FilesDontExistHealthCheck( files_to_check, check_id="quiesce file doesn't exist")) ok, details = HealthChecker(checks)() if not ok: return HttpResponseServerError((json.dumps(details))) return HttpResponse(json.dumps(details)) <commit_msg>Fix content_type for JSON responses<commit_after>
import json from django.conf import settings from django.views.decorators.http import require_http_methods from django.http import HttpResponse from healthcheck.healthcheck import ( DjangoDBsHealthCheck, FilesDontExistHealthCheck, HealthChecker) class JsonResponse(HttpResponse): def __init__(self, data, **kwargs): kwargs.setdefault('content_type', 'application/json') data = json.dumps(data) super(JsonResponse, self).__init__(content=data, **kwargs) class JsonResponseServerError(JsonResponse): status_code = 500 @require_http_methods(['GET']) def status(request): checks = [] if getattr(settings, 'STATUS_CHECK_DBS', True): checks.append(DjangoDBsHealthCheck()) files_to_check = getattr(settings, 'STATUS_CHECK_FILES') if files_to_check: checks.append(FilesDontExistHealthCheck( files_to_check, check_id="quiesce file doesn't exist")) ok, details = HealthChecker(checks)() if not ok: return JsonResponseServerError(json.dumps(details)) return JsonResponse(details)
import json from django.conf import settings from django.views.decorators.http import require_http_methods from django.http import HttpResponse, HttpResponseServerError from healthcheck.healthcheck import ( DjangoDBsHealthCheck, FilesDontExistHealthCheck, HealthChecker) @require_http_methods(['GET']) def status(request): checks = [] if getattr(settings, 'STATUS_CHECK_DBS', True): checks.append(DjangoDBsHealthCheck()) files_to_check = getattr( settings, 'STATUS_CHECK_FILES') if files_to_check: checks.append( FilesDontExistHealthCheck( files_to_check, check_id="quiesce file doesn't exist")) ok, details = HealthChecker(checks)() if not ok: return HttpResponseServerError((json.dumps(details))) return HttpResponse(json.dumps(details)) Fix content_type for JSON responsesimport json from django.conf import settings from django.views.decorators.http import require_http_methods from django.http import HttpResponse from healthcheck.healthcheck import ( DjangoDBsHealthCheck, FilesDontExistHealthCheck, HealthChecker) class JsonResponse(HttpResponse): def __init__(self, data, **kwargs): kwargs.setdefault('content_type', 'application/json') data = json.dumps(data) super(JsonResponse, self).__init__(content=data, **kwargs) class JsonResponseServerError(JsonResponse): status_code = 500 @require_http_methods(['GET']) def status(request): checks = [] if getattr(settings, 'STATUS_CHECK_DBS', True): checks.append(DjangoDBsHealthCheck()) files_to_check = getattr(settings, 'STATUS_CHECK_FILES') if files_to_check: checks.append(FilesDontExistHealthCheck( files_to_check, check_id="quiesce file doesn't exist")) ok, details = HealthChecker(checks)() if not ok: return JsonResponseServerError(json.dumps(details)) return JsonResponse(details)
<commit_before>import json from django.conf import settings from django.views.decorators.http import require_http_methods from django.http import HttpResponse, HttpResponseServerError from healthcheck.healthcheck import ( DjangoDBsHealthCheck, FilesDontExistHealthCheck, HealthChecker) @require_http_methods(['GET']) def status(request): checks = [] if getattr(settings, 'STATUS_CHECK_DBS', True): checks.append(DjangoDBsHealthCheck()) files_to_check = getattr( settings, 'STATUS_CHECK_FILES') if files_to_check: checks.append( FilesDontExistHealthCheck( files_to_check, check_id="quiesce file doesn't exist")) ok, details = HealthChecker(checks)() if not ok: return HttpResponseServerError((json.dumps(details))) return HttpResponse(json.dumps(details)) <commit_msg>Fix content_type for JSON responses<commit_after>import json from django.conf import settings from django.views.decorators.http import require_http_methods from django.http import HttpResponse from healthcheck.healthcheck import ( DjangoDBsHealthCheck, FilesDontExistHealthCheck, HealthChecker) class JsonResponse(HttpResponse): def __init__(self, data, **kwargs): kwargs.setdefault('content_type', 'application/json') data = json.dumps(data) super(JsonResponse, self).__init__(content=data, **kwargs) class JsonResponseServerError(JsonResponse): status_code = 500 @require_http_methods(['GET']) def status(request): checks = [] if getattr(settings, 'STATUS_CHECK_DBS', True): checks.append(DjangoDBsHealthCheck()) files_to_check = getattr(settings, 'STATUS_CHECK_FILES') if files_to_check: checks.append(FilesDontExistHealthCheck( files_to_check, check_id="quiesce file doesn't exist")) ok, details = HealthChecker(checks)() if not ok: return JsonResponseServerError(json.dumps(details)) return JsonResponse(details)
14a59d3b1d440d20c1df9b6e43c45657d8a60774
setup.py
setup.py
#!/usr/bin/env python try: from setuptools import setup, find_packages from setuptools.command.test import test except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test import os here = os.path.dirname(os.path.abspath(__file__)) f = open(os.path.join(here, 'README.rst')) long_description = f.read().strip() f.close() setup( name='django-common-helpers', version='0.1', author='Sumit Chachra', author_email='chachra@tivix.com', url='http://github.com/tivix/django-common', description = 'Common things every Django app needs!', packages=find_packages(), long_description=long_description, keywords = 'django', zip_safe=False, install_requires=[ 'Django>=1.2.3', 'South>=0.7.2' ], # test_suite = 'django_common.tests', include_package_data=True, # cmdclass={}, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
#!/usr/bin/env python try: from setuptools import setup, find_packages from setuptools.command.test import test except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test import os here = os.path.dirname(os.path.abspath(__file__)) f = open(os.path.join(here, 'README.rst')) long_description = f.read().strip() f.close() setup( name='django-common-tivix', version='0.1.2', author='Matthew Farver', author_email='mfarver@tivix.com', url='http://github.com/tivix/django-common', description = 'Common things every Django app needs!', packages=find_packages(), long_description=long_description, keywords = 'django', zip_safe=False, install_requires=[ 'Django>=1.2.3', 'South>=0.7.2' ], # test_suite = 'django_common.tests', include_package_data=True, # cmdclass={}, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
Change it to upload to pypi
Change it to upload to pypi
Python
mit
Tivix/django-common,Tivix/django-common,WikiRealtyInc/django-common,WikiRealtyInc/django-common,Tivix/django-common,WikiRealtyInc/django-common
#!/usr/bin/env python try: from setuptools import setup, find_packages from setuptools.command.test import test except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test import os here = os.path.dirname(os.path.abspath(__file__)) f = open(os.path.join(here, 'README.rst')) long_description = f.read().strip() f.close() setup( name='django-common-helpers', version='0.1', author='Sumit Chachra', author_email='chachra@tivix.com', url='http://github.com/tivix/django-common', description = 'Common things every Django app needs!', packages=find_packages(), long_description=long_description, keywords = 'django', zip_safe=False, install_requires=[ 'Django>=1.2.3', 'South>=0.7.2' ], # test_suite = 'django_common.tests', include_package_data=True, # cmdclass={}, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], ) Change it to upload to pypi
#!/usr/bin/env python try: from setuptools import setup, find_packages from setuptools.command.test import test except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test import os here = os.path.dirname(os.path.abspath(__file__)) f = open(os.path.join(here, 'README.rst')) long_description = f.read().strip() f.close() setup( name='django-common-tivix', version='0.1.2', author='Matthew Farver', author_email='mfarver@tivix.com', url='http://github.com/tivix/django-common', description = 'Common things every Django app needs!', packages=find_packages(), long_description=long_description, keywords = 'django', zip_safe=False, install_requires=[ 'Django>=1.2.3', 'South>=0.7.2' ], # test_suite = 'django_common.tests', include_package_data=True, # cmdclass={}, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
<commit_before>#!/usr/bin/env python try: from setuptools import setup, find_packages from setuptools.command.test import test except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test import os here = os.path.dirname(os.path.abspath(__file__)) f = open(os.path.join(here, 'README.rst')) long_description = f.read().strip() f.close() setup( name='django-common-helpers', version='0.1', author='Sumit Chachra', author_email='chachra@tivix.com', url='http://github.com/tivix/django-common', description = 'Common things every Django app needs!', packages=find_packages(), long_description=long_description, keywords = 'django', zip_safe=False, install_requires=[ 'Django>=1.2.3', 'South>=0.7.2' ], # test_suite = 'django_common.tests', include_package_data=True, # cmdclass={}, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], ) <commit_msg>Change it to upload to pypi<commit_after>
#!/usr/bin/env python try: from setuptools import setup, find_packages from setuptools.command.test import test except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test import os here = os.path.dirname(os.path.abspath(__file__)) f = open(os.path.join(here, 'README.rst')) long_description = f.read().strip() f.close() setup( name='django-common-tivix', version='0.1.2', author='Matthew Farver', author_email='mfarver@tivix.com', url='http://github.com/tivix/django-common', description = 'Common things every Django app needs!', packages=find_packages(), long_description=long_description, keywords = 'django', zip_safe=False, install_requires=[ 'Django>=1.2.3', 'South>=0.7.2' ], # test_suite = 'django_common.tests', include_package_data=True, # cmdclass={}, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
#!/usr/bin/env python try: from setuptools import setup, find_packages from setuptools.command.test import test except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test import os here = os.path.dirname(os.path.abspath(__file__)) f = open(os.path.join(here, 'README.rst')) long_description = f.read().strip() f.close() setup( name='django-common-helpers', version='0.1', author='Sumit Chachra', author_email='chachra@tivix.com', url='http://github.com/tivix/django-common', description = 'Common things every Django app needs!', packages=find_packages(), long_description=long_description, keywords = 'django', zip_safe=False, install_requires=[ 'Django>=1.2.3', 'South>=0.7.2' ], # test_suite = 'django_common.tests', include_package_data=True, # cmdclass={}, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], ) Change it to upload to pypi#!/usr/bin/env python try: from setuptools import setup, find_packages from setuptools.command.test import test except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test import os here = os.path.dirname(os.path.abspath(__file__)) f = open(os.path.join(here, 'README.rst')) long_description = f.read().strip() f.close() setup( name='django-common-tivix', version='0.1.2', author='Matthew Farver', author_email='mfarver@tivix.com', url='http://github.com/tivix/django-common', description = 'Common things every Django app needs!', packages=find_packages(), long_description=long_description, keywords = 'django', zip_safe=False, install_requires=[ 'Django>=1.2.3', 'South>=0.7.2' ], # test_suite = 'django_common.tests', include_package_data=True, # cmdclass={}, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
<commit_before>#!/usr/bin/env python try: from setuptools import setup, find_packages from setuptools.command.test import test except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test import os here = os.path.dirname(os.path.abspath(__file__)) f = open(os.path.join(here, 'README.rst')) long_description = f.read().strip() f.close() setup( name='django-common-helpers', version='0.1', author='Sumit Chachra', author_email='chachra@tivix.com', url='http://github.com/tivix/django-common', description = 'Common things every Django app needs!', packages=find_packages(), long_description=long_description, keywords = 'django', zip_safe=False, install_requires=[ 'Django>=1.2.3', 'South>=0.7.2' ], # test_suite = 'django_common.tests', include_package_data=True, # cmdclass={}, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], ) <commit_msg>Change it to upload to pypi<commit_after>#!/usr/bin/env python try: from setuptools import setup, find_packages from setuptools.command.test import test except ImportError: from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test import os here = os.path.dirname(os.path.abspath(__file__)) f = open(os.path.join(here, 'README.rst')) long_description = f.read().strip() f.close() setup( name='django-common-tivix', version='0.1.2', author='Matthew Farver', author_email='mfarver@tivix.com', url='http://github.com/tivix/django-common', description = 'Common things every Django app needs!', packages=find_packages(), long_description=long_description, keywords = 'django', zip_safe=False, install_requires=[ 'Django>=1.2.3', 'South>=0.7.2' ], # test_suite = 'django_common.tests', include_package_data=True, # cmdclass={}, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
cf49c86c6e0966061ab7756a9922c12430ade35f
setup.py
setup.py
from setuptools import find_packages, setup setup( version='0.4.0', name='incuna-groups', packages=find_packages(), include_package_data=True, install_requires=[ 'django_crispy_forms==1.4.0', 'django-polymorphic==0.6.1', ], description='Generic group/forum framework.', author='Incuna Ltd', author_email='admin@incuna.com', url='https://github.com/incuna/incuna-groups', )
from setuptools import find_packages, setup setup( version='0.4.0', name='incuna-groups', packages=find_packages(), include_package_data=True, install_requires=[ 'django_crispy_forms>=1.4.0,<2', 'django-polymorphic>=0.6.1,<1', ], description='Generic group/forum framework.', author='Incuna Ltd', author_email='admin@incuna.com', url='https://github.com/incuna/incuna-groups', )
Allow wider range of dependencies
Allow wider range of dependencies
Python
bsd-2-clause
incuna/incuna-groups,incuna/incuna-groups
from setuptools import find_packages, setup setup( version='0.4.0', name='incuna-groups', packages=find_packages(), include_package_data=True, install_requires=[ 'django_crispy_forms==1.4.0', 'django-polymorphic==0.6.1', ], description='Generic group/forum framework.', author='Incuna Ltd', author_email='admin@incuna.com', url='https://github.com/incuna/incuna-groups', ) Allow wider range of dependencies
from setuptools import find_packages, setup setup( version='0.4.0', name='incuna-groups', packages=find_packages(), include_package_data=True, install_requires=[ 'django_crispy_forms>=1.4.0,<2', 'django-polymorphic>=0.6.1,<1', ], description='Generic group/forum framework.', author='Incuna Ltd', author_email='admin@incuna.com', url='https://github.com/incuna/incuna-groups', )
<commit_before>from setuptools import find_packages, setup setup( version='0.4.0', name='incuna-groups', packages=find_packages(), include_package_data=True, install_requires=[ 'django_crispy_forms==1.4.0', 'django-polymorphic==0.6.1', ], description='Generic group/forum framework.', author='Incuna Ltd', author_email='admin@incuna.com', url='https://github.com/incuna/incuna-groups', ) <commit_msg>Allow wider range of dependencies<commit_after>
from setuptools import find_packages, setup setup( version='0.4.0', name='incuna-groups', packages=find_packages(), include_package_data=True, install_requires=[ 'django_crispy_forms>=1.4.0,<2', 'django-polymorphic>=0.6.1,<1', ], description='Generic group/forum framework.', author='Incuna Ltd', author_email='admin@incuna.com', url='https://github.com/incuna/incuna-groups', )
from setuptools import find_packages, setup setup( version='0.4.0', name='incuna-groups', packages=find_packages(), include_package_data=True, install_requires=[ 'django_crispy_forms==1.4.0', 'django-polymorphic==0.6.1', ], description='Generic group/forum framework.', author='Incuna Ltd', author_email='admin@incuna.com', url='https://github.com/incuna/incuna-groups', ) Allow wider range of dependenciesfrom setuptools import find_packages, setup setup( version='0.4.0', name='incuna-groups', packages=find_packages(), include_package_data=True, install_requires=[ 'django_crispy_forms>=1.4.0,<2', 'django-polymorphic>=0.6.1,<1', ], description='Generic group/forum framework.', author='Incuna Ltd', author_email='admin@incuna.com', url='https://github.com/incuna/incuna-groups', )
<commit_before>from setuptools import find_packages, setup setup( version='0.4.0', name='incuna-groups', packages=find_packages(), include_package_data=True, install_requires=[ 'django_crispy_forms==1.4.0', 'django-polymorphic==0.6.1', ], description='Generic group/forum framework.', author='Incuna Ltd', author_email='admin@incuna.com', url='https://github.com/incuna/incuna-groups', ) <commit_msg>Allow wider range of dependencies<commit_after>from setuptools import find_packages, setup setup( version='0.4.0', name='incuna-groups', packages=find_packages(), include_package_data=True, install_requires=[ 'django_crispy_forms>=1.4.0,<2', 'django-polymorphic>=0.6.1,<1', ], description='Generic group/forum framework.', author='Incuna Ltd', author_email='admin@incuna.com', url='https://github.com/incuna/incuna-groups', )
710dbd42142f4f3ea0f02b1924f518b3cb38b79f
setup.py
setup.py
"""Setuptools file for a MultiMarkdown Python wrapper.""" from codecs import open from os import path from distutils.core import setup from setuptools import find_packages here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'README.md'), encoding='utf-8') as f: long_description = f.read() setup( name='scriptorium', version='2.0.0', description='Multimarkdown and LaTeX framework for academic papers.', long_description=long_description, license='MIT', author='Jason Ziglar', author_email='jasedit@gmail.com', url="https://github.com/jasedit/scriptorium", classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Topic :: Text Processing :: Markup', 'Topic :: Text Processing :: Filters', 'Topic :: Software Development :: Libraries :: Python Modules', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3' ], packages=find_packages(), entry_points = { 'console_scripts': ['scriptorium = scriptorium:main'], }, package_data={'scriptorium': ['data/gitignore']} )
"""Setuptools file for a MultiMarkdown Python wrapper.""" from codecs import open from os import path from distutils.core import setup from setuptools import find_packages import pypandoc here = path.abspath(path.dirname(__file__)) long_description = pypandoc.convert_file('README.md', 'rst') setup( name='scriptorium', version='2.0.1', description='Multimarkdown and LaTeX framework for academic papers.', long_description=long_description, license='MIT', author='Jason Ziglar', author_email='jasedit@gmail.com', url="https://github.com/jasedit/scriptorium", classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Topic :: Text Processing :: Markup', 'Topic :: Text Processing :: Filters', 'Topic :: Software Development :: Libraries :: Python Modules', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3' ], packages=find_packages(), entry_points = { 'console_scripts': ['scriptorium = scriptorium:main'], }, package_data={'scriptorium': ['data/gitignore']} )
Convert README to Restructured Text for distribution.
Convert README to Restructured Text for distribution.
Python
mit
jasedit/scriptorium,jasedit/papers_base
"""Setuptools file for a MultiMarkdown Python wrapper.""" from codecs import open from os import path from distutils.core import setup from setuptools import find_packages here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'README.md'), encoding='utf-8') as f: long_description = f.read() setup( name='scriptorium', version='2.0.0', description='Multimarkdown and LaTeX framework for academic papers.', long_description=long_description, license='MIT', author='Jason Ziglar', author_email='jasedit@gmail.com', url="https://github.com/jasedit/scriptorium", classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Topic :: Text Processing :: Markup', 'Topic :: Text Processing :: Filters', 'Topic :: Software Development :: Libraries :: Python Modules', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3' ], packages=find_packages(), entry_points = { 'console_scripts': ['scriptorium = scriptorium:main'], }, package_data={'scriptorium': ['data/gitignore']} ) Convert README to Restructured Text for distribution.
"""Setuptools file for a MultiMarkdown Python wrapper.""" from codecs import open from os import path from distutils.core import setup from setuptools import find_packages import pypandoc here = path.abspath(path.dirname(__file__)) long_description = pypandoc.convert_file('README.md', 'rst') setup( name='scriptorium', version='2.0.1', description='Multimarkdown and LaTeX framework for academic papers.', long_description=long_description, license='MIT', author='Jason Ziglar', author_email='jasedit@gmail.com', url="https://github.com/jasedit/scriptorium", classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Topic :: Text Processing :: Markup', 'Topic :: Text Processing :: Filters', 'Topic :: Software Development :: Libraries :: Python Modules', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3' ], packages=find_packages(), entry_points = { 'console_scripts': ['scriptorium = scriptorium:main'], }, package_data={'scriptorium': ['data/gitignore']} )
<commit_before>"""Setuptools file for a MultiMarkdown Python wrapper.""" from codecs import open from os import path from distutils.core import setup from setuptools import find_packages here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'README.md'), encoding='utf-8') as f: long_description = f.read() setup( name='scriptorium', version='2.0.0', description='Multimarkdown and LaTeX framework for academic papers.', long_description=long_description, license='MIT', author='Jason Ziglar', author_email='jasedit@gmail.com', url="https://github.com/jasedit/scriptorium", classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Topic :: Text Processing :: Markup', 'Topic :: Text Processing :: Filters', 'Topic :: Software Development :: Libraries :: Python Modules', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3' ], packages=find_packages(), entry_points = { 'console_scripts': ['scriptorium = scriptorium:main'], }, package_data={'scriptorium': ['data/gitignore']} ) <commit_msg>Convert README to Restructured Text for distribution.<commit_after>
"""Setuptools file for a MultiMarkdown Python wrapper.""" from codecs import open from os import path from distutils.core import setup from setuptools import find_packages import pypandoc here = path.abspath(path.dirname(__file__)) long_description = pypandoc.convert_file('README.md', 'rst') setup( name='scriptorium', version='2.0.1', description='Multimarkdown and LaTeX framework for academic papers.', long_description=long_description, license='MIT', author='Jason Ziglar', author_email='jasedit@gmail.com', url="https://github.com/jasedit/scriptorium", classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Topic :: Text Processing :: Markup', 'Topic :: Text Processing :: Filters', 'Topic :: Software Development :: Libraries :: Python Modules', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3' ], packages=find_packages(), entry_points = { 'console_scripts': ['scriptorium = scriptorium:main'], }, package_data={'scriptorium': ['data/gitignore']} )
"""Setuptools file for a MultiMarkdown Python wrapper.""" from codecs import open from os import path from distutils.core import setup from setuptools import find_packages here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'README.md'), encoding='utf-8') as f: long_description = f.read() setup( name='scriptorium', version='2.0.0', description='Multimarkdown and LaTeX framework for academic papers.', long_description=long_description, license='MIT', author='Jason Ziglar', author_email='jasedit@gmail.com', url="https://github.com/jasedit/scriptorium", classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Topic :: Text Processing :: Markup', 'Topic :: Text Processing :: Filters', 'Topic :: Software Development :: Libraries :: Python Modules', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3' ], packages=find_packages(), entry_points = { 'console_scripts': ['scriptorium = scriptorium:main'], }, package_data={'scriptorium': ['data/gitignore']} ) Convert README to Restructured Text for distribution."""Setuptools file for a MultiMarkdown Python wrapper.""" from codecs import open from os import path from distutils.core import setup from setuptools import find_packages import pypandoc here = path.abspath(path.dirname(__file__)) long_description = pypandoc.convert_file('README.md', 'rst') setup( name='scriptorium', version='2.0.1', description='Multimarkdown and LaTeX framework for academic papers.', long_description=long_description, license='MIT', author='Jason Ziglar', author_email='jasedit@gmail.com', url="https://github.com/jasedit/scriptorium", classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Topic :: Text Processing :: Markup', 'Topic :: Text Processing :: Filters', 'Topic :: Software Development :: Libraries :: Python Modules', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3' ], packages=find_packages(), entry_points = { 'console_scripts': ['scriptorium = scriptorium:main'], }, package_data={'scriptorium': ['data/gitignore']} )
<commit_before>"""Setuptools file for a MultiMarkdown Python wrapper.""" from codecs import open from os import path from distutils.core import setup from setuptools import find_packages here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'README.md'), encoding='utf-8') as f: long_description = f.read() setup( name='scriptorium', version='2.0.0', description='Multimarkdown and LaTeX framework for academic papers.', long_description=long_description, license='MIT', author='Jason Ziglar', author_email='jasedit@gmail.com', url="https://github.com/jasedit/scriptorium", classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Topic :: Text Processing :: Markup', 'Topic :: Text Processing :: Filters', 'Topic :: Software Development :: Libraries :: Python Modules', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3' ], packages=find_packages(), entry_points = { 'console_scripts': ['scriptorium = scriptorium:main'], }, package_data={'scriptorium': ['data/gitignore']} ) <commit_msg>Convert README to Restructured Text for distribution.<commit_after>"""Setuptools file for a MultiMarkdown Python wrapper.""" from codecs import open from os import path from distutils.core import setup from setuptools import find_packages import pypandoc here = path.abspath(path.dirname(__file__)) long_description = pypandoc.convert_file('README.md', 'rst') setup( name='scriptorium', version='2.0.1', description='Multimarkdown and LaTeX framework for academic papers.', long_description=long_description, license='MIT', author='Jason Ziglar', author_email='jasedit@gmail.com', url="https://github.com/jasedit/scriptorium", classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Topic :: Text Processing :: Markup', 'Topic :: Text Processing :: Filters', 'Topic :: Software Development :: Libraries :: Python Modules', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3' ], packages=find_packages(), entry_points = { 'console_scripts': ['scriptorium = scriptorium:main'], }, package_data={'scriptorium': ['data/gitignore']} )
9eb09bcece4a92f06c55078d2cd0e19e872c2794
setup.py
setup.py
import astral import os from setuptools import find_packages, setup ROOT_PATH = os.path.abspath(os.path.dirname(__file__)) long_description = open(os.path.join(ROOT_PATH, 'README.rst')).read() setup(name='astral', version=astral.__version__, description='Astral Streaming P2P Client', long_description=long_description, author='Astral Project Group', author_email='astral@bueda.com', url='http://github.com/peplin/astral', test_suite='nose.collector', setup_requires=['nose>=0.11',], test_requires=['unittest2>=0.5.1', 'mockito==0.5.1', 'python-faker==0.2.3', 'factory-boy==1.0.0',], install_requires=[ 'tornado>=1.2.1', 'importlib>=1.0.2', 'sqlalchemy>=0.6.6', 'Elixir>=0.7.1', 'restkit>=3.2.0', ], packages=find_packages(), entry_points={ 'console_scripts': [ 'astralnode = astral.bin.astralnode:main',], }, )
import astral import os from setuptools import find_packages, setup ROOT_PATH = os.path.abspath(os.path.dirname(__file__)) long_description = open(os.path.join(ROOT_PATH, 'README.rst')).read() setup(name='astral', version=astral.__version__, description='Astral Streaming P2P Client', long_description=long_description, author='Astral Project Group', author_email='astral@bueda.com', url='http://github.com/peplin/astral', test_suite='nose.collector', setup_requires=['nose>=0.11',], test_requires=['unittest2>=0.5.1', 'mockito==0.5.1', 'python-faker==0.2.3', 'factory-boy==1.0.0',], install_requires=[ 'tornado>=1.2.1', 'importlib>=1.0.2', 'sqlalchemy>=0.6.6', 'Elixir>=0.7.1', 'restkit>=3.2.0', ], packages=find_packages(), entry_points={ 'console_scripts': [ 'astralnode = astral.bin.astralnode:main', 'astralctl = astral.bin.astralctl:main', ], }, )
Add astralctl script to console scripts section of package config.
Add astralctl script to console scripts section of package config.
Python
mit
peplin/astral
import astral import os from setuptools import find_packages, setup ROOT_PATH = os.path.abspath(os.path.dirname(__file__)) long_description = open(os.path.join(ROOT_PATH, 'README.rst')).read() setup(name='astral', version=astral.__version__, description='Astral Streaming P2P Client', long_description=long_description, author='Astral Project Group', author_email='astral@bueda.com', url='http://github.com/peplin/astral', test_suite='nose.collector', setup_requires=['nose>=0.11',], test_requires=['unittest2>=0.5.1', 'mockito==0.5.1', 'python-faker==0.2.3', 'factory-boy==1.0.0',], install_requires=[ 'tornado>=1.2.1', 'importlib>=1.0.2', 'sqlalchemy>=0.6.6', 'Elixir>=0.7.1', 'restkit>=3.2.0', ], packages=find_packages(), entry_points={ 'console_scripts': [ 'astralnode = astral.bin.astralnode:main',], }, ) Add astralctl script to console scripts section of package config.
import astral import os from setuptools import find_packages, setup ROOT_PATH = os.path.abspath(os.path.dirname(__file__)) long_description = open(os.path.join(ROOT_PATH, 'README.rst')).read() setup(name='astral', version=astral.__version__, description='Astral Streaming P2P Client', long_description=long_description, author='Astral Project Group', author_email='astral@bueda.com', url='http://github.com/peplin/astral', test_suite='nose.collector', setup_requires=['nose>=0.11',], test_requires=['unittest2>=0.5.1', 'mockito==0.5.1', 'python-faker==0.2.3', 'factory-boy==1.0.0',], install_requires=[ 'tornado>=1.2.1', 'importlib>=1.0.2', 'sqlalchemy>=0.6.6', 'Elixir>=0.7.1', 'restkit>=3.2.0', ], packages=find_packages(), entry_points={ 'console_scripts': [ 'astralnode = astral.bin.astralnode:main', 'astralctl = astral.bin.astralctl:main', ], }, )
<commit_before>import astral import os from setuptools import find_packages, setup ROOT_PATH = os.path.abspath(os.path.dirname(__file__)) long_description = open(os.path.join(ROOT_PATH, 'README.rst')).read() setup(name='astral', version=astral.__version__, description='Astral Streaming P2P Client', long_description=long_description, author='Astral Project Group', author_email='astral@bueda.com', url='http://github.com/peplin/astral', test_suite='nose.collector', setup_requires=['nose>=0.11',], test_requires=['unittest2>=0.5.1', 'mockito==0.5.1', 'python-faker==0.2.3', 'factory-boy==1.0.0',], install_requires=[ 'tornado>=1.2.1', 'importlib>=1.0.2', 'sqlalchemy>=0.6.6', 'Elixir>=0.7.1', 'restkit>=3.2.0', ], packages=find_packages(), entry_points={ 'console_scripts': [ 'astralnode = astral.bin.astralnode:main',], }, ) <commit_msg>Add astralctl script to console scripts section of package config.<commit_after>
import astral import os from setuptools import find_packages, setup ROOT_PATH = os.path.abspath(os.path.dirname(__file__)) long_description = open(os.path.join(ROOT_PATH, 'README.rst')).read() setup(name='astral', version=astral.__version__, description='Astral Streaming P2P Client', long_description=long_description, author='Astral Project Group', author_email='astral@bueda.com', url='http://github.com/peplin/astral', test_suite='nose.collector', setup_requires=['nose>=0.11',], test_requires=['unittest2>=0.5.1', 'mockito==0.5.1', 'python-faker==0.2.3', 'factory-boy==1.0.0',], install_requires=[ 'tornado>=1.2.1', 'importlib>=1.0.2', 'sqlalchemy>=0.6.6', 'Elixir>=0.7.1', 'restkit>=3.2.0', ], packages=find_packages(), entry_points={ 'console_scripts': [ 'astralnode = astral.bin.astralnode:main', 'astralctl = astral.bin.astralctl:main', ], }, )
import astral import os from setuptools import find_packages, setup ROOT_PATH = os.path.abspath(os.path.dirname(__file__)) long_description = open(os.path.join(ROOT_PATH, 'README.rst')).read() setup(name='astral', version=astral.__version__, description='Astral Streaming P2P Client', long_description=long_description, author='Astral Project Group', author_email='astral@bueda.com', url='http://github.com/peplin/astral', test_suite='nose.collector', setup_requires=['nose>=0.11',], test_requires=['unittest2>=0.5.1', 'mockito==0.5.1', 'python-faker==0.2.3', 'factory-boy==1.0.0',], install_requires=[ 'tornado>=1.2.1', 'importlib>=1.0.2', 'sqlalchemy>=0.6.6', 'Elixir>=0.7.1', 'restkit>=3.2.0', ], packages=find_packages(), entry_points={ 'console_scripts': [ 'astralnode = astral.bin.astralnode:main',], }, ) Add astralctl script to console scripts section of package config.import astral import os from setuptools import find_packages, setup ROOT_PATH = os.path.abspath(os.path.dirname(__file__)) long_description = open(os.path.join(ROOT_PATH, 'README.rst')).read() setup(name='astral', version=astral.__version__, description='Astral Streaming P2P Client', long_description=long_description, author='Astral Project Group', author_email='astral@bueda.com', url='http://github.com/peplin/astral', test_suite='nose.collector', setup_requires=['nose>=0.11',], test_requires=['unittest2>=0.5.1', 'mockito==0.5.1', 'python-faker==0.2.3', 'factory-boy==1.0.0',], install_requires=[ 'tornado>=1.2.1', 'importlib>=1.0.2', 'sqlalchemy>=0.6.6', 'Elixir>=0.7.1', 'restkit>=3.2.0', ], packages=find_packages(), entry_points={ 'console_scripts': [ 'astralnode = astral.bin.astralnode:main', 'astralctl = astral.bin.astralctl:main', ], }, )
<commit_before>import astral import os from setuptools import find_packages, setup ROOT_PATH = os.path.abspath(os.path.dirname(__file__)) long_description = open(os.path.join(ROOT_PATH, 'README.rst')).read() setup(name='astral', version=astral.__version__, description='Astral Streaming P2P Client', long_description=long_description, author='Astral Project Group', author_email='astral@bueda.com', url='http://github.com/peplin/astral', test_suite='nose.collector', setup_requires=['nose>=0.11',], test_requires=['unittest2>=0.5.1', 'mockito==0.5.1', 'python-faker==0.2.3', 'factory-boy==1.0.0',], install_requires=[ 'tornado>=1.2.1', 'importlib>=1.0.2', 'sqlalchemy>=0.6.6', 'Elixir>=0.7.1', 'restkit>=3.2.0', ], packages=find_packages(), entry_points={ 'console_scripts': [ 'astralnode = astral.bin.astralnode:main',], }, ) <commit_msg>Add astralctl script to console scripts section of package config.<commit_after>import astral import os from setuptools import find_packages, setup ROOT_PATH = os.path.abspath(os.path.dirname(__file__)) long_description = open(os.path.join(ROOT_PATH, 'README.rst')).read() setup(name='astral', version=astral.__version__, description='Astral Streaming P2P Client', long_description=long_description, author='Astral Project Group', author_email='astral@bueda.com', url='http://github.com/peplin/astral', test_suite='nose.collector', setup_requires=['nose>=0.11',], test_requires=['unittest2>=0.5.1', 'mockito==0.5.1', 'python-faker==0.2.3', 'factory-boy==1.0.0',], install_requires=[ 'tornado>=1.2.1', 'importlib>=1.0.2', 'sqlalchemy>=0.6.6', 'Elixir>=0.7.1', 'restkit>=3.2.0', ], packages=find_packages(), entry_points={ 'console_scripts': [ 'astralnode = astral.bin.astralnode:main', 'astralctl = astral.bin.astralctl:main', ], }, )
3ed73321bb70817faa87d31ec8b588857b204634
setup.py
setup.py
from setuptools import setup, find_packages setup(name='pypif', version='2.0.1', url='http://github.com/CitrineInformatics/pypif', description='Python tools for working with the Physical Information File (PIF)', author='Kyle Michel', author_email='kyle@citrine.io', packages=find_packages(), install_requires=[ 'six>=1.10.0,<2' ])
from setuptools import setup, find_packages setup(name='pypif', version='2.1.0', url='http://github.com/CitrineInformatics/pypif', description='Python tools for working with the Physical Information File (PIF)', author='Kyle Michel', author_email='kyle@citrine.io', packages=find_packages(), install_requires=[ 'six>=1.10.0,<2' ])
Add url field to Person object
Add url field to Person object
Python
apache-2.0
CitrineInformatics/pypif
from setuptools import setup, find_packages setup(name='pypif', version='2.0.1', url='http://github.com/CitrineInformatics/pypif', description='Python tools for working with the Physical Information File (PIF)', author='Kyle Michel', author_email='kyle@citrine.io', packages=find_packages(), install_requires=[ 'six>=1.10.0,<2' ]) Add url field to Person object
from setuptools import setup, find_packages setup(name='pypif', version='2.1.0', url='http://github.com/CitrineInformatics/pypif', description='Python tools for working with the Physical Information File (PIF)', author='Kyle Michel', author_email='kyle@citrine.io', packages=find_packages(), install_requires=[ 'six>=1.10.0,<2' ])
<commit_before>from setuptools import setup, find_packages setup(name='pypif', version='2.0.1', url='http://github.com/CitrineInformatics/pypif', description='Python tools for working with the Physical Information File (PIF)', author='Kyle Michel', author_email='kyle@citrine.io', packages=find_packages(), install_requires=[ 'six>=1.10.0,<2' ]) <commit_msg>Add url field to Person object<commit_after>
from setuptools import setup, find_packages setup(name='pypif', version='2.1.0', url='http://github.com/CitrineInformatics/pypif', description='Python tools for working with the Physical Information File (PIF)', author='Kyle Michel', author_email='kyle@citrine.io', packages=find_packages(), install_requires=[ 'six>=1.10.0,<2' ])
from setuptools import setup, find_packages setup(name='pypif', version='2.0.1', url='http://github.com/CitrineInformatics/pypif', description='Python tools for working with the Physical Information File (PIF)', author='Kyle Michel', author_email='kyle@citrine.io', packages=find_packages(), install_requires=[ 'six>=1.10.0,<2' ]) Add url field to Person objectfrom setuptools import setup, find_packages setup(name='pypif', version='2.1.0', url='http://github.com/CitrineInformatics/pypif', description='Python tools for working with the Physical Information File (PIF)', author='Kyle Michel', author_email='kyle@citrine.io', packages=find_packages(), install_requires=[ 'six>=1.10.0,<2' ])
<commit_before>from setuptools import setup, find_packages setup(name='pypif', version='2.0.1', url='http://github.com/CitrineInformatics/pypif', description='Python tools for working with the Physical Information File (PIF)', author='Kyle Michel', author_email='kyle@citrine.io', packages=find_packages(), install_requires=[ 'six>=1.10.0,<2' ]) <commit_msg>Add url field to Person object<commit_after>from setuptools import setup, find_packages setup(name='pypif', version='2.1.0', url='http://github.com/CitrineInformatics/pypif', description='Python tools for working with the Physical Information File (PIF)', author='Kyle Michel', author_email='kyle@citrine.io', packages=find_packages(), install_requires=[ 'six>=1.10.0,<2' ])
b6b8c8e12b18515591c0b053cac9e515314b819e
setup.py
setup.py
""" setup script """ from setuptools import setup, find_packages _VERSION = '0.7' setup( name='jut-tools', version=_VERSION, author='Rodney Gomes', author_email='rodney@jut.io', url='https://github.com/jut-io/jut-python-tools', download_url='https://github.com/jut-io/jut-python-tools/tarball/%s' % _VERSION, install_requires=[ 'requests==2.7.0', 'websocket-client==0.32.0', 'memoized==0.2', 'tabulate==0.7.5' ], test_suite='tests', keywords=[''], packages=find_packages(exclude=['tests']), license='MIT License', description='jut command line tools', # pypi doesn't support markdown so we can't push the README.md as is long_description='https://github.com/jut-io/jut-python-tools/blob/master/README.md', include_package_data=True, zip_safe=False, entry_points={ 'console_scripts': [ 'jut = jut.cli:main' ] }, )
""" setup script """ from setuptools import setup, find_packages _VERSION = '0.7' setup( name='jut-tools', version=_VERSION, author='Rodney Gomes', author_email='rodney@jut.io', url='https://github.com/jut-io/jut-python-tools', download_url='https://github.com/jut-io/jut-python-tools/tarball/%s' % _VERSION, install_requires=[ 'requests==2.7.0', 'websocket-client==0.32.0', 'memoized==0.2', 'tabulate==0.7.5' ], test_suite='tests', tests_install=[ 'sh==1.11' ], keywords=[''], packages=find_packages(exclude=['tests']), license='MIT License', description='jut command line tools', # pypi doesn't support markdown so we can't push the README.md as is long_description='https://github.com/jut-io/jut-python-tools/blob/master/README.md', include_package_data=True, zip_safe=False, entry_points={ 'console_scripts': [ 'jut = jut.cli:main' ] }, )
Revert "cleaning up a few unnecessary modules"
Revert "cleaning up a few unnecessary modules" This reverts commit 6b6911ca54a8bb61a1715c19be71729a55497278.
Python
mit
jut-io/jut-python-tools
""" setup script """ from setuptools import setup, find_packages _VERSION = '0.7' setup( name='jut-tools', version=_VERSION, author='Rodney Gomes', author_email='rodney@jut.io', url='https://github.com/jut-io/jut-python-tools', download_url='https://github.com/jut-io/jut-python-tools/tarball/%s' % _VERSION, install_requires=[ 'requests==2.7.0', 'websocket-client==0.32.0', 'memoized==0.2', 'tabulate==0.7.5' ], test_suite='tests', keywords=[''], packages=find_packages(exclude=['tests']), license='MIT License', description='jut command line tools', # pypi doesn't support markdown so we can't push the README.md as is long_description='https://github.com/jut-io/jut-python-tools/blob/master/README.md', include_package_data=True, zip_safe=False, entry_points={ 'console_scripts': [ 'jut = jut.cli:main' ] }, ) Revert "cleaning up a few unnecessary modules" This reverts commit 6b6911ca54a8bb61a1715c19be71729a55497278.
""" setup script """ from setuptools import setup, find_packages _VERSION = '0.7' setup( name='jut-tools', version=_VERSION, author='Rodney Gomes', author_email='rodney@jut.io', url='https://github.com/jut-io/jut-python-tools', download_url='https://github.com/jut-io/jut-python-tools/tarball/%s' % _VERSION, install_requires=[ 'requests==2.7.0', 'websocket-client==0.32.0', 'memoized==0.2', 'tabulate==0.7.5' ], test_suite='tests', tests_install=[ 'sh==1.11' ], keywords=[''], packages=find_packages(exclude=['tests']), license='MIT License', description='jut command line tools', # pypi doesn't support markdown so we can't push the README.md as is long_description='https://github.com/jut-io/jut-python-tools/blob/master/README.md', include_package_data=True, zip_safe=False, entry_points={ 'console_scripts': [ 'jut = jut.cli:main' ] }, )
<commit_before>""" setup script """ from setuptools import setup, find_packages _VERSION = '0.7' setup( name='jut-tools', version=_VERSION, author='Rodney Gomes', author_email='rodney@jut.io', url='https://github.com/jut-io/jut-python-tools', download_url='https://github.com/jut-io/jut-python-tools/tarball/%s' % _VERSION, install_requires=[ 'requests==2.7.0', 'websocket-client==0.32.0', 'memoized==0.2', 'tabulate==0.7.5' ], test_suite='tests', keywords=[''], packages=find_packages(exclude=['tests']), license='MIT License', description='jut command line tools', # pypi doesn't support markdown so we can't push the README.md as is long_description='https://github.com/jut-io/jut-python-tools/blob/master/README.md', include_package_data=True, zip_safe=False, entry_points={ 'console_scripts': [ 'jut = jut.cli:main' ] }, ) <commit_msg>Revert "cleaning up a few unnecessary modules" This reverts commit 6b6911ca54a8bb61a1715c19be71729a55497278.<commit_after>
""" setup script """ from setuptools import setup, find_packages _VERSION = '0.7' setup( name='jut-tools', version=_VERSION, author='Rodney Gomes', author_email='rodney@jut.io', url='https://github.com/jut-io/jut-python-tools', download_url='https://github.com/jut-io/jut-python-tools/tarball/%s' % _VERSION, install_requires=[ 'requests==2.7.0', 'websocket-client==0.32.0', 'memoized==0.2', 'tabulate==0.7.5' ], test_suite='tests', tests_install=[ 'sh==1.11' ], keywords=[''], packages=find_packages(exclude=['tests']), license='MIT License', description='jut command line tools', # pypi doesn't support markdown so we can't push the README.md as is long_description='https://github.com/jut-io/jut-python-tools/blob/master/README.md', include_package_data=True, zip_safe=False, entry_points={ 'console_scripts': [ 'jut = jut.cli:main' ] }, )
""" setup script """ from setuptools import setup, find_packages _VERSION = '0.7' setup( name='jut-tools', version=_VERSION, author='Rodney Gomes', author_email='rodney@jut.io', url='https://github.com/jut-io/jut-python-tools', download_url='https://github.com/jut-io/jut-python-tools/tarball/%s' % _VERSION, install_requires=[ 'requests==2.7.0', 'websocket-client==0.32.0', 'memoized==0.2', 'tabulate==0.7.5' ], test_suite='tests', keywords=[''], packages=find_packages(exclude=['tests']), license='MIT License', description='jut command line tools', # pypi doesn't support markdown so we can't push the README.md as is long_description='https://github.com/jut-io/jut-python-tools/blob/master/README.md', include_package_data=True, zip_safe=False, entry_points={ 'console_scripts': [ 'jut = jut.cli:main' ] }, ) Revert "cleaning up a few unnecessary modules" This reverts commit 6b6911ca54a8bb61a1715c19be71729a55497278.""" setup script """ from setuptools import setup, find_packages _VERSION = '0.7' setup( name='jut-tools', version=_VERSION, author='Rodney Gomes', author_email='rodney@jut.io', url='https://github.com/jut-io/jut-python-tools', download_url='https://github.com/jut-io/jut-python-tools/tarball/%s' % _VERSION, install_requires=[ 'requests==2.7.0', 'websocket-client==0.32.0', 'memoized==0.2', 'tabulate==0.7.5' ], test_suite='tests', tests_install=[ 'sh==1.11' ], keywords=[''], packages=find_packages(exclude=['tests']), license='MIT License', description='jut command line tools', # pypi doesn't support markdown so we can't push the README.md as is long_description='https://github.com/jut-io/jut-python-tools/blob/master/README.md', include_package_data=True, zip_safe=False, entry_points={ 'console_scripts': [ 'jut = jut.cli:main' ] }, )
<commit_before>""" setup script """ from setuptools import setup, find_packages _VERSION = '0.7' setup( name='jut-tools', version=_VERSION, author='Rodney Gomes', author_email='rodney@jut.io', url='https://github.com/jut-io/jut-python-tools', download_url='https://github.com/jut-io/jut-python-tools/tarball/%s' % _VERSION, install_requires=[ 'requests==2.7.0', 'websocket-client==0.32.0', 'memoized==0.2', 'tabulate==0.7.5' ], test_suite='tests', keywords=[''], packages=find_packages(exclude=['tests']), license='MIT License', description='jut command line tools', # pypi doesn't support markdown so we can't push the README.md as is long_description='https://github.com/jut-io/jut-python-tools/blob/master/README.md', include_package_data=True, zip_safe=False, entry_points={ 'console_scripts': [ 'jut = jut.cli:main' ] }, ) <commit_msg>Revert "cleaning up a few unnecessary modules" This reverts commit 6b6911ca54a8bb61a1715c19be71729a55497278.<commit_after>""" setup script """ from setuptools import setup, find_packages _VERSION = '0.7' setup( name='jut-tools', version=_VERSION, author='Rodney Gomes', author_email='rodney@jut.io', url='https://github.com/jut-io/jut-python-tools', download_url='https://github.com/jut-io/jut-python-tools/tarball/%s' % _VERSION, install_requires=[ 'requests==2.7.0', 'websocket-client==0.32.0', 'memoized==0.2', 'tabulate==0.7.5' ], test_suite='tests', tests_install=[ 'sh==1.11' ], keywords=[''], packages=find_packages(exclude=['tests']), license='MIT License', description='jut command line tools', # pypi doesn't support markdown so we can't push the README.md as is long_description='https://github.com/jut-io/jut-python-tools/blob/master/README.md', include_package_data=True, zip_safe=False, entry_points={ 'console_scripts': [ 'jut = jut.cli:main' ] }, )
6dd3946a0ecf63d913efc71150d753895710001e
setup.py
setup.py
#! /usr/bin/env python # -*- coding:utf-8 -*- """ Install the `arte_plus7` script """ from setuptools import setup NAME = 'arte_plus7' def get_version(module): """ Extract package version without importing file Importing cause issues with coverage, (modules can be removed from sys.modules to prevent this) Inspired from pep8 setup.py """ with open('%s.py' % module) as module_fd: for line in module_fd: if line.startswith('__version__'): return eval(line.split('=')[-1]) # pylint:disable=eval-used setup( name=NAME, version=get_version(NAME), description='CLI script to get videos from Arte plus 7 using their URL', author='cladmi', download_url='https://github.com/cladmi/arte_plus7', py_modules=NAME, entry_points={ 'console_scripts': ['{name} = {name}:main'.format(name=NAME)], }, classifiers=[ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Intended Audience :: End Users/Desktop', 'Topic :: utilities', ], install_requires=['argparse', 'beautifulsoup4'], )
#! /usr/bin/env python # -*- coding:utf-8 -*- """ Install the `arte_plus7` script """ from setuptools import setup NAME = 'arte_plus7' def get_version(module): """ Extract package version without importing file Importing cause issues with coverage, (modules can be removed from sys.modules to prevent this) Inspired from pep8 setup.py """ with open('%s.py' % module) as module_fd: for line in module_fd: if line.startswith('__version__'): return eval(line.split('=')[-1]) # pylint:disable=eval-used setup( name=NAME, version=get_version(NAME), description='CLI script to get videos from Arte plus 7 using their URL', author='cladmi', download_url='https://github.com/cladmi/arte_plus7', py_modules=[NAME], entry_points={ 'console_scripts': ['{name} = {name}:main'.format(name=NAME)], }, classifiers=[ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Intended Audience :: End Users/Desktop', 'Topic :: utilities', ], install_requires=['argparse', 'beautifulsoup4'], )
Correct py_modules as a list
Correct py_modules as a list
Python
mit
cladmi/arte_plus7
#! /usr/bin/env python # -*- coding:utf-8 -*- """ Install the `arte_plus7` script """ from setuptools import setup NAME = 'arte_plus7' def get_version(module): """ Extract package version without importing file Importing cause issues with coverage, (modules can be removed from sys.modules to prevent this) Inspired from pep8 setup.py """ with open('%s.py' % module) as module_fd: for line in module_fd: if line.startswith('__version__'): return eval(line.split('=')[-1]) # pylint:disable=eval-used setup( name=NAME, version=get_version(NAME), description='CLI script to get videos from Arte plus 7 using their URL', author='cladmi', download_url='https://github.com/cladmi/arte_plus7', py_modules=NAME, entry_points={ 'console_scripts': ['{name} = {name}:main'.format(name=NAME)], }, classifiers=[ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Intended Audience :: End Users/Desktop', 'Topic :: utilities', ], install_requires=['argparse', 'beautifulsoup4'], ) Correct py_modules as a list
#! /usr/bin/env python # -*- coding:utf-8 -*- """ Install the `arte_plus7` script """ from setuptools import setup NAME = 'arte_plus7' def get_version(module): """ Extract package version without importing file Importing cause issues with coverage, (modules can be removed from sys.modules to prevent this) Inspired from pep8 setup.py """ with open('%s.py' % module) as module_fd: for line in module_fd: if line.startswith('__version__'): return eval(line.split('=')[-1]) # pylint:disable=eval-used setup( name=NAME, version=get_version(NAME), description='CLI script to get videos from Arte plus 7 using their URL', author='cladmi', download_url='https://github.com/cladmi/arte_plus7', py_modules=[NAME], entry_points={ 'console_scripts': ['{name} = {name}:main'.format(name=NAME)], }, classifiers=[ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Intended Audience :: End Users/Desktop', 'Topic :: utilities', ], install_requires=['argparse', 'beautifulsoup4'], )
<commit_before>#! /usr/bin/env python # -*- coding:utf-8 -*- """ Install the `arte_plus7` script """ from setuptools import setup NAME = 'arte_plus7' def get_version(module): """ Extract package version without importing file Importing cause issues with coverage, (modules can be removed from sys.modules to prevent this) Inspired from pep8 setup.py """ with open('%s.py' % module) as module_fd: for line in module_fd: if line.startswith('__version__'): return eval(line.split('=')[-1]) # pylint:disable=eval-used setup( name=NAME, version=get_version(NAME), description='CLI script to get videos from Arte plus 7 using their URL', author='cladmi', download_url='https://github.com/cladmi/arte_plus7', py_modules=NAME, entry_points={ 'console_scripts': ['{name} = {name}:main'.format(name=NAME)], }, classifiers=[ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Intended Audience :: End Users/Desktop', 'Topic :: utilities', ], install_requires=['argparse', 'beautifulsoup4'], ) <commit_msg>Correct py_modules as a list<commit_after>
#! /usr/bin/env python # -*- coding:utf-8 -*- """ Install the `arte_plus7` script """ from setuptools import setup NAME = 'arte_plus7' def get_version(module): """ Extract package version without importing file Importing cause issues with coverage, (modules can be removed from sys.modules to prevent this) Inspired from pep8 setup.py """ with open('%s.py' % module) as module_fd: for line in module_fd: if line.startswith('__version__'): return eval(line.split('=')[-1]) # pylint:disable=eval-used setup( name=NAME, version=get_version(NAME), description='CLI script to get videos from Arte plus 7 using their URL', author='cladmi', download_url='https://github.com/cladmi/arte_plus7', py_modules=[NAME], entry_points={ 'console_scripts': ['{name} = {name}:main'.format(name=NAME)], }, classifiers=[ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Intended Audience :: End Users/Desktop', 'Topic :: utilities', ], install_requires=['argparse', 'beautifulsoup4'], )
#! /usr/bin/env python # -*- coding:utf-8 -*- """ Install the `arte_plus7` script """ from setuptools import setup NAME = 'arte_plus7' def get_version(module): """ Extract package version without importing file Importing cause issues with coverage, (modules can be removed from sys.modules to prevent this) Inspired from pep8 setup.py """ with open('%s.py' % module) as module_fd: for line in module_fd: if line.startswith('__version__'): return eval(line.split('=')[-1]) # pylint:disable=eval-used setup( name=NAME, version=get_version(NAME), description='CLI script to get videos from Arte plus 7 using their URL', author='cladmi', download_url='https://github.com/cladmi/arte_plus7', py_modules=NAME, entry_points={ 'console_scripts': ['{name} = {name}:main'.format(name=NAME)], }, classifiers=[ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Intended Audience :: End Users/Desktop', 'Topic :: utilities', ], install_requires=['argparse', 'beautifulsoup4'], ) Correct py_modules as a list#! /usr/bin/env python # -*- coding:utf-8 -*- """ Install the `arte_plus7` script """ from setuptools import setup NAME = 'arte_plus7' def get_version(module): """ Extract package version without importing file Importing cause issues with coverage, (modules can be removed from sys.modules to prevent this) Inspired from pep8 setup.py """ with open('%s.py' % module) as module_fd: for line in module_fd: if line.startswith('__version__'): return eval(line.split('=')[-1]) # pylint:disable=eval-used setup( name=NAME, version=get_version(NAME), description='CLI script to get videos from Arte plus 7 using their URL', author='cladmi', download_url='https://github.com/cladmi/arte_plus7', py_modules=[NAME], entry_points={ 'console_scripts': ['{name} = {name}:main'.format(name=NAME)], }, classifiers=[ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Intended Audience :: End Users/Desktop', 'Topic :: utilities', ], install_requires=['argparse', 'beautifulsoup4'], )
<commit_before>#! /usr/bin/env python # -*- coding:utf-8 -*- """ Install the `arte_plus7` script """ from setuptools import setup NAME = 'arte_plus7' def get_version(module): """ Extract package version without importing file Importing cause issues with coverage, (modules can be removed from sys.modules to prevent this) Inspired from pep8 setup.py """ with open('%s.py' % module) as module_fd: for line in module_fd: if line.startswith('__version__'): return eval(line.split('=')[-1]) # pylint:disable=eval-used setup( name=NAME, version=get_version(NAME), description='CLI script to get videos from Arte plus 7 using their URL', author='cladmi', download_url='https://github.com/cladmi/arte_plus7', py_modules=NAME, entry_points={ 'console_scripts': ['{name} = {name}:main'.format(name=NAME)], }, classifiers=[ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Intended Audience :: End Users/Desktop', 'Topic :: utilities', ], install_requires=['argparse', 'beautifulsoup4'], ) <commit_msg>Correct py_modules as a list<commit_after>#! /usr/bin/env python # -*- coding:utf-8 -*- """ Install the `arte_plus7` script """ from setuptools import setup NAME = 'arte_plus7' def get_version(module): """ Extract package version without importing file Importing cause issues with coverage, (modules can be removed from sys.modules to prevent this) Inspired from pep8 setup.py """ with open('%s.py' % module) as module_fd: for line in module_fd: if line.startswith('__version__'): return eval(line.split('=')[-1]) # pylint:disable=eval-used setup( name=NAME, version=get_version(NAME), description='CLI script to get videos from Arte plus 7 using their URL', author='cladmi', download_url='https://github.com/cladmi/arte_plus7', py_modules=[NAME], entry_points={ 'console_scripts': ['{name} = {name}:main'.format(name=NAME)], }, classifiers=[ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Intended Audience :: End Users/Desktop', 'Topic :: utilities', ], install_requires=['argparse', 'beautifulsoup4'], )
be2cd54386c0fb9c407ac5dc7da467547b0b426e
aldryn_apphooks_config/utils.py
aldryn_apphooks_config/utils.py
# -*- coding: utf-8 -*- from app_data import AppDataContainer, app_registry from cms.apphook_pool import apphook_pool from django.core.urlresolvers import resolve def get_app_instance(request): """ Returns a tuple containing the current namespace and the AppHookConfig instance :param request: request object :return: namespace, config """ app = None if getattr(request, 'current_page', None): app = apphook_pool.get_apphook(request.current_page.application_urls) config = None namespace = resolve(request.path_info).namespace if app and app.app_config: config = app.get_config(namespace) return namespace, config def setup_config(form_class, config_model): """ Register the provided form as config form for the provided config model :param form_class: Form class derived from AppDataForm :param config_model: Model class derived from AppHookConfig :return: """ app_registry.register('config', AppDataContainer.from_form(form_class), config_model)
# -*- coding: utf-8 -*- from app_data import AppDataContainer, app_registry from cms.apphook_pool import apphook_pool from django.core.urlresolvers import resolve, Resolver404 def get_app_instance(request): """ Returns a tuple containing the current namespace and the AppHookConfig instance :param request: request object :return: namespace, config """ app = None if getattr(request, 'current_page', None): app = apphook_pool.get_apphook(request.current_page.application_urls) if app and app.app_config: try: config = None namespace = resolve(request.path).namespace if app and app.app_config: config = app.get_config(namespace) return namespace, config except Resolver404: pass return '', None def setup_config(form_class, config_model): """ Register the provided form as config form for the provided config model :param form_class: Form class derived from AppDataForm :param config_model: Model class derived from AppHookConfig :return: """ app_registry.register('config', AppDataContainer.from_form(form_class), config_model)
Add checks to get_app_instance to avoid Resolver404 even if namespace does not exists
Add checks to get_app_instance to avoid Resolver404 even if namespace does not exists
Python
bsd-3-clause
aldryn/aldryn-apphooks-config,aldryn/aldryn-apphooks-config,aldryn/aldryn-apphooks-config
# -*- coding: utf-8 -*- from app_data import AppDataContainer, app_registry from cms.apphook_pool import apphook_pool from django.core.urlresolvers import resolve def get_app_instance(request): """ Returns a tuple containing the current namespace and the AppHookConfig instance :param request: request object :return: namespace, config """ app = None if getattr(request, 'current_page', None): app = apphook_pool.get_apphook(request.current_page.application_urls) config = None namespace = resolve(request.path_info).namespace if app and app.app_config: config = app.get_config(namespace) return namespace, config def setup_config(form_class, config_model): """ Register the provided form as config form for the provided config model :param form_class: Form class derived from AppDataForm :param config_model: Model class derived from AppHookConfig :return: """ app_registry.register('config', AppDataContainer.from_form(form_class), config_model)Add checks to get_app_instance to avoid Resolver404 even if namespace does not exists
# -*- coding: utf-8 -*- from app_data import AppDataContainer, app_registry from cms.apphook_pool import apphook_pool from django.core.urlresolvers import resolve, Resolver404 def get_app_instance(request): """ Returns a tuple containing the current namespace and the AppHookConfig instance :param request: request object :return: namespace, config """ app = None if getattr(request, 'current_page', None): app = apphook_pool.get_apphook(request.current_page.application_urls) if app and app.app_config: try: config = None namespace = resolve(request.path).namespace if app and app.app_config: config = app.get_config(namespace) return namespace, config except Resolver404: pass return '', None def setup_config(form_class, config_model): """ Register the provided form as config form for the provided config model :param form_class: Form class derived from AppDataForm :param config_model: Model class derived from AppHookConfig :return: """ app_registry.register('config', AppDataContainer.from_form(form_class), config_model)
<commit_before># -*- coding: utf-8 -*- from app_data import AppDataContainer, app_registry from cms.apphook_pool import apphook_pool from django.core.urlresolvers import resolve def get_app_instance(request): """ Returns a tuple containing the current namespace and the AppHookConfig instance :param request: request object :return: namespace, config """ app = None if getattr(request, 'current_page', None): app = apphook_pool.get_apphook(request.current_page.application_urls) config = None namespace = resolve(request.path_info).namespace if app and app.app_config: config = app.get_config(namespace) return namespace, config def setup_config(form_class, config_model): """ Register the provided form as config form for the provided config model :param form_class: Form class derived from AppDataForm :param config_model: Model class derived from AppHookConfig :return: """ app_registry.register('config', AppDataContainer.from_form(form_class), config_model)<commit_msg>Add checks to get_app_instance to avoid Resolver404 even if namespace does not exists<commit_after>
# -*- coding: utf-8 -*- from app_data import AppDataContainer, app_registry from cms.apphook_pool import apphook_pool from django.core.urlresolvers import resolve, Resolver404 def get_app_instance(request): """ Returns a tuple containing the current namespace and the AppHookConfig instance :param request: request object :return: namespace, config """ app = None if getattr(request, 'current_page', None): app = apphook_pool.get_apphook(request.current_page.application_urls) if app and app.app_config: try: config = None namespace = resolve(request.path).namespace if app and app.app_config: config = app.get_config(namespace) return namespace, config except Resolver404: pass return '', None def setup_config(form_class, config_model): """ Register the provided form as config form for the provided config model :param form_class: Form class derived from AppDataForm :param config_model: Model class derived from AppHookConfig :return: """ app_registry.register('config', AppDataContainer.from_form(form_class), config_model)
# -*- coding: utf-8 -*- from app_data import AppDataContainer, app_registry from cms.apphook_pool import apphook_pool from django.core.urlresolvers import resolve def get_app_instance(request): """ Returns a tuple containing the current namespace and the AppHookConfig instance :param request: request object :return: namespace, config """ app = None if getattr(request, 'current_page', None): app = apphook_pool.get_apphook(request.current_page.application_urls) config = None namespace = resolve(request.path_info).namespace if app and app.app_config: config = app.get_config(namespace) return namespace, config def setup_config(form_class, config_model): """ Register the provided form as config form for the provided config model :param form_class: Form class derived from AppDataForm :param config_model: Model class derived from AppHookConfig :return: """ app_registry.register('config', AppDataContainer.from_form(form_class), config_model)Add checks to get_app_instance to avoid Resolver404 even if namespace does not exists# -*- coding: utf-8 -*- from app_data import AppDataContainer, app_registry from cms.apphook_pool import apphook_pool from django.core.urlresolvers import resolve, Resolver404 def get_app_instance(request): """ Returns a tuple containing the current namespace and the AppHookConfig instance :param request: request object :return: namespace, config """ app = None if getattr(request, 'current_page', None): app = apphook_pool.get_apphook(request.current_page.application_urls) if app and app.app_config: try: config = None namespace = resolve(request.path).namespace if app and app.app_config: config = app.get_config(namespace) return namespace, config except Resolver404: pass return '', None def setup_config(form_class, config_model): """ Register the provided form as config form for the provided config model :param form_class: Form class derived from AppDataForm :param config_model: Model class derived from AppHookConfig :return: """ app_registry.register('config', AppDataContainer.from_form(form_class), config_model)
<commit_before># -*- coding: utf-8 -*- from app_data import AppDataContainer, app_registry from cms.apphook_pool import apphook_pool from django.core.urlresolvers import resolve def get_app_instance(request): """ Returns a tuple containing the current namespace and the AppHookConfig instance :param request: request object :return: namespace, config """ app = None if getattr(request, 'current_page', None): app = apphook_pool.get_apphook(request.current_page.application_urls) config = None namespace = resolve(request.path_info).namespace if app and app.app_config: config = app.get_config(namespace) return namespace, config def setup_config(form_class, config_model): """ Register the provided form as config form for the provided config model :param form_class: Form class derived from AppDataForm :param config_model: Model class derived from AppHookConfig :return: """ app_registry.register('config', AppDataContainer.from_form(form_class), config_model)<commit_msg>Add checks to get_app_instance to avoid Resolver404 even if namespace does not exists<commit_after># -*- coding: utf-8 -*- from app_data import AppDataContainer, app_registry from cms.apphook_pool import apphook_pool from django.core.urlresolvers import resolve, Resolver404 def get_app_instance(request): """ Returns a tuple containing the current namespace and the AppHookConfig instance :param request: request object :return: namespace, config """ app = None if getattr(request, 'current_page', None): app = apphook_pool.get_apphook(request.current_page.application_urls) if app and app.app_config: try: config = None namespace = resolve(request.path).namespace if app and app.app_config: config = app.get_config(namespace) return namespace, config except Resolver404: pass return '', None def setup_config(form_class, config_model): """ Register the provided form as config form for the provided config model :param form_class: Form class derived from AppDataForm :param config_model: Model class derived from AppHookConfig :return: """ app_registry.register('config', AppDataContainer.from_form(form_class), config_model)
300415c30dab2f8d2622c8b8de03f433bf6d5960
setup.py
setup.py
from setuptools import setup, find_packages from jamo import __version__ import sys if sys.version_info <= (3, 0): print("ERROR: jamo requires Python 3.0 or later " "(bleeding edge preferred)", file=sys.stderr) sys.exit(1) with open('README.rst') as f: long_description = f.read() setup( name="jamo", version=__version__, description="A Hangul syllable and jamo analyzer.", long_description=long_description, url="https://github.com/jdong820/python-jamo", author="Joshua Dong", author_email="jdong42@gmail.com", license="http://www.apache.org/licenses/LICENSE-2.0", classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", ], keywords="Korean Hangul jamo syllable nlp", packages=find_packages(), package_dir={'jamo': 'jamo'}, package_data={'jamo': ['data/*.json']}, )
from setuptools import setup, find_packages from jamo import __version__ import sys if sys.version_info <= (3, 0): print("ERROR: jamo requires Python 3.0 or later " "(bleeding edge preferred)", file=sys.stderr) sys.exit(1) with open('README.rst', encoding='utf8') as f: long_description = f.read() setup( name="jamo", version=__version__, description="A Hangul syllable and jamo analyzer.", long_description=long_description, url="https://github.com/jdong820/python-jamo", author="Joshua Dong", author_email="jdong42@gmail.com", license="http://www.apache.org/licenses/LICENSE-2.0", classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", ], keywords="Korean Hangul jamo syllable nlp", packages=find_packages(), package_dir={'jamo': 'jamo'}, package_data={'jamo': ['data/*.json']}, )
Fix for encoding bug during installation on Windows
Fix for encoding bug during installation on Windows
Python
apache-2.0
JDongian/python-jamo
from setuptools import setup, find_packages from jamo import __version__ import sys if sys.version_info <= (3, 0): print("ERROR: jamo requires Python 3.0 or later " "(bleeding edge preferred)", file=sys.stderr) sys.exit(1) with open('README.rst') as f: long_description = f.read() setup( name="jamo", version=__version__, description="A Hangul syllable and jamo analyzer.", long_description=long_description, url="https://github.com/jdong820/python-jamo", author="Joshua Dong", author_email="jdong42@gmail.com", license="http://www.apache.org/licenses/LICENSE-2.0", classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", ], keywords="Korean Hangul jamo syllable nlp", packages=find_packages(), package_dir={'jamo': 'jamo'}, package_data={'jamo': ['data/*.json']}, ) Fix for encoding bug during installation on Windows
from setuptools import setup, find_packages from jamo import __version__ import sys if sys.version_info <= (3, 0): print("ERROR: jamo requires Python 3.0 or later " "(bleeding edge preferred)", file=sys.stderr) sys.exit(1) with open('README.rst', encoding='utf8') as f: long_description = f.read() setup( name="jamo", version=__version__, description="A Hangul syllable and jamo analyzer.", long_description=long_description, url="https://github.com/jdong820/python-jamo", author="Joshua Dong", author_email="jdong42@gmail.com", license="http://www.apache.org/licenses/LICENSE-2.0", classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", ], keywords="Korean Hangul jamo syllable nlp", packages=find_packages(), package_dir={'jamo': 'jamo'}, package_data={'jamo': ['data/*.json']}, )
<commit_before>from setuptools import setup, find_packages from jamo import __version__ import sys if sys.version_info <= (3, 0): print("ERROR: jamo requires Python 3.0 or later " "(bleeding edge preferred)", file=sys.stderr) sys.exit(1) with open('README.rst') as f: long_description = f.read() setup( name="jamo", version=__version__, description="A Hangul syllable and jamo analyzer.", long_description=long_description, url="https://github.com/jdong820/python-jamo", author="Joshua Dong", author_email="jdong42@gmail.com", license="http://www.apache.org/licenses/LICENSE-2.0", classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", ], keywords="Korean Hangul jamo syllable nlp", packages=find_packages(), package_dir={'jamo': 'jamo'}, package_data={'jamo': ['data/*.json']}, ) <commit_msg>Fix for encoding bug during installation on Windows<commit_after>
from setuptools import setup, find_packages from jamo import __version__ import sys if sys.version_info <= (3, 0): print("ERROR: jamo requires Python 3.0 or later " "(bleeding edge preferred)", file=sys.stderr) sys.exit(1) with open('README.rst', encoding='utf8') as f: long_description = f.read() setup( name="jamo", version=__version__, description="A Hangul syllable and jamo analyzer.", long_description=long_description, url="https://github.com/jdong820/python-jamo", author="Joshua Dong", author_email="jdong42@gmail.com", license="http://www.apache.org/licenses/LICENSE-2.0", classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", ], keywords="Korean Hangul jamo syllable nlp", packages=find_packages(), package_dir={'jamo': 'jamo'}, package_data={'jamo': ['data/*.json']}, )
from setuptools import setup, find_packages from jamo import __version__ import sys if sys.version_info <= (3, 0): print("ERROR: jamo requires Python 3.0 or later " "(bleeding edge preferred)", file=sys.stderr) sys.exit(1) with open('README.rst') as f: long_description = f.read() setup( name="jamo", version=__version__, description="A Hangul syllable and jamo analyzer.", long_description=long_description, url="https://github.com/jdong820/python-jamo", author="Joshua Dong", author_email="jdong42@gmail.com", license="http://www.apache.org/licenses/LICENSE-2.0", classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", ], keywords="Korean Hangul jamo syllable nlp", packages=find_packages(), package_dir={'jamo': 'jamo'}, package_data={'jamo': ['data/*.json']}, ) Fix for encoding bug during installation on Windowsfrom setuptools import setup, find_packages from jamo import __version__ import sys if sys.version_info <= (3, 0): print("ERROR: jamo requires Python 3.0 or later " "(bleeding edge preferred)", file=sys.stderr) sys.exit(1) with open('README.rst', encoding='utf8') as f: long_description = f.read() setup( name="jamo", version=__version__, description="A Hangul syllable and jamo analyzer.", long_description=long_description, url="https://github.com/jdong820/python-jamo", author="Joshua Dong", author_email="jdong42@gmail.com", license="http://www.apache.org/licenses/LICENSE-2.0", classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", ], keywords="Korean Hangul jamo syllable nlp", packages=find_packages(), package_dir={'jamo': 'jamo'}, package_data={'jamo': ['data/*.json']}, )
<commit_before>from setuptools import setup, find_packages from jamo import __version__ import sys if sys.version_info <= (3, 0): print("ERROR: jamo requires Python 3.0 or later " "(bleeding edge preferred)", file=sys.stderr) sys.exit(1) with open('README.rst') as f: long_description = f.read() setup( name="jamo", version=__version__, description="A Hangul syllable and jamo analyzer.", long_description=long_description, url="https://github.com/jdong820/python-jamo", author="Joshua Dong", author_email="jdong42@gmail.com", license="http://www.apache.org/licenses/LICENSE-2.0", classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", ], keywords="Korean Hangul jamo syllable nlp", packages=find_packages(), package_dir={'jamo': 'jamo'}, package_data={'jamo': ['data/*.json']}, ) <commit_msg>Fix for encoding bug during installation on Windows<commit_after>from setuptools import setup, find_packages from jamo import __version__ import sys if sys.version_info <= (3, 0): print("ERROR: jamo requires Python 3.0 or later " "(bleeding edge preferred)", file=sys.stderr) sys.exit(1) with open('README.rst', encoding='utf8') as f: long_description = f.read() setup( name="jamo", version=__version__, description="A Hangul syllable and jamo analyzer.", long_description=long_description, url="https://github.com/jdong820/python-jamo", author="Joshua Dong", author_email="jdong42@gmail.com", license="http://www.apache.org/licenses/LICENSE-2.0", classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", ], keywords="Korean Hangul jamo syllable nlp", packages=find_packages(), package_dir={'jamo': 'jamo'}, package_data={'jamo': ['data/*.json']}, )
7941c60ba5a52ead654e5816ee39e48b9e927a21
setup.py
setup.py
#!/usr/bin/env python3 from distutils.core import setup setup( name='LoadStone', version='0.1', description='Interface for FFXIV Lodestone', author='Sami Elahmadie', author_email='s.elahmadie@gmail.com', url='https://github.com/Demotivated/loadstone/', packages=['api'], install_requires=[ 'flask==0.10.1', 'flask_sqlalchemy==2.0', 'lxml==3.4.4', 'psycopg2==2.6.1', 'pytest==2.8.2', 'pytest-flask==0.10.0', 'requests==2.8.1', ] )
#!/usr/bin/env python3 from distutils.core import setup setup( name='LoadStone', version='0.1', description='Interface for FFXIV Lodestone', author='Sami Elahmadie', author_email='s.elahmadie@gmail.com', url='https://github.com/Demotivated/loadstone/', packages=['api'], install_requires=[ 'flask==0.10.1', 'flask_sqlalchemy==2.0', 'lxml==3.4.4', 'psycopg2==2.6.1', 'pytest==2.8.2', 'pytest-flask==0.10.0', 'requests==2.8.1', 'sphinx==1.3.1', 'sphinx-rtd-theme==0.1.9' ] )
Add sphinx & theme to requirements
Add sphinx & theme to requirements
Python
mit
Demotivated/loadstone
#!/usr/bin/env python3 from distutils.core import setup setup( name='LoadStone', version='0.1', description='Interface for FFXIV Lodestone', author='Sami Elahmadie', author_email='s.elahmadie@gmail.com', url='https://github.com/Demotivated/loadstone/', packages=['api'], install_requires=[ 'flask==0.10.1', 'flask_sqlalchemy==2.0', 'lxml==3.4.4', 'psycopg2==2.6.1', 'pytest==2.8.2', 'pytest-flask==0.10.0', 'requests==2.8.1', ] ) Add sphinx & theme to requirements
#!/usr/bin/env python3 from distutils.core import setup setup( name='LoadStone', version='0.1', description='Interface for FFXIV Lodestone', author='Sami Elahmadie', author_email='s.elahmadie@gmail.com', url='https://github.com/Demotivated/loadstone/', packages=['api'], install_requires=[ 'flask==0.10.1', 'flask_sqlalchemy==2.0', 'lxml==3.4.4', 'psycopg2==2.6.1', 'pytest==2.8.2', 'pytest-flask==0.10.0', 'requests==2.8.1', 'sphinx==1.3.1', 'sphinx-rtd-theme==0.1.9' ] )
<commit_before>#!/usr/bin/env python3 from distutils.core import setup setup( name='LoadStone', version='0.1', description='Interface for FFXIV Lodestone', author='Sami Elahmadie', author_email='s.elahmadie@gmail.com', url='https://github.com/Demotivated/loadstone/', packages=['api'], install_requires=[ 'flask==0.10.1', 'flask_sqlalchemy==2.0', 'lxml==3.4.4', 'psycopg2==2.6.1', 'pytest==2.8.2', 'pytest-flask==0.10.0', 'requests==2.8.1', ] ) <commit_msg>Add sphinx & theme to requirements<commit_after>
#!/usr/bin/env python3 from distutils.core import setup setup( name='LoadStone', version='0.1', description='Interface for FFXIV Lodestone', author='Sami Elahmadie', author_email='s.elahmadie@gmail.com', url='https://github.com/Demotivated/loadstone/', packages=['api'], install_requires=[ 'flask==0.10.1', 'flask_sqlalchemy==2.0', 'lxml==3.4.4', 'psycopg2==2.6.1', 'pytest==2.8.2', 'pytest-flask==0.10.0', 'requests==2.8.1', 'sphinx==1.3.1', 'sphinx-rtd-theme==0.1.9' ] )
#!/usr/bin/env python3 from distutils.core import setup setup( name='LoadStone', version='0.1', description='Interface for FFXIV Lodestone', author='Sami Elahmadie', author_email='s.elahmadie@gmail.com', url='https://github.com/Demotivated/loadstone/', packages=['api'], install_requires=[ 'flask==0.10.1', 'flask_sqlalchemy==2.0', 'lxml==3.4.4', 'psycopg2==2.6.1', 'pytest==2.8.2', 'pytest-flask==0.10.0', 'requests==2.8.1', ] ) Add sphinx & theme to requirements#!/usr/bin/env python3 from distutils.core import setup setup( name='LoadStone', version='0.1', description='Interface for FFXIV Lodestone', author='Sami Elahmadie', author_email='s.elahmadie@gmail.com', url='https://github.com/Demotivated/loadstone/', packages=['api'], install_requires=[ 'flask==0.10.1', 'flask_sqlalchemy==2.0', 'lxml==3.4.4', 'psycopg2==2.6.1', 'pytest==2.8.2', 'pytest-flask==0.10.0', 'requests==2.8.1', 'sphinx==1.3.1', 'sphinx-rtd-theme==0.1.9' ] )
<commit_before>#!/usr/bin/env python3 from distutils.core import setup setup( name='LoadStone', version='0.1', description='Interface for FFXIV Lodestone', author='Sami Elahmadie', author_email='s.elahmadie@gmail.com', url='https://github.com/Demotivated/loadstone/', packages=['api'], install_requires=[ 'flask==0.10.1', 'flask_sqlalchemy==2.0', 'lxml==3.4.4', 'psycopg2==2.6.1', 'pytest==2.8.2', 'pytest-flask==0.10.0', 'requests==2.8.1', ] ) <commit_msg>Add sphinx & theme to requirements<commit_after>#!/usr/bin/env python3 from distutils.core import setup setup( name='LoadStone', version='0.1', description='Interface for FFXIV Lodestone', author='Sami Elahmadie', author_email='s.elahmadie@gmail.com', url='https://github.com/Demotivated/loadstone/', packages=['api'], install_requires=[ 'flask==0.10.1', 'flask_sqlalchemy==2.0', 'lxml==3.4.4', 'psycopg2==2.6.1', 'pytest==2.8.2', 'pytest-flask==0.10.0', 'requests==2.8.1', 'sphinx==1.3.1', 'sphinx-rtd-theme==0.1.9' ] )
433c041e3016fce0f1e3edbb9dae980f69071f3b
setup.py
setup.py
import os from setuptools import setup setup( name = "pyscribe", version = "0.1.1", author = "Alexander Wang", author_email = "alexanderw@berkeley.edu", description = ("PyScribe makes print debugging easier and more efficient"), license = "MIT", keywords = "python pyscribe debug print", url = "https://github.com/alixander/pyscribe", download_url = "https://github.com/alixander/pyscribe/tarbell/0.1.1", entry_points={ 'console_scripts': [ 'pyscribe = pyscribe.pyscribe:main', ], }, packages=['pyscribe'], classifiers=[ "Development Status :: 3 - Alpha" ], )
import os from setuptools import setup setup( name = "pyscribe", version = "0.1.1", author = "Alexander Wang", author_email = "alexanderw@berkeley.edu", description = ("PyScribe makes print debugging easier and more efficient"), license = "MIT", keywords = "python pyscribe debug print", url = "https://github.com/alixander/pyscribe", download_url = "https://github.com/alixander/pyscribe/tarbell/0.1.2", entry_points={ 'console_scripts': [ 'pyscribe = pyscribe.pyscribe:main', ], }, packages=['pyscribe'], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', ], )
Add classifiers since python3 now supported
Add classifiers since python3 now supported
Python
mit
alixander/PyScribe
import os from setuptools import setup setup( name = "pyscribe", version = "0.1.1", author = "Alexander Wang", author_email = "alexanderw@berkeley.edu", description = ("PyScribe makes print debugging easier and more efficient"), license = "MIT", keywords = "python pyscribe debug print", url = "https://github.com/alixander/pyscribe", download_url = "https://github.com/alixander/pyscribe/tarbell/0.1.1", entry_points={ 'console_scripts': [ 'pyscribe = pyscribe.pyscribe:main', ], }, packages=['pyscribe'], classifiers=[ "Development Status :: 3 - Alpha" ], ) Add classifiers since python3 now supported
import os from setuptools import setup setup( name = "pyscribe", version = "0.1.1", author = "Alexander Wang", author_email = "alexanderw@berkeley.edu", description = ("PyScribe makes print debugging easier and more efficient"), license = "MIT", keywords = "python pyscribe debug print", url = "https://github.com/alixander/pyscribe", download_url = "https://github.com/alixander/pyscribe/tarbell/0.1.2", entry_points={ 'console_scripts': [ 'pyscribe = pyscribe.pyscribe:main', ], }, packages=['pyscribe'], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', ], )
<commit_before>import os from setuptools import setup setup( name = "pyscribe", version = "0.1.1", author = "Alexander Wang", author_email = "alexanderw@berkeley.edu", description = ("PyScribe makes print debugging easier and more efficient"), license = "MIT", keywords = "python pyscribe debug print", url = "https://github.com/alixander/pyscribe", download_url = "https://github.com/alixander/pyscribe/tarbell/0.1.1", entry_points={ 'console_scripts': [ 'pyscribe = pyscribe.pyscribe:main', ], }, packages=['pyscribe'], classifiers=[ "Development Status :: 3 - Alpha" ], ) <commit_msg>Add classifiers since python3 now supported<commit_after>
import os from setuptools import setup setup( name = "pyscribe", version = "0.1.1", author = "Alexander Wang", author_email = "alexanderw@berkeley.edu", description = ("PyScribe makes print debugging easier and more efficient"), license = "MIT", keywords = "python pyscribe debug print", url = "https://github.com/alixander/pyscribe", download_url = "https://github.com/alixander/pyscribe/tarbell/0.1.2", entry_points={ 'console_scripts': [ 'pyscribe = pyscribe.pyscribe:main', ], }, packages=['pyscribe'], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', ], )
import os from setuptools import setup setup( name = "pyscribe", version = "0.1.1", author = "Alexander Wang", author_email = "alexanderw@berkeley.edu", description = ("PyScribe makes print debugging easier and more efficient"), license = "MIT", keywords = "python pyscribe debug print", url = "https://github.com/alixander/pyscribe", download_url = "https://github.com/alixander/pyscribe/tarbell/0.1.1", entry_points={ 'console_scripts': [ 'pyscribe = pyscribe.pyscribe:main', ], }, packages=['pyscribe'], classifiers=[ "Development Status :: 3 - Alpha" ], ) Add classifiers since python3 now supportedimport os from setuptools import setup setup( name = "pyscribe", version = "0.1.1", author = "Alexander Wang", author_email = "alexanderw@berkeley.edu", description = ("PyScribe makes print debugging easier and more efficient"), license = "MIT", keywords = "python pyscribe debug print", url = "https://github.com/alixander/pyscribe", download_url = "https://github.com/alixander/pyscribe/tarbell/0.1.2", entry_points={ 'console_scripts': [ 'pyscribe = pyscribe.pyscribe:main', ], }, packages=['pyscribe'], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', ], )
<commit_before>import os from setuptools import setup setup( name = "pyscribe", version = "0.1.1", author = "Alexander Wang", author_email = "alexanderw@berkeley.edu", description = ("PyScribe makes print debugging easier and more efficient"), license = "MIT", keywords = "python pyscribe debug print", url = "https://github.com/alixander/pyscribe", download_url = "https://github.com/alixander/pyscribe/tarbell/0.1.1", entry_points={ 'console_scripts': [ 'pyscribe = pyscribe.pyscribe:main', ], }, packages=['pyscribe'], classifiers=[ "Development Status :: 3 - Alpha" ], ) <commit_msg>Add classifiers since python3 now supported<commit_after>import os from setuptools import setup setup( name = "pyscribe", version = "0.1.1", author = "Alexander Wang", author_email = "alexanderw@berkeley.edu", description = ("PyScribe makes print debugging easier and more efficient"), license = "MIT", keywords = "python pyscribe debug print", url = "https://github.com/alixander/pyscribe", download_url = "https://github.com/alixander/pyscribe/tarbell/0.1.2", entry_points={ 'console_scripts': [ 'pyscribe = pyscribe.pyscribe:main', ], }, packages=['pyscribe'], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', ], )
9cb532115f44b25101efa7d448328e35ca51c37c
setup.py
setup.py
from setuptools import setup, find_packages from os import path here = path.abspath(path.dirname(__file__)) setup( name='publish-aws-lambda', version='0.3', description='Publish a Python module as a set of AWS lambda functions', url='https://github.com/ophirh/publish-aws-lambda', author='Ophir', author_email='opensource@itculate.io', license='MIT', keywords=['aws', 'lambda', 'publish'], packages=find_packages(), )
from setuptools import setup, find_packages from os import path here = path.abspath(path.dirname(__file__)) setup( name='publish-aws-lambda', version='0.3.1', description='Publish a Python module as a set of AWS lambda functions', url='https://github.com/ophirh/publish-aws-lambda', author='Ophir', author_email='opensource@itculate.io', license='MIT', keywords=['aws', 'lambda', 'publish'], packages=find_packages(), )
Fix issue with deleting boto3 from ZIP file
Fix issue with deleting boto3 from ZIP file
Python
mit
ophirh/publish-aws-lambda
from setuptools import setup, find_packages from os import path here = path.abspath(path.dirname(__file__)) setup( name='publish-aws-lambda', version='0.3', description='Publish a Python module as a set of AWS lambda functions', url='https://github.com/ophirh/publish-aws-lambda', author='Ophir', author_email='opensource@itculate.io', license='MIT', keywords=['aws', 'lambda', 'publish'], packages=find_packages(), ) Fix issue with deleting boto3 from ZIP file
from setuptools import setup, find_packages from os import path here = path.abspath(path.dirname(__file__)) setup( name='publish-aws-lambda', version='0.3.1', description='Publish a Python module as a set of AWS lambda functions', url='https://github.com/ophirh/publish-aws-lambda', author='Ophir', author_email='opensource@itculate.io', license='MIT', keywords=['aws', 'lambda', 'publish'], packages=find_packages(), )
<commit_before>from setuptools import setup, find_packages from os import path here = path.abspath(path.dirname(__file__)) setup( name='publish-aws-lambda', version='0.3', description='Publish a Python module as a set of AWS lambda functions', url='https://github.com/ophirh/publish-aws-lambda', author='Ophir', author_email='opensource@itculate.io', license='MIT', keywords=['aws', 'lambda', 'publish'], packages=find_packages(), ) <commit_msg>Fix issue with deleting boto3 from ZIP file<commit_after>
from setuptools import setup, find_packages from os import path here = path.abspath(path.dirname(__file__)) setup( name='publish-aws-lambda', version='0.3.1', description='Publish a Python module as a set of AWS lambda functions', url='https://github.com/ophirh/publish-aws-lambda', author='Ophir', author_email='opensource@itculate.io', license='MIT', keywords=['aws', 'lambda', 'publish'], packages=find_packages(), )
from setuptools import setup, find_packages from os import path here = path.abspath(path.dirname(__file__)) setup( name='publish-aws-lambda', version='0.3', description='Publish a Python module as a set of AWS lambda functions', url='https://github.com/ophirh/publish-aws-lambda', author='Ophir', author_email='opensource@itculate.io', license='MIT', keywords=['aws', 'lambda', 'publish'], packages=find_packages(), ) Fix issue with deleting boto3 from ZIP filefrom setuptools import setup, find_packages from os import path here = path.abspath(path.dirname(__file__)) setup( name='publish-aws-lambda', version='0.3.1', description='Publish a Python module as a set of AWS lambda functions', url='https://github.com/ophirh/publish-aws-lambda', author='Ophir', author_email='opensource@itculate.io', license='MIT', keywords=['aws', 'lambda', 'publish'], packages=find_packages(), )
<commit_before>from setuptools import setup, find_packages from os import path here = path.abspath(path.dirname(__file__)) setup( name='publish-aws-lambda', version='0.3', description='Publish a Python module as a set of AWS lambda functions', url='https://github.com/ophirh/publish-aws-lambda', author='Ophir', author_email='opensource@itculate.io', license='MIT', keywords=['aws', 'lambda', 'publish'], packages=find_packages(), ) <commit_msg>Fix issue with deleting boto3 from ZIP file<commit_after>from setuptools import setup, find_packages from os import path here = path.abspath(path.dirname(__file__)) setup( name='publish-aws-lambda', version='0.3.1', description='Publish a Python module as a set of AWS lambda functions', url='https://github.com/ophirh/publish-aws-lambda', author='Ophir', author_email='opensource@itculate.io', license='MIT', keywords=['aws', 'lambda', 'publish'], packages=find_packages(), )
f80ef1738608ececd6fb2b6a034ca9263f06ae17
setup.py
setup.py
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-ical', version='1.4', description="iCal feeds for Django based on Django's syndication feed " "framework.", long_description=(open('README.rst').read() + '\n' + open('CHANGES.rst').read()), author='Ian Lewis', author_email='IanMLewis@gmail.com', license='MIT License', url='https://github.com/Pinkerton/django-ical', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Plugins', 'Framework :: Django', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', ], install_requires=[ 'Django>=1.3.4', 'icalendar>=3.1', ], packages=find_packages(), test_suite='tests.main', )
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-ical', version='1.4', description="iCal feeds for Django based on Django's syndication feed " "framework.", long_description=(open('README.rst').read() + '\n' + open('CHANGES.rst').read()), author='Ian Lewis', author_email='IanMLewis@gmail.com', license='MIT License', url='https://github.com/Pinkerton/django-ical', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Plugins', 'Framework :: Django', 'Framework :: Django :: 1.3', 'Framework :: Django :: 1.4', 'Framework :: Django :: 1.5', 'Framework :: Django :: 1.6', 'Framework :: Django :: 1.7', 'Framework :: Django :: 1.8', 'Framework :: Django :: 1.9', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Software Development :: Libraries :: Python Modules', ], install_requires=[ 'Django>=1.3.4', 'icalendar>=3.1', ], packages=find_packages(), test_suite='tests.main', )
Add Python & Django classifiers
Add Python & Django classifiers
Python
mit
Pinkerton/django-ical
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-ical', version='1.4', description="iCal feeds for Django based on Django's syndication feed " "framework.", long_description=(open('README.rst').read() + '\n' + open('CHANGES.rst').read()), author='Ian Lewis', author_email='IanMLewis@gmail.com', license='MIT License', url='https://github.com/Pinkerton/django-ical', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Plugins', 'Framework :: Django', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', ], install_requires=[ 'Django>=1.3.4', 'icalendar>=3.1', ], packages=find_packages(), test_suite='tests.main', ) Add Python & Django classifiers
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-ical', version='1.4', description="iCal feeds for Django based on Django's syndication feed " "framework.", long_description=(open('README.rst').read() + '\n' + open('CHANGES.rst').read()), author='Ian Lewis', author_email='IanMLewis@gmail.com', license='MIT License', url='https://github.com/Pinkerton/django-ical', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Plugins', 'Framework :: Django', 'Framework :: Django :: 1.3', 'Framework :: Django :: 1.4', 'Framework :: Django :: 1.5', 'Framework :: Django :: 1.6', 'Framework :: Django :: 1.7', 'Framework :: Django :: 1.8', 'Framework :: Django :: 1.9', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Software Development :: Libraries :: Python Modules', ], install_requires=[ 'Django>=1.3.4', 'icalendar>=3.1', ], packages=find_packages(), test_suite='tests.main', )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-ical', version='1.4', description="iCal feeds for Django based on Django's syndication feed " "framework.", long_description=(open('README.rst').read() + '\n' + open('CHANGES.rst').read()), author='Ian Lewis', author_email='IanMLewis@gmail.com', license='MIT License', url='https://github.com/Pinkerton/django-ical', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Plugins', 'Framework :: Django', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', ], install_requires=[ 'Django>=1.3.4', 'icalendar>=3.1', ], packages=find_packages(), test_suite='tests.main', ) <commit_msg>Add Python & Django classifiers<commit_after>
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-ical', version='1.4', description="iCal feeds for Django based on Django's syndication feed " "framework.", long_description=(open('README.rst').read() + '\n' + open('CHANGES.rst').read()), author='Ian Lewis', author_email='IanMLewis@gmail.com', license='MIT License', url='https://github.com/Pinkerton/django-ical', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Plugins', 'Framework :: Django', 'Framework :: Django :: 1.3', 'Framework :: Django :: 1.4', 'Framework :: Django :: 1.5', 'Framework :: Django :: 1.6', 'Framework :: Django :: 1.7', 'Framework :: Django :: 1.8', 'Framework :: Django :: 1.9', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Software Development :: Libraries :: Python Modules', ], install_requires=[ 'Django>=1.3.4', 'icalendar>=3.1', ], packages=find_packages(), test_suite='tests.main', )
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-ical', version='1.4', description="iCal feeds for Django based on Django's syndication feed " "framework.", long_description=(open('README.rst').read() + '\n' + open('CHANGES.rst').read()), author='Ian Lewis', author_email='IanMLewis@gmail.com', license='MIT License', url='https://github.com/Pinkerton/django-ical', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Plugins', 'Framework :: Django', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', ], install_requires=[ 'Django>=1.3.4', 'icalendar>=3.1', ], packages=find_packages(), test_suite='tests.main', ) Add Python & Django classifiers#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-ical', version='1.4', description="iCal feeds for Django based on Django's syndication feed " "framework.", long_description=(open('README.rst').read() + '\n' + open('CHANGES.rst').read()), author='Ian Lewis', author_email='IanMLewis@gmail.com', license='MIT License', url='https://github.com/Pinkerton/django-ical', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Plugins', 'Framework :: Django', 'Framework :: Django :: 1.3', 'Framework :: Django :: 1.4', 'Framework :: Django :: 1.5', 'Framework :: Django :: 1.6', 'Framework :: Django :: 1.7', 'Framework :: Django :: 1.8', 'Framework :: Django :: 1.9', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Software Development :: Libraries :: Python Modules', ], install_requires=[ 'Django>=1.3.4', 'icalendar>=3.1', ], packages=find_packages(), test_suite='tests.main', )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-ical', version='1.4', description="iCal feeds for Django based on Django's syndication feed " "framework.", long_description=(open('README.rst').read() + '\n' + open('CHANGES.rst').read()), author='Ian Lewis', author_email='IanMLewis@gmail.com', license='MIT License', url='https://github.com/Pinkerton/django-ical', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Plugins', 'Framework :: Django', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Topic :: Software Development :: Libraries :: Python Modules', ], install_requires=[ 'Django>=1.3.4', 'icalendar>=3.1', ], packages=find_packages(), test_suite='tests.main', ) <commit_msg>Add Python & Django classifiers<commit_after>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-ical', version='1.4', description="iCal feeds for Django based on Django's syndication feed " "framework.", long_description=(open('README.rst').read() + '\n' + open('CHANGES.rst').read()), author='Ian Lewis', author_email='IanMLewis@gmail.com', license='MIT License', url='https://github.com/Pinkerton/django-ical', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Plugins', 'Framework :: Django', 'Framework :: Django :: 1.3', 'Framework :: Django :: 1.4', 'Framework :: Django :: 1.5', 'Framework :: Django :: 1.6', 'Framework :: Django :: 1.7', 'Framework :: Django :: 1.8', 'Framework :: Django :: 1.9', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Topic :: Software Development :: Libraries :: Python Modules', ], install_requires=[ 'Django>=1.3.4', 'icalendar>=3.1', ], packages=find_packages(), test_suite='tests.main', )
535b9f95c89974714eacece9bf5109fd1ec5116f
setup.py
setup.py
from setuptools import setup, find_packages with open('requirements.txt', 'r') as f: requirements = f.readlines() setup( name='resultsdb-updater', version='3.0.0', description=('A micro-service that listens for messages on the message ' 'bus and updates ResultsDB'), license='GPLv2+', author='Matt Prahl', author_email='mprahl@redhat.com', url='https://github.com/release-engineering/resultsdb-updater', install_requires=requirements, packages=find_packages(), include_data=True, entry_points=""" [moksha.consumer] ciconsumer = resultsdbupdater.consumer:CIConsumer """, )
from setuptools import setup, find_packages with open('requirements.txt', 'r') as f: requirements = f.readlines() setup( name='resultsdb-updater', version='3.0.0', description=('A micro-service that listens for messages on the message ' 'bus and updates ResultsDB'), license='GPLv2+', author='Matt Prahl', author_email='mprahl@redhat.com', url='https://github.com/release-engineering/resultsdb-updater', install_requires=requirements, packages=find_packages(), entry_points=""" [moksha.consumer] ciconsumer = resultsdbupdater.consumer:CIConsumer """, )
Resolve warning "Unknown distribution option: 'include_data'"
Resolve warning "Unknown distribution option: 'include_data'"
Python
lgpl-2.1
release-engineering/resultsdb-updater,release-engineering/resultsdb-updater
from setuptools import setup, find_packages with open('requirements.txt', 'r') as f: requirements = f.readlines() setup( name='resultsdb-updater', version='3.0.0', description=('A micro-service that listens for messages on the message ' 'bus and updates ResultsDB'), license='GPLv2+', author='Matt Prahl', author_email='mprahl@redhat.com', url='https://github.com/release-engineering/resultsdb-updater', install_requires=requirements, packages=find_packages(), include_data=True, entry_points=""" [moksha.consumer] ciconsumer = resultsdbupdater.consumer:CIConsumer """, ) Resolve warning "Unknown distribution option: 'include_data'"
from setuptools import setup, find_packages with open('requirements.txt', 'r') as f: requirements = f.readlines() setup( name='resultsdb-updater', version='3.0.0', description=('A micro-service that listens for messages on the message ' 'bus and updates ResultsDB'), license='GPLv2+', author='Matt Prahl', author_email='mprahl@redhat.com', url='https://github.com/release-engineering/resultsdb-updater', install_requires=requirements, packages=find_packages(), entry_points=""" [moksha.consumer] ciconsumer = resultsdbupdater.consumer:CIConsumer """, )
<commit_before>from setuptools import setup, find_packages with open('requirements.txt', 'r') as f: requirements = f.readlines() setup( name='resultsdb-updater', version='3.0.0', description=('A micro-service that listens for messages on the message ' 'bus and updates ResultsDB'), license='GPLv2+', author='Matt Prahl', author_email='mprahl@redhat.com', url='https://github.com/release-engineering/resultsdb-updater', install_requires=requirements, packages=find_packages(), include_data=True, entry_points=""" [moksha.consumer] ciconsumer = resultsdbupdater.consumer:CIConsumer """, ) <commit_msg>Resolve warning "Unknown distribution option: 'include_data'"<commit_after>
from setuptools import setup, find_packages with open('requirements.txt', 'r') as f: requirements = f.readlines() setup( name='resultsdb-updater', version='3.0.0', description=('A micro-service that listens for messages on the message ' 'bus and updates ResultsDB'), license='GPLv2+', author='Matt Prahl', author_email='mprahl@redhat.com', url='https://github.com/release-engineering/resultsdb-updater', install_requires=requirements, packages=find_packages(), entry_points=""" [moksha.consumer] ciconsumer = resultsdbupdater.consumer:CIConsumer """, )
from setuptools import setup, find_packages with open('requirements.txt', 'r') as f: requirements = f.readlines() setup( name='resultsdb-updater', version='3.0.0', description=('A micro-service that listens for messages on the message ' 'bus and updates ResultsDB'), license='GPLv2+', author='Matt Prahl', author_email='mprahl@redhat.com', url='https://github.com/release-engineering/resultsdb-updater', install_requires=requirements, packages=find_packages(), include_data=True, entry_points=""" [moksha.consumer] ciconsumer = resultsdbupdater.consumer:CIConsumer """, ) Resolve warning "Unknown distribution option: 'include_data'"from setuptools import setup, find_packages with open('requirements.txt', 'r') as f: requirements = f.readlines() setup( name='resultsdb-updater', version='3.0.0', description=('A micro-service that listens for messages on the message ' 'bus and updates ResultsDB'), license='GPLv2+', author='Matt Prahl', author_email='mprahl@redhat.com', url='https://github.com/release-engineering/resultsdb-updater', install_requires=requirements, packages=find_packages(), entry_points=""" [moksha.consumer] ciconsumer = resultsdbupdater.consumer:CIConsumer """, )
<commit_before>from setuptools import setup, find_packages with open('requirements.txt', 'r') as f: requirements = f.readlines() setup( name='resultsdb-updater', version='3.0.0', description=('A micro-service that listens for messages on the message ' 'bus and updates ResultsDB'), license='GPLv2+', author='Matt Prahl', author_email='mprahl@redhat.com', url='https://github.com/release-engineering/resultsdb-updater', install_requires=requirements, packages=find_packages(), include_data=True, entry_points=""" [moksha.consumer] ciconsumer = resultsdbupdater.consumer:CIConsumer """, ) <commit_msg>Resolve warning "Unknown distribution option: 'include_data'"<commit_after>from setuptools import setup, find_packages with open('requirements.txt', 'r') as f: requirements = f.readlines() setup( name='resultsdb-updater', version='3.0.0', description=('A micro-service that listens for messages on the message ' 'bus and updates ResultsDB'), license='GPLv2+', author='Matt Prahl', author_email='mprahl@redhat.com', url='https://github.com/release-engineering/resultsdb-updater', install_requires=requirements, packages=find_packages(), entry_points=""" [moksha.consumer] ciconsumer = resultsdbupdater.consumer:CIConsumer """, )
a5c0f5c46c64e56e0a4a0791b86b820e8ed0241b
setup.py
setup.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup setup(name='sophiabus230', version='0.4', description='Module to get the timetable of the Sophia Antipolis bus line 230', url='http://github.com/paraita/sophiabus230', author='Paraita Wohler', author_email='paraita.wohler@gmail.com', license='MIT', packages=['sophiabus230'], install_requires=[ 'beautifulsoup4', 'python-dateutil', 'future' ], test_suite='nose.collector', tests_require=[ 'mock', 'nose', 'coverage', 'coveralls' ], zip_safe=False)
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup setup(name='sophiabus230', version='0.5', description='Module to get the timetable of the Sophia Antipolis bus line 230', url='http://github.com/paraita/sophiabus230', author='Paraita Wohler', author_email='paraita.wohler@gmail.com', license='MIT', packages=['sophiabus230'], install_requires=[ 'beautifulsoup4', 'python-dateutil', 'future' ], test_suite='nose.collector', tests_require=[ 'mock', 'nose', 'coverage', 'coveralls' ], zip_safe=False)
Update package version for Pypi
Update package version for Pypi
Python
mit
paraita/sophiabus230,paraita/sophiabus230
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup setup(name='sophiabus230', version='0.4', description='Module to get the timetable of the Sophia Antipolis bus line 230', url='http://github.com/paraita/sophiabus230', author='Paraita Wohler', author_email='paraita.wohler@gmail.com', license='MIT', packages=['sophiabus230'], install_requires=[ 'beautifulsoup4', 'python-dateutil', 'future' ], test_suite='nose.collector', tests_require=[ 'mock', 'nose', 'coverage', 'coveralls' ], zip_safe=False) Update package version for Pypi
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup setup(name='sophiabus230', version='0.5', description='Module to get the timetable of the Sophia Antipolis bus line 230', url='http://github.com/paraita/sophiabus230', author='Paraita Wohler', author_email='paraita.wohler@gmail.com', license='MIT', packages=['sophiabus230'], install_requires=[ 'beautifulsoup4', 'python-dateutil', 'future' ], test_suite='nose.collector', tests_require=[ 'mock', 'nose', 'coverage', 'coveralls' ], zip_safe=False)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup setup(name='sophiabus230', version='0.4', description='Module to get the timetable of the Sophia Antipolis bus line 230', url='http://github.com/paraita/sophiabus230', author='Paraita Wohler', author_email='paraita.wohler@gmail.com', license='MIT', packages=['sophiabus230'], install_requires=[ 'beautifulsoup4', 'python-dateutil', 'future' ], test_suite='nose.collector', tests_require=[ 'mock', 'nose', 'coverage', 'coveralls' ], zip_safe=False) <commit_msg>Update package version for Pypi<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup setup(name='sophiabus230', version='0.5', description='Module to get the timetable of the Sophia Antipolis bus line 230', url='http://github.com/paraita/sophiabus230', author='Paraita Wohler', author_email='paraita.wohler@gmail.com', license='MIT', packages=['sophiabus230'], install_requires=[ 'beautifulsoup4', 'python-dateutil', 'future' ], test_suite='nose.collector', tests_require=[ 'mock', 'nose', 'coverage', 'coveralls' ], zip_safe=False)
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup setup(name='sophiabus230', version='0.4', description='Module to get the timetable of the Sophia Antipolis bus line 230', url='http://github.com/paraita/sophiabus230', author='Paraita Wohler', author_email='paraita.wohler@gmail.com', license='MIT', packages=['sophiabus230'], install_requires=[ 'beautifulsoup4', 'python-dateutil', 'future' ], test_suite='nose.collector', tests_require=[ 'mock', 'nose', 'coverage', 'coveralls' ], zip_safe=False) Update package version for Pypi#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup setup(name='sophiabus230', version='0.5', description='Module to get the timetable of the Sophia Antipolis bus line 230', url='http://github.com/paraita/sophiabus230', author='Paraita Wohler', author_email='paraita.wohler@gmail.com', license='MIT', packages=['sophiabus230'], install_requires=[ 'beautifulsoup4', 'python-dateutil', 'future' ], test_suite='nose.collector', tests_require=[ 'mock', 'nose', 'coverage', 'coveralls' ], zip_safe=False)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup setup(name='sophiabus230', version='0.4', description='Module to get the timetable of the Sophia Antipolis bus line 230', url='http://github.com/paraita/sophiabus230', author='Paraita Wohler', author_email='paraita.wohler@gmail.com', license='MIT', packages=['sophiabus230'], install_requires=[ 'beautifulsoup4', 'python-dateutil', 'future' ], test_suite='nose.collector', tests_require=[ 'mock', 'nose', 'coverage', 'coveralls' ], zip_safe=False) <commit_msg>Update package version for Pypi<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup setup(name='sophiabus230', version='0.5', description='Module to get the timetable of the Sophia Antipolis bus line 230', url='http://github.com/paraita/sophiabus230', author='Paraita Wohler', author_email='paraita.wohler@gmail.com', license='MIT', packages=['sophiabus230'], install_requires=[ 'beautifulsoup4', 'python-dateutil', 'future' ], test_suite='nose.collector', tests_require=[ 'mock', 'nose', 'coverage', 'coveralls' ], zip_safe=False)
d96b8f237e79c924865b429b5e8a725630537d2b
setup.py
setup.py
from setuptools import setup with open('README.rst') as README: long_description = README.read() long_description = long_description[long_description.index('Description'):] setup(name='wos', version='0.1.2', description='Web of Science client using API v3.', long_description=long_description, install_requires=['suds'], url='http://github.com/enricobacis/wos', author='Enrico Bacis', author_email='enrico.bacis@gmail.com', license='MIT', packages=['wos'], scripts=['scripts/wos'], keywords='wos isi web of science knowledge api client' )
from setuptools import setup with open('README.rst') as README: long_description = README.read() long_description = long_description[long_description.index('Description'):] setup(name='wos', version='0.1.2', description='Web of Science client using API v3.', long_description=long_description, extras_require={ ':python_version <= "2.7"': ['suds'], ':python_version >= "3.0"': ['suds-jurko'], }, url='http://github.com/enricobacis/wos', author='Enrico Bacis', author_email='enrico.bacis@gmail.com', license='MIT', packages=['wos'], scripts=['scripts/wos'], keywords='wos isi web of science knowledge api client' )
Use suds-jurko for python 3
Use suds-jurko for python 3
Python
mit
enricobacis/wos
from setuptools import setup with open('README.rst') as README: long_description = README.read() long_description = long_description[long_description.index('Description'):] setup(name='wos', version='0.1.2', description='Web of Science client using API v3.', long_description=long_description, install_requires=['suds'], url='http://github.com/enricobacis/wos', author='Enrico Bacis', author_email='enrico.bacis@gmail.com', license='MIT', packages=['wos'], scripts=['scripts/wos'], keywords='wos isi web of science knowledge api client' ) Use suds-jurko for python 3
from setuptools import setup with open('README.rst') as README: long_description = README.read() long_description = long_description[long_description.index('Description'):] setup(name='wos', version='0.1.2', description='Web of Science client using API v3.', long_description=long_description, extras_require={ ':python_version <= "2.7"': ['suds'], ':python_version >= "3.0"': ['suds-jurko'], }, url='http://github.com/enricobacis/wos', author='Enrico Bacis', author_email='enrico.bacis@gmail.com', license='MIT', packages=['wos'], scripts=['scripts/wos'], keywords='wos isi web of science knowledge api client' )
<commit_before>from setuptools import setup with open('README.rst') as README: long_description = README.read() long_description = long_description[long_description.index('Description'):] setup(name='wos', version='0.1.2', description='Web of Science client using API v3.', long_description=long_description, install_requires=['suds'], url='http://github.com/enricobacis/wos', author='Enrico Bacis', author_email='enrico.bacis@gmail.com', license='MIT', packages=['wos'], scripts=['scripts/wos'], keywords='wos isi web of science knowledge api client' ) <commit_msg>Use suds-jurko for python 3<commit_after>
from setuptools import setup with open('README.rst') as README: long_description = README.read() long_description = long_description[long_description.index('Description'):] setup(name='wos', version='0.1.2', description='Web of Science client using API v3.', long_description=long_description, extras_require={ ':python_version <= "2.7"': ['suds'], ':python_version >= "3.0"': ['suds-jurko'], }, url='http://github.com/enricobacis/wos', author='Enrico Bacis', author_email='enrico.bacis@gmail.com', license='MIT', packages=['wos'], scripts=['scripts/wos'], keywords='wos isi web of science knowledge api client' )
from setuptools import setup with open('README.rst') as README: long_description = README.read() long_description = long_description[long_description.index('Description'):] setup(name='wos', version='0.1.2', description='Web of Science client using API v3.', long_description=long_description, install_requires=['suds'], url='http://github.com/enricobacis/wos', author='Enrico Bacis', author_email='enrico.bacis@gmail.com', license='MIT', packages=['wos'], scripts=['scripts/wos'], keywords='wos isi web of science knowledge api client' ) Use suds-jurko for python 3from setuptools import setup with open('README.rst') as README: long_description = README.read() long_description = long_description[long_description.index('Description'):] setup(name='wos', version='0.1.2', description='Web of Science client using API v3.', long_description=long_description, extras_require={ ':python_version <= "2.7"': ['suds'], ':python_version >= "3.0"': ['suds-jurko'], }, url='http://github.com/enricobacis/wos', author='Enrico Bacis', author_email='enrico.bacis@gmail.com', license='MIT', packages=['wos'], scripts=['scripts/wos'], keywords='wos isi web of science knowledge api client' )
<commit_before>from setuptools import setup with open('README.rst') as README: long_description = README.read() long_description = long_description[long_description.index('Description'):] setup(name='wos', version='0.1.2', description='Web of Science client using API v3.', long_description=long_description, install_requires=['suds'], url='http://github.com/enricobacis/wos', author='Enrico Bacis', author_email='enrico.bacis@gmail.com', license='MIT', packages=['wos'], scripts=['scripts/wos'], keywords='wos isi web of science knowledge api client' ) <commit_msg>Use suds-jurko for python 3<commit_after>from setuptools import setup with open('README.rst') as README: long_description = README.read() long_description = long_description[long_description.index('Description'):] setup(name='wos', version='0.1.2', description='Web of Science client using API v3.', long_description=long_description, extras_require={ ':python_version <= "2.7"': ['suds'], ':python_version >= "3.0"': ['suds-jurko'], }, url='http://github.com/enricobacis/wos', author='Enrico Bacis', author_email='enrico.bacis@gmail.com', license='MIT', packages=['wos'], scripts=['scripts/wos'], keywords='wos isi web of science knowledge api client' )
8ccc50c222163b76e2284438ab3779de13ba5c6d
setup.py
setup.py
# -*- coding: utf-8 -*- from os import path from setuptools import find_packages, setup README_rst = path.join(path.abspath(path.dirname(__file__)), 'README.rst') with open(README_rst, 'r') as f: long_description = f.read() setup( name="pyee", vcversioner={}, packages=find_packages(), setup_requires=[ 'pytest-runner', 'pytest-asyncio;python_version>"3.4"', 'vcversioner' ], tests_require=['twisted'], include_package_data=True, description="A port of node.js's EventEmitter to python.", long_description=long_description, author="Joshua Holbrook", author_email="josh.holbrook@gmail.com", url="https://github.com/jfhbrook/pyee", keywords=[ "events", "emitter", "node.js", "node", "eventemitter", "event_emitter" ], classifiers=[ "Programming Language :: Python", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Topic :: Other/Nonlisted Topic" ] )
# -*- coding: utf-8 -*- from os import path from setuptools import find_packages, setup README_rst = path.join(path.abspath(path.dirname(__file__)), 'README.rst') with open(README_rst, 'r') as f: long_description = f.read() setup( name="pyee", vcversioner={}, packages=find_packages(), setup_requires=[ 'pytest-runner', 'pytest-asyncio; python_version > "3.4"', 'vcversioner' ], tests_require=['twisted'], include_package_data=True, description="A port of node.js's EventEmitter to python.", long_description=long_description, author="Joshua Holbrook", author_email="josh.holbrook@gmail.com", url="https://github.com/jfhbrook/pyee", keywords=[ "events", "emitter", "node.js", "node", "eventemitter", "event_emitter" ], classifiers=[ "Programming Language :: Python", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Topic :: Other/Nonlisted Topic" ] )
Add whitespace to version spec?
Add whitespace to version spec?
Python
mit
jfhbrook/pyee
# -*- coding: utf-8 -*- from os import path from setuptools import find_packages, setup README_rst = path.join(path.abspath(path.dirname(__file__)), 'README.rst') with open(README_rst, 'r') as f: long_description = f.read() setup( name="pyee", vcversioner={}, packages=find_packages(), setup_requires=[ 'pytest-runner', 'pytest-asyncio;python_version>"3.4"', 'vcversioner' ], tests_require=['twisted'], include_package_data=True, description="A port of node.js's EventEmitter to python.", long_description=long_description, author="Joshua Holbrook", author_email="josh.holbrook@gmail.com", url="https://github.com/jfhbrook/pyee", keywords=[ "events", "emitter", "node.js", "node", "eventemitter", "event_emitter" ], classifiers=[ "Programming Language :: Python", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Topic :: Other/Nonlisted Topic" ] ) Add whitespace to version spec?
# -*- coding: utf-8 -*- from os import path from setuptools import find_packages, setup README_rst = path.join(path.abspath(path.dirname(__file__)), 'README.rst') with open(README_rst, 'r') as f: long_description = f.read() setup( name="pyee", vcversioner={}, packages=find_packages(), setup_requires=[ 'pytest-runner', 'pytest-asyncio; python_version > "3.4"', 'vcversioner' ], tests_require=['twisted'], include_package_data=True, description="A port of node.js's EventEmitter to python.", long_description=long_description, author="Joshua Holbrook", author_email="josh.holbrook@gmail.com", url="https://github.com/jfhbrook/pyee", keywords=[ "events", "emitter", "node.js", "node", "eventemitter", "event_emitter" ], classifiers=[ "Programming Language :: Python", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Topic :: Other/Nonlisted Topic" ] )
<commit_before># -*- coding: utf-8 -*- from os import path from setuptools import find_packages, setup README_rst = path.join(path.abspath(path.dirname(__file__)), 'README.rst') with open(README_rst, 'r') as f: long_description = f.read() setup( name="pyee", vcversioner={}, packages=find_packages(), setup_requires=[ 'pytest-runner', 'pytest-asyncio;python_version>"3.4"', 'vcversioner' ], tests_require=['twisted'], include_package_data=True, description="A port of node.js's EventEmitter to python.", long_description=long_description, author="Joshua Holbrook", author_email="josh.holbrook@gmail.com", url="https://github.com/jfhbrook/pyee", keywords=[ "events", "emitter", "node.js", "node", "eventemitter", "event_emitter" ], classifiers=[ "Programming Language :: Python", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Topic :: Other/Nonlisted Topic" ] ) <commit_msg>Add whitespace to version spec?<commit_after>
# -*- coding: utf-8 -*- from os import path from setuptools import find_packages, setup README_rst = path.join(path.abspath(path.dirname(__file__)), 'README.rst') with open(README_rst, 'r') as f: long_description = f.read() setup( name="pyee", vcversioner={}, packages=find_packages(), setup_requires=[ 'pytest-runner', 'pytest-asyncio; python_version > "3.4"', 'vcversioner' ], tests_require=['twisted'], include_package_data=True, description="A port of node.js's EventEmitter to python.", long_description=long_description, author="Joshua Holbrook", author_email="josh.holbrook@gmail.com", url="https://github.com/jfhbrook/pyee", keywords=[ "events", "emitter", "node.js", "node", "eventemitter", "event_emitter" ], classifiers=[ "Programming Language :: Python", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Topic :: Other/Nonlisted Topic" ] )
# -*- coding: utf-8 -*- from os import path from setuptools import find_packages, setup README_rst = path.join(path.abspath(path.dirname(__file__)), 'README.rst') with open(README_rst, 'r') as f: long_description = f.read() setup( name="pyee", vcversioner={}, packages=find_packages(), setup_requires=[ 'pytest-runner', 'pytest-asyncio;python_version>"3.4"', 'vcversioner' ], tests_require=['twisted'], include_package_data=True, description="A port of node.js's EventEmitter to python.", long_description=long_description, author="Joshua Holbrook", author_email="josh.holbrook@gmail.com", url="https://github.com/jfhbrook/pyee", keywords=[ "events", "emitter", "node.js", "node", "eventemitter", "event_emitter" ], classifiers=[ "Programming Language :: Python", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Topic :: Other/Nonlisted Topic" ] ) Add whitespace to version spec?# -*- coding: utf-8 -*- from os import path from setuptools import find_packages, setup README_rst = path.join(path.abspath(path.dirname(__file__)), 'README.rst') with open(README_rst, 'r') as f: long_description = f.read() setup( name="pyee", vcversioner={}, packages=find_packages(), setup_requires=[ 'pytest-runner', 'pytest-asyncio; python_version > "3.4"', 'vcversioner' ], tests_require=['twisted'], include_package_data=True, description="A port of node.js's EventEmitter to python.", long_description=long_description, author="Joshua Holbrook", author_email="josh.holbrook@gmail.com", url="https://github.com/jfhbrook/pyee", keywords=[ "events", "emitter", "node.js", "node", "eventemitter", "event_emitter" ], classifiers=[ "Programming Language :: Python", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Topic :: Other/Nonlisted Topic" ] )
<commit_before># -*- coding: utf-8 -*- from os import path from setuptools import find_packages, setup README_rst = path.join(path.abspath(path.dirname(__file__)), 'README.rst') with open(README_rst, 'r') as f: long_description = f.read() setup( name="pyee", vcversioner={}, packages=find_packages(), setup_requires=[ 'pytest-runner', 'pytest-asyncio;python_version>"3.4"', 'vcversioner' ], tests_require=['twisted'], include_package_data=True, description="A port of node.js's EventEmitter to python.", long_description=long_description, author="Joshua Holbrook", author_email="josh.holbrook@gmail.com", url="https://github.com/jfhbrook/pyee", keywords=[ "events", "emitter", "node.js", "node", "eventemitter", "event_emitter" ], classifiers=[ "Programming Language :: Python", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Topic :: Other/Nonlisted Topic" ] ) <commit_msg>Add whitespace to version spec?<commit_after># -*- coding: utf-8 -*- from os import path from setuptools import find_packages, setup README_rst = path.join(path.abspath(path.dirname(__file__)), 'README.rst') with open(README_rst, 'r') as f: long_description = f.read() setup( name="pyee", vcversioner={}, packages=find_packages(), setup_requires=[ 'pytest-runner', 'pytest-asyncio; python_version > "3.4"', 'vcversioner' ], tests_require=['twisted'], include_package_data=True, description="A port of node.js's EventEmitter to python.", long_description=long_description, author="Joshua Holbrook", author_email="josh.holbrook@gmail.com", url="https://github.com/jfhbrook/pyee", keywords=[ "events", "emitter", "node.js", "node", "eventemitter", "event_emitter" ], classifiers=[ "Programming Language :: Python", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Topic :: Other/Nonlisted Topic" ] )