commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
364731a5986a629f934ae8f82743385c5e4b7226
main.py
main.py
import argparse from microproxy import proxy from microproxy.viewer import log as log_viewer def proxy_handler(args): proxy.start_proxy_server(args.host, args.port, args.proxy_mode) def viewer_handler(args): if args.viewer_type == "log": log_viewer.start() def main(): parser = argparse.ArgumentParser(description="") subparser = parser.add_subparsers() proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server") proxy_parser.add_argument("--host", default="127.0.0.1") proxy_parser.add_argument("--port", type=int, default=5580) proxy_parser.add_argument("--proxy_mode", choices=["socks", "transparent"], default="socks") proxy_parser.set_defaults(func=proxy_handler) viewer_parser = subparser.add_parser("viewer", help="Open Viewer") viewer_parser.add_argument("--viewer_type", choices=["log"], default="log") viewer_parser.set_defaults(func=viewer_handler) args = parser.parse_args() args.func(args) if __name__ == "__main__": main()
import argparse from microproxy import proxy from microproxy.viewer import log as log_viewer def proxy_handler(args): proxy.start_proxy_server(args.host, args.port, args.proxy_mode) def viewer_handler(args): if args.viewer_mode == "log": log_viewer.start() def main(): parser = argparse.ArgumentParser(description="") subparser = parser.add_subparsers() proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server") proxy_parser.add_argument("--host", default="127.0.0.1") proxy_parser.add_argument("--port", type=int, default=5580) proxy_parser.add_argument("--proxy-mode", choices=["socks", "transparent"], default="socks") proxy_parser.set_defaults(func=proxy_handler) viewer_parser = subparser.add_parser("viewer", help="Open Viewer") viewer_parser.add_argument("--viewer-mode", choices=["log"], default="log") viewer_parser.set_defaults(func=viewer_handler) args = parser.parse_args() args.func(args) if __name__ == "__main__": main()
Change command option format. - proxy_mode into proxy-mode - viewer_type into viewer-mode
Change command option format. - proxy_mode into proxy-mode - viewer_type into viewer-mode
Python
mit
mike820324/microProxy,mike820324/microProxy
import argparse from microproxy import proxy from microproxy.viewer import log as log_viewer def proxy_handler(args): proxy.start_proxy_server(args.host, args.port, args.proxy_mode) def viewer_handler(args): if args.viewer_type == "log": log_viewer.start() def main(): parser = argparse.ArgumentParser(description="") subparser = parser.add_subparsers() proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server") proxy_parser.add_argument("--host", default="127.0.0.1") proxy_parser.add_argument("--port", type=int, default=5580) proxy_parser.add_argument("--proxy_mode", choices=["socks", "transparent"], default="socks") proxy_parser.set_defaults(func=proxy_handler) viewer_parser = subparser.add_parser("viewer", help="Open Viewer") viewer_parser.add_argument("--viewer_type", choices=["log"], default="log") viewer_parser.set_defaults(func=viewer_handler) args = parser.parse_args() args.func(args) if __name__ == "__main__": main() Change command option format. - proxy_mode into proxy-mode - viewer_type into viewer-mode
import argparse from microproxy import proxy from microproxy.viewer import log as log_viewer def proxy_handler(args): proxy.start_proxy_server(args.host, args.port, args.proxy_mode) def viewer_handler(args): if args.viewer_mode == "log": log_viewer.start() def main(): parser = argparse.ArgumentParser(description="") subparser = parser.add_subparsers() proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server") proxy_parser.add_argument("--host", default="127.0.0.1") proxy_parser.add_argument("--port", type=int, default=5580) proxy_parser.add_argument("--proxy-mode", choices=["socks", "transparent"], default="socks") proxy_parser.set_defaults(func=proxy_handler) viewer_parser = subparser.add_parser("viewer", help="Open Viewer") viewer_parser.add_argument("--viewer-mode", choices=["log"], default="log") viewer_parser.set_defaults(func=viewer_handler) args = parser.parse_args() args.func(args) if __name__ == "__main__": main()
<commit_before>import argparse from microproxy import proxy from microproxy.viewer import log as log_viewer def proxy_handler(args): proxy.start_proxy_server(args.host, args.port, args.proxy_mode) def viewer_handler(args): if args.viewer_type == "log": log_viewer.start() def main(): parser = argparse.ArgumentParser(description="") subparser = parser.add_subparsers() proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server") proxy_parser.add_argument("--host", default="127.0.0.1") proxy_parser.add_argument("--port", type=int, default=5580) proxy_parser.add_argument("--proxy_mode", choices=["socks", "transparent"], default="socks") proxy_parser.set_defaults(func=proxy_handler) viewer_parser = subparser.add_parser("viewer", help="Open Viewer") viewer_parser.add_argument("--viewer_type", choices=["log"], default="log") viewer_parser.set_defaults(func=viewer_handler) args = parser.parse_args() args.func(args) if __name__ == "__main__": main() <commit_msg>Change command option format. - proxy_mode into proxy-mode - viewer_type into viewer-mode<commit_after>
import argparse from microproxy import proxy from microproxy.viewer import log as log_viewer def proxy_handler(args): proxy.start_proxy_server(args.host, args.port, args.proxy_mode) def viewer_handler(args): if args.viewer_mode == "log": log_viewer.start() def main(): parser = argparse.ArgumentParser(description="") subparser = parser.add_subparsers() proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server") proxy_parser.add_argument("--host", default="127.0.0.1") proxy_parser.add_argument("--port", type=int, default=5580) proxy_parser.add_argument("--proxy-mode", choices=["socks", "transparent"], default="socks") proxy_parser.set_defaults(func=proxy_handler) viewer_parser = subparser.add_parser("viewer", help="Open Viewer") viewer_parser.add_argument("--viewer-mode", choices=["log"], default="log") viewer_parser.set_defaults(func=viewer_handler) args = parser.parse_args() args.func(args) if __name__ == "__main__": main()
import argparse from microproxy import proxy from microproxy.viewer import log as log_viewer def proxy_handler(args): proxy.start_proxy_server(args.host, args.port, args.proxy_mode) def viewer_handler(args): if args.viewer_type == "log": log_viewer.start() def main(): parser = argparse.ArgumentParser(description="") subparser = parser.add_subparsers() proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server") proxy_parser.add_argument("--host", default="127.0.0.1") proxy_parser.add_argument("--port", type=int, default=5580) proxy_parser.add_argument("--proxy_mode", choices=["socks", "transparent"], default="socks") proxy_parser.set_defaults(func=proxy_handler) viewer_parser = subparser.add_parser("viewer", help="Open Viewer") viewer_parser.add_argument("--viewer_type", choices=["log"], default="log") viewer_parser.set_defaults(func=viewer_handler) args = parser.parse_args() args.func(args) if __name__ == "__main__": main() Change command option format. - proxy_mode into proxy-mode - viewer_type into viewer-modeimport argparse from microproxy import proxy from microproxy.viewer import log as log_viewer def proxy_handler(args): proxy.start_proxy_server(args.host, args.port, args.proxy_mode) def viewer_handler(args): if args.viewer_mode == "log": log_viewer.start() def main(): parser = argparse.ArgumentParser(description="") subparser = parser.add_subparsers() proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server") proxy_parser.add_argument("--host", default="127.0.0.1") proxy_parser.add_argument("--port", type=int, default=5580) proxy_parser.add_argument("--proxy-mode", choices=["socks", "transparent"], default="socks") proxy_parser.set_defaults(func=proxy_handler) viewer_parser = subparser.add_parser("viewer", help="Open Viewer") viewer_parser.add_argument("--viewer-mode", choices=["log"], default="log") viewer_parser.set_defaults(func=viewer_handler) args = parser.parse_args() args.func(args) if __name__ == "__main__": main()
<commit_before>import argparse from microproxy import proxy from microproxy.viewer import log as log_viewer def proxy_handler(args): proxy.start_proxy_server(args.host, args.port, args.proxy_mode) def viewer_handler(args): if args.viewer_type == "log": log_viewer.start() def main(): parser = argparse.ArgumentParser(description="") subparser = parser.add_subparsers() proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server") proxy_parser.add_argument("--host", default="127.0.0.1") proxy_parser.add_argument("--port", type=int, default=5580) proxy_parser.add_argument("--proxy_mode", choices=["socks", "transparent"], default="socks") proxy_parser.set_defaults(func=proxy_handler) viewer_parser = subparser.add_parser("viewer", help="Open Viewer") viewer_parser.add_argument("--viewer_type", choices=["log"], default="log") viewer_parser.set_defaults(func=viewer_handler) args = parser.parse_args() args.func(args) if __name__ == "__main__": main() <commit_msg>Change command option format. - proxy_mode into proxy-mode - viewer_type into viewer-mode<commit_after>import argparse from microproxy import proxy from microproxy.viewer import log as log_viewer def proxy_handler(args): proxy.start_proxy_server(args.host, args.port, args.proxy_mode) def viewer_handler(args): if args.viewer_mode == "log": log_viewer.start() def main(): parser = argparse.ArgumentParser(description="") subparser = parser.add_subparsers() proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server") proxy_parser.add_argument("--host", default="127.0.0.1") proxy_parser.add_argument("--port", type=int, default=5580) proxy_parser.add_argument("--proxy-mode", choices=["socks", "transparent"], default="socks") proxy_parser.set_defaults(func=proxy_handler) viewer_parser = subparser.add_parser("viewer", help="Open Viewer") viewer_parser.add_argument("--viewer-mode", choices=["log"], default="log") viewer_parser.set_defaults(func=viewer_handler) args = parser.parse_args() args.func(args) if __name__ == "__main__": main()
aa34f571e93d298884f08014865a86a4c92dfcbd
main.py
main.py
__author__ = 'Chad Peterson' __email__ = 'chapeter@cisco.com' from CHROnIC_Portal import app app.secret_key = '1234' app.run(host='0.0.0.0', port=5001, debug=True)
__author__ = 'Chad Peterson' __email__ = 'chapeter@cisco.com' from CHROnIC_Portal import app app.secret_key = '1234' app.run(host='0.0.0.0', port=5000, debug=True)
Put port back to 5000. Missed change from integration testing...
Put port back to 5000. Missed change from integration testing...
Python
mit
chapeter/CHROnIC_Portal,chapeter/CHROnIC_Portal
__author__ = 'Chad Peterson' __email__ = 'chapeter@cisco.com' from CHROnIC_Portal import app app.secret_key = '1234' app.run(host='0.0.0.0', port=5001, debug=True) Put port back to 5000. Missed change from integration testing...
__author__ = 'Chad Peterson' __email__ = 'chapeter@cisco.com' from CHROnIC_Portal import app app.secret_key = '1234' app.run(host='0.0.0.0', port=5000, debug=True)
<commit_before>__author__ = 'Chad Peterson' __email__ = 'chapeter@cisco.com' from CHROnIC_Portal import app app.secret_key = '1234' app.run(host='0.0.0.0', port=5001, debug=True) <commit_msg>Put port back to 5000. Missed change from integration testing...<commit_after>
__author__ = 'Chad Peterson' __email__ = 'chapeter@cisco.com' from CHROnIC_Portal import app app.secret_key = '1234' app.run(host='0.0.0.0', port=5000, debug=True)
__author__ = 'Chad Peterson' __email__ = 'chapeter@cisco.com' from CHROnIC_Portal import app app.secret_key = '1234' app.run(host='0.0.0.0', port=5001, debug=True) Put port back to 5000. Missed change from integration testing...__author__ = 'Chad Peterson' __email__ = 'chapeter@cisco.com' from CHROnIC_Portal import app app.secret_key = '1234' app.run(host='0.0.0.0', port=5000, debug=True)
<commit_before>__author__ = 'Chad Peterson' __email__ = 'chapeter@cisco.com' from CHROnIC_Portal import app app.secret_key = '1234' app.run(host='0.0.0.0', port=5001, debug=True) <commit_msg>Put port back to 5000. Missed change from integration testing...<commit_after>__author__ = 'Chad Peterson' __email__ = 'chapeter@cisco.com' from CHROnIC_Portal import app app.secret_key = '1234' app.run(host='0.0.0.0', port=5000, debug=True)
a4f010ed53615dcbe48c08a445e7d64045001133
base_comment_template/tests/test_base_comment_template.py
base_comment_template/tests/test_base_comment_template.py
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl). from odoo.tests.common import TransactionCase class TestResPartner(TransactionCase): def setUp(self): self.template_id = self.env['base.comment.template'].create({ 'name': 'Comment before lines', 'position': 'before_lines', 'text': 'Text before lines', }) def test_commercial_partner_fields(self): # Azure Interior partner_id = self.env.ref('base.res_partner_12') partner_id.property_comment_template_id = self.template_id.id # Test childs propagation of commercial partner field for child_id in partner_id.child_ids: self.assertEqual( child_id.property_comment_template_id == self.template_id)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl). from odoo.tests.common import TransactionCase class TestResPartner(TransactionCase): def setUp(self): super(TestResPartner, self).setUp() self.template_id = self.env['base.comment.template'].create({ 'name': 'Comment before lines', 'position': 'before_lines', 'text': 'Text before lines', }) def test_commercial_partner_fields(self): # Azure Interior partner_id = self.env.ref('base.res_partner_12') partner_id.property_comment_template_id = self.template_id.id # Test childs propagation of commercial partner field for child_id in partner_id.child_ids: self.assertEqual( child_id.property_comment_template_id, self.template_id)
Move comment_template_id field to the Invoicing tab
[IMP] account_invoice_comment_template: Move comment_template_id field to the Invoicing tab [IMP] account_invoice_comment_template: rename partner field name from comment_template_id to invoice_comment_template_id [IMP] account_invoice_comment_template: Make partner field company_dependant and move domain definition of invoice fields from the view to the model [MOV] account_invoice_comment_template: comment_template_id to base_comment_template [IMP] account_invoice_comment_template: Translate templates when partner changes
Python
agpl-3.0
OCA/reporting-engine,OCA/reporting-engine,OCA/reporting-engine,OCA/reporting-engine
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl). from odoo.tests.common import TransactionCase class TestResPartner(TransactionCase): def setUp(self): self.template_id = self.env['base.comment.template'].create({ 'name': 'Comment before lines', 'position': 'before_lines', 'text': 'Text before lines', }) def test_commercial_partner_fields(self): # Azure Interior partner_id = self.env.ref('base.res_partner_12') partner_id.property_comment_template_id = self.template_id.id # Test childs propagation of commercial partner field for child_id in partner_id.child_ids: self.assertEqual( child_id.property_comment_template_id == self.template_id) [IMP] account_invoice_comment_template: Move comment_template_id field to the Invoicing tab [IMP] account_invoice_comment_template: rename partner field name from comment_template_id to invoice_comment_template_id [IMP] account_invoice_comment_template: Make partner field company_dependant and move domain definition of invoice fields from the view to the model [MOV] account_invoice_comment_template: comment_template_id to base_comment_template [IMP] account_invoice_comment_template: Translate templates when partner changes
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl). from odoo.tests.common import TransactionCase class TestResPartner(TransactionCase): def setUp(self): super(TestResPartner, self).setUp() self.template_id = self.env['base.comment.template'].create({ 'name': 'Comment before lines', 'position': 'before_lines', 'text': 'Text before lines', }) def test_commercial_partner_fields(self): # Azure Interior partner_id = self.env.ref('base.res_partner_12') partner_id.property_comment_template_id = self.template_id.id # Test childs propagation of commercial partner field for child_id in partner_id.child_ids: self.assertEqual( child_id.property_comment_template_id, self.template_id)
<commit_before># License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl). from odoo.tests.common import TransactionCase class TestResPartner(TransactionCase): def setUp(self): self.template_id = self.env['base.comment.template'].create({ 'name': 'Comment before lines', 'position': 'before_lines', 'text': 'Text before lines', }) def test_commercial_partner_fields(self): # Azure Interior partner_id = self.env.ref('base.res_partner_12') partner_id.property_comment_template_id = self.template_id.id # Test childs propagation of commercial partner field for child_id in partner_id.child_ids: self.assertEqual( child_id.property_comment_template_id == self.template_id) <commit_msg>[IMP] account_invoice_comment_template: Move comment_template_id field to the Invoicing tab [IMP] account_invoice_comment_template: rename partner field name from comment_template_id to invoice_comment_template_id [IMP] account_invoice_comment_template: Make partner field company_dependant and move domain definition of invoice fields from the view to the model [MOV] account_invoice_comment_template: comment_template_id to base_comment_template [IMP] account_invoice_comment_template: Translate templates when partner changes<commit_after>
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl). from odoo.tests.common import TransactionCase class TestResPartner(TransactionCase): def setUp(self): super(TestResPartner, self).setUp() self.template_id = self.env['base.comment.template'].create({ 'name': 'Comment before lines', 'position': 'before_lines', 'text': 'Text before lines', }) def test_commercial_partner_fields(self): # Azure Interior partner_id = self.env.ref('base.res_partner_12') partner_id.property_comment_template_id = self.template_id.id # Test childs propagation of commercial partner field for child_id in partner_id.child_ids: self.assertEqual( child_id.property_comment_template_id, self.template_id)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl). from odoo.tests.common import TransactionCase class TestResPartner(TransactionCase): def setUp(self): self.template_id = self.env['base.comment.template'].create({ 'name': 'Comment before lines', 'position': 'before_lines', 'text': 'Text before lines', }) def test_commercial_partner_fields(self): # Azure Interior partner_id = self.env.ref('base.res_partner_12') partner_id.property_comment_template_id = self.template_id.id # Test childs propagation of commercial partner field for child_id in partner_id.child_ids: self.assertEqual( child_id.property_comment_template_id == self.template_id) [IMP] account_invoice_comment_template: Move comment_template_id field to the Invoicing tab [IMP] account_invoice_comment_template: rename partner field name from comment_template_id to invoice_comment_template_id [IMP] account_invoice_comment_template: Make partner field company_dependant and move domain definition of invoice fields from the view to the model [MOV] account_invoice_comment_template: comment_template_id to base_comment_template [IMP] account_invoice_comment_template: Translate templates when partner changes# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl). from odoo.tests.common import TransactionCase class TestResPartner(TransactionCase): def setUp(self): super(TestResPartner, self).setUp() self.template_id = self.env['base.comment.template'].create({ 'name': 'Comment before lines', 'position': 'before_lines', 'text': 'Text before lines', }) def test_commercial_partner_fields(self): # Azure Interior partner_id = self.env.ref('base.res_partner_12') partner_id.property_comment_template_id = self.template_id.id # Test childs propagation of commercial partner field for child_id in partner_id.child_ids: self.assertEqual( child_id.property_comment_template_id, self.template_id)
<commit_before># License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl). from odoo.tests.common import TransactionCase class TestResPartner(TransactionCase): def setUp(self): self.template_id = self.env['base.comment.template'].create({ 'name': 'Comment before lines', 'position': 'before_lines', 'text': 'Text before lines', }) def test_commercial_partner_fields(self): # Azure Interior partner_id = self.env.ref('base.res_partner_12') partner_id.property_comment_template_id = self.template_id.id # Test childs propagation of commercial partner field for child_id in partner_id.child_ids: self.assertEqual( child_id.property_comment_template_id == self.template_id) <commit_msg>[IMP] account_invoice_comment_template: Move comment_template_id field to the Invoicing tab [IMP] account_invoice_comment_template: rename partner field name from comment_template_id to invoice_comment_template_id [IMP] account_invoice_comment_template: Make partner field company_dependant and move domain definition of invoice fields from the view to the model [MOV] account_invoice_comment_template: comment_template_id to base_comment_template [IMP] account_invoice_comment_template: Translate templates when partner changes<commit_after># License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl). from odoo.tests.common import TransactionCase class TestResPartner(TransactionCase): def setUp(self): super(TestResPartner, self).setUp() self.template_id = self.env['base.comment.template'].create({ 'name': 'Comment before lines', 'position': 'before_lines', 'text': 'Text before lines', }) def test_commercial_partner_fields(self): # Azure Interior partner_id = self.env.ref('base.res_partner_12') partner_id.property_comment_template_id = self.template_id.id # Test childs propagation of commercial partner field for child_id in partner_id.child_ids: self.assertEqual( child_id.property_comment_template_id, self.template_id)
0fd4ba14a16c6bfc100856dec0af6b17eb6917f2
codewars/valid_braces.py
codewars/valid_braces.py
# Valid Braces # http://www.codewars.com/kata/5277c8a221e209d3f6000b56/train/python import unittest def valid_braces(string: str) -> bool: stack = [] braces = {')': '(', '}': '{', ']': '['} for l in string: if l in braces.values(): stack.append(l) elif stack: if stack.pop() != braces[l]: # We are closing a brace which we didn't open last. return False else: # The stack was empty and we tried to remove a brace which is not valid. return False if stack: # List is not empty. return False else: return True class ValidBracesTest(unittest.TestCase): def test(self): self.assertEqual(True, valid_braces("()")) self.assertEqual(valid_braces("[(])"), False) if __name__ == '__main__': unittest.main()
Add solution for `Valid Braces`
Add solution for `Valid Braces`
Python
mit
davidlukac/codekata-python
Add solution for `Valid Braces`
# Valid Braces # http://www.codewars.com/kata/5277c8a221e209d3f6000b56/train/python import unittest def valid_braces(string: str) -> bool: stack = [] braces = {')': '(', '}': '{', ']': '['} for l in string: if l in braces.values(): stack.append(l) elif stack: if stack.pop() != braces[l]: # We are closing a brace which we didn't open last. return False else: # The stack was empty and we tried to remove a brace which is not valid. return False if stack: # List is not empty. return False else: return True class ValidBracesTest(unittest.TestCase): def test(self): self.assertEqual(True, valid_braces("()")) self.assertEqual(valid_braces("[(])"), False) if __name__ == '__main__': unittest.main()
<commit_before><commit_msg>Add solution for `Valid Braces`<commit_after>
# Valid Braces # http://www.codewars.com/kata/5277c8a221e209d3f6000b56/train/python import unittest def valid_braces(string: str) -> bool: stack = [] braces = {')': '(', '}': '{', ']': '['} for l in string: if l in braces.values(): stack.append(l) elif stack: if stack.pop() != braces[l]: # We are closing a brace which we didn't open last. return False else: # The stack was empty and we tried to remove a brace which is not valid. return False if stack: # List is not empty. return False else: return True class ValidBracesTest(unittest.TestCase): def test(self): self.assertEqual(True, valid_braces("()")) self.assertEqual(valid_braces("[(])"), False) if __name__ == '__main__': unittest.main()
Add solution for `Valid Braces`# Valid Braces # http://www.codewars.com/kata/5277c8a221e209d3f6000b56/train/python import unittest def valid_braces(string: str) -> bool: stack = [] braces = {')': '(', '}': '{', ']': '['} for l in string: if l in braces.values(): stack.append(l) elif stack: if stack.pop() != braces[l]: # We are closing a brace which we didn't open last. return False else: # The stack was empty and we tried to remove a brace which is not valid. return False if stack: # List is not empty. return False else: return True class ValidBracesTest(unittest.TestCase): def test(self): self.assertEqual(True, valid_braces("()")) self.assertEqual(valid_braces("[(])"), False) if __name__ == '__main__': unittest.main()
<commit_before><commit_msg>Add solution for `Valid Braces`<commit_after># Valid Braces # http://www.codewars.com/kata/5277c8a221e209d3f6000b56/train/python import unittest def valid_braces(string: str) -> bool: stack = [] braces = {')': '(', '}': '{', ']': '['} for l in string: if l in braces.values(): stack.append(l) elif stack: if stack.pop() != braces[l]: # We are closing a brace which we didn't open last. return False else: # The stack was empty and we tried to remove a brace which is not valid. return False if stack: # List is not empty. return False else: return True class ValidBracesTest(unittest.TestCase): def test(self): self.assertEqual(True, valid_braces("()")) self.assertEqual(valid_braces("[(])"), False) if __name__ == '__main__': unittest.main()
886d6e56e53742ec6cd2c59440459b17b093f4e0
blockbuster/__init__.py
blockbuster/__init__.py
__author__ = 'Matt Stibbs' __version__ = '1.26.04' target_schema_version = '1.25.00' from flask import Flask app = Flask(__name__) def startup(): import blockbuster.bb_dbconnector_factory import blockbuster.bb_logging as log import blockbuster.bb_auditlogger as audit blockbuster.app.debug = blockbuster.config.debug_mode blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}", blockbuster.__version__, target_schema_version)) blockbuster.bb_logging.logger.info( 'Application Setting - Time Restriction Disabled') \ if not blockbuster.config.timerestriction else blockbuster.bb_logging.logger.info( 'Application Setting - Time Restriction Enabled') if blockbuster.config.debug_mode: blockbuster.bb_logging.logger.info("========= APPLICATION IS RUNNING IN DEBUG MODE ==========") try: if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check(): import blockbuster.bb_routes print("Running...") else: raise RuntimeError("Incorrect database schema version. Wanted ") except RuntimeError, e: log.logger.exception(e) audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e)) startup()
__author__ = 'Matt Stibbs' __version__ = '1.26.04' target_schema_version = '1.25.00' from flask import Flask app = Flask(__name__) def startup(): import blockbuster.bb_dbconnector_factory import blockbuster.bb_logging as log import blockbuster.bb_auditlogger as audit blockbuster.app.debug = blockbuster.config.debug_mode blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}", blockbuster.__version__, target_schema_version)) time_setting = "Application Setting - Time Restriction Disabled" if not blockbuster.config.timerestriction else "Application Setting - Time Restriction Enabled" print(time_setting) if blockbuster.config.debug_mode: print("========= APPLICATION IS RUNNING IN DEBUG MODE ==========") try: if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check(): import blockbuster.bb_routes print("Running...") else: raise RuntimeError("Incorrect database schema version. Wanted ") except RuntimeError, e: log.logger.exception(e) audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e)) startup()
Tweak console output on startup
Tweak console output on startup
Python
mit
mattstibbs/blockbuster-server,mattstibbs/blockbuster-server
__author__ = 'Matt Stibbs' __version__ = '1.26.04' target_schema_version = '1.25.00' from flask import Flask app = Flask(__name__) def startup(): import blockbuster.bb_dbconnector_factory import blockbuster.bb_logging as log import blockbuster.bb_auditlogger as audit blockbuster.app.debug = blockbuster.config.debug_mode blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}", blockbuster.__version__, target_schema_version)) blockbuster.bb_logging.logger.info( 'Application Setting - Time Restriction Disabled') \ if not blockbuster.config.timerestriction else blockbuster.bb_logging.logger.info( 'Application Setting - Time Restriction Enabled') if blockbuster.config.debug_mode: blockbuster.bb_logging.logger.info("========= APPLICATION IS RUNNING IN DEBUG MODE ==========") try: if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check(): import blockbuster.bb_routes print("Running...") else: raise RuntimeError("Incorrect database schema version. Wanted ") except RuntimeError, e: log.logger.exception(e) audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e)) startup() Tweak console output on startup
__author__ = 'Matt Stibbs' __version__ = '1.26.04' target_schema_version = '1.25.00' from flask import Flask app = Flask(__name__) def startup(): import blockbuster.bb_dbconnector_factory import blockbuster.bb_logging as log import blockbuster.bb_auditlogger as audit blockbuster.app.debug = blockbuster.config.debug_mode blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}", blockbuster.__version__, target_schema_version)) time_setting = "Application Setting - Time Restriction Disabled" if not blockbuster.config.timerestriction else "Application Setting - Time Restriction Enabled" print(time_setting) if blockbuster.config.debug_mode: print("========= APPLICATION IS RUNNING IN DEBUG MODE ==========") try: if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check(): import blockbuster.bb_routes print("Running...") else: raise RuntimeError("Incorrect database schema version. Wanted ") except RuntimeError, e: log.logger.exception(e) audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e)) startup()
<commit_before>__author__ = 'Matt Stibbs' __version__ = '1.26.04' target_schema_version = '1.25.00' from flask import Flask app = Flask(__name__) def startup(): import blockbuster.bb_dbconnector_factory import blockbuster.bb_logging as log import blockbuster.bb_auditlogger as audit blockbuster.app.debug = blockbuster.config.debug_mode blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}", blockbuster.__version__, target_schema_version)) blockbuster.bb_logging.logger.info( 'Application Setting - Time Restriction Disabled') \ if not blockbuster.config.timerestriction else blockbuster.bb_logging.logger.info( 'Application Setting - Time Restriction Enabled') if blockbuster.config.debug_mode: blockbuster.bb_logging.logger.info("========= APPLICATION IS RUNNING IN DEBUG MODE ==========") try: if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check(): import blockbuster.bb_routes print("Running...") else: raise RuntimeError("Incorrect database schema version. Wanted ") except RuntimeError, e: log.logger.exception(e) audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e)) startup() <commit_msg>Tweak console output on startup<commit_after>
__author__ = 'Matt Stibbs' __version__ = '1.26.04' target_schema_version = '1.25.00' from flask import Flask app = Flask(__name__) def startup(): import blockbuster.bb_dbconnector_factory import blockbuster.bb_logging as log import blockbuster.bb_auditlogger as audit blockbuster.app.debug = blockbuster.config.debug_mode blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}", blockbuster.__version__, target_schema_version)) time_setting = "Application Setting - Time Restriction Disabled" if not blockbuster.config.timerestriction else "Application Setting - Time Restriction Enabled" print(time_setting) if blockbuster.config.debug_mode: print("========= APPLICATION IS RUNNING IN DEBUG MODE ==========") try: if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check(): import blockbuster.bb_routes print("Running...") else: raise RuntimeError("Incorrect database schema version. Wanted ") except RuntimeError, e: log.logger.exception(e) audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e)) startup()
__author__ = 'Matt Stibbs' __version__ = '1.26.04' target_schema_version = '1.25.00' from flask import Flask app = Flask(__name__) def startup(): import blockbuster.bb_dbconnector_factory import blockbuster.bb_logging as log import blockbuster.bb_auditlogger as audit blockbuster.app.debug = blockbuster.config.debug_mode blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}", blockbuster.__version__, target_schema_version)) blockbuster.bb_logging.logger.info( 'Application Setting - Time Restriction Disabled') \ if not blockbuster.config.timerestriction else blockbuster.bb_logging.logger.info( 'Application Setting - Time Restriction Enabled') if blockbuster.config.debug_mode: blockbuster.bb_logging.logger.info("========= APPLICATION IS RUNNING IN DEBUG MODE ==========") try: if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check(): import blockbuster.bb_routes print("Running...") else: raise RuntimeError("Incorrect database schema version. Wanted ") except RuntimeError, e: log.logger.exception(e) audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e)) startup() Tweak console output on startup__author__ = 'Matt Stibbs' __version__ = '1.26.04' target_schema_version = '1.25.00' from flask import Flask app = Flask(__name__) def startup(): import blockbuster.bb_dbconnector_factory import blockbuster.bb_logging as log import blockbuster.bb_auditlogger as audit blockbuster.app.debug = blockbuster.config.debug_mode blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}", blockbuster.__version__, target_schema_version)) time_setting = "Application Setting - Time Restriction Disabled" if not blockbuster.config.timerestriction else "Application Setting - Time Restriction Enabled" print(time_setting) if blockbuster.config.debug_mode: print("========= APPLICATION IS RUNNING IN DEBUG MODE ==========") try: if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check(): import blockbuster.bb_routes print("Running...") else: raise RuntimeError("Incorrect database schema version. Wanted ") except RuntimeError, e: log.logger.exception(e) audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e)) startup()
<commit_before>__author__ = 'Matt Stibbs' __version__ = '1.26.04' target_schema_version = '1.25.00' from flask import Flask app = Flask(__name__) def startup(): import blockbuster.bb_dbconnector_factory import blockbuster.bb_logging as log import blockbuster.bb_auditlogger as audit blockbuster.app.debug = blockbuster.config.debug_mode blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}", blockbuster.__version__, target_schema_version)) blockbuster.bb_logging.logger.info( 'Application Setting - Time Restriction Disabled') \ if not blockbuster.config.timerestriction else blockbuster.bb_logging.logger.info( 'Application Setting - Time Restriction Enabled') if blockbuster.config.debug_mode: blockbuster.bb_logging.logger.info("========= APPLICATION IS RUNNING IN DEBUG MODE ==========") try: if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check(): import blockbuster.bb_routes print("Running...") else: raise RuntimeError("Incorrect database schema version. Wanted ") except RuntimeError, e: log.logger.exception(e) audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e)) startup() <commit_msg>Tweak console output on startup<commit_after>__author__ = 'Matt Stibbs' __version__ = '1.26.04' target_schema_version = '1.25.00' from flask import Flask app = Flask(__name__) def startup(): import blockbuster.bb_dbconnector_factory import blockbuster.bb_logging as log import blockbuster.bb_auditlogger as audit blockbuster.app.debug = blockbuster.config.debug_mode blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}", blockbuster.__version__, target_schema_version)) time_setting = "Application Setting - Time Restriction Disabled" if not blockbuster.config.timerestriction else "Application Setting - Time Restriction Enabled" print(time_setting) if blockbuster.config.debug_mode: print("========= APPLICATION IS RUNNING IN DEBUG MODE ==========") try: if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check(): import blockbuster.bb_routes print("Running...") else: raise RuntimeError("Incorrect database schema version. Wanted ") except RuntimeError, e: log.logger.exception(e) audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e)) startup()
40095b001ab95fda4cc80bcc807508e9580ebf2d
fireplace/cards/gvg/neutral_legendary.py
fireplace/cards/gvg/neutral_legendary.py
from ..utils import * ## # Minions # Dr. Boom class GVG_110: action = [Summon(CONTROLLER, "GVG_110t") * 2] # Boom Bot class GVG_110t: def deathrattle(self): return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))] # Sneed's Old Shredder class GVG_114: def deathrattle(self): legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY) return [Summon(CONTROLLER, legendary)]
from ..utils import * ## # Minions # Dr. Boom class GVG_110: action = [Summon(CONTROLLER, "GVG_110t") * 2] # Boom Bot class GVG_110t: def deathrattle(self): return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))] # Sneed's Old Shredder class GVG_114: def deathrattle(self): legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY) return [Summon(CONTROLLER, legendary)] # Toshley class GVG_115: action = [GiveSparePart(CONTROLLER)] deathrattle = [GiveSparePart(CONTROLLER)] # Mekgineer Thermaplugg class GVG_116: def MINION_DESTROY(self, minion): if minion.controller is not self.controller: return [Summon(CONTROLLER, "EX1_029")] # Gazlowe class GVG_117: def OWN_CARD_PLAYED(self, card): if card.type == CardType.SPELL and card.cost == 1: return [Give(CONTROLLER, randomCollectible(race=Race.MECHANICAL))]
Implement Toshley, Mekgineer Thermaplugg and Gazlowe
Implement Toshley, Mekgineer Thermaplugg and Gazlowe
Python
agpl-3.0
amw2104/fireplace,oftc-ftw/fireplace,smallnamespace/fireplace,smallnamespace/fireplace,liujimj/fireplace,oftc-ftw/fireplace,amw2104/fireplace,butozerca/fireplace,Ragowit/fireplace,liujimj/fireplace,jleclanche/fireplace,butozerca/fireplace,Ragowit/fireplace,NightKev/fireplace,beheh/fireplace,Meerkov/fireplace,Meerkov/fireplace
from ..utils import * ## # Minions # Dr. Boom class GVG_110: action = [Summon(CONTROLLER, "GVG_110t") * 2] # Boom Bot class GVG_110t: def deathrattle(self): return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))] # Sneed's Old Shredder class GVG_114: def deathrattle(self): legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY) return [Summon(CONTROLLER, legendary)] Implement Toshley, Mekgineer Thermaplugg and Gazlowe
from ..utils import * ## # Minions # Dr. Boom class GVG_110: action = [Summon(CONTROLLER, "GVG_110t") * 2] # Boom Bot class GVG_110t: def deathrattle(self): return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))] # Sneed's Old Shredder class GVG_114: def deathrattle(self): legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY) return [Summon(CONTROLLER, legendary)] # Toshley class GVG_115: action = [GiveSparePart(CONTROLLER)] deathrattle = [GiveSparePart(CONTROLLER)] # Mekgineer Thermaplugg class GVG_116: def MINION_DESTROY(self, minion): if minion.controller is not self.controller: return [Summon(CONTROLLER, "EX1_029")] # Gazlowe class GVG_117: def OWN_CARD_PLAYED(self, card): if card.type == CardType.SPELL and card.cost == 1: return [Give(CONTROLLER, randomCollectible(race=Race.MECHANICAL))]
<commit_before>from ..utils import * ## # Minions # Dr. Boom class GVG_110: action = [Summon(CONTROLLER, "GVG_110t") * 2] # Boom Bot class GVG_110t: def deathrattle(self): return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))] # Sneed's Old Shredder class GVG_114: def deathrattle(self): legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY) return [Summon(CONTROLLER, legendary)] <commit_msg>Implement Toshley, Mekgineer Thermaplugg and Gazlowe<commit_after>
from ..utils import * ## # Minions # Dr. Boom class GVG_110: action = [Summon(CONTROLLER, "GVG_110t") * 2] # Boom Bot class GVG_110t: def deathrattle(self): return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))] # Sneed's Old Shredder class GVG_114: def deathrattle(self): legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY) return [Summon(CONTROLLER, legendary)] # Toshley class GVG_115: action = [GiveSparePart(CONTROLLER)] deathrattle = [GiveSparePart(CONTROLLER)] # Mekgineer Thermaplugg class GVG_116: def MINION_DESTROY(self, minion): if minion.controller is not self.controller: return [Summon(CONTROLLER, "EX1_029")] # Gazlowe class GVG_117: def OWN_CARD_PLAYED(self, card): if card.type == CardType.SPELL and card.cost == 1: return [Give(CONTROLLER, randomCollectible(race=Race.MECHANICAL))]
from ..utils import * ## # Minions # Dr. Boom class GVG_110: action = [Summon(CONTROLLER, "GVG_110t") * 2] # Boom Bot class GVG_110t: def deathrattle(self): return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))] # Sneed's Old Shredder class GVG_114: def deathrattle(self): legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY) return [Summon(CONTROLLER, legendary)] Implement Toshley, Mekgineer Thermaplugg and Gazlowefrom ..utils import * ## # Minions # Dr. Boom class GVG_110: action = [Summon(CONTROLLER, "GVG_110t") * 2] # Boom Bot class GVG_110t: def deathrattle(self): return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))] # Sneed's Old Shredder class GVG_114: def deathrattle(self): legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY) return [Summon(CONTROLLER, legendary)] # Toshley class GVG_115: action = [GiveSparePart(CONTROLLER)] deathrattle = [GiveSparePart(CONTROLLER)] # Mekgineer Thermaplugg class GVG_116: def MINION_DESTROY(self, minion): if minion.controller is not self.controller: return [Summon(CONTROLLER, "EX1_029")] # Gazlowe class GVG_117: def OWN_CARD_PLAYED(self, card): if card.type == CardType.SPELL and card.cost == 1: return [Give(CONTROLLER, randomCollectible(race=Race.MECHANICAL))]
<commit_before>from ..utils import * ## # Minions # Dr. Boom class GVG_110: action = [Summon(CONTROLLER, "GVG_110t") * 2] # Boom Bot class GVG_110t: def deathrattle(self): return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))] # Sneed's Old Shredder class GVG_114: def deathrattle(self): legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY) return [Summon(CONTROLLER, legendary)] <commit_msg>Implement Toshley, Mekgineer Thermaplugg and Gazlowe<commit_after>from ..utils import * ## # Minions # Dr. Boom class GVG_110: action = [Summon(CONTROLLER, "GVG_110t") * 2] # Boom Bot class GVG_110t: def deathrattle(self): return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))] # Sneed's Old Shredder class GVG_114: def deathrattle(self): legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY) return [Summon(CONTROLLER, legendary)] # Toshley class GVG_115: action = [GiveSparePart(CONTROLLER)] deathrattle = [GiveSparePart(CONTROLLER)] # Mekgineer Thermaplugg class GVG_116: def MINION_DESTROY(self, minion): if minion.controller is not self.controller: return [Summon(CONTROLLER, "EX1_029")] # Gazlowe class GVG_117: def OWN_CARD_PLAYED(self, card): if card.type == CardType.SPELL and card.cost == 1: return [Give(CONTROLLER, randomCollectible(race=Race.MECHANICAL))]
dd445cbf33268ece3a6b006d3d31d6169fec03b8
acoustid/scripts/backfill_meta_created.py
acoustid/scripts/backfill_meta_created.py
#!/usr/bin/env python # Copyright (C) 2019 Lukas Lalinsky # Distributed under the MIT license, see the LICENSE file for details. import logging logger = logging.getLogger(__name__) def run_backfill_meta_created(script, opts, args): if script.config.cluster.role != 'master': logger.info('Not running backfill_meta_created in slave mode') return query = """ WITH meta_created AS ( SELECT meta_id, min(created) AS created FROM track_meta WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000) GROUP BY meta_id ) UPDATE meta SET created = meta_created.created FROM meta_created WHERE meta.id = meta_created.meta_id AND meta.created IS NULL """ for i in range(10): with script.context() as ctx: fingerprint_db = ctx.db.get_fingerprint_db() fingerprint_db.execute(query) ctx.db.session.commit()
#!/usr/bin/env python # Copyright (C) 2019 Lukas Lalinsky # Distributed under the MIT license, see the LICENSE file for details. import logging logger = logging.getLogger(__name__) def run_backfill_meta_created(script, opts, args): if script.config.cluster.role != 'master': logger.info('Not running backfill_meta_created in slave mode') return query = """ WITH meta_created AS ( SELECT meta_id, min(created) AS created FROM track_meta WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000) GROUP BY meta_id ) UPDATE meta SET created = meta_created.created FROM meta_created WHERE meta.id = meta_created.meta_id AND meta.created IS NULL """ for i in range(100): with script.context() as ctx: fingerprint_db = ctx.db.get_fingerprint_db() fingerprint_db.execute(query) ctx.db.session.commit()
Increase the number of backill_meta_created iterations
Increase the number of backill_meta_created iterations
Python
mit
lalinsky/acoustid-server,lalinsky/acoustid-server,lalinsky/acoustid-server,lalinsky/acoustid-server
#!/usr/bin/env python # Copyright (C) 2019 Lukas Lalinsky # Distributed under the MIT license, see the LICENSE file for details. import logging logger = logging.getLogger(__name__) def run_backfill_meta_created(script, opts, args): if script.config.cluster.role != 'master': logger.info('Not running backfill_meta_created in slave mode') return query = """ WITH meta_created AS ( SELECT meta_id, min(created) AS created FROM track_meta WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000) GROUP BY meta_id ) UPDATE meta SET created = meta_created.created FROM meta_created WHERE meta.id = meta_created.meta_id AND meta.created IS NULL """ for i in range(10): with script.context() as ctx: fingerprint_db = ctx.db.get_fingerprint_db() fingerprint_db.execute(query) ctx.db.session.commit() Increase the number of backill_meta_created iterations
#!/usr/bin/env python # Copyright (C) 2019 Lukas Lalinsky # Distributed under the MIT license, see the LICENSE file for details. import logging logger = logging.getLogger(__name__) def run_backfill_meta_created(script, opts, args): if script.config.cluster.role != 'master': logger.info('Not running backfill_meta_created in slave mode') return query = """ WITH meta_created AS ( SELECT meta_id, min(created) AS created FROM track_meta WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000) GROUP BY meta_id ) UPDATE meta SET created = meta_created.created FROM meta_created WHERE meta.id = meta_created.meta_id AND meta.created IS NULL """ for i in range(100): with script.context() as ctx: fingerprint_db = ctx.db.get_fingerprint_db() fingerprint_db.execute(query) ctx.db.session.commit()
<commit_before>#!/usr/bin/env python # Copyright (C) 2019 Lukas Lalinsky # Distributed under the MIT license, see the LICENSE file for details. import logging logger = logging.getLogger(__name__) def run_backfill_meta_created(script, opts, args): if script.config.cluster.role != 'master': logger.info('Not running backfill_meta_created in slave mode') return query = """ WITH meta_created AS ( SELECT meta_id, min(created) AS created FROM track_meta WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000) GROUP BY meta_id ) UPDATE meta SET created = meta_created.created FROM meta_created WHERE meta.id = meta_created.meta_id AND meta.created IS NULL """ for i in range(10): with script.context() as ctx: fingerprint_db = ctx.db.get_fingerprint_db() fingerprint_db.execute(query) ctx.db.session.commit() <commit_msg>Increase the number of backill_meta_created iterations<commit_after>
#!/usr/bin/env python # Copyright (C) 2019 Lukas Lalinsky # Distributed under the MIT license, see the LICENSE file for details. import logging logger = logging.getLogger(__name__) def run_backfill_meta_created(script, opts, args): if script.config.cluster.role != 'master': logger.info('Not running backfill_meta_created in slave mode') return query = """ WITH meta_created AS ( SELECT meta_id, min(created) AS created FROM track_meta WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000) GROUP BY meta_id ) UPDATE meta SET created = meta_created.created FROM meta_created WHERE meta.id = meta_created.meta_id AND meta.created IS NULL """ for i in range(100): with script.context() as ctx: fingerprint_db = ctx.db.get_fingerprint_db() fingerprint_db.execute(query) ctx.db.session.commit()
#!/usr/bin/env python # Copyright (C) 2019 Lukas Lalinsky # Distributed under the MIT license, see the LICENSE file for details. import logging logger = logging.getLogger(__name__) def run_backfill_meta_created(script, opts, args): if script.config.cluster.role != 'master': logger.info('Not running backfill_meta_created in slave mode') return query = """ WITH meta_created AS ( SELECT meta_id, min(created) AS created FROM track_meta WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000) GROUP BY meta_id ) UPDATE meta SET created = meta_created.created FROM meta_created WHERE meta.id = meta_created.meta_id AND meta.created IS NULL """ for i in range(10): with script.context() as ctx: fingerprint_db = ctx.db.get_fingerprint_db() fingerprint_db.execute(query) ctx.db.session.commit() Increase the number of backill_meta_created iterations#!/usr/bin/env python # Copyright (C) 2019 Lukas Lalinsky # Distributed under the MIT license, see the LICENSE file for details. import logging logger = logging.getLogger(__name__) def run_backfill_meta_created(script, opts, args): if script.config.cluster.role != 'master': logger.info('Not running backfill_meta_created in slave mode') return query = """ WITH meta_created AS ( SELECT meta_id, min(created) AS created FROM track_meta WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000) GROUP BY meta_id ) UPDATE meta SET created = meta_created.created FROM meta_created WHERE meta.id = meta_created.meta_id AND meta.created IS NULL """ for i in range(100): with script.context() as ctx: fingerprint_db = ctx.db.get_fingerprint_db() fingerprint_db.execute(query) ctx.db.session.commit()
<commit_before>#!/usr/bin/env python # Copyright (C) 2019 Lukas Lalinsky # Distributed under the MIT license, see the LICENSE file for details. import logging logger = logging.getLogger(__name__) def run_backfill_meta_created(script, opts, args): if script.config.cluster.role != 'master': logger.info('Not running backfill_meta_created in slave mode') return query = """ WITH meta_created AS ( SELECT meta_id, min(created) AS created FROM track_meta WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000) GROUP BY meta_id ) UPDATE meta SET created = meta_created.created FROM meta_created WHERE meta.id = meta_created.meta_id AND meta.created IS NULL """ for i in range(10): with script.context() as ctx: fingerprint_db = ctx.db.get_fingerprint_db() fingerprint_db.execute(query) ctx.db.session.commit() <commit_msg>Increase the number of backill_meta_created iterations<commit_after>#!/usr/bin/env python # Copyright (C) 2019 Lukas Lalinsky # Distributed under the MIT license, see the LICENSE file for details. import logging logger = logging.getLogger(__name__) def run_backfill_meta_created(script, opts, args): if script.config.cluster.role != 'master': logger.info('Not running backfill_meta_created in slave mode') return query = """ WITH meta_created AS ( SELECT meta_id, min(created) AS created FROM track_meta WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000) GROUP BY meta_id ) UPDATE meta SET created = meta_created.created FROM meta_created WHERE meta.id = meta_created.meta_id AND meta.created IS NULL """ for i in range(100): with script.context() as ctx: fingerprint_db = ctx.db.get_fingerprint_db() fingerprint_db.execute(query) ctx.db.session.commit()
1ad03769569d86d1eda45f7c6582234ed455ea88
src/main.py
src/main.py
"""Where player runs the game""" import random import time import board import conversion import games if __name__ == '__main__': NUMBER_OF_TRIALS = 1 for i in range(NUMBER_OF_TRIALS): X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square() LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS) ORIENTATIONS = (True, False) GAMES = games.GamePositions() BOARD_DESCRIPTION = GAMES.random_game() BOARD = board.Board(BOARD_DESCRIPTION, orientation=random.choice(ORIENTATIONS)) BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS) BOARD.update_board_string() print(BOARD) start = time.time() ATTEMPT = input('Enter highlighted square: ') end = time.time() if ATTEMPT == LOCATION: print("Correct!") else: print("Wrong, the answer was {0}".format(LOCATION)) print("Attempt took {0} seconds".format(end - start))
"""Where player runs the game""" import random import time import board import conversion import games from settings import RECORD_FILE def write_record_to_file(a_string, file_name): with open(file_name, 'w') as f: f.write(a_string) def get_record_from_file(file_name): with open(file_name, 'r') as f: result = f.readline() return result if __name__ == '__main__': NUMBER_OF_TRIALS = 1 for i in range(NUMBER_OF_TRIALS): X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square() LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS) ORIENTATIONS = (True, False) GAMES = games.GamePositions() BOARD_DESCRIPTION = GAMES.random_game() BOARD = board.Board(BOARD_DESCRIPTION, orientation=random.choice(ORIENTATIONS)) BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS) BOARD.update_board_string() print(BOARD) start = time.time() ATTEMPT = input('Enter highlighted square: ') end = time.time() if ATTEMPT == LOCATION: print("Correct!") else: print("Wrong, the answer was {0}".format(LOCATION)) print("Attempt took {0} seconds".format(end - start))
Add functions to read/write to a record file
Add functions to read/write to a record file
Python
mit
blairck/chess_notation
"""Where player runs the game""" import random import time import board import conversion import games if __name__ == '__main__': NUMBER_OF_TRIALS = 1 for i in range(NUMBER_OF_TRIALS): X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square() LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS) ORIENTATIONS = (True, False) GAMES = games.GamePositions() BOARD_DESCRIPTION = GAMES.random_game() BOARD = board.Board(BOARD_DESCRIPTION, orientation=random.choice(ORIENTATIONS)) BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS) BOARD.update_board_string() print(BOARD) start = time.time() ATTEMPT = input('Enter highlighted square: ') end = time.time() if ATTEMPT == LOCATION: print("Correct!") else: print("Wrong, the answer was {0}".format(LOCATION)) print("Attempt took {0} seconds".format(end - start)) Add functions to read/write to a record file
"""Where player runs the game""" import random import time import board import conversion import games from settings import RECORD_FILE def write_record_to_file(a_string, file_name): with open(file_name, 'w') as f: f.write(a_string) def get_record_from_file(file_name): with open(file_name, 'r') as f: result = f.readline() return result if __name__ == '__main__': NUMBER_OF_TRIALS = 1 for i in range(NUMBER_OF_TRIALS): X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square() LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS) ORIENTATIONS = (True, False) GAMES = games.GamePositions() BOARD_DESCRIPTION = GAMES.random_game() BOARD = board.Board(BOARD_DESCRIPTION, orientation=random.choice(ORIENTATIONS)) BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS) BOARD.update_board_string() print(BOARD) start = time.time() ATTEMPT = input('Enter highlighted square: ') end = time.time() if ATTEMPT == LOCATION: print("Correct!") else: print("Wrong, the answer was {0}".format(LOCATION)) print("Attempt took {0} seconds".format(end - start))
<commit_before>"""Where player runs the game""" import random import time import board import conversion import games if __name__ == '__main__': NUMBER_OF_TRIALS = 1 for i in range(NUMBER_OF_TRIALS): X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square() LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS) ORIENTATIONS = (True, False) GAMES = games.GamePositions() BOARD_DESCRIPTION = GAMES.random_game() BOARD = board.Board(BOARD_DESCRIPTION, orientation=random.choice(ORIENTATIONS)) BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS) BOARD.update_board_string() print(BOARD) start = time.time() ATTEMPT = input('Enter highlighted square: ') end = time.time() if ATTEMPT == LOCATION: print("Correct!") else: print("Wrong, the answer was {0}".format(LOCATION)) print("Attempt took {0} seconds".format(end - start)) <commit_msg>Add functions to read/write to a record file<commit_after>
"""Where player runs the game""" import random import time import board import conversion import games from settings import RECORD_FILE def write_record_to_file(a_string, file_name): with open(file_name, 'w') as f: f.write(a_string) def get_record_from_file(file_name): with open(file_name, 'r') as f: result = f.readline() return result if __name__ == '__main__': NUMBER_OF_TRIALS = 1 for i in range(NUMBER_OF_TRIALS): X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square() LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS) ORIENTATIONS = (True, False) GAMES = games.GamePositions() BOARD_DESCRIPTION = GAMES.random_game() BOARD = board.Board(BOARD_DESCRIPTION, orientation=random.choice(ORIENTATIONS)) BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS) BOARD.update_board_string() print(BOARD) start = time.time() ATTEMPT = input('Enter highlighted square: ') end = time.time() if ATTEMPT == LOCATION: print("Correct!") else: print("Wrong, the answer was {0}".format(LOCATION)) print("Attempt took {0} seconds".format(end - start))
"""Where player runs the game""" import random import time import board import conversion import games if __name__ == '__main__': NUMBER_OF_TRIALS = 1 for i in range(NUMBER_OF_TRIALS): X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square() LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS) ORIENTATIONS = (True, False) GAMES = games.GamePositions() BOARD_DESCRIPTION = GAMES.random_game() BOARD = board.Board(BOARD_DESCRIPTION, orientation=random.choice(ORIENTATIONS)) BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS) BOARD.update_board_string() print(BOARD) start = time.time() ATTEMPT = input('Enter highlighted square: ') end = time.time() if ATTEMPT == LOCATION: print("Correct!") else: print("Wrong, the answer was {0}".format(LOCATION)) print("Attempt took {0} seconds".format(end - start)) Add functions to read/write to a record file"""Where player runs the game""" import random import time import board import conversion import games from settings import RECORD_FILE def write_record_to_file(a_string, file_name): with open(file_name, 'w') as f: f.write(a_string) def get_record_from_file(file_name): with open(file_name, 'r') as f: result = f.readline() return result if __name__ == '__main__': NUMBER_OF_TRIALS = 1 for i in range(NUMBER_OF_TRIALS): X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square() LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS) ORIENTATIONS = (True, False) GAMES = games.GamePositions() BOARD_DESCRIPTION = GAMES.random_game() BOARD = board.Board(BOARD_DESCRIPTION, orientation=random.choice(ORIENTATIONS)) BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS) BOARD.update_board_string() print(BOARD) start = time.time() ATTEMPT = input('Enter highlighted square: ') end = time.time() if ATTEMPT == LOCATION: print("Correct!") else: print("Wrong, the answer was {0}".format(LOCATION)) print("Attempt took {0} seconds".format(end - start))
<commit_before>"""Where player runs the game""" import random import time import board import conversion import games if __name__ == '__main__': NUMBER_OF_TRIALS = 1 for i in range(NUMBER_OF_TRIALS): X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square() LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS) ORIENTATIONS = (True, False) GAMES = games.GamePositions() BOARD_DESCRIPTION = GAMES.random_game() BOARD = board.Board(BOARD_DESCRIPTION, orientation=random.choice(ORIENTATIONS)) BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS) BOARD.update_board_string() print(BOARD) start = time.time() ATTEMPT = input('Enter highlighted square: ') end = time.time() if ATTEMPT == LOCATION: print("Correct!") else: print("Wrong, the answer was {0}".format(LOCATION)) print("Attempt took {0} seconds".format(end - start)) <commit_msg>Add functions to read/write to a record file<commit_after>"""Where player runs the game""" import random import time import board import conversion import games from settings import RECORD_FILE def write_record_to_file(a_string, file_name): with open(file_name, 'w') as f: f.write(a_string) def get_record_from_file(file_name): with open(file_name, 'r') as f: result = f.readline() return result if __name__ == '__main__': NUMBER_OF_TRIALS = 1 for i in range(NUMBER_OF_TRIALS): X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square() LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS) ORIENTATIONS = (True, False) GAMES = games.GamePositions() BOARD_DESCRIPTION = GAMES.random_game() BOARD = board.Board(BOARD_DESCRIPTION, orientation=random.choice(ORIENTATIONS)) BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS) BOARD.update_board_string() print(BOARD) start = time.time() ATTEMPT = input('Enter highlighted square: ') end = time.time() if ATTEMPT == LOCATION: print("Correct!") else: print("Wrong, the answer was {0}".format(LOCATION)) print("Attempt took {0} seconds".format(end - start))
96d12496e425806a635ba345a534c0ca2790754d
satchmo/apps/payment/modules/giftcertificate/processor.py
satchmo/apps/payment/modules/giftcertificate/processor.py
""" GiftCertificate processor """ from django.utils.translation import ugettext as _ from l10n.utils import moneyfmt from models import GiftCertificate from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET class PaymentProcessor(BasePaymentProcessor): def __init__(self, settings): super(PaymentProcessor, self).__init__('giftcertificate', settings) def capture_payment(self, testing=False, order=None, amount=NOTSET): """ Process the transaction and return a ProcessorResponse """ if not order: order = self.order if amount==NOTSET: amount = order.balance payment = None if self.order.paid_in_full: success = True reason_code = "0" response_text = _("No balance to pay") else: try: gc = GiftCertificate.objects.from_order(self.order) except GiftCertificate.DoesNotExist: success = False reason_code="1" response_text = _("No such Gift Certificate") if not gc.valid: success = False reason_code="2" response_text = _("Bad Gift Certificate") else: gc.apply_to_order(self.order) payment = gc.orderpayment reason_code = "0" response_text = _("Success") success = True if not self.order.paid_in_full: response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance) return ProcessorResult(self.key, success, response_text, payment=payment)
""" GiftCertificate processor """ from django.utils.translation import ugettext as _ from l10n.utils import moneyfmt from models import GiftCertificate from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET class PaymentProcessor(BasePaymentProcessor): def __init__(self, settings): super(PaymentProcessor, self).__init__('giftcertificate', settings) def capture_payment(self, testing=False, order=None, amount=NOTSET): """ Process the transaction and return a ProcessorResponse """ if not order: order = self.order if amount==NOTSET: amount = order.balance payment = None valid_gc = False if self.order.paid_in_full: success = True reason_code = "0" response_text = _("No balance to pay") else: try: gc = GiftCertificate.objects.from_order(self.order) valid_gc = gc.valid except GiftCertificate.DoesNotExist: success = False reason_code="1" response_text = _("No such Gift Certificate") if not valid_gc: success = False reason_code="2" response_text = _("Bad Gift Certificate") else: gc.apply_to_order(self.order) payment = gc.orderpayment reason_code = "0" response_text = _("Success") success = True if not self.order.paid_in_full: response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance) return ProcessorResult(self.key, success, response_text, payment=payment)
Fix the gift certificate module so that an invalid code won't throw an exception.
Fix the gift certificate module so that an invalid code won't throw an exception.
Python
bsd-3-clause
twidi/satchmo,ringemup/satchmo,ringemup/satchmo,dokterbob/satchmo,twidi/satchmo,dokterbob/satchmo,Ryati/satchmo,Ryati/satchmo
""" GiftCertificate processor """ from django.utils.translation import ugettext as _ from l10n.utils import moneyfmt from models import GiftCertificate from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET class PaymentProcessor(BasePaymentProcessor): def __init__(self, settings): super(PaymentProcessor, self).__init__('giftcertificate', settings) def capture_payment(self, testing=False, order=None, amount=NOTSET): """ Process the transaction and return a ProcessorResponse """ if not order: order = self.order if amount==NOTSET: amount = order.balance payment = None if self.order.paid_in_full: success = True reason_code = "0" response_text = _("No balance to pay") else: try: gc = GiftCertificate.objects.from_order(self.order) except GiftCertificate.DoesNotExist: success = False reason_code="1" response_text = _("No such Gift Certificate") if not gc.valid: success = False reason_code="2" response_text = _("Bad Gift Certificate") else: gc.apply_to_order(self.order) payment = gc.orderpayment reason_code = "0" response_text = _("Success") success = True if not self.order.paid_in_full: response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance) return ProcessorResult(self.key, success, response_text, payment=payment) Fix the gift certificate module so that an invalid code won't throw an exception.
""" GiftCertificate processor """ from django.utils.translation import ugettext as _ from l10n.utils import moneyfmt from models import GiftCertificate from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET class PaymentProcessor(BasePaymentProcessor): def __init__(self, settings): super(PaymentProcessor, self).__init__('giftcertificate', settings) def capture_payment(self, testing=False, order=None, amount=NOTSET): """ Process the transaction and return a ProcessorResponse """ if not order: order = self.order if amount==NOTSET: amount = order.balance payment = None valid_gc = False if self.order.paid_in_full: success = True reason_code = "0" response_text = _("No balance to pay") else: try: gc = GiftCertificate.objects.from_order(self.order) valid_gc = gc.valid except GiftCertificate.DoesNotExist: success = False reason_code="1" response_text = _("No such Gift Certificate") if not valid_gc: success = False reason_code="2" response_text = _("Bad Gift Certificate") else: gc.apply_to_order(self.order) payment = gc.orderpayment reason_code = "0" response_text = _("Success") success = True if not self.order.paid_in_full: response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance) return ProcessorResult(self.key, success, response_text, payment=payment)
<commit_before>""" GiftCertificate processor """ from django.utils.translation import ugettext as _ from l10n.utils import moneyfmt from models import GiftCertificate from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET class PaymentProcessor(BasePaymentProcessor): def __init__(self, settings): super(PaymentProcessor, self).__init__('giftcertificate', settings) def capture_payment(self, testing=False, order=None, amount=NOTSET): """ Process the transaction and return a ProcessorResponse """ if not order: order = self.order if amount==NOTSET: amount = order.balance payment = None if self.order.paid_in_full: success = True reason_code = "0" response_text = _("No balance to pay") else: try: gc = GiftCertificate.objects.from_order(self.order) except GiftCertificate.DoesNotExist: success = False reason_code="1" response_text = _("No such Gift Certificate") if not gc.valid: success = False reason_code="2" response_text = _("Bad Gift Certificate") else: gc.apply_to_order(self.order) payment = gc.orderpayment reason_code = "0" response_text = _("Success") success = True if not self.order.paid_in_full: response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance) return ProcessorResult(self.key, success, response_text, payment=payment) <commit_msg>Fix the gift certificate module so that an invalid code won't throw an exception.<commit_after>
""" GiftCertificate processor """ from django.utils.translation import ugettext as _ from l10n.utils import moneyfmt from models import GiftCertificate from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET class PaymentProcessor(BasePaymentProcessor): def __init__(self, settings): super(PaymentProcessor, self).__init__('giftcertificate', settings) def capture_payment(self, testing=False, order=None, amount=NOTSET): """ Process the transaction and return a ProcessorResponse """ if not order: order = self.order if amount==NOTSET: amount = order.balance payment = None valid_gc = False if self.order.paid_in_full: success = True reason_code = "0" response_text = _("No balance to pay") else: try: gc = GiftCertificate.objects.from_order(self.order) valid_gc = gc.valid except GiftCertificate.DoesNotExist: success = False reason_code="1" response_text = _("No such Gift Certificate") if not valid_gc: success = False reason_code="2" response_text = _("Bad Gift Certificate") else: gc.apply_to_order(self.order) payment = gc.orderpayment reason_code = "0" response_text = _("Success") success = True if not self.order.paid_in_full: response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance) return ProcessorResult(self.key, success, response_text, payment=payment)
""" GiftCertificate processor """ from django.utils.translation import ugettext as _ from l10n.utils import moneyfmt from models import GiftCertificate from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET class PaymentProcessor(BasePaymentProcessor): def __init__(self, settings): super(PaymentProcessor, self).__init__('giftcertificate', settings) def capture_payment(self, testing=False, order=None, amount=NOTSET): """ Process the transaction and return a ProcessorResponse """ if not order: order = self.order if amount==NOTSET: amount = order.balance payment = None if self.order.paid_in_full: success = True reason_code = "0" response_text = _("No balance to pay") else: try: gc = GiftCertificate.objects.from_order(self.order) except GiftCertificate.DoesNotExist: success = False reason_code="1" response_text = _("No such Gift Certificate") if not gc.valid: success = False reason_code="2" response_text = _("Bad Gift Certificate") else: gc.apply_to_order(self.order) payment = gc.orderpayment reason_code = "0" response_text = _("Success") success = True if not self.order.paid_in_full: response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance) return ProcessorResult(self.key, success, response_text, payment=payment) Fix the gift certificate module so that an invalid code won't throw an exception.""" GiftCertificate processor """ from django.utils.translation import ugettext as _ from l10n.utils import moneyfmt from models import GiftCertificate from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET class PaymentProcessor(BasePaymentProcessor): def __init__(self, settings): super(PaymentProcessor, self).__init__('giftcertificate', settings) def capture_payment(self, testing=False, order=None, amount=NOTSET): """ Process the transaction and return a ProcessorResponse """ if not order: order = self.order if amount==NOTSET: amount = order.balance payment = None valid_gc = False if self.order.paid_in_full: success = True reason_code = "0" response_text = _("No balance to pay") else: try: gc = GiftCertificate.objects.from_order(self.order) valid_gc = gc.valid except GiftCertificate.DoesNotExist: success = False reason_code="1" response_text = _("No such Gift Certificate") if not valid_gc: success = False reason_code="2" response_text = _("Bad Gift Certificate") else: gc.apply_to_order(self.order) payment = gc.orderpayment reason_code = "0" response_text = _("Success") success = True if not self.order.paid_in_full: response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance) return ProcessorResult(self.key, success, response_text, payment=payment)
<commit_before>""" GiftCertificate processor """ from django.utils.translation import ugettext as _ from l10n.utils import moneyfmt from models import GiftCertificate from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET class PaymentProcessor(BasePaymentProcessor): def __init__(self, settings): super(PaymentProcessor, self).__init__('giftcertificate', settings) def capture_payment(self, testing=False, order=None, amount=NOTSET): """ Process the transaction and return a ProcessorResponse """ if not order: order = self.order if amount==NOTSET: amount = order.balance payment = None if self.order.paid_in_full: success = True reason_code = "0" response_text = _("No balance to pay") else: try: gc = GiftCertificate.objects.from_order(self.order) except GiftCertificate.DoesNotExist: success = False reason_code="1" response_text = _("No such Gift Certificate") if not gc.valid: success = False reason_code="2" response_text = _("Bad Gift Certificate") else: gc.apply_to_order(self.order) payment = gc.orderpayment reason_code = "0" response_text = _("Success") success = True if not self.order.paid_in_full: response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance) return ProcessorResult(self.key, success, response_text, payment=payment) <commit_msg>Fix the gift certificate module so that an invalid code won't throw an exception.<commit_after>""" GiftCertificate processor """ from django.utils.translation import ugettext as _ from l10n.utils import moneyfmt from models import GiftCertificate from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET class PaymentProcessor(BasePaymentProcessor): def __init__(self, settings): super(PaymentProcessor, self).__init__('giftcertificate', settings) def capture_payment(self, testing=False, order=None, amount=NOTSET): """ Process the transaction and return a ProcessorResponse """ if not order: order = self.order if amount==NOTSET: amount = order.balance payment = None valid_gc = False if self.order.paid_in_full: success = True reason_code = "0" response_text = _("No balance to pay") else: try: gc = GiftCertificate.objects.from_order(self.order) valid_gc = gc.valid except GiftCertificate.DoesNotExist: success = False reason_code="1" response_text = _("No such Gift Certificate") if not valid_gc: success = False reason_code="2" response_text = _("Bad Gift Certificate") else: gc.apply_to_order(self.order) payment = gc.orderpayment reason_code = "0" response_text = _("Success") success = True if not self.order.paid_in_full: response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance) return ProcessorResult(self.key, success, response_text, payment=payment)
056d82002c133736a800b08bd071b71c9f5615f8
ci/generate_pipeline_yml.py
ci/generate_pipeline_yml.py
#!/usr/bin/env python import os from jinja2 import Template clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml")
#!/usr/bin/env python import os from jinja2 import Template clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml")
Update TAS versions we test against
Update TAS versions we test against
Python
apache-2.0
cf-platform-eng/tile-generator,cf-platform-eng/tile-generator,cf-platform-eng/tile-generator,cf-platform-eng/tile-generator
#!/usr/bin/env python import os from jinja2 import Template clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml") Update TAS versions we test against
#!/usr/bin/env python import os from jinja2 import Template clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml")
<commit_before>#!/usr/bin/env python import os from jinja2 import Template clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml") <commit_msg>Update TAS versions we test against<commit_after>
#!/usr/bin/env python import os from jinja2 import Template clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml")
#!/usr/bin/env python import os from jinja2 import Template clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml") Update TAS versions we test against#!/usr/bin/env python import os from jinja2 import Template clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml")
<commit_before>#!/usr/bin/env python import os from jinja2 import Template clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml") <commit_msg>Update TAS versions we test against<commit_after>#!/usr/bin/env python import os from jinja2 import Template clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13'] # Commenting out this as we only have one example and it breaks tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))] with open('pipeline.yml.jinja2', 'r') as f: t = Template(f.read()); with open('pipeline.yml', 'w') as f: f.write(t.render(clusters=clusters, tiles=tiles)) print("Successfully generated pipeline.yml")
bc8d7a7572fcde45ae95176301522979fa54aa87
carnifex/test/unit/mocks.py
carnifex/test/unit/mocks.py
from twisted.internet._baseprocess import BaseProcess from carnifex.inductor import ProcessInductor from twisted.internet.error import ProcessTerminated, ProcessDone class MockProcess(BaseProcess): def run(self, fauxProcessData): for childFd, data in fauxProcessData: self.proto.childDataReceived(childFd, data) def terminate(self, signal): """Simulate that the process was terminated with a signal """ self.processEnded((None, signal)) def exit(self, exitCode): """Simulate that the process exited """ self.processEnded((exitCode, None)) def _getReason(self, status): exitCode, signal = status if exitCode or signal: return ProcessTerminated(exitCode, signal, status) return ProcessDone(status) class MockProcessInductor(ProcessInductor): def __init__(self, reactor, fauxProcessData): self.reactor = reactor self.fauxProcessData = fauxProcessData def execute(self, processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): process = MockProcess(processProtocol) process.run(self.fauxProcessData) processProtocol.makeConnection(process) process.exit(0) return process
from twisted.internet._baseprocess import BaseProcess from carnifex.inductor import ProcessInductor from twisted.internet.error import ProcessTerminated, ProcessDone class MockProcess(BaseProcess): def run(self, fauxProcessData): for childFd, data in fauxProcessData: self.proto.childDataReceived(childFd, data) def terminate(self, signal): """Simulate that the process was terminated with a signal """ self.processEnded((None, signal)) def exit(self, exitCode): """Simulate that the process exited """ self.processEnded((exitCode, None)) def _getReason(self, status): exitCode, signal = status if exitCode or signal: return ProcessTerminated(exitCode, signal, status) return ProcessDone(status) class MockProcessInductor(ProcessInductor): def __init__(self, reactor, fauxProcessData, exitCode=0): self.reactor = reactor self.fauxProcessData = fauxProcessData self.exitCode = exitCode def execute(self, processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): process = MockProcess(processProtocol) process.run(self.fauxProcessData) processProtocol.makeConnection(process) process.exit(self.exitCode) return process
Allow specifying what exit code to use when emulating process exit
Allow specifying what exit code to use when emulating process exit
Python
mit
sporsh/carnifex
from twisted.internet._baseprocess import BaseProcess from carnifex.inductor import ProcessInductor from twisted.internet.error import ProcessTerminated, ProcessDone class MockProcess(BaseProcess): def run(self, fauxProcessData): for childFd, data in fauxProcessData: self.proto.childDataReceived(childFd, data) def terminate(self, signal): """Simulate that the process was terminated with a signal """ self.processEnded((None, signal)) def exit(self, exitCode): """Simulate that the process exited """ self.processEnded((exitCode, None)) def _getReason(self, status): exitCode, signal = status if exitCode or signal: return ProcessTerminated(exitCode, signal, status) return ProcessDone(status) class MockProcessInductor(ProcessInductor): def __init__(self, reactor, fauxProcessData): self.reactor = reactor self.fauxProcessData = fauxProcessData def execute(self, processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): process = MockProcess(processProtocol) process.run(self.fauxProcessData) processProtocol.makeConnection(process) process.exit(0) return process Allow specifying what exit code to use when emulating process exit
from twisted.internet._baseprocess import BaseProcess from carnifex.inductor import ProcessInductor from twisted.internet.error import ProcessTerminated, ProcessDone class MockProcess(BaseProcess): def run(self, fauxProcessData): for childFd, data in fauxProcessData: self.proto.childDataReceived(childFd, data) def terminate(self, signal): """Simulate that the process was terminated with a signal """ self.processEnded((None, signal)) def exit(self, exitCode): """Simulate that the process exited """ self.processEnded((exitCode, None)) def _getReason(self, status): exitCode, signal = status if exitCode or signal: return ProcessTerminated(exitCode, signal, status) return ProcessDone(status) class MockProcessInductor(ProcessInductor): def __init__(self, reactor, fauxProcessData, exitCode=0): self.reactor = reactor self.fauxProcessData = fauxProcessData self.exitCode = exitCode def execute(self, processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): process = MockProcess(processProtocol) process.run(self.fauxProcessData) processProtocol.makeConnection(process) process.exit(self.exitCode) return process
<commit_before>from twisted.internet._baseprocess import BaseProcess from carnifex.inductor import ProcessInductor from twisted.internet.error import ProcessTerminated, ProcessDone class MockProcess(BaseProcess): def run(self, fauxProcessData): for childFd, data in fauxProcessData: self.proto.childDataReceived(childFd, data) def terminate(self, signal): """Simulate that the process was terminated with a signal """ self.processEnded((None, signal)) def exit(self, exitCode): """Simulate that the process exited """ self.processEnded((exitCode, None)) def _getReason(self, status): exitCode, signal = status if exitCode or signal: return ProcessTerminated(exitCode, signal, status) return ProcessDone(status) class MockProcessInductor(ProcessInductor): def __init__(self, reactor, fauxProcessData): self.reactor = reactor self.fauxProcessData = fauxProcessData def execute(self, processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): process = MockProcess(processProtocol) process.run(self.fauxProcessData) processProtocol.makeConnection(process) process.exit(0) return process <commit_msg>Allow specifying what exit code to use when emulating process exit<commit_after>
from twisted.internet._baseprocess import BaseProcess from carnifex.inductor import ProcessInductor from twisted.internet.error import ProcessTerminated, ProcessDone class MockProcess(BaseProcess): def run(self, fauxProcessData): for childFd, data in fauxProcessData: self.proto.childDataReceived(childFd, data) def terminate(self, signal): """Simulate that the process was terminated with a signal """ self.processEnded((None, signal)) def exit(self, exitCode): """Simulate that the process exited """ self.processEnded((exitCode, None)) def _getReason(self, status): exitCode, signal = status if exitCode or signal: return ProcessTerminated(exitCode, signal, status) return ProcessDone(status) class MockProcessInductor(ProcessInductor): def __init__(self, reactor, fauxProcessData, exitCode=0): self.reactor = reactor self.fauxProcessData = fauxProcessData self.exitCode = exitCode def execute(self, processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): process = MockProcess(processProtocol) process.run(self.fauxProcessData) processProtocol.makeConnection(process) process.exit(self.exitCode) return process
from twisted.internet._baseprocess import BaseProcess from carnifex.inductor import ProcessInductor from twisted.internet.error import ProcessTerminated, ProcessDone class MockProcess(BaseProcess): def run(self, fauxProcessData): for childFd, data in fauxProcessData: self.proto.childDataReceived(childFd, data) def terminate(self, signal): """Simulate that the process was terminated with a signal """ self.processEnded((None, signal)) def exit(self, exitCode): """Simulate that the process exited """ self.processEnded((exitCode, None)) def _getReason(self, status): exitCode, signal = status if exitCode or signal: return ProcessTerminated(exitCode, signal, status) return ProcessDone(status) class MockProcessInductor(ProcessInductor): def __init__(self, reactor, fauxProcessData): self.reactor = reactor self.fauxProcessData = fauxProcessData def execute(self, processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): process = MockProcess(processProtocol) process.run(self.fauxProcessData) processProtocol.makeConnection(process) process.exit(0) return process Allow specifying what exit code to use when emulating process exitfrom twisted.internet._baseprocess import BaseProcess from carnifex.inductor import ProcessInductor from twisted.internet.error import ProcessTerminated, ProcessDone class MockProcess(BaseProcess): def run(self, fauxProcessData): for childFd, data in fauxProcessData: self.proto.childDataReceived(childFd, data) def terminate(self, signal): """Simulate that the process was terminated with a signal """ self.processEnded((None, signal)) def exit(self, exitCode): """Simulate that the process exited """ self.processEnded((exitCode, None)) def _getReason(self, status): exitCode, signal = status if exitCode or signal: return ProcessTerminated(exitCode, signal, status) return ProcessDone(status) class MockProcessInductor(ProcessInductor): def __init__(self, reactor, fauxProcessData, exitCode=0): self.reactor = reactor self.fauxProcessData = fauxProcessData self.exitCode = exitCode def execute(self, processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): process = MockProcess(processProtocol) process.run(self.fauxProcessData) processProtocol.makeConnection(process) process.exit(self.exitCode) return process
<commit_before>from twisted.internet._baseprocess import BaseProcess from carnifex.inductor import ProcessInductor from twisted.internet.error import ProcessTerminated, ProcessDone class MockProcess(BaseProcess): def run(self, fauxProcessData): for childFd, data in fauxProcessData: self.proto.childDataReceived(childFd, data) def terminate(self, signal): """Simulate that the process was terminated with a signal """ self.processEnded((None, signal)) def exit(self, exitCode): """Simulate that the process exited """ self.processEnded((exitCode, None)) def _getReason(self, status): exitCode, signal = status if exitCode or signal: return ProcessTerminated(exitCode, signal, status) return ProcessDone(status) class MockProcessInductor(ProcessInductor): def __init__(self, reactor, fauxProcessData): self.reactor = reactor self.fauxProcessData = fauxProcessData def execute(self, processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): process = MockProcess(processProtocol) process.run(self.fauxProcessData) processProtocol.makeConnection(process) process.exit(0) return process <commit_msg>Allow specifying what exit code to use when emulating process exit<commit_after>from twisted.internet._baseprocess import BaseProcess from carnifex.inductor import ProcessInductor from twisted.internet.error import ProcessTerminated, ProcessDone class MockProcess(BaseProcess): def run(self, fauxProcessData): for childFd, data in fauxProcessData: self.proto.childDataReceived(childFd, data) def terminate(self, signal): """Simulate that the process was terminated with a signal """ self.processEnded((None, signal)) def exit(self, exitCode): """Simulate that the process exited """ self.processEnded((exitCode, None)) def _getReason(self, status): exitCode, signal = status if exitCode or signal: return ProcessTerminated(exitCode, signal, status) return ProcessDone(status) class MockProcessInductor(ProcessInductor): def __init__(self, reactor, fauxProcessData, exitCode=0): self.reactor = reactor self.fauxProcessData = fauxProcessData self.exitCode = exitCode def execute(self, processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): process = MockProcess(processProtocol) process.run(self.fauxProcessData) processProtocol.makeConnection(process) process.exit(self.exitCode) return process
fdaabeaa3694103153c81a18971e6b55597cd66e
Sketches/JT/Jam/library/trunk/Kamaelia/Apps/Jam/Audio/Synth.py
Sketches/JT/Jam/library/trunk/Kamaelia/Apps/Jam/Audio/Synth.py
import Axon from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser from Kamaelia.Apps.Jam.Audio.Mixer import MonoMixer class Synth(Axon.Component.component): polyphony = 8 polyphoniser = Polyphoniser def __init__(self, voiceGenerator, **argd): super(Synth, self).__init__(**argd) polyphoniser = self.polyphoniser(**argd).activate() mixer = MonoMixer(channels=self.polyphony, **argd).activate() self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1) self.link((mixer, "outbox"), (self, "outbox"), passthrough=2) for index, voice in enumerate(voiceGenerator()): voice = voice.activate() self.link((polyphoniser, "voice%i" % index), (voice, "inbox")) self.link((voice, "outbox"), (mixer, "in%i" % index)) def main(self): while 1: if not self.anyReady(): self.pause() yield 1
import Axon from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser class Synth(Axon.Component.component): polyphony = 8 polyphoniser = Polyphoniser def __init__(self, voiceGenerator, **argd): super(Synth, self).__init__(**argd) polyphoniser = self.polyphoniser(**argd).activate() self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1) for index, voice in enumerate(voiceGenerator()): voice = voice.activate() self.link((polyphoniser, "voice%i" % index), (voice, "inbox")) def main(self): while 1: if not self.anyReady(): self.pause() yield 1
Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.
Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.
Python
apache-2.0
sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia
import Axon from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser from Kamaelia.Apps.Jam.Audio.Mixer import MonoMixer class Synth(Axon.Component.component): polyphony = 8 polyphoniser = Polyphoniser def __init__(self, voiceGenerator, **argd): super(Synth, self).__init__(**argd) polyphoniser = self.polyphoniser(**argd).activate() mixer = MonoMixer(channels=self.polyphony, **argd).activate() self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1) self.link((mixer, "outbox"), (self, "outbox"), passthrough=2) for index, voice in enumerate(voiceGenerator()): voice = voice.activate() self.link((polyphoniser, "voice%i" % index), (voice, "inbox")) self.link((voice, "outbox"), (mixer, "in%i" % index)) def main(self): while 1: if not self.anyReady(): self.pause() yield 1 Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.
import Axon from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser class Synth(Axon.Component.component): polyphony = 8 polyphoniser = Polyphoniser def __init__(self, voiceGenerator, **argd): super(Synth, self).__init__(**argd) polyphoniser = self.polyphoniser(**argd).activate() self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1) for index, voice in enumerate(voiceGenerator()): voice = voice.activate() self.link((polyphoniser, "voice%i" % index), (voice, "inbox")) def main(self): while 1: if not self.anyReady(): self.pause() yield 1
<commit_before>import Axon from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser from Kamaelia.Apps.Jam.Audio.Mixer import MonoMixer class Synth(Axon.Component.component): polyphony = 8 polyphoniser = Polyphoniser def __init__(self, voiceGenerator, **argd): super(Synth, self).__init__(**argd) polyphoniser = self.polyphoniser(**argd).activate() mixer = MonoMixer(channels=self.polyphony, **argd).activate() self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1) self.link((mixer, "outbox"), (self, "outbox"), passthrough=2) for index, voice in enumerate(voiceGenerator()): voice = voice.activate() self.link((polyphoniser, "voice%i" % index), (voice, "inbox")) self.link((voice, "outbox"), (mixer, "in%i" % index)) def main(self): while 1: if not self.anyReady(): self.pause() yield 1 <commit_msg>Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.<commit_after>
import Axon from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser class Synth(Axon.Component.component): polyphony = 8 polyphoniser = Polyphoniser def __init__(self, voiceGenerator, **argd): super(Synth, self).__init__(**argd) polyphoniser = self.polyphoniser(**argd).activate() self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1) for index, voice in enumerate(voiceGenerator()): voice = voice.activate() self.link((polyphoniser, "voice%i" % index), (voice, "inbox")) def main(self): while 1: if not self.anyReady(): self.pause() yield 1
import Axon from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser from Kamaelia.Apps.Jam.Audio.Mixer import MonoMixer class Synth(Axon.Component.component): polyphony = 8 polyphoniser = Polyphoniser def __init__(self, voiceGenerator, **argd): super(Synth, self).__init__(**argd) polyphoniser = self.polyphoniser(**argd).activate() mixer = MonoMixer(channels=self.polyphony, **argd).activate() self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1) self.link((mixer, "outbox"), (self, "outbox"), passthrough=2) for index, voice in enumerate(voiceGenerator()): voice = voice.activate() self.link((polyphoniser, "voice%i" % index), (voice, "inbox")) self.link((voice, "outbox"), (mixer, "in%i" % index)) def main(self): while 1: if not self.anyReady(): self.pause() yield 1 Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.import Axon from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser class Synth(Axon.Component.component): polyphony = 8 polyphoniser = Polyphoniser def __init__(self, voiceGenerator, **argd): super(Synth, self).__init__(**argd) polyphoniser = self.polyphoniser(**argd).activate() self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1) for index, voice in enumerate(voiceGenerator()): voice = voice.activate() self.link((polyphoniser, "voice%i" % index), (voice, "inbox")) def main(self): while 1: if not self.anyReady(): self.pause() yield 1
<commit_before>import Axon from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser from Kamaelia.Apps.Jam.Audio.Mixer import MonoMixer class Synth(Axon.Component.component): polyphony = 8 polyphoniser = Polyphoniser def __init__(self, voiceGenerator, **argd): super(Synth, self).__init__(**argd) polyphoniser = self.polyphoniser(**argd).activate() mixer = MonoMixer(channels=self.polyphony, **argd).activate() self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1) self.link((mixer, "outbox"), (self, "outbox"), passthrough=2) for index, voice in enumerate(voiceGenerator()): voice = voice.activate() self.link((polyphoniser, "voice%i" % index), (voice, "inbox")) self.link((voice, "outbox"), (mixer, "in%i" % index)) def main(self): while 1: if not self.anyReady(): self.pause() yield 1 <commit_msg>Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.<commit_after>import Axon from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser class Synth(Axon.Component.component): polyphony = 8 polyphoniser = Polyphoniser def __init__(self, voiceGenerator, **argd): super(Synth, self).__init__(**argd) polyphoniser = self.polyphoniser(**argd).activate() self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1) for index, voice in enumerate(voiceGenerator()): voice = voice.activate() self.link((polyphoniser, "voice%i" % index), (voice, "inbox")) def main(self): while 1: if not self.anyReady(): self.pause() yield 1
75e14847fe2c0f0c40897e449bab093f4be1b17c
cineapp/jinja_filters.py
cineapp/jinja_filters.py
# -*- coding: utf-8 -*- from cineapp import app @app.template_filter() def minutes_to_human_duration(minutes_duration): """ Convert a duration in minutes into a duration in a cool format human readable """ try: hours,minutes = divmod(minutes_duration,60) return "%sh %smin" %(hours,minutes) except TypeError: return None @app.template_filter() def date_format(date,format_date): """ Convert a date object into a custom format """ try: return date.strftime(format_date) except: return None
# -*- coding: utf-8 -*- from cineapp import app import datetime @app.template_filter() def minutes_to_human_duration(minutes_duration): """ Convert a duration in minutes into a duration in a cool format human readable """ try: hours,minutes = divmod(minutes_duration,60) return "%sh %smin" %(hours,minutes) except TypeError: return None @app.template_filter() def date_format(date,format_date): """ Convert a date object into a custom format """ try: if isinstance(date, basestring): date_array=date.split('-') date_to_convert=datetime.datetime(int(date_array[0]),int(date_array[1]),int(date_array[2])) return date_to_convert.strftime(format_date) else: return date.strftime(format_date) except Exception,e: print e return None
Improve jinja filter date converter
Improve jinja filter date converter The filter now can convert date which are strings and not datetime objects.
Python
mit
ptitoliv/cineapp,ptitoliv/cineapp,ptitoliv/cineapp
# -*- coding: utf-8 -*- from cineapp import app @app.template_filter() def minutes_to_human_duration(minutes_duration): """ Convert a duration in minutes into a duration in a cool format human readable """ try: hours,minutes = divmod(minutes_duration,60) return "%sh %smin" %(hours,minutes) except TypeError: return None @app.template_filter() def date_format(date,format_date): """ Convert a date object into a custom format """ try: return date.strftime(format_date) except: return None Improve jinja filter date converter The filter now can convert date which are strings and not datetime objects.
# -*- coding: utf-8 -*- from cineapp import app import datetime @app.template_filter() def minutes_to_human_duration(minutes_duration): """ Convert a duration in minutes into a duration in a cool format human readable """ try: hours,minutes = divmod(minutes_duration,60) return "%sh %smin" %(hours,minutes) except TypeError: return None @app.template_filter() def date_format(date,format_date): """ Convert a date object into a custom format """ try: if isinstance(date, basestring): date_array=date.split('-') date_to_convert=datetime.datetime(int(date_array[0]),int(date_array[1]),int(date_array[2])) return date_to_convert.strftime(format_date) else: return date.strftime(format_date) except Exception,e: print e return None
<commit_before># -*- coding: utf-8 -*- from cineapp import app @app.template_filter() def minutes_to_human_duration(minutes_duration): """ Convert a duration in minutes into a duration in a cool format human readable """ try: hours,minutes = divmod(minutes_duration,60) return "%sh %smin" %(hours,minutes) except TypeError: return None @app.template_filter() def date_format(date,format_date): """ Convert a date object into a custom format """ try: return date.strftime(format_date) except: return None <commit_msg>Improve jinja filter date converter The filter now can convert date which are strings and not datetime objects.<commit_after>
# -*- coding: utf-8 -*- from cineapp import app import datetime @app.template_filter() def minutes_to_human_duration(minutes_duration): """ Convert a duration in minutes into a duration in a cool format human readable """ try: hours,minutes = divmod(minutes_duration,60) return "%sh %smin" %(hours,minutes) except TypeError: return None @app.template_filter() def date_format(date,format_date): """ Convert a date object into a custom format """ try: if isinstance(date, basestring): date_array=date.split('-') date_to_convert=datetime.datetime(int(date_array[0]),int(date_array[1]),int(date_array[2])) return date_to_convert.strftime(format_date) else: return date.strftime(format_date) except Exception,e: print e return None
# -*- coding: utf-8 -*- from cineapp import app @app.template_filter() def minutes_to_human_duration(minutes_duration): """ Convert a duration in minutes into a duration in a cool format human readable """ try: hours,minutes = divmod(minutes_duration,60) return "%sh %smin" %(hours,minutes) except TypeError: return None @app.template_filter() def date_format(date,format_date): """ Convert a date object into a custom format """ try: return date.strftime(format_date) except: return None Improve jinja filter date converter The filter now can convert date which are strings and not datetime objects.# -*- coding: utf-8 -*- from cineapp import app import datetime @app.template_filter() def minutes_to_human_duration(minutes_duration): """ Convert a duration in minutes into a duration in a cool format human readable """ try: hours,minutes = divmod(minutes_duration,60) return "%sh %smin" %(hours,minutes) except TypeError: return None @app.template_filter() def date_format(date,format_date): """ Convert a date object into a custom format """ try: if isinstance(date, basestring): date_array=date.split('-') date_to_convert=datetime.datetime(int(date_array[0]),int(date_array[1]),int(date_array[2])) return date_to_convert.strftime(format_date) else: return date.strftime(format_date) except Exception,e: print e return None
<commit_before># -*- coding: utf-8 -*- from cineapp import app @app.template_filter() def minutes_to_human_duration(minutes_duration): """ Convert a duration in minutes into a duration in a cool format human readable """ try: hours,minutes = divmod(minutes_duration,60) return "%sh %smin" %(hours,minutes) except TypeError: return None @app.template_filter() def date_format(date,format_date): """ Convert a date object into a custom format """ try: return date.strftime(format_date) except: return None <commit_msg>Improve jinja filter date converter The filter now can convert date which are strings and not datetime objects.<commit_after># -*- coding: utf-8 -*- from cineapp import app import datetime @app.template_filter() def minutes_to_human_duration(minutes_duration): """ Convert a duration in minutes into a duration in a cool format human readable """ try: hours,minutes = divmod(minutes_duration,60) return "%sh %smin" %(hours,minutes) except TypeError: return None @app.template_filter() def date_format(date,format_date): """ Convert a date object into a custom format """ try: if isinstance(date, basestring): date_array=date.split('-') date_to_convert=datetime.datetime(int(date_array[0]),int(date_array[1]),int(date_array[2])) return date_to_convert.strftime(format_date) else: return date.strftime(format_date) except Exception,e: print e return None
4974f83d9ed1e085ef2daaeba4db56a4001055cf
comics/comics/ctrlaltdel.py
comics/comics/ctrlaltdel.py
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Ctrl+Alt+Del" language = "en" url = "http://www.cad-comic.com/cad/" start_date = "2002-10-23" rights = "Tim Buckley" class Crawler(CrawlerBase): history_capable_date = "2002-10-23" schedule = "Mo,We,Fr" time_zone = "US/Eastern" # Without User-Agent set, the server returns empty responses headers = {"User-Agent": "Mozilla/4.0"} def crawl(self, pub_date): page = self.parse_page( "http://www.cad-comic.com/cad/%s" % pub_date.strftime("%Y%m%d") ) url = page.src('img[src*="/comics/"]') title = page.alt('img[src*="/comics/"]') return CrawlerImage(url, title)
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Ctrl+Alt+Del" language = "en" url = "https://cad-comic.com/category/ctrl-alt-del/" start_date = "2002-10-23" rights = "Tim Buckley" class Crawler(CrawlerBase): # history_capable_date = "2002-10-23" history_capable_days = 20 schedule = "Mo,We,Fr" time_zone = "US/Eastern" # Without User-Agent set, the server returns empty responses headers = {"User-Agent": "Mozilla/4.0"} def crawl(self, pub_date): feed = self.parse_feed("https://cad-comic.com/feed/") for entry in feed.for_date(pub_date): url = entry.summary.src("img") title = entry.title return CrawlerImage(url, title)
Update "Ctrl+Alt+Del" after site change
Update "Ctrl+Alt+Del" after site change
Python
agpl-3.0
datagutten/comics,jodal/comics,datagutten/comics,datagutten/comics,datagutten/comics,jodal/comics,jodal/comics,jodal/comics
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Ctrl+Alt+Del" language = "en" url = "http://www.cad-comic.com/cad/" start_date = "2002-10-23" rights = "Tim Buckley" class Crawler(CrawlerBase): history_capable_date = "2002-10-23" schedule = "Mo,We,Fr" time_zone = "US/Eastern" # Without User-Agent set, the server returns empty responses headers = {"User-Agent": "Mozilla/4.0"} def crawl(self, pub_date): page = self.parse_page( "http://www.cad-comic.com/cad/%s" % pub_date.strftime("%Y%m%d") ) url = page.src('img[src*="/comics/"]') title = page.alt('img[src*="/comics/"]') return CrawlerImage(url, title) Update "Ctrl+Alt+Del" after site change
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Ctrl+Alt+Del" language = "en" url = "https://cad-comic.com/category/ctrl-alt-del/" start_date = "2002-10-23" rights = "Tim Buckley" class Crawler(CrawlerBase): # history_capable_date = "2002-10-23" history_capable_days = 20 schedule = "Mo,We,Fr" time_zone = "US/Eastern" # Without User-Agent set, the server returns empty responses headers = {"User-Agent": "Mozilla/4.0"} def crawl(self, pub_date): feed = self.parse_feed("https://cad-comic.com/feed/") for entry in feed.for_date(pub_date): url = entry.summary.src("img") title = entry.title return CrawlerImage(url, title)
<commit_before>from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Ctrl+Alt+Del" language = "en" url = "http://www.cad-comic.com/cad/" start_date = "2002-10-23" rights = "Tim Buckley" class Crawler(CrawlerBase): history_capable_date = "2002-10-23" schedule = "Mo,We,Fr" time_zone = "US/Eastern" # Without User-Agent set, the server returns empty responses headers = {"User-Agent": "Mozilla/4.0"} def crawl(self, pub_date): page = self.parse_page( "http://www.cad-comic.com/cad/%s" % pub_date.strftime("%Y%m%d") ) url = page.src('img[src*="/comics/"]') title = page.alt('img[src*="/comics/"]') return CrawlerImage(url, title) <commit_msg>Update "Ctrl+Alt+Del" after site change<commit_after>
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Ctrl+Alt+Del" language = "en" url = "https://cad-comic.com/category/ctrl-alt-del/" start_date = "2002-10-23" rights = "Tim Buckley" class Crawler(CrawlerBase): # history_capable_date = "2002-10-23" history_capable_days = 20 schedule = "Mo,We,Fr" time_zone = "US/Eastern" # Without User-Agent set, the server returns empty responses headers = {"User-Agent": "Mozilla/4.0"} def crawl(self, pub_date): feed = self.parse_feed("https://cad-comic.com/feed/") for entry in feed.for_date(pub_date): url = entry.summary.src("img") title = entry.title return CrawlerImage(url, title)
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Ctrl+Alt+Del" language = "en" url = "http://www.cad-comic.com/cad/" start_date = "2002-10-23" rights = "Tim Buckley" class Crawler(CrawlerBase): history_capable_date = "2002-10-23" schedule = "Mo,We,Fr" time_zone = "US/Eastern" # Without User-Agent set, the server returns empty responses headers = {"User-Agent": "Mozilla/4.0"} def crawl(self, pub_date): page = self.parse_page( "http://www.cad-comic.com/cad/%s" % pub_date.strftime("%Y%m%d") ) url = page.src('img[src*="/comics/"]') title = page.alt('img[src*="/comics/"]') return CrawlerImage(url, title) Update "Ctrl+Alt+Del" after site changefrom comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Ctrl+Alt+Del" language = "en" url = "https://cad-comic.com/category/ctrl-alt-del/" start_date = "2002-10-23" rights = "Tim Buckley" class Crawler(CrawlerBase): # history_capable_date = "2002-10-23" history_capable_days = 20 schedule = "Mo,We,Fr" time_zone = "US/Eastern" # Without User-Agent set, the server returns empty responses headers = {"User-Agent": "Mozilla/4.0"} def crawl(self, pub_date): feed = self.parse_feed("https://cad-comic.com/feed/") for entry in feed.for_date(pub_date): url = entry.summary.src("img") title = entry.title return CrawlerImage(url, title)
<commit_before>from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Ctrl+Alt+Del" language = "en" url = "http://www.cad-comic.com/cad/" start_date = "2002-10-23" rights = "Tim Buckley" class Crawler(CrawlerBase): history_capable_date = "2002-10-23" schedule = "Mo,We,Fr" time_zone = "US/Eastern" # Without User-Agent set, the server returns empty responses headers = {"User-Agent": "Mozilla/4.0"} def crawl(self, pub_date): page = self.parse_page( "http://www.cad-comic.com/cad/%s" % pub_date.strftime("%Y%m%d") ) url = page.src('img[src*="/comics/"]') title = page.alt('img[src*="/comics/"]') return CrawlerImage(url, title) <commit_msg>Update "Ctrl+Alt+Del" after site change<commit_after>from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.core.comic_data import ComicDataBase class ComicData(ComicDataBase): name = "Ctrl+Alt+Del" language = "en" url = "https://cad-comic.com/category/ctrl-alt-del/" start_date = "2002-10-23" rights = "Tim Buckley" class Crawler(CrawlerBase): # history_capable_date = "2002-10-23" history_capable_days = 20 schedule = "Mo,We,Fr" time_zone = "US/Eastern" # Without User-Agent set, the server returns empty responses headers = {"User-Agent": "Mozilla/4.0"} def crawl(self, pub_date): feed = self.parse_feed("https://cad-comic.com/feed/") for entry in feed.for_date(pub_date): url = entry.summary.src("img") title = entry.title return CrawlerImage(url, title)
4007ecdc66e361bcb81bb5b661e682eeef0a6ea5
remo/profiles/migrations/0011_groups_new_onboarding_group.py
remo/profiles/migrations/0011_groups_new_onboarding_group.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models def forwards(apps, schema_editor): """Create Onboarding group.""" Group = apps.get_model('auth', 'Group') Group.objects.create(name='Onboarding') def backwards(apps, schema_editor): """Delete Onboarding group.""" Group = apps.get_model('auth', 'Group') Group.objects.filter(name='Onboarding').delete() class Migration(migrations.Migration): dependencies = [ ('profiles', '0010_auto_20171221_0112'), ] operations = [ migrations.RunPython(forwards, backwards) ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models def forwards(apps, schema_editor): """Create Onboarding group.""" Group = apps.get_model('auth', 'Group') if not Group.objects.filter(name='Onboarding').exists(): Group.objects.create(name='Onboarding') def backwards(apps, schema_editor): """Delete Onboarding group.""" Group = apps.get_model('auth', 'Group') if Group.objects.filter(name='Onboarding').exists(): Group.objects.filter(name='Onboarding').delete() class Migration(migrations.Migration): dependencies = [ ('profiles', '0010_auto_20171221_0112'), ] operations = [ migrations.RunPython(forwards, backwards) ]
Check if Onboarding exists before creating.
Check if Onboarding exists before creating.
Python
bsd-3-clause
mozilla/remo,akatsoulas/remo,Mte90/remo,mozilla/remo,Mte90/remo,akatsoulas/remo,mozilla/remo,Mte90/remo,mozilla/remo,akatsoulas/remo,akatsoulas/remo,Mte90/remo
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models def forwards(apps, schema_editor): """Create Onboarding group.""" Group = apps.get_model('auth', 'Group') Group.objects.create(name='Onboarding') def backwards(apps, schema_editor): """Delete Onboarding group.""" Group = apps.get_model('auth', 'Group') Group.objects.filter(name='Onboarding').delete() class Migration(migrations.Migration): dependencies = [ ('profiles', '0010_auto_20171221_0112'), ] operations = [ migrations.RunPython(forwards, backwards) ] Check if Onboarding exists before creating.
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models def forwards(apps, schema_editor): """Create Onboarding group.""" Group = apps.get_model('auth', 'Group') if not Group.objects.filter(name='Onboarding').exists(): Group.objects.create(name='Onboarding') def backwards(apps, schema_editor): """Delete Onboarding group.""" Group = apps.get_model('auth', 'Group') if Group.objects.filter(name='Onboarding').exists(): Group.objects.filter(name='Onboarding').delete() class Migration(migrations.Migration): dependencies = [ ('profiles', '0010_auto_20171221_0112'), ] operations = [ migrations.RunPython(forwards, backwards) ]
<commit_before># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models def forwards(apps, schema_editor): """Create Onboarding group.""" Group = apps.get_model('auth', 'Group') Group.objects.create(name='Onboarding') def backwards(apps, schema_editor): """Delete Onboarding group.""" Group = apps.get_model('auth', 'Group') Group.objects.filter(name='Onboarding').delete() class Migration(migrations.Migration): dependencies = [ ('profiles', '0010_auto_20171221_0112'), ] operations = [ migrations.RunPython(forwards, backwards) ] <commit_msg>Check if Onboarding exists before creating.<commit_after>
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models def forwards(apps, schema_editor): """Create Onboarding group.""" Group = apps.get_model('auth', 'Group') if not Group.objects.filter(name='Onboarding').exists(): Group.objects.create(name='Onboarding') def backwards(apps, schema_editor): """Delete Onboarding group.""" Group = apps.get_model('auth', 'Group') if Group.objects.filter(name='Onboarding').exists(): Group.objects.filter(name='Onboarding').delete() class Migration(migrations.Migration): dependencies = [ ('profiles', '0010_auto_20171221_0112'), ] operations = [ migrations.RunPython(forwards, backwards) ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models def forwards(apps, schema_editor): """Create Onboarding group.""" Group = apps.get_model('auth', 'Group') Group.objects.create(name='Onboarding') def backwards(apps, schema_editor): """Delete Onboarding group.""" Group = apps.get_model('auth', 'Group') Group.objects.filter(name='Onboarding').delete() class Migration(migrations.Migration): dependencies = [ ('profiles', '0010_auto_20171221_0112'), ] operations = [ migrations.RunPython(forwards, backwards) ] Check if Onboarding exists before creating.# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models def forwards(apps, schema_editor): """Create Onboarding group.""" Group = apps.get_model('auth', 'Group') if not Group.objects.filter(name='Onboarding').exists(): Group.objects.create(name='Onboarding') def backwards(apps, schema_editor): """Delete Onboarding group.""" Group = apps.get_model('auth', 'Group') if Group.objects.filter(name='Onboarding').exists(): Group.objects.filter(name='Onboarding').delete() class Migration(migrations.Migration): dependencies = [ ('profiles', '0010_auto_20171221_0112'), ] operations = [ migrations.RunPython(forwards, backwards) ]
<commit_before># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models def forwards(apps, schema_editor): """Create Onboarding group.""" Group = apps.get_model('auth', 'Group') Group.objects.create(name='Onboarding') def backwards(apps, schema_editor): """Delete Onboarding group.""" Group = apps.get_model('auth', 'Group') Group.objects.filter(name='Onboarding').delete() class Migration(migrations.Migration): dependencies = [ ('profiles', '0010_auto_20171221_0112'), ] operations = [ migrations.RunPython(forwards, backwards) ] <commit_msg>Check if Onboarding exists before creating.<commit_after># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models def forwards(apps, schema_editor): """Create Onboarding group.""" Group = apps.get_model('auth', 'Group') if not Group.objects.filter(name='Onboarding').exists(): Group.objects.create(name='Onboarding') def backwards(apps, schema_editor): """Delete Onboarding group.""" Group = apps.get_model('auth', 'Group') if Group.objects.filter(name='Onboarding').exists(): Group.objects.filter(name='Onboarding').delete() class Migration(migrations.Migration): dependencies = [ ('profiles', '0010_auto_20171221_0112'), ] operations = [ migrations.RunPython(forwards, backwards) ]
5e3be1d123063495f21d0c0068c7132d43fd9724
account/models.py
account/models.py
from django.db import models from django.db.models import signals from django.contrib.auth.models import User from course.models import Term class Profile(models.Model): user = models.OneToOneField(User) student_id = models.CharField(max_length=10, null=True) default_term = models.ForeignKey(Term, null=True) facebook_id = models.CharField(max_length=50, null=True) def create_profile(sender, instance, created, **kwargs): profile = Profile.objects.get(user=instance) if not profile: Profile(user=instance).save() signals.post_save.connect(create_profile, sender=User)
from django.db import models from django.db.models import signals from django.contrib.auth.models import User from course.models import Term class Profile(models.Model): user = models.OneToOneField(User) student_id = models.CharField(max_length=10, null=True) default_term = models.ForeignKey(Term, null=True) facebook_id = models.CharField(max_length=50, null=True) def create_profile(sender, instance, created, **kwargs): try: profile = Profile.objects.get(user=instance) except Profile.DoesNotExist: Profile(user=instance).save() signals.post_save.connect(create_profile, sender=User)
Fix login error for new accounts where a profile doesn't exist
Fix login error for new accounts where a profile doesn't exist
Python
apache-2.0
OpenCourseProject/OpenCourse,gravitylow/OpenCourse,gravitylow/OpenCourse,gravitylow/OpenCourse,OpenCourseProject/OpenCourse,OpenCourseProject/OpenCourse
from django.db import models from django.db.models import signals from django.contrib.auth.models import User from course.models import Term class Profile(models.Model): user = models.OneToOneField(User) student_id = models.CharField(max_length=10, null=True) default_term = models.ForeignKey(Term, null=True) facebook_id = models.CharField(max_length=50, null=True) def create_profile(sender, instance, created, **kwargs): profile = Profile.objects.get(user=instance) if not profile: Profile(user=instance).save() signals.post_save.connect(create_profile, sender=User) Fix login error for new accounts where a profile doesn't exist
from django.db import models from django.db.models import signals from django.contrib.auth.models import User from course.models import Term class Profile(models.Model): user = models.OneToOneField(User) student_id = models.CharField(max_length=10, null=True) default_term = models.ForeignKey(Term, null=True) facebook_id = models.CharField(max_length=50, null=True) def create_profile(sender, instance, created, **kwargs): try: profile = Profile.objects.get(user=instance) except Profile.DoesNotExist: Profile(user=instance).save() signals.post_save.connect(create_profile, sender=User)
<commit_before>from django.db import models from django.db.models import signals from django.contrib.auth.models import User from course.models import Term class Profile(models.Model): user = models.OneToOneField(User) student_id = models.CharField(max_length=10, null=True) default_term = models.ForeignKey(Term, null=True) facebook_id = models.CharField(max_length=50, null=True) def create_profile(sender, instance, created, **kwargs): profile = Profile.objects.get(user=instance) if not profile: Profile(user=instance).save() signals.post_save.connect(create_profile, sender=User) <commit_msg>Fix login error for new accounts where a profile doesn't exist<commit_after>
from django.db import models from django.db.models import signals from django.contrib.auth.models import User from course.models import Term class Profile(models.Model): user = models.OneToOneField(User) student_id = models.CharField(max_length=10, null=True) default_term = models.ForeignKey(Term, null=True) facebook_id = models.CharField(max_length=50, null=True) def create_profile(sender, instance, created, **kwargs): try: profile = Profile.objects.get(user=instance) except Profile.DoesNotExist: Profile(user=instance).save() signals.post_save.connect(create_profile, sender=User)
from django.db import models from django.db.models import signals from django.contrib.auth.models import User from course.models import Term class Profile(models.Model): user = models.OneToOneField(User) student_id = models.CharField(max_length=10, null=True) default_term = models.ForeignKey(Term, null=True) facebook_id = models.CharField(max_length=50, null=True) def create_profile(sender, instance, created, **kwargs): profile = Profile.objects.get(user=instance) if not profile: Profile(user=instance).save() signals.post_save.connect(create_profile, sender=User) Fix login error for new accounts where a profile doesn't existfrom django.db import models from django.db.models import signals from django.contrib.auth.models import User from course.models import Term class Profile(models.Model): user = models.OneToOneField(User) student_id = models.CharField(max_length=10, null=True) default_term = models.ForeignKey(Term, null=True) facebook_id = models.CharField(max_length=50, null=True) def create_profile(sender, instance, created, **kwargs): try: profile = Profile.objects.get(user=instance) except Profile.DoesNotExist: Profile(user=instance).save() signals.post_save.connect(create_profile, sender=User)
<commit_before>from django.db import models from django.db.models import signals from django.contrib.auth.models import User from course.models import Term class Profile(models.Model): user = models.OneToOneField(User) student_id = models.CharField(max_length=10, null=True) default_term = models.ForeignKey(Term, null=True) facebook_id = models.CharField(max_length=50, null=True) def create_profile(sender, instance, created, **kwargs): profile = Profile.objects.get(user=instance) if not profile: Profile(user=instance).save() signals.post_save.connect(create_profile, sender=User) <commit_msg>Fix login error for new accounts where a profile doesn't exist<commit_after>from django.db import models from django.db.models import signals from django.contrib.auth.models import User from course.models import Term class Profile(models.Model): user = models.OneToOneField(User) student_id = models.CharField(max_length=10, null=True) default_term = models.ForeignKey(Term, null=True) facebook_id = models.CharField(max_length=50, null=True) def create_profile(sender, instance, created, **kwargs): try: profile = Profile.objects.get(user=instance) except Profile.DoesNotExist: Profile(user=instance).save() signals.post_save.connect(create_profile, sender=User)
7a07a89d59250127fce21b5f1b68492046b3eb60
pyshelf/search/metadata.py
pyshelf/search/metadata.py
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer # Required for case sensitivity metadata_analyzer = analyzer("metadata_analyzer", tokenizer=tokenizer("keyword")) class Metadata(DocType): property_list = Nested( properties={ "name": String(), "value": String(analyzer=metadata_analyzer), "immutable": Boolean() } ) def update_all(self, metadata): """ Updates all metadata related to an artifact. Args metadata(dict): collection of metadata for document. """ self.property_list = metadata.values()
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer # Required for case sensitivity # To add an analyzer to an existing mapping requires mapping to be "closed" case_sensitive_analyzer = analyzer("case_sensitive_analyzer", tokenizer=tokenizer("keyword")) class Metadata(DocType): property_list = Nested( properties={ "name": String(analyzer=case_sensitive_analyzer), "value": String(analyzer=case_sensitive_analyzer), "immutable": Boolean() } ) def update_all(self, metadata): """ Updates all metadata related to an artifact. Args metadata(dict): collection of metadata for document. """ self.property_list = metadata.values()
Add case sensitivity to field and clarify analyzer.
Add case sensitivity to field and clarify analyzer.
Python
mit
not-nexus/shelf,kyle-long/pyshelf,kyle-long/pyshelf,not-nexus/shelf
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer # Required for case sensitivity metadata_analyzer = analyzer("metadata_analyzer", tokenizer=tokenizer("keyword")) class Metadata(DocType): property_list = Nested( properties={ "name": String(), "value": String(analyzer=metadata_analyzer), "immutable": Boolean() } ) def update_all(self, metadata): """ Updates all metadata related to an artifact. Args metadata(dict): collection of metadata for document. """ self.property_list = metadata.values() Add case sensitivity to field and clarify analyzer.
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer # Required for case sensitivity # To add an analyzer to an existing mapping requires mapping to be "closed" case_sensitive_analyzer = analyzer("case_sensitive_analyzer", tokenizer=tokenizer("keyword")) class Metadata(DocType): property_list = Nested( properties={ "name": String(analyzer=case_sensitive_analyzer), "value": String(analyzer=case_sensitive_analyzer), "immutable": Boolean() } ) def update_all(self, metadata): """ Updates all metadata related to an artifact. Args metadata(dict): collection of metadata for document. """ self.property_list = metadata.values()
<commit_before>from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer # Required for case sensitivity metadata_analyzer = analyzer("metadata_analyzer", tokenizer=tokenizer("keyword")) class Metadata(DocType): property_list = Nested( properties={ "name": String(), "value": String(analyzer=metadata_analyzer), "immutable": Boolean() } ) def update_all(self, metadata): """ Updates all metadata related to an artifact. Args metadata(dict): collection of metadata for document. """ self.property_list = metadata.values() <commit_msg>Add case sensitivity to field and clarify analyzer.<commit_after>
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer # Required for case sensitivity # To add an analyzer to an existing mapping requires mapping to be "closed" case_sensitive_analyzer = analyzer("case_sensitive_analyzer", tokenizer=tokenizer("keyword")) class Metadata(DocType): property_list = Nested( properties={ "name": String(analyzer=case_sensitive_analyzer), "value": String(analyzer=case_sensitive_analyzer), "immutable": Boolean() } ) def update_all(self, metadata): """ Updates all metadata related to an artifact. Args metadata(dict): collection of metadata for document. """ self.property_list = metadata.values()
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer # Required for case sensitivity metadata_analyzer = analyzer("metadata_analyzer", tokenizer=tokenizer("keyword")) class Metadata(DocType): property_list = Nested( properties={ "name": String(), "value": String(analyzer=metadata_analyzer), "immutable": Boolean() } ) def update_all(self, metadata): """ Updates all metadata related to an artifact. Args metadata(dict): collection of metadata for document. """ self.property_list = metadata.values() Add case sensitivity to field and clarify analyzer.from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer # Required for case sensitivity # To add an analyzer to an existing mapping requires mapping to be "closed" case_sensitive_analyzer = analyzer("case_sensitive_analyzer", tokenizer=tokenizer("keyword")) class Metadata(DocType): property_list = Nested( properties={ "name": String(analyzer=case_sensitive_analyzer), "value": String(analyzer=case_sensitive_analyzer), "immutable": Boolean() } ) def update_all(self, metadata): """ Updates all metadata related to an artifact. Args metadata(dict): collection of metadata for document. """ self.property_list = metadata.values()
<commit_before>from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer # Required for case sensitivity metadata_analyzer = analyzer("metadata_analyzer", tokenizer=tokenizer("keyword")) class Metadata(DocType): property_list = Nested( properties={ "name": String(), "value": String(analyzer=metadata_analyzer), "immutable": Boolean() } ) def update_all(self, metadata): """ Updates all metadata related to an artifact. Args metadata(dict): collection of metadata for document. """ self.property_list = metadata.values() <commit_msg>Add case sensitivity to field and clarify analyzer.<commit_after>from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer # Required for case sensitivity # To add an analyzer to an existing mapping requires mapping to be "closed" case_sensitive_analyzer = analyzer("case_sensitive_analyzer", tokenizer=tokenizer("keyword")) class Metadata(DocType): property_list = Nested( properties={ "name": String(analyzer=case_sensitive_analyzer), "value": String(analyzer=case_sensitive_analyzer), "immutable": Boolean() } ) def update_all(self, metadata): """ Updates all metadata related to an artifact. Args metadata(dict): collection of metadata for document. """ self.property_list = metadata.values()
7bc777a5e9fb15720dd6b41aa5e1fbcfd7d3141b
tests/test_postgres_processor.py
tests/test_postgres_processor.py
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres @pytest.mark.django_db def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
Add django pytest mark to process raw test
Add django pytest mark to process raw test
Python
apache-2.0
CenterForOpenScience/scrapi,felliott/scrapi,mehanig/scrapi,CenterForOpenScience/scrapi,fabianvf/scrapi,erinspace/scrapi,erinspace/scrapi,felliott/scrapi,mehanig/scrapi,fabianvf/scrapi
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] Add django pytest mark to process raw test
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres @pytest.mark.django_db def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
<commit_before>import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] <commit_msg>Add django pytest mark to process raw test<commit_after>
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres @pytest.mark.django_db def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] Add django pytest mark to process raw testimport pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres @pytest.mark.django_db def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
<commit_before>import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] <commit_msg>Add django pytest mark to process raw test<commit_after>import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres @pytest.mark.django_db def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
53646da453a4aa6d0e559ee3069626458f2fef78
common/urls.py
common/urls.py
import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): return re.sub("\\(.+\\)", "{id}", pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): _pattern = pattern.replace('^', '').replace('$', '') return re.sub("\\(.+\\)", "{id}", _pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
Fix backend home page url generator
Fix backend home page url generator
Python
mit
DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange
import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): return re.sub("\\(.+\\)", "{id}", pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)Fix backend home page url generator
import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): _pattern = pattern.replace('^', '').replace('$', '') return re.sub("\\(.+\\)", "{id}", _pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
<commit_before>import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): return re.sub("\\(.+\\)", "{id}", pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)<commit_msg>Fix backend home page url generator<commit_after>
import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): _pattern = pattern.replace('^', '').replace('$', '') return re.sub("\\(.+\\)", "{id}", _pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): return re.sub("\\(.+\\)", "{id}", pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)Fix backend home page url generatorimport json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): _pattern = pattern.replace('^', '').replace('$', '') return re.sub("\\(.+\\)", "{id}", _pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
<commit_before>import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): return re.sub("\\(.+\\)", "{id}", pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)<commit_msg>Fix backend home page url generator<commit_after>import json import os import re from django.urls import re_path from civictechprojects import views def url_generator_from_pattern(pattern): _pattern = pattern.replace('^', '').replace('$', '') return re.sub("\\(.+\\)", "{id}", _pattern) def generate_url_patterns(spec_path, set_url_generators=False): # Read json file base_dir = os.path.dirname(__file__) filename = os.path.join(base_dir, spec_path) url_patterns = [] with open(filename, 'r', encoding='utf-8') as f: urls_json = json.load(f) for url_spec_json in urls_json: url_patterns.append(re_path(url_spec_json['pattern'], views.index)) if set_url_generators: url_generators[url_spec_json['name']] = { 'section': url_spec_json['name'], 'regex': re.compile(url_spec_json['pattern']), 'generator': url_generator_from_pattern(url_spec_json['pattern']) } return url_patterns url_generators = {} v1_urls = generate_url_patterns('./components/urls/urls_v1.json') v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
0f8e2313d6f0ec06806ea05e861d1fc47d3c3016
utils/internal/zz_parse.py
utils/internal/zz_parse.py
import sys sys.path.insert(0, '../..') from pycparser import c_parser, c_ast, parse_file if __name__ == "__main__": #ast = parse_file('zc_pp.c', use_cpp=True, cpp_path="../cpp.exe") parser = c_parser.CParser() #code = r'''int ar[30];''' code = r''' char ***arr3d[40]; ''' #code = r''' #int foo(int a, int arr[*]); #''' print(code) ast = parser.parse(code) ast.show(attrnames=True, nodenames=True) print(ast.ext[0].__slots__) print(dir(ast.ext[0]))
from __future__ import print_function import sys from pycparser import c_parser, c_generator, c_ast, parse_file if __name__ == "__main__": parser = c_parser.CParser() code = r''' void* ptr = (int[ ]){0}; ''' print(code) ast = parser.parse(code) ast.show(attrnames=True, nodenames=True) print(ast.ext[0].__slots__) print(dir(ast.ext[0])) print("==== From C generator:") generator = c_generator.CGenerator() print(generator.visit(ast))
Clean up internal hacking util
Clean up internal hacking util
Python
bsd-3-clause
CtheSky/pycparser,CtheSky/pycparser,CtheSky/pycparser
import sys sys.path.insert(0, '../..') from pycparser import c_parser, c_ast, parse_file if __name__ == "__main__": #ast = parse_file('zc_pp.c', use_cpp=True, cpp_path="../cpp.exe") parser = c_parser.CParser() #code = r'''int ar[30];''' code = r''' char ***arr3d[40]; ''' #code = r''' #int foo(int a, int arr[*]); #''' print(code) ast = parser.parse(code) ast.show(attrnames=True, nodenames=True) print(ast.ext[0].__slots__) print(dir(ast.ext[0])) Clean up internal hacking util
from __future__ import print_function import sys from pycparser import c_parser, c_generator, c_ast, parse_file if __name__ == "__main__": parser = c_parser.CParser() code = r''' void* ptr = (int[ ]){0}; ''' print(code) ast = parser.parse(code) ast.show(attrnames=True, nodenames=True) print(ast.ext[0].__slots__) print(dir(ast.ext[0])) print("==== From C generator:") generator = c_generator.CGenerator() print(generator.visit(ast))
<commit_before>import sys sys.path.insert(0, '../..') from pycparser import c_parser, c_ast, parse_file if __name__ == "__main__": #ast = parse_file('zc_pp.c', use_cpp=True, cpp_path="../cpp.exe") parser = c_parser.CParser() #code = r'''int ar[30];''' code = r''' char ***arr3d[40]; ''' #code = r''' #int foo(int a, int arr[*]); #''' print(code) ast = parser.parse(code) ast.show(attrnames=True, nodenames=True) print(ast.ext[0].__slots__) print(dir(ast.ext[0])) <commit_msg>Clean up internal hacking util<commit_after>
from __future__ import print_function import sys from pycparser import c_parser, c_generator, c_ast, parse_file if __name__ == "__main__": parser = c_parser.CParser() code = r''' void* ptr = (int[ ]){0}; ''' print(code) ast = parser.parse(code) ast.show(attrnames=True, nodenames=True) print(ast.ext[0].__slots__) print(dir(ast.ext[0])) print("==== From C generator:") generator = c_generator.CGenerator() print(generator.visit(ast))
import sys sys.path.insert(0, '../..') from pycparser import c_parser, c_ast, parse_file if __name__ == "__main__": #ast = parse_file('zc_pp.c', use_cpp=True, cpp_path="../cpp.exe") parser = c_parser.CParser() #code = r'''int ar[30];''' code = r''' char ***arr3d[40]; ''' #code = r''' #int foo(int a, int arr[*]); #''' print(code) ast = parser.parse(code) ast.show(attrnames=True, nodenames=True) print(ast.ext[0].__slots__) print(dir(ast.ext[0])) Clean up internal hacking utilfrom __future__ import print_function import sys from pycparser import c_parser, c_generator, c_ast, parse_file if __name__ == "__main__": parser = c_parser.CParser() code = r''' void* ptr = (int[ ]){0}; ''' print(code) ast = parser.parse(code) ast.show(attrnames=True, nodenames=True) print(ast.ext[0].__slots__) print(dir(ast.ext[0])) print("==== From C generator:") generator = c_generator.CGenerator() print(generator.visit(ast))
<commit_before>import sys sys.path.insert(0, '../..') from pycparser import c_parser, c_ast, parse_file if __name__ == "__main__": #ast = parse_file('zc_pp.c', use_cpp=True, cpp_path="../cpp.exe") parser = c_parser.CParser() #code = r'''int ar[30];''' code = r''' char ***arr3d[40]; ''' #code = r''' #int foo(int a, int arr[*]); #''' print(code) ast = parser.parse(code) ast.show(attrnames=True, nodenames=True) print(ast.ext[0].__slots__) print(dir(ast.ext[0])) <commit_msg>Clean up internal hacking util<commit_after>from __future__ import print_function import sys from pycparser import c_parser, c_generator, c_ast, parse_file if __name__ == "__main__": parser = c_parser.CParser() code = r''' void* ptr = (int[ ]){0}; ''' print(code) ast = parser.parse(code) ast.show(attrnames=True, nodenames=True) print(ast.ext[0].__slots__) print(dir(ast.ext[0])) print("==== From C generator:") generator = c_generator.CGenerator() print(generator.visit(ast))
044a051c637f256613ff307caf3ae0126d09b049
backend/unichat/views.py
backend/unichat/views.py
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest from django.views.decorators.csrf import csrf_exempt import json from helpers import get_school_list, check_signup_email def get_schools(request): resp = JsonResponse({'schools': get_school_list()}) resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp @csrf_exempt def signup(request): if request.method == "POST": signup_parameters = json.loads(request.body.decode('utf-8')) if check_signup_email(signup_parameters['email']): resp = HttpResponse('Signup OK') else: resp = HttpResponseBadRequest('Invalid univesity email') elif request.method == "OPTIONS": resp = HttpResponse('') resp['Access-Control-Allow-Headers'] = 'Content-Type' else: resp = HttpResponseBadRequest('') resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest from django.views.decorators.csrf import csrf_exempt import json from helpers import get_school_list, check_signup_email def get_schools(request): resp = JsonResponse({'schools': get_school_list()}) resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp @csrf_exempt def signup(request): if request.method == "POST": signup_parameters = json.loads(request.body.decode('utf-8')) if check_signup_email(signup_parameters['email']): resp = HttpResponse('Signup OK') else: resp = HttpResponseBadRequest('Invalid univesity email') elif request.method == "OPTIONS": resp = HttpResponse('') resp['Access-Control-Allow-Headers'] = 'Content-Type' else: resp = HttpResponseBadRequest('Invalid request method') resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp
Add error message to BadRequest signup response for invalid method
Add error message to BadRequest signup response for invalid method
Python
mit
dimkarakostas/unimeet,dimkarakostas/unimeet,dimkarakostas/unimeet,dimkarakostas/unimeet
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest from django.views.decorators.csrf import csrf_exempt import json from helpers import get_school_list, check_signup_email def get_schools(request): resp = JsonResponse({'schools': get_school_list()}) resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp @csrf_exempt def signup(request): if request.method == "POST": signup_parameters = json.loads(request.body.decode('utf-8')) if check_signup_email(signup_parameters['email']): resp = HttpResponse('Signup OK') else: resp = HttpResponseBadRequest('Invalid univesity email') elif request.method == "OPTIONS": resp = HttpResponse('') resp['Access-Control-Allow-Headers'] = 'Content-Type' else: resp = HttpResponseBadRequest('') resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp Add error message to BadRequest signup response for invalid method
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest from django.views.decorators.csrf import csrf_exempt import json from helpers import get_school_list, check_signup_email def get_schools(request): resp = JsonResponse({'schools': get_school_list()}) resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp @csrf_exempt def signup(request): if request.method == "POST": signup_parameters = json.loads(request.body.decode('utf-8')) if check_signup_email(signup_parameters['email']): resp = HttpResponse('Signup OK') else: resp = HttpResponseBadRequest('Invalid univesity email') elif request.method == "OPTIONS": resp = HttpResponse('') resp['Access-Control-Allow-Headers'] = 'Content-Type' else: resp = HttpResponseBadRequest('Invalid request method') resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp
<commit_before>from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest from django.views.decorators.csrf import csrf_exempt import json from helpers import get_school_list, check_signup_email def get_schools(request): resp = JsonResponse({'schools': get_school_list()}) resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp @csrf_exempt def signup(request): if request.method == "POST": signup_parameters = json.loads(request.body.decode('utf-8')) if check_signup_email(signup_parameters['email']): resp = HttpResponse('Signup OK') else: resp = HttpResponseBadRequest('Invalid univesity email') elif request.method == "OPTIONS": resp = HttpResponse('') resp['Access-Control-Allow-Headers'] = 'Content-Type' else: resp = HttpResponseBadRequest('') resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp <commit_msg>Add error message to BadRequest signup response for invalid method<commit_after>
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest from django.views.decorators.csrf import csrf_exempt import json from helpers import get_school_list, check_signup_email def get_schools(request): resp = JsonResponse({'schools': get_school_list()}) resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp @csrf_exempt def signup(request): if request.method == "POST": signup_parameters = json.loads(request.body.decode('utf-8')) if check_signup_email(signup_parameters['email']): resp = HttpResponse('Signup OK') else: resp = HttpResponseBadRequest('Invalid univesity email') elif request.method == "OPTIONS": resp = HttpResponse('') resp['Access-Control-Allow-Headers'] = 'Content-Type' else: resp = HttpResponseBadRequest('Invalid request method') resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest from django.views.decorators.csrf import csrf_exempt import json from helpers import get_school_list, check_signup_email def get_schools(request): resp = JsonResponse({'schools': get_school_list()}) resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp @csrf_exempt def signup(request): if request.method == "POST": signup_parameters = json.loads(request.body.decode('utf-8')) if check_signup_email(signup_parameters['email']): resp = HttpResponse('Signup OK') else: resp = HttpResponseBadRequest('Invalid univesity email') elif request.method == "OPTIONS": resp = HttpResponse('') resp['Access-Control-Allow-Headers'] = 'Content-Type' else: resp = HttpResponseBadRequest('') resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp Add error message to BadRequest signup response for invalid methodfrom django.http import JsonResponse, HttpResponse, HttpResponseBadRequest from django.views.decorators.csrf import csrf_exempt import json from helpers import get_school_list, check_signup_email def get_schools(request): resp = JsonResponse({'schools': get_school_list()}) resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp @csrf_exempt def signup(request): if request.method == "POST": signup_parameters = json.loads(request.body.decode('utf-8')) if check_signup_email(signup_parameters['email']): resp = HttpResponse('Signup OK') else: resp = HttpResponseBadRequest('Invalid univesity email') elif request.method == "OPTIONS": resp = HttpResponse('') resp['Access-Control-Allow-Headers'] = 'Content-Type' else: resp = HttpResponseBadRequest('Invalid request method') resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp
<commit_before>from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest from django.views.decorators.csrf import csrf_exempt import json from helpers import get_school_list, check_signup_email def get_schools(request): resp = JsonResponse({'schools': get_school_list()}) resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp @csrf_exempt def signup(request): if request.method == "POST": signup_parameters = json.loads(request.body.decode('utf-8')) if check_signup_email(signup_parameters['email']): resp = HttpResponse('Signup OK') else: resp = HttpResponseBadRequest('Invalid univesity email') elif request.method == "OPTIONS": resp = HttpResponse('') resp['Access-Control-Allow-Headers'] = 'Content-Type' else: resp = HttpResponseBadRequest('') resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp <commit_msg>Add error message to BadRequest signup response for invalid method<commit_after>from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest from django.views.decorators.csrf import csrf_exempt import json from helpers import get_school_list, check_signup_email def get_schools(request): resp = JsonResponse({'schools': get_school_list()}) resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp @csrf_exempt def signup(request): if request.method == "POST": signup_parameters = json.loads(request.body.decode('utf-8')) if check_signup_email(signup_parameters['email']): resp = HttpResponse('Signup OK') else: resp = HttpResponseBadRequest('Invalid univesity email') elif request.method == "OPTIONS": resp = HttpResponse('') resp['Access-Control-Allow-Headers'] = 'Content-Type' else: resp = HttpResponseBadRequest('Invalid request method') resp['Access-Control-Allow-Origin'] = 'http://localhost:3000' return resp
4c1c902010096d6d87d93b865d9c68794da51414
trex/parsers.py
trex/parsers.py
# -*- coding: utf-8 -*- # # (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com> # # See LICENSE comming with the source of 'trex' for details. # from io import TextIOWrapper from rest_framework.parsers import BaseParser class PlainTextParser(BaseParser): media_type = "text/plain" def parse(self, stream, media_type=None, parser_context=None): print "Running PlainTextParser" charset = self.get_charset(media_type) if charset: stream = TextIOWrapper(stream, encoding=charset) return stream def get_charset(self, media_type): if not media_type: return None charset = None msplit = media_type.split(" "); for m in msplit: m = m.strip() if "charset" in m: csplit = m.split("=") if len(csplit) > 1: charset = csplit[1] return charset.strip().lower() return None
# -*- coding: utf-8 -*- # # (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com> # # See LICENSE comming with the source of 'trex' for details. # from io import TextIOWrapper, BytesIO from django.core.handlers.wsgi import WSGIRequest from rest_framework.parsers import BaseParser class PlainTextParser(BaseParser): media_type = "text/plain" def parse(self, stream, media_type=None, parser_context=None): print "Running PlainTextParser" charset = self.get_charset(media_type) if charset: if isinstance(stream, WSGIRequest): stream = BytesIO(stream.read()) stream = TextIOWrapper(stream, encoding=charset) return stream def get_charset(self, media_type): if not media_type: return None charset = None msplit = media_type.split(" "); for m in msplit: m = m.strip() if "charset" in m: csplit = m.split("=") if len(csplit) > 1: charset = csplit[1] return charset.strip().lower() return None
Fix parsing data from request
Fix parsing data from request The object passed to the parser method is not a real IOBase stream. It may only be a Request object which has read, etc. methods. Therefore the real data must be encapsulated in a BytesIO stream before changing the content type.
Python
mit
bjoernricks/trex,bjoernricks/trex
# -*- coding: utf-8 -*- # # (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com> # # See LICENSE comming with the source of 'trex' for details. # from io import TextIOWrapper from rest_framework.parsers import BaseParser class PlainTextParser(BaseParser): media_type = "text/plain" def parse(self, stream, media_type=None, parser_context=None): print "Running PlainTextParser" charset = self.get_charset(media_type) if charset: stream = TextIOWrapper(stream, encoding=charset) return stream def get_charset(self, media_type): if not media_type: return None charset = None msplit = media_type.split(" "); for m in msplit: m = m.strip() if "charset" in m: csplit = m.split("=") if len(csplit) > 1: charset = csplit[1] return charset.strip().lower() return None Fix parsing data from request The object passed to the parser method is not a real IOBase stream. It may only be a Request object which has read, etc. methods. Therefore the real data must be encapsulated in a BytesIO stream before changing the content type.
# -*- coding: utf-8 -*- # # (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com> # # See LICENSE comming with the source of 'trex' for details. # from io import TextIOWrapper, BytesIO from django.core.handlers.wsgi import WSGIRequest from rest_framework.parsers import BaseParser class PlainTextParser(BaseParser): media_type = "text/plain" def parse(self, stream, media_type=None, parser_context=None): print "Running PlainTextParser" charset = self.get_charset(media_type) if charset: if isinstance(stream, WSGIRequest): stream = BytesIO(stream.read()) stream = TextIOWrapper(stream, encoding=charset) return stream def get_charset(self, media_type): if not media_type: return None charset = None msplit = media_type.split(" "); for m in msplit: m = m.strip() if "charset" in m: csplit = m.split("=") if len(csplit) > 1: charset = csplit[1] return charset.strip().lower() return None
<commit_before># -*- coding: utf-8 -*- # # (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com> # # See LICENSE comming with the source of 'trex' for details. # from io import TextIOWrapper from rest_framework.parsers import BaseParser class PlainTextParser(BaseParser): media_type = "text/plain" def parse(self, stream, media_type=None, parser_context=None): print "Running PlainTextParser" charset = self.get_charset(media_type) if charset: stream = TextIOWrapper(stream, encoding=charset) return stream def get_charset(self, media_type): if not media_type: return None charset = None msplit = media_type.split(" "); for m in msplit: m = m.strip() if "charset" in m: csplit = m.split("=") if len(csplit) > 1: charset = csplit[1] return charset.strip().lower() return None <commit_msg>Fix parsing data from request The object passed to the parser method is not a real IOBase stream. It may only be a Request object which has read, etc. methods. Therefore the real data must be encapsulated in a BytesIO stream before changing the content type.<commit_after>
# -*- coding: utf-8 -*- # # (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com> # # See LICENSE comming with the source of 'trex' for details. # from io import TextIOWrapper, BytesIO from django.core.handlers.wsgi import WSGIRequest from rest_framework.parsers import BaseParser class PlainTextParser(BaseParser): media_type = "text/plain" def parse(self, stream, media_type=None, parser_context=None): print "Running PlainTextParser" charset = self.get_charset(media_type) if charset: if isinstance(stream, WSGIRequest): stream = BytesIO(stream.read()) stream = TextIOWrapper(stream, encoding=charset) return stream def get_charset(self, media_type): if not media_type: return None charset = None msplit = media_type.split(" "); for m in msplit: m = m.strip() if "charset" in m: csplit = m.split("=") if len(csplit) > 1: charset = csplit[1] return charset.strip().lower() return None
# -*- coding: utf-8 -*- # # (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com> # # See LICENSE comming with the source of 'trex' for details. # from io import TextIOWrapper from rest_framework.parsers import BaseParser class PlainTextParser(BaseParser): media_type = "text/plain" def parse(self, stream, media_type=None, parser_context=None): print "Running PlainTextParser" charset = self.get_charset(media_type) if charset: stream = TextIOWrapper(stream, encoding=charset) return stream def get_charset(self, media_type): if not media_type: return None charset = None msplit = media_type.split(" "); for m in msplit: m = m.strip() if "charset" in m: csplit = m.split("=") if len(csplit) > 1: charset = csplit[1] return charset.strip().lower() return None Fix parsing data from request The object passed to the parser method is not a real IOBase stream. It may only be a Request object which has read, etc. methods. Therefore the real data must be encapsulated in a BytesIO stream before changing the content type.# -*- coding: utf-8 -*- # # (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com> # # See LICENSE comming with the source of 'trex' for details. # from io import TextIOWrapper, BytesIO from django.core.handlers.wsgi import WSGIRequest from rest_framework.parsers import BaseParser class PlainTextParser(BaseParser): media_type = "text/plain" def parse(self, stream, media_type=None, parser_context=None): print "Running PlainTextParser" charset = self.get_charset(media_type) if charset: if isinstance(stream, WSGIRequest): stream = BytesIO(stream.read()) stream = TextIOWrapper(stream, encoding=charset) return stream def get_charset(self, media_type): if not media_type: return None charset = None msplit = media_type.split(" "); for m in msplit: m = m.strip() if "charset" in m: csplit = m.split("=") if len(csplit) > 1: charset = csplit[1] return charset.strip().lower() return None
<commit_before># -*- coding: utf-8 -*- # # (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com> # # See LICENSE comming with the source of 'trex' for details. # from io import TextIOWrapper from rest_framework.parsers import BaseParser class PlainTextParser(BaseParser): media_type = "text/plain" def parse(self, stream, media_type=None, parser_context=None): print "Running PlainTextParser" charset = self.get_charset(media_type) if charset: stream = TextIOWrapper(stream, encoding=charset) return stream def get_charset(self, media_type): if not media_type: return None charset = None msplit = media_type.split(" "); for m in msplit: m = m.strip() if "charset" in m: csplit = m.split("=") if len(csplit) > 1: charset = csplit[1] return charset.strip().lower() return None <commit_msg>Fix parsing data from request The object passed to the parser method is not a real IOBase stream. It may only be a Request object which has read, etc. methods. Therefore the real data must be encapsulated in a BytesIO stream before changing the content type.<commit_after># -*- coding: utf-8 -*- # # (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com> # # See LICENSE comming with the source of 'trex' for details. # from io import TextIOWrapper, BytesIO from django.core.handlers.wsgi import WSGIRequest from rest_framework.parsers import BaseParser class PlainTextParser(BaseParser): media_type = "text/plain" def parse(self, stream, media_type=None, parser_context=None): print "Running PlainTextParser" charset = self.get_charset(media_type) if charset: if isinstance(stream, WSGIRequest): stream = BytesIO(stream.read()) stream = TextIOWrapper(stream, encoding=charset) return stream def get_charset(self, media_type): if not media_type: return None charset = None msplit = media_type.split(" "); for m in msplit: m = m.strip() if "charset" in m: csplit = m.split("=") if len(csplit) > 1: charset = csplit[1] return charset.strip().lower() return None
b86c53c388c39baee1ddfe3a615cdad20d272055
antcolony/util.py
antcolony/util.py
import json def avg(iterable): return sum(iterable) / len(iterable) def nice_json_dump(data, filepath): with open(filepath, 'w') as f: json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
import json def avg(iterable): sum_ = 0 element_count = 0 for element in iterable: sum_ += element element_count += 1 return sum_ / element_count def nice_json_dump(data, filepath): with open(filepath, 'w') as f: json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
Make avg() work with iterators
Make avg() work with iterators
Python
bsd-3-clause
ppolewicz/ant-colony,ppolewicz/ant-colony
import json def avg(iterable): return sum(iterable) / len(iterable) def nice_json_dump(data, filepath): with open(filepath, 'w') as f: json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': ')) Make avg() work with iterators
import json def avg(iterable): sum_ = 0 element_count = 0 for element in iterable: sum_ += element element_count += 1 return sum_ / element_count def nice_json_dump(data, filepath): with open(filepath, 'w') as f: json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
<commit_before>import json def avg(iterable): return sum(iterable) / len(iterable) def nice_json_dump(data, filepath): with open(filepath, 'w') as f: json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': ')) <commit_msg>Make avg() work with iterators<commit_after>
import json def avg(iterable): sum_ = 0 element_count = 0 for element in iterable: sum_ += element element_count += 1 return sum_ / element_count def nice_json_dump(data, filepath): with open(filepath, 'w') as f: json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
import json def avg(iterable): return sum(iterable) / len(iterable) def nice_json_dump(data, filepath): with open(filepath, 'w') as f: json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': ')) Make avg() work with iteratorsimport json def avg(iterable): sum_ = 0 element_count = 0 for element in iterable: sum_ += element element_count += 1 return sum_ / element_count def nice_json_dump(data, filepath): with open(filepath, 'w') as f: json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
<commit_before>import json def avg(iterable): return sum(iterable) / len(iterable) def nice_json_dump(data, filepath): with open(filepath, 'w') as f: json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': ')) <commit_msg>Make avg() work with iterators<commit_after>import json def avg(iterable): sum_ = 0 element_count = 0 for element in iterable: sum_ += element element_count += 1 return sum_ / element_count def nice_json_dump(data, filepath): with open(filepath, 'w') as f: json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
50a30ded8705b343478f85ea1c6c60e827982d37
auwsssp/urls.py
auwsssp/urls.py
from django.conf.urls import patterns, include, url from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: # url(r'^$', 'auwsssp.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), )
from django.conf.urls import patterns, include, url from django.conf import settings from django.conf.urls.static import static from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'signups.views.home', name='home'), # url(r'^blog/', include('blog.urls')), #url(r'^admin/doc/', include('django.contrib.admindocs.urls')), url(r'^admin/', include(admin.site.urls)), ) if settings.DEBUG: urlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_ROOT) urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
Modify index url and static folders
Modify index url and static folders
Python
mit
eyassug/au-water-sanitation-template,eyassug/au-water-sanitation-template,eyassug/au-water-sanitation-template
from django.conf.urls import patterns, include, url from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: # url(r'^$', 'auwsssp.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), ) Modify index url and static folders
from django.conf.urls import patterns, include, url from django.conf import settings from django.conf.urls.static import static from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'signups.views.home', name='home'), # url(r'^blog/', include('blog.urls')), #url(r'^admin/doc/', include('django.contrib.admindocs.urls')), url(r'^admin/', include(admin.site.urls)), ) if settings.DEBUG: urlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_ROOT) urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
<commit_before>from django.conf.urls import patterns, include, url from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: # url(r'^$', 'auwsssp.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), ) <commit_msg>Modify index url and static folders<commit_after>
from django.conf.urls import patterns, include, url from django.conf import settings from django.conf.urls.static import static from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'signups.views.home', name='home'), # url(r'^blog/', include('blog.urls')), #url(r'^admin/doc/', include('django.contrib.admindocs.urls')), url(r'^admin/', include(admin.site.urls)), ) if settings.DEBUG: urlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_ROOT) urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
from django.conf.urls import patterns, include, url from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: # url(r'^$', 'auwsssp.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), ) Modify index url and static foldersfrom django.conf.urls import patterns, include, url from django.conf import settings from django.conf.urls.static import static from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'signups.views.home', name='home'), # url(r'^blog/', include('blog.urls')), #url(r'^admin/doc/', include('django.contrib.admindocs.urls')), url(r'^admin/', include(admin.site.urls)), ) if settings.DEBUG: urlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_ROOT) urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
<commit_before>from django.conf.urls import patterns, include, url from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: # url(r'^$', 'auwsssp.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), ) <commit_msg>Modify index url and static folders<commit_after>from django.conf.urls import patterns, include, url from django.conf import settings from django.conf.urls.static import static from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'signups.views.home', name='home'), # url(r'^blog/', include('blog.urls')), #url(r'^admin/doc/', include('django.contrib.admindocs.urls')), url(r'^admin/', include(admin.site.urls)), ) if settings.DEBUG: urlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_ROOT) urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
b43c6604163e18ae03c6ef206c4892e0beb873f7
django_cradmin/demo/uimock_demo/urls.py
django_cradmin/demo/uimock_demo/urls.py
from django.urls import path from django_cradmin import viewhelpers from .views import overview urlpatterns = [ path('simple/<str:mockname>', viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'), name='cradmin_uimock_demo_simple'), path('', overview.Overview.as_view(), name='cradmin_uimock_demo'), ]
from django.urls import path, re_path from django_cradmin import viewhelpers from .views import overview urlpatterns = [ re_path(r'^simple/(?P<mockname>.+)?$', viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'), name='cradmin_uimock_demo_simple'), path('', overview.Overview.as_view(), name='cradmin_uimock_demo'), ]
Fix url that was wrongly converted to django3.
Fix url that was wrongly converted to django3.
Python
bsd-3-clause
appressoas/django_cradmin,appressoas/django_cradmin,appressoas/django_cradmin
from django.urls import path from django_cradmin import viewhelpers from .views import overview urlpatterns = [ path('simple/<str:mockname>', viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'), name='cradmin_uimock_demo_simple'), path('', overview.Overview.as_view(), name='cradmin_uimock_demo'), ] Fix url that was wrongly converted to django3.
from django.urls import path, re_path from django_cradmin import viewhelpers from .views import overview urlpatterns = [ re_path(r'^simple/(?P<mockname>.+)?$', viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'), name='cradmin_uimock_demo_simple'), path('', overview.Overview.as_view(), name='cradmin_uimock_demo'), ]
<commit_before>from django.urls import path from django_cradmin import viewhelpers from .views import overview urlpatterns = [ path('simple/<str:mockname>', viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'), name='cradmin_uimock_demo_simple'), path('', overview.Overview.as_view(), name='cradmin_uimock_demo'), ] <commit_msg>Fix url that was wrongly converted to django3.<commit_after>
from django.urls import path, re_path from django_cradmin import viewhelpers from .views import overview urlpatterns = [ re_path(r'^simple/(?P<mockname>.+)?$', viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'), name='cradmin_uimock_demo_simple'), path('', overview.Overview.as_view(), name='cradmin_uimock_demo'), ]
from django.urls import path from django_cradmin import viewhelpers from .views import overview urlpatterns = [ path('simple/<str:mockname>', viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'), name='cradmin_uimock_demo_simple'), path('', overview.Overview.as_view(), name='cradmin_uimock_demo'), ] Fix url that was wrongly converted to django3.from django.urls import path, re_path from django_cradmin import viewhelpers from .views import overview urlpatterns = [ re_path(r'^simple/(?P<mockname>.+)?$', viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'), name='cradmin_uimock_demo_simple'), path('', overview.Overview.as_view(), name='cradmin_uimock_demo'), ]
<commit_before>from django.urls import path from django_cradmin import viewhelpers from .views import overview urlpatterns = [ path('simple/<str:mockname>', viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'), name='cradmin_uimock_demo_simple'), path('', overview.Overview.as_view(), name='cradmin_uimock_demo'), ] <commit_msg>Fix url that was wrongly converted to django3.<commit_after>from django.urls import path, re_path from django_cradmin import viewhelpers from .views import overview urlpatterns = [ re_path(r'^simple/(?P<mockname>.+)?$', viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'), name='cradmin_uimock_demo_simple'), path('', overview.Overview.as_view(), name='cradmin_uimock_demo'), ]
ac3c8155abae010fb79866addd1e9cd50f5cae78
tests/test_impersonation.py
tests/test_impersonation.py
from django.core.urlresolvers import reverse import pytest from saleor.userprofile.impersonate import can_impersonate from saleor.userprofile.models import User def test_staff_with_permission_can_impersonate( staff_client, customer_user, staff_user, permission_impersonate_user, staff_group): staff_group.permissions.add(permission_impersonate_user) staff_user.groups.add(staff_group) staff_user = User.objects.get(pk=staff_user.pk) response = staff_client.get('/impersonate/{}'.format(customer_user.pk), follow=True) assert response.context['user'] == customer_user assert response.context['user'].is_impersonate assert response.context['request'].impersonator == staff_user
from django.core.urlresolvers import reverse import pytest from saleor.userprofile.impersonate import can_impersonate from saleor.userprofile.models import User def test_staff_with_permission_can_impersonate( staff_client, customer_user, staff_user, permission_impersonate_user, staff_group): staff_group.permissions.add(permission_impersonate_user) staff_user.groups.add(staff_group) staff_user = User.objects.get(pk=staff_user.pk) response = staff_client.get(reverse('impersonate-start', args=[customer_user.pk]), follow=True) assert response.context['user'] == customer_user assert response.context['user'].is_impersonate assert response.context['request'].impersonator == staff_user
Use reverse function in tests
Use reverse function in tests
Python
bsd-3-clause
UITools/saleor,maferelo/saleor,mociepka/saleor,UITools/saleor,mociepka/saleor,maferelo/saleor,UITools/saleor,UITools/saleor,maferelo/saleor,UITools/saleor,mociepka/saleor
from django.core.urlresolvers import reverse import pytest from saleor.userprofile.impersonate import can_impersonate from saleor.userprofile.models import User def test_staff_with_permission_can_impersonate( staff_client, customer_user, staff_user, permission_impersonate_user, staff_group): staff_group.permissions.add(permission_impersonate_user) staff_user.groups.add(staff_group) staff_user = User.objects.get(pk=staff_user.pk) response = staff_client.get('/impersonate/{}'.format(customer_user.pk), follow=True) assert response.context['user'] == customer_user assert response.context['user'].is_impersonate assert response.context['request'].impersonator == staff_user Use reverse function in tests
from django.core.urlresolvers import reverse import pytest from saleor.userprofile.impersonate import can_impersonate from saleor.userprofile.models import User def test_staff_with_permission_can_impersonate( staff_client, customer_user, staff_user, permission_impersonate_user, staff_group): staff_group.permissions.add(permission_impersonate_user) staff_user.groups.add(staff_group) staff_user = User.objects.get(pk=staff_user.pk) response = staff_client.get(reverse('impersonate-start', args=[customer_user.pk]), follow=True) assert response.context['user'] == customer_user assert response.context['user'].is_impersonate assert response.context['request'].impersonator == staff_user
<commit_before>from django.core.urlresolvers import reverse import pytest from saleor.userprofile.impersonate import can_impersonate from saleor.userprofile.models import User def test_staff_with_permission_can_impersonate( staff_client, customer_user, staff_user, permission_impersonate_user, staff_group): staff_group.permissions.add(permission_impersonate_user) staff_user.groups.add(staff_group) staff_user = User.objects.get(pk=staff_user.pk) response = staff_client.get('/impersonate/{}'.format(customer_user.pk), follow=True) assert response.context['user'] == customer_user assert response.context['user'].is_impersonate assert response.context['request'].impersonator == staff_user <commit_msg>Use reverse function in tests<commit_after>
from django.core.urlresolvers import reverse import pytest from saleor.userprofile.impersonate import can_impersonate from saleor.userprofile.models import User def test_staff_with_permission_can_impersonate( staff_client, customer_user, staff_user, permission_impersonate_user, staff_group): staff_group.permissions.add(permission_impersonate_user) staff_user.groups.add(staff_group) staff_user = User.objects.get(pk=staff_user.pk) response = staff_client.get(reverse('impersonate-start', args=[customer_user.pk]), follow=True) assert response.context['user'] == customer_user assert response.context['user'].is_impersonate assert response.context['request'].impersonator == staff_user
from django.core.urlresolvers import reverse import pytest from saleor.userprofile.impersonate import can_impersonate from saleor.userprofile.models import User def test_staff_with_permission_can_impersonate( staff_client, customer_user, staff_user, permission_impersonate_user, staff_group): staff_group.permissions.add(permission_impersonate_user) staff_user.groups.add(staff_group) staff_user = User.objects.get(pk=staff_user.pk) response = staff_client.get('/impersonate/{}'.format(customer_user.pk), follow=True) assert response.context['user'] == customer_user assert response.context['user'].is_impersonate assert response.context['request'].impersonator == staff_user Use reverse function in testsfrom django.core.urlresolvers import reverse import pytest from saleor.userprofile.impersonate import can_impersonate from saleor.userprofile.models import User def test_staff_with_permission_can_impersonate( staff_client, customer_user, staff_user, permission_impersonate_user, staff_group): staff_group.permissions.add(permission_impersonate_user) staff_user.groups.add(staff_group) staff_user = User.objects.get(pk=staff_user.pk) response = staff_client.get(reverse('impersonate-start', args=[customer_user.pk]), follow=True) assert response.context['user'] == customer_user assert response.context['user'].is_impersonate assert response.context['request'].impersonator == staff_user
<commit_before>from django.core.urlresolvers import reverse import pytest from saleor.userprofile.impersonate import can_impersonate from saleor.userprofile.models import User def test_staff_with_permission_can_impersonate( staff_client, customer_user, staff_user, permission_impersonate_user, staff_group): staff_group.permissions.add(permission_impersonate_user) staff_user.groups.add(staff_group) staff_user = User.objects.get(pk=staff_user.pk) response = staff_client.get('/impersonate/{}'.format(customer_user.pk), follow=True) assert response.context['user'] == customer_user assert response.context['user'].is_impersonate assert response.context['request'].impersonator == staff_user <commit_msg>Use reverse function in tests<commit_after>from django.core.urlresolvers import reverse import pytest from saleor.userprofile.impersonate import can_impersonate from saleor.userprofile.models import User def test_staff_with_permission_can_impersonate( staff_client, customer_user, staff_user, permission_impersonate_user, staff_group): staff_group.permissions.add(permission_impersonate_user) staff_user.groups.add(staff_group) staff_user = User.objects.get(pk=staff_user.pk) response = staff_client.get(reverse('impersonate-start', args=[customer_user.pk]), follow=True) assert response.context['user'] == customer_user assert response.context['user'].is_impersonate assert response.context['request'].impersonator == staff_user
8b545ee63ec695a77ba08fa5ff45b7d6dd3d94f8
cuteshop/downloaders/git.py
cuteshop/downloaders/git.py
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess from ..utils import DEVNULL, change_working_directory from .base import DOWNLOAD_CONTAINER def download(source_info): url = source_info['git'] subprocess.call( ('git', 'clone', url, DOWNLOAD_CONTAINER), stdout=DEVNULL, stderr=subprocess.STDOUT, ) if 'tag' in source_info: with change_working_directory(DOWNLOAD_CONTAINER): subprocess.call( ('git', 'checkout', source_info['tag']), stdout=DEVNULL, stderr=subprocess.STDOUT, )
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess from ..utils import DEVNULL, change_working_directory from .base import DOWNLOAD_CONTAINER def _checkout(name): with change_working_directory(DOWNLOAD_CONTAINER): subprocess.call( ('git', 'checkout', name), stdout=DEVNULL, stderr=subprocess.STDOUT, ) def download(source_info): url = source_info['git'] subprocess.call( ('git', 'clone', url, DOWNLOAD_CONTAINER), stdout=DEVNULL, stderr=subprocess.STDOUT, ) if 'tag' in source_info: _checkout(source_info['tag']) elif 'branch' in source_info: _checkout(source_info['branch'])
Add auto branch checkout functionality
Add auto branch checkout functionality
Python
mit
uranusjr/cuteshop
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess from ..utils import DEVNULL, change_working_directory from .base import DOWNLOAD_CONTAINER def download(source_info): url = source_info['git'] subprocess.call( ('git', 'clone', url, DOWNLOAD_CONTAINER), stdout=DEVNULL, stderr=subprocess.STDOUT, ) if 'tag' in source_info: with change_working_directory(DOWNLOAD_CONTAINER): subprocess.call( ('git', 'checkout', source_info['tag']), stdout=DEVNULL, stderr=subprocess.STDOUT, ) Add auto branch checkout functionality
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess from ..utils import DEVNULL, change_working_directory from .base import DOWNLOAD_CONTAINER def _checkout(name): with change_working_directory(DOWNLOAD_CONTAINER): subprocess.call( ('git', 'checkout', name), stdout=DEVNULL, stderr=subprocess.STDOUT, ) def download(source_info): url = source_info['git'] subprocess.call( ('git', 'clone', url, DOWNLOAD_CONTAINER), stdout=DEVNULL, stderr=subprocess.STDOUT, ) if 'tag' in source_info: _checkout(source_info['tag']) elif 'branch' in source_info: _checkout(source_info['branch'])
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess from ..utils import DEVNULL, change_working_directory from .base import DOWNLOAD_CONTAINER def download(source_info): url = source_info['git'] subprocess.call( ('git', 'clone', url, DOWNLOAD_CONTAINER), stdout=DEVNULL, stderr=subprocess.STDOUT, ) if 'tag' in source_info: with change_working_directory(DOWNLOAD_CONTAINER): subprocess.call( ('git', 'checkout', source_info['tag']), stdout=DEVNULL, stderr=subprocess.STDOUT, ) <commit_msg>Add auto branch checkout functionality<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess from ..utils import DEVNULL, change_working_directory from .base import DOWNLOAD_CONTAINER def _checkout(name): with change_working_directory(DOWNLOAD_CONTAINER): subprocess.call( ('git', 'checkout', name), stdout=DEVNULL, stderr=subprocess.STDOUT, ) def download(source_info): url = source_info['git'] subprocess.call( ('git', 'clone', url, DOWNLOAD_CONTAINER), stdout=DEVNULL, stderr=subprocess.STDOUT, ) if 'tag' in source_info: _checkout(source_info['tag']) elif 'branch' in source_info: _checkout(source_info['branch'])
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess from ..utils import DEVNULL, change_working_directory from .base import DOWNLOAD_CONTAINER def download(source_info): url = source_info['git'] subprocess.call( ('git', 'clone', url, DOWNLOAD_CONTAINER), stdout=DEVNULL, stderr=subprocess.STDOUT, ) if 'tag' in source_info: with change_working_directory(DOWNLOAD_CONTAINER): subprocess.call( ('git', 'checkout', source_info['tag']), stdout=DEVNULL, stderr=subprocess.STDOUT, ) Add auto branch checkout functionality#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess from ..utils import DEVNULL, change_working_directory from .base import DOWNLOAD_CONTAINER def _checkout(name): with change_working_directory(DOWNLOAD_CONTAINER): subprocess.call( ('git', 'checkout', name), stdout=DEVNULL, stderr=subprocess.STDOUT, ) def download(source_info): url = source_info['git'] subprocess.call( ('git', 'clone', url, DOWNLOAD_CONTAINER), stdout=DEVNULL, stderr=subprocess.STDOUT, ) if 'tag' in source_info: _checkout(source_info['tag']) elif 'branch' in source_info: _checkout(source_info['branch'])
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess from ..utils import DEVNULL, change_working_directory from .base import DOWNLOAD_CONTAINER def download(source_info): url = source_info['git'] subprocess.call( ('git', 'clone', url, DOWNLOAD_CONTAINER), stdout=DEVNULL, stderr=subprocess.STDOUT, ) if 'tag' in source_info: with change_working_directory(DOWNLOAD_CONTAINER): subprocess.call( ('git', 'checkout', source_info['tag']), stdout=DEVNULL, stderr=subprocess.STDOUT, ) <commit_msg>Add auto branch checkout functionality<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess from ..utils import DEVNULL, change_working_directory from .base import DOWNLOAD_CONTAINER def _checkout(name): with change_working_directory(DOWNLOAD_CONTAINER): subprocess.call( ('git', 'checkout', name), stdout=DEVNULL, stderr=subprocess.STDOUT, ) def download(source_info): url = source_info['git'] subprocess.call( ('git', 'clone', url, DOWNLOAD_CONTAINER), stdout=DEVNULL, stderr=subprocess.STDOUT, ) if 'tag' in source_info: _checkout(source_info['tag']) elif 'branch' in source_info: _checkout(source_info['branch'])
687f48ca94b67321a1576a1dbb1d7ae89fe6f0b7
tests/test_pubannotation.py
tests/test_pubannotation.py
import kindred def test_pubannotation_groST(): corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 50 assert relationCount == 1454 assert entityCount == 2657 def test_pubannotation_wikiPain(): corpus = kindred.pubannotation.load('WikiPainGoldStandard') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 49 assert relationCount == 715 assert entityCount == 878 if __name__ == '__main__': test_pubannotation()
import kindred def test_pubannotation(): corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 50 assert relationCount == 1454 assert entityCount == 2657 if __name__ == '__main__': test_pubannotation()
Remove one of the pubannotation tests as their data seems to change
Remove one of the pubannotation tests as their data seems to change
Python
mit
jakelever/kindred,jakelever/kindred
import kindred def test_pubannotation_groST(): corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 50 assert relationCount == 1454 assert entityCount == 2657 def test_pubannotation_wikiPain(): corpus = kindred.pubannotation.load('WikiPainGoldStandard') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 49 assert relationCount == 715 assert entityCount == 878 if __name__ == '__main__': test_pubannotation() Remove one of the pubannotation tests as their data seems to change
import kindred def test_pubannotation(): corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 50 assert relationCount == 1454 assert entityCount == 2657 if __name__ == '__main__': test_pubannotation()
<commit_before> import kindred def test_pubannotation_groST(): corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 50 assert relationCount == 1454 assert entityCount == 2657 def test_pubannotation_wikiPain(): corpus = kindred.pubannotation.load('WikiPainGoldStandard') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 49 assert relationCount == 715 assert entityCount == 878 if __name__ == '__main__': test_pubannotation() <commit_msg>Remove one of the pubannotation tests as their data seems to change<commit_after>
import kindred def test_pubannotation(): corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 50 assert relationCount == 1454 assert entityCount == 2657 if __name__ == '__main__': test_pubannotation()
import kindred def test_pubannotation_groST(): corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 50 assert relationCount == 1454 assert entityCount == 2657 def test_pubannotation_wikiPain(): corpus = kindred.pubannotation.load('WikiPainGoldStandard') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 49 assert relationCount == 715 assert entityCount == 878 if __name__ == '__main__': test_pubannotation() Remove one of the pubannotation tests as their data seems to change import kindred def test_pubannotation(): corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 50 assert relationCount == 1454 assert entityCount == 2657 if __name__ == '__main__': test_pubannotation()
<commit_before> import kindred def test_pubannotation_groST(): corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 50 assert relationCount == 1454 assert entityCount == 2657 def test_pubannotation_wikiPain(): corpus = kindred.pubannotation.load('WikiPainGoldStandard') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 49 assert relationCount == 715 assert entityCount == 878 if __name__ == '__main__': test_pubannotation() <commit_msg>Remove one of the pubannotation tests as their data seems to change<commit_after> import kindred def test_pubannotation(): corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development') assert isinstance(corpus,kindred.Corpus) fileCount = len(corpus.documents) entityCount = sum([ len(d.entities) for d in corpus.documents ]) relationCount = sum([ len(d.relations) for d in corpus.documents ]) assert fileCount == 50 assert relationCount == 1454 assert entityCount == 2657 if __name__ == '__main__': test_pubannotation()
462700e3b1158fef187732007125a0930841dafd
bugsy/errors.py
bugsy/errors.py
class BugsyException(Exception): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ def __init__(self, msg, error_code=None): self.msg = msg self.code = error_code def __str__(self): return "Message: {message} Code: {code}".format(message=self.msg, code=self.code) class LoginException(BugsyException): """ If a username and password are passed in but we don't receive a token then this error will be raised. """ pass class BugException(BugsyException): """ If we try do something that is not allowed to a bug then this error is raised """ pass class SearchException(BugsyException): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ pass
class BugsyException(Exception): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ def __init__(self, msg, error_code=None): self.msg = msg self.code = error_code def __str__(self): return "Message: {message} Code: {code}".format(message=self.msg, code=self.code) class LoginException(BugsyException): """ If a username and password are passed in but we don't receive a token then this error will be raised. """ pass class AttachmentException(BugsyException): """ If we try do something that is not allowed to an attachment then this error is raised """ pass class BugException(BugsyException): """ If we try do something that is not allowed to a bug then this error is raised """ pass class SearchException(BugsyException): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ pass
Add exception handler for attachment related operations
Add exception handler for attachment related operations
Python
apache-2.0
AutomatedTester/Bugsy
class BugsyException(Exception): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ def __init__(self, msg, error_code=None): self.msg = msg self.code = error_code def __str__(self): return "Message: {message} Code: {code}".format(message=self.msg, code=self.code) class LoginException(BugsyException): """ If a username and password are passed in but we don't receive a token then this error will be raised. """ pass class BugException(BugsyException): """ If we try do something that is not allowed to a bug then this error is raised """ pass class SearchException(BugsyException): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ pass Add exception handler for attachment related operations
class BugsyException(Exception): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ def __init__(self, msg, error_code=None): self.msg = msg self.code = error_code def __str__(self): return "Message: {message} Code: {code}".format(message=self.msg, code=self.code) class LoginException(BugsyException): """ If a username and password are passed in but we don't receive a token then this error will be raised. """ pass class AttachmentException(BugsyException): """ If we try do something that is not allowed to an attachment then this error is raised """ pass class BugException(BugsyException): """ If we try do something that is not allowed to a bug then this error is raised """ pass class SearchException(BugsyException): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ pass
<commit_before>class BugsyException(Exception): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ def __init__(self, msg, error_code=None): self.msg = msg self.code = error_code def __str__(self): return "Message: {message} Code: {code}".format(message=self.msg, code=self.code) class LoginException(BugsyException): """ If a username and password are passed in but we don't receive a token then this error will be raised. """ pass class BugException(BugsyException): """ If we try do something that is not allowed to a bug then this error is raised """ pass class SearchException(BugsyException): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ pass <commit_msg>Add exception handler for attachment related operations<commit_after>
class BugsyException(Exception): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ def __init__(self, msg, error_code=None): self.msg = msg self.code = error_code def __str__(self): return "Message: {message} Code: {code}".format(message=self.msg, code=self.code) class LoginException(BugsyException): """ If a username and password are passed in but we don't receive a token then this error will be raised. """ pass class AttachmentException(BugsyException): """ If we try do something that is not allowed to an attachment then this error is raised """ pass class BugException(BugsyException): """ If we try do something that is not allowed to a bug then this error is raised """ pass class SearchException(BugsyException): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ pass
class BugsyException(Exception): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ def __init__(self, msg, error_code=None): self.msg = msg self.code = error_code def __str__(self): return "Message: {message} Code: {code}".format(message=self.msg, code=self.code) class LoginException(BugsyException): """ If a username and password are passed in but we don't receive a token then this error will be raised. """ pass class BugException(BugsyException): """ If we try do something that is not allowed to a bug then this error is raised """ pass class SearchException(BugsyException): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ pass Add exception handler for attachment related operationsclass BugsyException(Exception): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ def __init__(self, msg, error_code=None): self.msg = msg self.code = error_code def __str__(self): return "Message: {message} Code: {code}".format(message=self.msg, code=self.code) class LoginException(BugsyException): """ If a username and password are passed in but we don't receive a token then this error will be raised. """ pass class AttachmentException(BugsyException): """ If we try do something that is not allowed to an attachment then this error is raised """ pass class BugException(BugsyException): """ If we try do something that is not allowed to a bug then this error is raised """ pass class SearchException(BugsyException): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ pass
<commit_before>class BugsyException(Exception): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ def __init__(self, msg, error_code=None): self.msg = msg self.code = error_code def __str__(self): return "Message: {message} Code: {code}".format(message=self.msg, code=self.code) class LoginException(BugsyException): """ If a username and password are passed in but we don't receive a token then this error will be raised. """ pass class BugException(BugsyException): """ If we try do something that is not allowed to a bug then this error is raised """ pass class SearchException(BugsyException): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ pass <commit_msg>Add exception handler for attachment related operations<commit_after>class BugsyException(Exception): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ def __init__(self, msg, error_code=None): self.msg = msg self.code = error_code def __str__(self): return "Message: {message} Code: {code}".format(message=self.msg, code=self.code) class LoginException(BugsyException): """ If a username and password are passed in but we don't receive a token then this error will be raised. """ pass class AttachmentException(BugsyException): """ If we try do something that is not allowed to an attachment then this error is raised """ pass class BugException(BugsyException): """ If we try do something that is not allowed to a bug then this error is raised """ pass class SearchException(BugsyException): """ If while interacting with Bugzilla and we try do something that is not supported this error will be raised. """ pass
31c921f0f88df5bc532db0f326ba9ef53318feb9
codejail/django_integration.py
codejail/django_integration.py
"""Django integration for codejail""" from django.core.exceptions import MiddlewareNotUsed from django.conf import settings import codejail.jail_code class ConfigureCodeJailMiddleware(object): """Middleware to configure codejail on startup.""" def __init__(self): python_bin = settings.CODE_JAIL.get('python_bin') if python_bin: user = settings.CODE_JAIL['user'] codejail.jail_code.configure("python", python_bin, user=user) raise MiddlewareNotUsed
"""Django integration for codejail. Code to glue codejail into a Django environment. """ from django.core.exceptions import MiddlewareNotUsed from django.conf import settings import codejail.jail_code class ConfigureCodeJailMiddleware(object): """ Middleware to configure codejail on startup. This is a Django idiom to have code run once on server startup: put the code in the `__init__` of some middleware, and have it do the work, then raise `MiddlewareNotUsed` to disable the middleware. """ def __init__(self): python_bin = settings.CODE_JAIL.get('python_bin') if python_bin: user = settings.CODE_JAIL['user'] codejail.jail_code.configure("python", python_bin, user=user) raise MiddlewareNotUsed
Add more detail in docstring
Add more detail in docstring
Python
agpl-3.0
StepicOrg/codejail,edx/codejail
"""Django integration for codejail""" from django.core.exceptions import MiddlewareNotUsed from django.conf import settings import codejail.jail_code class ConfigureCodeJailMiddleware(object): """Middleware to configure codejail on startup.""" def __init__(self): python_bin = settings.CODE_JAIL.get('python_bin') if python_bin: user = settings.CODE_JAIL['user'] codejail.jail_code.configure("python", python_bin, user=user) raise MiddlewareNotUsed Add more detail in docstring
"""Django integration for codejail. Code to glue codejail into a Django environment. """ from django.core.exceptions import MiddlewareNotUsed from django.conf import settings import codejail.jail_code class ConfigureCodeJailMiddleware(object): """ Middleware to configure codejail on startup. This is a Django idiom to have code run once on server startup: put the code in the `__init__` of some middleware, and have it do the work, then raise `MiddlewareNotUsed` to disable the middleware. """ def __init__(self): python_bin = settings.CODE_JAIL.get('python_bin') if python_bin: user = settings.CODE_JAIL['user'] codejail.jail_code.configure("python", python_bin, user=user) raise MiddlewareNotUsed
<commit_before>"""Django integration for codejail""" from django.core.exceptions import MiddlewareNotUsed from django.conf import settings import codejail.jail_code class ConfigureCodeJailMiddleware(object): """Middleware to configure codejail on startup.""" def __init__(self): python_bin = settings.CODE_JAIL.get('python_bin') if python_bin: user = settings.CODE_JAIL['user'] codejail.jail_code.configure("python", python_bin, user=user) raise MiddlewareNotUsed <commit_msg>Add more detail in docstring<commit_after>
"""Django integration for codejail. Code to glue codejail into a Django environment. """ from django.core.exceptions import MiddlewareNotUsed from django.conf import settings import codejail.jail_code class ConfigureCodeJailMiddleware(object): """ Middleware to configure codejail on startup. This is a Django idiom to have code run once on server startup: put the code in the `__init__` of some middleware, and have it do the work, then raise `MiddlewareNotUsed` to disable the middleware. """ def __init__(self): python_bin = settings.CODE_JAIL.get('python_bin') if python_bin: user = settings.CODE_JAIL['user'] codejail.jail_code.configure("python", python_bin, user=user) raise MiddlewareNotUsed
"""Django integration for codejail""" from django.core.exceptions import MiddlewareNotUsed from django.conf import settings import codejail.jail_code class ConfigureCodeJailMiddleware(object): """Middleware to configure codejail on startup.""" def __init__(self): python_bin = settings.CODE_JAIL.get('python_bin') if python_bin: user = settings.CODE_JAIL['user'] codejail.jail_code.configure("python", python_bin, user=user) raise MiddlewareNotUsed Add more detail in docstring"""Django integration for codejail. Code to glue codejail into a Django environment. """ from django.core.exceptions import MiddlewareNotUsed from django.conf import settings import codejail.jail_code class ConfigureCodeJailMiddleware(object): """ Middleware to configure codejail on startup. This is a Django idiom to have code run once on server startup: put the code in the `__init__` of some middleware, and have it do the work, then raise `MiddlewareNotUsed` to disable the middleware. """ def __init__(self): python_bin = settings.CODE_JAIL.get('python_bin') if python_bin: user = settings.CODE_JAIL['user'] codejail.jail_code.configure("python", python_bin, user=user) raise MiddlewareNotUsed
<commit_before>"""Django integration for codejail""" from django.core.exceptions import MiddlewareNotUsed from django.conf import settings import codejail.jail_code class ConfigureCodeJailMiddleware(object): """Middleware to configure codejail on startup.""" def __init__(self): python_bin = settings.CODE_JAIL.get('python_bin') if python_bin: user = settings.CODE_JAIL['user'] codejail.jail_code.configure("python", python_bin, user=user) raise MiddlewareNotUsed <commit_msg>Add more detail in docstring<commit_after>"""Django integration for codejail. Code to glue codejail into a Django environment. """ from django.core.exceptions import MiddlewareNotUsed from django.conf import settings import codejail.jail_code class ConfigureCodeJailMiddleware(object): """ Middleware to configure codejail on startup. This is a Django idiom to have code run once on server startup: put the code in the `__init__` of some middleware, and have it do the work, then raise `MiddlewareNotUsed` to disable the middleware. """ def __init__(self): python_bin = settings.CODE_JAIL.get('python_bin') if python_bin: user = settings.CODE_JAIL['user'] codejail.jail_code.configure("python", python_bin, user=user) raise MiddlewareNotUsed
55ba2c2310a0f3a4a413801ce8edf52e001c9ffd
tornado_srv.py
tornado_srv.py
import tornado.web import tornado.wsgi import tornado.httpserver import tornado.ioloop from mojibake.main import app from mojibake.settings import PORT container = tornado.wsgi.WSGIContainer(app) http_server = tornado.httpserver.HTTPServer(container) http_server.listen(PORT) tornado.ioloop.IOLoop.instance().start()
import tornado.web import tornado.wsgi import tornado.httpserver import tornado.ioloop import os from mojibake.main import app from mojibake.settings import PORT if os.name == 'posix': import setproctitle setproctitle.setproctitle('mojibake') # Set the process title to mojibake print('Starting Mojibake...') container = tornado.wsgi.WSGIContainer(app) http_server = tornado.httpserver.HTTPServer(container) http_server.listen(PORT) tornado.ioloop.IOLoop.instance().start()
Set the process title on posix systems
Set the process title on posix systems
Python
mit
ardinor/mojibake,ardinor/mojibake,ardinor/mojibake
import tornado.web import tornado.wsgi import tornado.httpserver import tornado.ioloop from mojibake.main import app from mojibake.settings import PORT container = tornado.wsgi.WSGIContainer(app) http_server = tornado.httpserver.HTTPServer(container) http_server.listen(PORT) tornado.ioloop.IOLoop.instance().start() Set the process title on posix systems
import tornado.web import tornado.wsgi import tornado.httpserver import tornado.ioloop import os from mojibake.main import app from mojibake.settings import PORT if os.name == 'posix': import setproctitle setproctitle.setproctitle('mojibake') # Set the process title to mojibake print('Starting Mojibake...') container = tornado.wsgi.WSGIContainer(app) http_server = tornado.httpserver.HTTPServer(container) http_server.listen(PORT) tornado.ioloop.IOLoop.instance().start()
<commit_before>import tornado.web import tornado.wsgi import tornado.httpserver import tornado.ioloop from mojibake.main import app from mojibake.settings import PORT container = tornado.wsgi.WSGIContainer(app) http_server = tornado.httpserver.HTTPServer(container) http_server.listen(PORT) tornado.ioloop.IOLoop.instance().start() <commit_msg>Set the process title on posix systems<commit_after>
import tornado.web import tornado.wsgi import tornado.httpserver import tornado.ioloop import os from mojibake.main import app from mojibake.settings import PORT if os.name == 'posix': import setproctitle setproctitle.setproctitle('mojibake') # Set the process title to mojibake print('Starting Mojibake...') container = tornado.wsgi.WSGIContainer(app) http_server = tornado.httpserver.HTTPServer(container) http_server.listen(PORT) tornado.ioloop.IOLoop.instance().start()
import tornado.web import tornado.wsgi import tornado.httpserver import tornado.ioloop from mojibake.main import app from mojibake.settings import PORT container = tornado.wsgi.WSGIContainer(app) http_server = tornado.httpserver.HTTPServer(container) http_server.listen(PORT) tornado.ioloop.IOLoop.instance().start() Set the process title on posix systemsimport tornado.web import tornado.wsgi import tornado.httpserver import tornado.ioloop import os from mojibake.main import app from mojibake.settings import PORT if os.name == 'posix': import setproctitle setproctitle.setproctitle('mojibake') # Set the process title to mojibake print('Starting Mojibake...') container = tornado.wsgi.WSGIContainer(app) http_server = tornado.httpserver.HTTPServer(container) http_server.listen(PORT) tornado.ioloop.IOLoop.instance().start()
<commit_before>import tornado.web import tornado.wsgi import tornado.httpserver import tornado.ioloop from mojibake.main import app from mojibake.settings import PORT container = tornado.wsgi.WSGIContainer(app) http_server = tornado.httpserver.HTTPServer(container) http_server.listen(PORT) tornado.ioloop.IOLoop.instance().start() <commit_msg>Set the process title on posix systems<commit_after>import tornado.web import tornado.wsgi import tornado.httpserver import tornado.ioloop import os from mojibake.main import app from mojibake.settings import PORT if os.name == 'posix': import setproctitle setproctitle.setproctitle('mojibake') # Set the process title to mojibake print('Starting Mojibake...') container = tornado.wsgi.WSGIContainer(app) http_server = tornado.httpserver.HTTPServer(container) http_server.listen(PORT) tornado.ioloop.IOLoop.instance().start()
2e9e14980d87239f861377d1dac45bb04d3f9712
tests/basics/array_intbig.py
tests/basics/array_intbig.py
# test array('q') and array('Q') try: from array import array except ImportError: import sys print("SKIP") sys.exit() print(array('q')) print(array('Q')) print(array('q', [0])) print(array('Q', [0])) print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1])) print(array('Q', [0, 1, 2, 2**64-1])) print(bytes(array('q', [-1]))) print(bytes(array('Q', [2**64-1])))
# test array types QqLl that require big-ints try: from array import array except ImportError: import sys print("SKIP") sys.exit() print(array('L', [0, 2**32-1])) print(array('l', [-2**31, 0, 2**31-1])) print(array('q')) print(array('Q')) print(array('q', [0])) print(array('Q', [0])) print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1])) print(array('Q', [0, 1, 2, 2**64-1])) print(bytes(array('q', [-1]))) print(bytes(array('Q', [2**64-1])))
Update array test for big-int with lL typecodes.
tests/basics: Update array test for big-int with lL typecodes.
Python
mit
TDAbboud/micropython,tralamazza/micropython,hiway/micropython,AriZuu/micropython,puuu/micropython,lowRISC/micropython,torwag/micropython,ryannathans/micropython,bvernoux/micropython,pozetroninc/micropython,pramasoul/micropython,deshipu/micropython,tralamazza/micropython,trezor/micropython,pramasoul/micropython,swegener/micropython,adafruit/micropython,adafruit/circuitpython,SHA2017-badge/micropython-esp32,pfalcon/micropython,alex-robbins/micropython,AriZuu/micropython,chrisdearman/micropython,adafruit/circuitpython,trezor/micropython,pozetroninc/micropython,HenrikSolver/micropython,bvernoux/micropython,pozetroninc/micropython,ryannathans/micropython,oopy/micropython,ryannathans/micropython,kerneltask/micropython,lowRISC/micropython,dmazzella/micropython,SHA2017-badge/micropython-esp32,infinnovation/micropython,oopy/micropython,lowRISC/micropython,kerneltask/micropython,blazewicz/micropython,Timmenem/micropython,pramasoul/micropython,trezor/micropython,oopy/micropython,tobbad/micropython,Timmenem/micropython,adafruit/circuitpython,toolmacher/micropython,deshipu/micropython,TDAbboud/micropython,trezor/micropython,chrisdearman/micropython,MrSurly/micropython,dmazzella/micropython,blazewicz/micropython,adafruit/micropython,henriknelson/micropython,micropython/micropython-esp32,pfalcon/micropython,puuu/micropython,PappaPeppar/micropython,pfalcon/micropython,adafruit/circuitpython,MrSurly/micropython-esp32,selste/micropython,pozetroninc/micropython,micropython/micropython-esp32,Timmenem/micropython,torwag/micropython,HenrikSolver/micropython,dmazzella/micropython,alex-robbins/micropython,swegener/micropython,alex-robbins/micropython,dmazzella/micropython,MrSurly/micropython,PappaPeppar/micropython,oopy/micropython,MrSurly/micropython,micropython/micropython-esp32,bvernoux/micropython,toolmacher/micropython,hiway/micropython,pfalcon/micropython,pramasoul/micropython,HenrikSolver/micropython,cwyark/micropython,pfalcon/micropython,tobbad/micropython,alex-robbins/micropython,lowRISC/micropython,cwyark/micropython,AriZuu/micropython,selste/micropython,PappaPeppar/micropython,infinnovation/micropython,deshipu/micropython,torwag/micropython,HenrikSolver/micropython,MrSurly/micropython,lowRISC/micropython,selste/micropython,henriknelson/micropython,swegener/micropython,kerneltask/micropython,cwyark/micropython,TDAbboud/micropython,deshipu/micropython,torwag/micropython,tobbad/micropython,adafruit/micropython,tobbad/micropython,MrSurly/micropython-esp32,cwyark/micropython,puuu/micropython,MrSurly/micropython-esp32,pramasoul/micropython,HenrikSolver/micropython,MrSurly/micropython-esp32,bvernoux/micropython,kerneltask/micropython,SHA2017-badge/micropython-esp32,tralamazza/micropython,deshipu/micropython,chrisdearman/micropython,puuu/micropython,TDAbboud/micropython,bvernoux/micropython,toolmacher/micropython,ryannathans/micropython,toolmacher/micropython,alex-robbins/micropython,adafruit/circuitpython,PappaPeppar/micropython,micropython/micropython-esp32,hiway/micropython,swegener/micropython,blazewicz/micropython,selste/micropython,swegener/micropython,blazewicz/micropython,henriknelson/micropython,pozetroninc/micropython,PappaPeppar/micropython,TDAbboud/micropython,selste/micropython,toolmacher/micropython,puuu/micropython,SHA2017-badge/micropython-esp32,adafruit/circuitpython,blazewicz/micropython,SHA2017-badge/micropython-esp32,chrisdearman/micropython,infinnovation/micropython,trezor/micropython,adafruit/micropython,Timmenem/micropython,henriknelson/micropython,AriZuu/micropython,hiway/micropython,infinnovation/micropython,adafruit/micropython,tralamazza/micropython,infinnovation/micropython,cwyark/micropython,AriZuu/micropython,kerneltask/micropython,henriknelson/micropython,chrisdearman/micropython,Timmenem/micropython,tobbad/micropython,hiway/micropython,MrSurly/micropython,micropython/micropython-esp32,MrSurly/micropython-esp32,ryannathans/micropython,oopy/micropython,torwag/micropython
# test array('q') and array('Q') try: from array import array except ImportError: import sys print("SKIP") sys.exit() print(array('q')) print(array('Q')) print(array('q', [0])) print(array('Q', [0])) print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1])) print(array('Q', [0, 1, 2, 2**64-1])) print(bytes(array('q', [-1]))) print(bytes(array('Q', [2**64-1]))) tests/basics: Update array test for big-int with lL typecodes.
# test array types QqLl that require big-ints try: from array import array except ImportError: import sys print("SKIP") sys.exit() print(array('L', [0, 2**32-1])) print(array('l', [-2**31, 0, 2**31-1])) print(array('q')) print(array('Q')) print(array('q', [0])) print(array('Q', [0])) print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1])) print(array('Q', [0, 1, 2, 2**64-1])) print(bytes(array('q', [-1]))) print(bytes(array('Q', [2**64-1])))
<commit_before># test array('q') and array('Q') try: from array import array except ImportError: import sys print("SKIP") sys.exit() print(array('q')) print(array('Q')) print(array('q', [0])) print(array('Q', [0])) print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1])) print(array('Q', [0, 1, 2, 2**64-1])) print(bytes(array('q', [-1]))) print(bytes(array('Q', [2**64-1]))) <commit_msg>tests/basics: Update array test for big-int with lL typecodes.<commit_after>
# test array types QqLl that require big-ints try: from array import array except ImportError: import sys print("SKIP") sys.exit() print(array('L', [0, 2**32-1])) print(array('l', [-2**31, 0, 2**31-1])) print(array('q')) print(array('Q')) print(array('q', [0])) print(array('Q', [0])) print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1])) print(array('Q', [0, 1, 2, 2**64-1])) print(bytes(array('q', [-1]))) print(bytes(array('Q', [2**64-1])))
# test array('q') and array('Q') try: from array import array except ImportError: import sys print("SKIP") sys.exit() print(array('q')) print(array('Q')) print(array('q', [0])) print(array('Q', [0])) print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1])) print(array('Q', [0, 1, 2, 2**64-1])) print(bytes(array('q', [-1]))) print(bytes(array('Q', [2**64-1]))) tests/basics: Update array test for big-int with lL typecodes.# test array types QqLl that require big-ints try: from array import array except ImportError: import sys print("SKIP") sys.exit() print(array('L', [0, 2**32-1])) print(array('l', [-2**31, 0, 2**31-1])) print(array('q')) print(array('Q')) print(array('q', [0])) print(array('Q', [0])) print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1])) print(array('Q', [0, 1, 2, 2**64-1])) print(bytes(array('q', [-1]))) print(bytes(array('Q', [2**64-1])))
<commit_before># test array('q') and array('Q') try: from array import array except ImportError: import sys print("SKIP") sys.exit() print(array('q')) print(array('Q')) print(array('q', [0])) print(array('Q', [0])) print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1])) print(array('Q', [0, 1, 2, 2**64-1])) print(bytes(array('q', [-1]))) print(bytes(array('Q', [2**64-1]))) <commit_msg>tests/basics: Update array test for big-int with lL typecodes.<commit_after># test array types QqLl that require big-ints try: from array import array except ImportError: import sys print("SKIP") sys.exit() print(array('L', [0, 2**32-1])) print(array('l', [-2**31, 0, 2**31-1])) print(array('q')) print(array('Q')) print(array('q', [0])) print(array('Q', [0])) print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1])) print(array('Q', [0, 1, 2, 2**64-1])) print(bytes(array('q', [-1]))) print(bytes(array('Q', [2**64-1])))
b9ef72138c5312fe8eb7cfa48abe48a8c477afdc
test/test_type_checker_creator.py
test/test_type_checker_creator.py
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import pytest from dataproperty._type_checker_creator import IntegerTypeCheckerCreator from dataproperty._type_checker_creator import FloatTypeCheckerCreator from dataproperty._type_checker_creator import DateTimeTypeCheckerCreator from dataproperty._type_checker import IntegerTypeChecker from dataproperty._type_checker import FloatTypeChecker from dataproperty._type_checker import DateTimeTypeChecker class Test_TypeCheckerCreator(object): @pytest.mark.parametrize(["value", "is_convert", "expected"], [ [IntegerTypeCheckerCreator, True, IntegerTypeChecker], [IntegerTypeCheckerCreator, False, IntegerTypeChecker], [FloatTypeCheckerCreator, True, FloatTypeChecker], [FloatTypeCheckerCreator, False, FloatTypeChecker], [DateTimeTypeCheckerCreator, True, DateTimeTypeChecker], [DateTimeTypeCheckerCreator, False, DateTimeTypeChecker], ]) def test_normal(self, value, is_convert, expected): creator = value() type_checker = creator.create(None, is_convert) assert isinstance(type_checker, expected)
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import pytest import dataproperty._type_checker_creator as tcc import dataproperty._type_checker as tc class Test_TypeCheckerCreator(object): @pytest.mark.parametrize(["value", "is_convert", "expected"], [ [tcc.NoneTypeCheckerCreator, True, tc.NoneTypeChecker], [tcc.NoneTypeCheckerCreator, False, tc.NoneTypeChecker], [tcc.IntegerTypeCheckerCreator, True, tc.IntegerTypeChecker], [tcc.IntegerTypeCheckerCreator, False, tc.IntegerTypeChecker], [tcc.FloatTypeCheckerCreator, True, tc.FloatTypeChecker], [tcc.FloatTypeCheckerCreator, False, tc.FloatTypeChecker], [tcc.DateTimeTypeCheckerCreator, True, tc.DateTimeTypeChecker], [tcc.DateTimeTypeCheckerCreator, False, tc.DateTimeTypeChecker], ]) def test_normal(self, value, is_convert, expected): creator = value() type_checker = creator.create(None, is_convert) assert isinstance(type_checker, expected)
Add tests for NoneTypeCheckerCreator class
Add tests for NoneTypeCheckerCreator class
Python
mit
thombashi/DataProperty
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import pytest from dataproperty._type_checker_creator import IntegerTypeCheckerCreator from dataproperty._type_checker_creator import FloatTypeCheckerCreator from dataproperty._type_checker_creator import DateTimeTypeCheckerCreator from dataproperty._type_checker import IntegerTypeChecker from dataproperty._type_checker import FloatTypeChecker from dataproperty._type_checker import DateTimeTypeChecker class Test_TypeCheckerCreator(object): @pytest.mark.parametrize(["value", "is_convert", "expected"], [ [IntegerTypeCheckerCreator, True, IntegerTypeChecker], [IntegerTypeCheckerCreator, False, IntegerTypeChecker], [FloatTypeCheckerCreator, True, FloatTypeChecker], [FloatTypeCheckerCreator, False, FloatTypeChecker], [DateTimeTypeCheckerCreator, True, DateTimeTypeChecker], [DateTimeTypeCheckerCreator, False, DateTimeTypeChecker], ]) def test_normal(self, value, is_convert, expected): creator = value() type_checker = creator.create(None, is_convert) assert isinstance(type_checker, expected) Add tests for NoneTypeCheckerCreator class
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import pytest import dataproperty._type_checker_creator as tcc import dataproperty._type_checker as tc class Test_TypeCheckerCreator(object): @pytest.mark.parametrize(["value", "is_convert", "expected"], [ [tcc.NoneTypeCheckerCreator, True, tc.NoneTypeChecker], [tcc.NoneTypeCheckerCreator, False, tc.NoneTypeChecker], [tcc.IntegerTypeCheckerCreator, True, tc.IntegerTypeChecker], [tcc.IntegerTypeCheckerCreator, False, tc.IntegerTypeChecker], [tcc.FloatTypeCheckerCreator, True, tc.FloatTypeChecker], [tcc.FloatTypeCheckerCreator, False, tc.FloatTypeChecker], [tcc.DateTimeTypeCheckerCreator, True, tc.DateTimeTypeChecker], [tcc.DateTimeTypeCheckerCreator, False, tc.DateTimeTypeChecker], ]) def test_normal(self, value, is_convert, expected): creator = value() type_checker = creator.create(None, is_convert) assert isinstance(type_checker, expected)
<commit_before># encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import pytest from dataproperty._type_checker_creator import IntegerTypeCheckerCreator from dataproperty._type_checker_creator import FloatTypeCheckerCreator from dataproperty._type_checker_creator import DateTimeTypeCheckerCreator from dataproperty._type_checker import IntegerTypeChecker from dataproperty._type_checker import FloatTypeChecker from dataproperty._type_checker import DateTimeTypeChecker class Test_TypeCheckerCreator(object): @pytest.mark.parametrize(["value", "is_convert", "expected"], [ [IntegerTypeCheckerCreator, True, IntegerTypeChecker], [IntegerTypeCheckerCreator, False, IntegerTypeChecker], [FloatTypeCheckerCreator, True, FloatTypeChecker], [FloatTypeCheckerCreator, False, FloatTypeChecker], [DateTimeTypeCheckerCreator, True, DateTimeTypeChecker], [DateTimeTypeCheckerCreator, False, DateTimeTypeChecker], ]) def test_normal(self, value, is_convert, expected): creator = value() type_checker = creator.create(None, is_convert) assert isinstance(type_checker, expected) <commit_msg>Add tests for NoneTypeCheckerCreator class<commit_after>
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import pytest import dataproperty._type_checker_creator as tcc import dataproperty._type_checker as tc class Test_TypeCheckerCreator(object): @pytest.mark.parametrize(["value", "is_convert", "expected"], [ [tcc.NoneTypeCheckerCreator, True, tc.NoneTypeChecker], [tcc.NoneTypeCheckerCreator, False, tc.NoneTypeChecker], [tcc.IntegerTypeCheckerCreator, True, tc.IntegerTypeChecker], [tcc.IntegerTypeCheckerCreator, False, tc.IntegerTypeChecker], [tcc.FloatTypeCheckerCreator, True, tc.FloatTypeChecker], [tcc.FloatTypeCheckerCreator, False, tc.FloatTypeChecker], [tcc.DateTimeTypeCheckerCreator, True, tc.DateTimeTypeChecker], [tcc.DateTimeTypeCheckerCreator, False, tc.DateTimeTypeChecker], ]) def test_normal(self, value, is_convert, expected): creator = value() type_checker = creator.create(None, is_convert) assert isinstance(type_checker, expected)
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import pytest from dataproperty._type_checker_creator import IntegerTypeCheckerCreator from dataproperty._type_checker_creator import FloatTypeCheckerCreator from dataproperty._type_checker_creator import DateTimeTypeCheckerCreator from dataproperty._type_checker import IntegerTypeChecker from dataproperty._type_checker import FloatTypeChecker from dataproperty._type_checker import DateTimeTypeChecker class Test_TypeCheckerCreator(object): @pytest.mark.parametrize(["value", "is_convert", "expected"], [ [IntegerTypeCheckerCreator, True, IntegerTypeChecker], [IntegerTypeCheckerCreator, False, IntegerTypeChecker], [FloatTypeCheckerCreator, True, FloatTypeChecker], [FloatTypeCheckerCreator, False, FloatTypeChecker], [DateTimeTypeCheckerCreator, True, DateTimeTypeChecker], [DateTimeTypeCheckerCreator, False, DateTimeTypeChecker], ]) def test_normal(self, value, is_convert, expected): creator = value() type_checker = creator.create(None, is_convert) assert isinstance(type_checker, expected) Add tests for NoneTypeCheckerCreator class# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import pytest import dataproperty._type_checker_creator as tcc import dataproperty._type_checker as tc class Test_TypeCheckerCreator(object): @pytest.mark.parametrize(["value", "is_convert", "expected"], [ [tcc.NoneTypeCheckerCreator, True, tc.NoneTypeChecker], [tcc.NoneTypeCheckerCreator, False, tc.NoneTypeChecker], [tcc.IntegerTypeCheckerCreator, True, tc.IntegerTypeChecker], [tcc.IntegerTypeCheckerCreator, False, tc.IntegerTypeChecker], [tcc.FloatTypeCheckerCreator, True, tc.FloatTypeChecker], [tcc.FloatTypeCheckerCreator, False, tc.FloatTypeChecker], [tcc.DateTimeTypeCheckerCreator, True, tc.DateTimeTypeChecker], [tcc.DateTimeTypeCheckerCreator, False, tc.DateTimeTypeChecker], ]) def test_normal(self, value, is_convert, expected): creator = value() type_checker = creator.create(None, is_convert) assert isinstance(type_checker, expected)
<commit_before># encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import pytest from dataproperty._type_checker_creator import IntegerTypeCheckerCreator from dataproperty._type_checker_creator import FloatTypeCheckerCreator from dataproperty._type_checker_creator import DateTimeTypeCheckerCreator from dataproperty._type_checker import IntegerTypeChecker from dataproperty._type_checker import FloatTypeChecker from dataproperty._type_checker import DateTimeTypeChecker class Test_TypeCheckerCreator(object): @pytest.mark.parametrize(["value", "is_convert", "expected"], [ [IntegerTypeCheckerCreator, True, IntegerTypeChecker], [IntegerTypeCheckerCreator, False, IntegerTypeChecker], [FloatTypeCheckerCreator, True, FloatTypeChecker], [FloatTypeCheckerCreator, False, FloatTypeChecker], [DateTimeTypeCheckerCreator, True, DateTimeTypeChecker], [DateTimeTypeCheckerCreator, False, DateTimeTypeChecker], ]) def test_normal(self, value, is_convert, expected): creator = value() type_checker = creator.create(None, is_convert) assert isinstance(type_checker, expected) <commit_msg>Add tests for NoneTypeCheckerCreator class<commit_after># encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import pytest import dataproperty._type_checker_creator as tcc import dataproperty._type_checker as tc class Test_TypeCheckerCreator(object): @pytest.mark.parametrize(["value", "is_convert", "expected"], [ [tcc.NoneTypeCheckerCreator, True, tc.NoneTypeChecker], [tcc.NoneTypeCheckerCreator, False, tc.NoneTypeChecker], [tcc.IntegerTypeCheckerCreator, True, tc.IntegerTypeChecker], [tcc.IntegerTypeCheckerCreator, False, tc.IntegerTypeChecker], [tcc.FloatTypeCheckerCreator, True, tc.FloatTypeChecker], [tcc.FloatTypeCheckerCreator, False, tc.FloatTypeChecker], [tcc.DateTimeTypeCheckerCreator, True, tc.DateTimeTypeChecker], [tcc.DateTimeTypeCheckerCreator, False, tc.DateTimeTypeChecker], ]) def test_normal(self, value, is_convert, expected): creator = value() type_checker = creator.create(None, is_convert) assert isinstance(type_checker, expected)
927915f11ce536074920c515fab6e6ec3134d390
tests/test_huckle_install.py
tests/test_huckle_install.py
from __future__ import absolute_import, division, print_function from subprocess import check_output import os def test_function(): setup = """ #!/bin/bash huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf echo '{"hello":"world"}' | jsonf go """ out = check_output(['bash', '-c', setup]) hello = """ #!/bin/bash export PATH=$PATH:~/.huckle/bin echo '{"hello":"world"}' | jsonf go """ out = check_output(['bash', '-c', hello]) assert('{\n "hello" : "world"\n}\n' in out)
from __future__ import absolute_import, division, print_function import subprocess import os def test_function(): setup = """ #!/bin/bash huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf echo '{"hello":"world"}' | jsonf go """ p1 = subprocess.Popen(['bash', '-c', setup], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, err = p1.communicate() hello = """ #!/bin/bash export PATH=$PATH:~/.huckle/bin echo '{"hello":"world"}' | jsonf go """ p2 = subprocess.Popen(['bash', '-c', hello], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, err = p2.communicate() result = out.decode('utf-8') assert('{\n "hello" : "world"\n}\n' in result)
Revert "fix test by switching to check_output"
Revert "fix test by switching to check_output" This reverts commit 6cfd9d01d68c2f7ff4a8bba3351ee618e770d315.
Python
mit
cometaj2/huckle
from __future__ import absolute_import, division, print_function from subprocess import check_output import os def test_function(): setup = """ #!/bin/bash huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf echo '{"hello":"world"}' | jsonf go """ out = check_output(['bash', '-c', setup]) hello = """ #!/bin/bash export PATH=$PATH:~/.huckle/bin echo '{"hello":"world"}' | jsonf go """ out = check_output(['bash', '-c', hello]) assert('{\n "hello" : "world"\n}\n' in out) Revert "fix test by switching to check_output" This reverts commit 6cfd9d01d68c2f7ff4a8bba3351ee618e770d315.
from __future__ import absolute_import, division, print_function import subprocess import os def test_function(): setup = """ #!/bin/bash huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf echo '{"hello":"world"}' | jsonf go """ p1 = subprocess.Popen(['bash', '-c', setup], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, err = p1.communicate() hello = """ #!/bin/bash export PATH=$PATH:~/.huckle/bin echo '{"hello":"world"}' | jsonf go """ p2 = subprocess.Popen(['bash', '-c', hello], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, err = p2.communicate() result = out.decode('utf-8') assert('{\n "hello" : "world"\n}\n' in result)
<commit_before>from __future__ import absolute_import, division, print_function from subprocess import check_output import os def test_function(): setup = """ #!/bin/bash huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf echo '{"hello":"world"}' | jsonf go """ out = check_output(['bash', '-c', setup]) hello = """ #!/bin/bash export PATH=$PATH:~/.huckle/bin echo '{"hello":"world"}' | jsonf go """ out = check_output(['bash', '-c', hello]) assert('{\n "hello" : "world"\n}\n' in out) <commit_msg>Revert "fix test by switching to check_output" This reverts commit 6cfd9d01d68c2f7ff4a8bba3351ee618e770d315.<commit_after>
from __future__ import absolute_import, division, print_function import subprocess import os def test_function(): setup = """ #!/bin/bash huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf echo '{"hello":"world"}' | jsonf go """ p1 = subprocess.Popen(['bash', '-c', setup], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, err = p1.communicate() hello = """ #!/bin/bash export PATH=$PATH:~/.huckle/bin echo '{"hello":"world"}' | jsonf go """ p2 = subprocess.Popen(['bash', '-c', hello], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, err = p2.communicate() result = out.decode('utf-8') assert('{\n "hello" : "world"\n}\n' in result)
from __future__ import absolute_import, division, print_function from subprocess import check_output import os def test_function(): setup = """ #!/bin/bash huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf echo '{"hello":"world"}' | jsonf go """ out = check_output(['bash', '-c', setup]) hello = """ #!/bin/bash export PATH=$PATH:~/.huckle/bin echo '{"hello":"world"}' | jsonf go """ out = check_output(['bash', '-c', hello]) assert('{\n "hello" : "world"\n}\n' in out) Revert "fix test by switching to check_output" This reverts commit 6cfd9d01d68c2f7ff4a8bba3351ee618e770d315.from __future__ import absolute_import, division, print_function import subprocess import os def test_function(): setup = """ #!/bin/bash huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf echo '{"hello":"world"}' | jsonf go """ p1 = subprocess.Popen(['bash', '-c', setup], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, err = p1.communicate() hello = """ #!/bin/bash export PATH=$PATH:~/.huckle/bin echo '{"hello":"world"}' | jsonf go """ p2 = subprocess.Popen(['bash', '-c', hello], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, err = p2.communicate() result = out.decode('utf-8') assert('{\n "hello" : "world"\n}\n' in result)
<commit_before>from __future__ import absolute_import, division, print_function from subprocess import check_output import os def test_function(): setup = """ #!/bin/bash huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf echo '{"hello":"world"}' | jsonf go """ out = check_output(['bash', '-c', setup]) hello = """ #!/bin/bash export PATH=$PATH:~/.huckle/bin echo '{"hello":"world"}' | jsonf go """ out = check_output(['bash', '-c', hello]) assert('{\n "hello" : "world"\n}\n' in out) <commit_msg>Revert "fix test by switching to check_output" This reverts commit 6cfd9d01d68c2f7ff4a8bba3351ee618e770d315.<commit_after>from __future__ import absolute_import, division, print_function import subprocess import os def test_function(): setup = """ #!/bin/bash huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf echo '{"hello":"world"}' | jsonf go """ p1 = subprocess.Popen(['bash', '-c', setup], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, err = p1.communicate() hello = """ #!/bin/bash export PATH=$PATH:~/.huckle/bin echo '{"hello":"world"}' | jsonf go """ p2 = subprocess.Popen(['bash', '-c', hello], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, err = p2.communicate() result = out.decode('utf-8') assert('{\n "hello" : "world"\n}\n' in result)
1c0f0decd5bdcea3174cee650ba08fb427b67016
tests/test_rover_instance.py
tests/test_rover_instance.py
from unittest import TestCase from rover import Rover class TestRover(TestCase): def setUp(self): self.rover = Rover() def test_rover_compass(self): assert self.rover.compass == ['N', 'E', 'S', 'W'] def test_rover_position(self): assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction) def test_rover_set_position(self): self.rover.set_position(4, 9, 'W') assert self.rover.position == (4, 9, 'W')
from unittest import TestCase from rover import Rover class TestRover(TestCase): def setUp(self): self.rover = Rover() def test_rover_compass(self): assert self.rover.compass == ['N', 'E', 'S', 'W'] def test_rover_position(self): assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction) def test_rover_set_position(self): self.rover.set_position(4, 9, 'W') assert self.rover.position == (4, 9, 'W') def test_rover_move_forward_north(self): self.rover.set_position(0, 0, 'N') self.rover.move('F') assert self.rover.position == (0, 1, 'N') def test_rover_move_forward_south(self): self.rover.set_position(0, 1, 'S') self.rover.move('F') assert self.rover.position == (0, 0, 'S') def test_rover_move_forward_east(self): self.rover.set_position(0, 0, 'E') self.rover.move('F') assert self.rover.position == (1, 0, 'E') def test_rover_move_forward_west(self): self.rover.set_position(1, 0, 'W') self.rover.move('F') assert self.rover.position == (0, 0, 'W')
Add failing tests for rover forward movement
Add failing tests for rover forward movement
Python
mit
authentik8/rover
from unittest import TestCase from rover import Rover class TestRover(TestCase): def setUp(self): self.rover = Rover() def test_rover_compass(self): assert self.rover.compass == ['N', 'E', 'S', 'W'] def test_rover_position(self): assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction) def test_rover_set_position(self): self.rover.set_position(4, 9, 'W') assert self.rover.position == (4, 9, 'W') Add failing tests for rover forward movement
from unittest import TestCase from rover import Rover class TestRover(TestCase): def setUp(self): self.rover = Rover() def test_rover_compass(self): assert self.rover.compass == ['N', 'E', 'S', 'W'] def test_rover_position(self): assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction) def test_rover_set_position(self): self.rover.set_position(4, 9, 'W') assert self.rover.position == (4, 9, 'W') def test_rover_move_forward_north(self): self.rover.set_position(0, 0, 'N') self.rover.move('F') assert self.rover.position == (0, 1, 'N') def test_rover_move_forward_south(self): self.rover.set_position(0, 1, 'S') self.rover.move('F') assert self.rover.position == (0, 0, 'S') def test_rover_move_forward_east(self): self.rover.set_position(0, 0, 'E') self.rover.move('F') assert self.rover.position == (1, 0, 'E') def test_rover_move_forward_west(self): self.rover.set_position(1, 0, 'W') self.rover.move('F') assert self.rover.position == (0, 0, 'W')
<commit_before> from unittest import TestCase from rover import Rover class TestRover(TestCase): def setUp(self): self.rover = Rover() def test_rover_compass(self): assert self.rover.compass == ['N', 'E', 'S', 'W'] def test_rover_position(self): assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction) def test_rover_set_position(self): self.rover.set_position(4, 9, 'W') assert self.rover.position == (4, 9, 'W') <commit_msg>Add failing tests for rover forward movement<commit_after>
from unittest import TestCase from rover import Rover class TestRover(TestCase): def setUp(self): self.rover = Rover() def test_rover_compass(self): assert self.rover.compass == ['N', 'E', 'S', 'W'] def test_rover_position(self): assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction) def test_rover_set_position(self): self.rover.set_position(4, 9, 'W') assert self.rover.position == (4, 9, 'W') def test_rover_move_forward_north(self): self.rover.set_position(0, 0, 'N') self.rover.move('F') assert self.rover.position == (0, 1, 'N') def test_rover_move_forward_south(self): self.rover.set_position(0, 1, 'S') self.rover.move('F') assert self.rover.position == (0, 0, 'S') def test_rover_move_forward_east(self): self.rover.set_position(0, 0, 'E') self.rover.move('F') assert self.rover.position == (1, 0, 'E') def test_rover_move_forward_west(self): self.rover.set_position(1, 0, 'W') self.rover.move('F') assert self.rover.position == (0, 0, 'W')
from unittest import TestCase from rover import Rover class TestRover(TestCase): def setUp(self): self.rover = Rover() def test_rover_compass(self): assert self.rover.compass == ['N', 'E', 'S', 'W'] def test_rover_position(self): assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction) def test_rover_set_position(self): self.rover.set_position(4, 9, 'W') assert self.rover.position == (4, 9, 'W') Add failing tests for rover forward movement from unittest import TestCase from rover import Rover class TestRover(TestCase): def setUp(self): self.rover = Rover() def test_rover_compass(self): assert self.rover.compass == ['N', 'E', 'S', 'W'] def test_rover_position(self): assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction) def test_rover_set_position(self): self.rover.set_position(4, 9, 'W') assert self.rover.position == (4, 9, 'W') def test_rover_move_forward_north(self): self.rover.set_position(0, 0, 'N') self.rover.move('F') assert self.rover.position == (0, 1, 'N') def test_rover_move_forward_south(self): self.rover.set_position(0, 1, 'S') self.rover.move('F') assert self.rover.position == (0, 0, 'S') def test_rover_move_forward_east(self): self.rover.set_position(0, 0, 'E') self.rover.move('F') assert self.rover.position == (1, 0, 'E') def test_rover_move_forward_west(self): self.rover.set_position(1, 0, 'W') self.rover.move('F') assert self.rover.position == (0, 0, 'W')
<commit_before> from unittest import TestCase from rover import Rover class TestRover(TestCase): def setUp(self): self.rover = Rover() def test_rover_compass(self): assert self.rover.compass == ['N', 'E', 'S', 'W'] def test_rover_position(self): assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction) def test_rover_set_position(self): self.rover.set_position(4, 9, 'W') assert self.rover.position == (4, 9, 'W') <commit_msg>Add failing tests for rover forward movement<commit_after> from unittest import TestCase from rover import Rover class TestRover(TestCase): def setUp(self): self.rover = Rover() def test_rover_compass(self): assert self.rover.compass == ['N', 'E', 'S', 'W'] def test_rover_position(self): assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction) def test_rover_set_position(self): self.rover.set_position(4, 9, 'W') assert self.rover.position == (4, 9, 'W') def test_rover_move_forward_north(self): self.rover.set_position(0, 0, 'N') self.rover.move('F') assert self.rover.position == (0, 1, 'N') def test_rover_move_forward_south(self): self.rover.set_position(0, 1, 'S') self.rover.move('F') assert self.rover.position == (0, 0, 'S') def test_rover_move_forward_east(self): self.rover.set_position(0, 0, 'E') self.rover.move('F') assert self.rover.position == (1, 0, 'E') def test_rover_move_forward_west(self): self.rover.set_position(1, 0, 'W') self.rover.move('F') assert self.rover.position == (0, 0, 'W')
05e61f1be4005edf2ff439ca2613bce8af217ff7
pubsubpull/models.py
pubsubpull/models.py
""" Models. """ from django.contrib.auth.models import User from django.core.exceptions import ValidationError from django.db import models from pubsubpull.fields import JSONB class Request(models.Model): """A web request. """ user = models.ForeignKey(User, null=True, blank=True, related_name='requests') method = models.CharField(max_length=20) path = models.TextField() OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE") class UpdateLog(models.Model): """Store a change to a single row in a table. """ table = models.CharField(max_length=200) type = models.CharField(max_length=1, choices=OPERATION_TYPE.items()) when = models.DateTimeField(auto_now_add=True) request = models.ForeignKey(Request, null=True, blank=True, related_name='changes') old = JSONB(null=True, blank=True) new = JSONB(null=True, blank=True) def save(self, **kw): raise ValidationError("Instances of this class cannot be using Django") def __unicode__(self): return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
""" Models. """ from django.contrib.auth.models import User from django.core.exceptions import ValidationError from django.db import models from pubsubpull.fields import JSONB class Request(models.Model): """A web request. """ user = models.ForeignKey(User, null=True, blank=True, related_name='requests') method = models.CharField(max_length=20) path = models.TextField() def __unicode__(self): return "%s %s" % (self.method, self.path) OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE") class UpdateLog(models.Model): """Store a change to a single row in a table. """ table = models.CharField(max_length=200) type = models.CharField(max_length=1, choices=OPERATION_TYPE.items()) when = models.DateTimeField(auto_now_add=True) request = models.ForeignKey(Request, null=True, blank=True, related_name='changes') old = JSONB(null=True, blank=True) new = JSONB(null=True, blank=True) def save(self, **kw): raise ValidationError("Instances of this class cannot be using Django") def __unicode__(self): return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
Add more useful display of the request data.
Add more useful display of the request data.
Python
mit
KayEss/django-pubsubpull,KayEss/django-pubsubpull,KayEss/django-pubsubpull
""" Models. """ from django.contrib.auth.models import User from django.core.exceptions import ValidationError from django.db import models from pubsubpull.fields import JSONB class Request(models.Model): """A web request. """ user = models.ForeignKey(User, null=True, blank=True, related_name='requests') method = models.CharField(max_length=20) path = models.TextField() OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE") class UpdateLog(models.Model): """Store a change to a single row in a table. """ table = models.CharField(max_length=200) type = models.CharField(max_length=1, choices=OPERATION_TYPE.items()) when = models.DateTimeField(auto_now_add=True) request = models.ForeignKey(Request, null=True, blank=True, related_name='changes') old = JSONB(null=True, blank=True) new = JSONB(null=True, blank=True) def save(self, **kw): raise ValidationError("Instances of this class cannot be using Django") def __unicode__(self): return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when) Add more useful display of the request data.
""" Models. """ from django.contrib.auth.models import User from django.core.exceptions import ValidationError from django.db import models from pubsubpull.fields import JSONB class Request(models.Model): """A web request. """ user = models.ForeignKey(User, null=True, blank=True, related_name='requests') method = models.CharField(max_length=20) path = models.TextField() def __unicode__(self): return "%s %s" % (self.method, self.path) OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE") class UpdateLog(models.Model): """Store a change to a single row in a table. """ table = models.CharField(max_length=200) type = models.CharField(max_length=1, choices=OPERATION_TYPE.items()) when = models.DateTimeField(auto_now_add=True) request = models.ForeignKey(Request, null=True, blank=True, related_name='changes') old = JSONB(null=True, blank=True) new = JSONB(null=True, blank=True) def save(self, **kw): raise ValidationError("Instances of this class cannot be using Django") def __unicode__(self): return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
<commit_before>""" Models. """ from django.contrib.auth.models import User from django.core.exceptions import ValidationError from django.db import models from pubsubpull.fields import JSONB class Request(models.Model): """A web request. """ user = models.ForeignKey(User, null=True, blank=True, related_name='requests') method = models.CharField(max_length=20) path = models.TextField() OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE") class UpdateLog(models.Model): """Store a change to a single row in a table. """ table = models.CharField(max_length=200) type = models.CharField(max_length=1, choices=OPERATION_TYPE.items()) when = models.DateTimeField(auto_now_add=True) request = models.ForeignKey(Request, null=True, blank=True, related_name='changes') old = JSONB(null=True, blank=True) new = JSONB(null=True, blank=True) def save(self, **kw): raise ValidationError("Instances of this class cannot be using Django") def __unicode__(self): return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when) <commit_msg>Add more useful display of the request data.<commit_after>
""" Models. """ from django.contrib.auth.models import User from django.core.exceptions import ValidationError from django.db import models from pubsubpull.fields import JSONB class Request(models.Model): """A web request. """ user = models.ForeignKey(User, null=True, blank=True, related_name='requests') method = models.CharField(max_length=20) path = models.TextField() def __unicode__(self): return "%s %s" % (self.method, self.path) OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE") class UpdateLog(models.Model): """Store a change to a single row in a table. """ table = models.CharField(max_length=200) type = models.CharField(max_length=1, choices=OPERATION_TYPE.items()) when = models.DateTimeField(auto_now_add=True) request = models.ForeignKey(Request, null=True, blank=True, related_name='changes') old = JSONB(null=True, blank=True) new = JSONB(null=True, blank=True) def save(self, **kw): raise ValidationError("Instances of this class cannot be using Django") def __unicode__(self): return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
""" Models. """ from django.contrib.auth.models import User from django.core.exceptions import ValidationError from django.db import models from pubsubpull.fields import JSONB class Request(models.Model): """A web request. """ user = models.ForeignKey(User, null=True, blank=True, related_name='requests') method = models.CharField(max_length=20) path = models.TextField() OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE") class UpdateLog(models.Model): """Store a change to a single row in a table. """ table = models.CharField(max_length=200) type = models.CharField(max_length=1, choices=OPERATION_TYPE.items()) when = models.DateTimeField(auto_now_add=True) request = models.ForeignKey(Request, null=True, blank=True, related_name='changes') old = JSONB(null=True, blank=True) new = JSONB(null=True, blank=True) def save(self, **kw): raise ValidationError("Instances of this class cannot be using Django") def __unicode__(self): return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when) Add more useful display of the request data.""" Models. """ from django.contrib.auth.models import User from django.core.exceptions import ValidationError from django.db import models from pubsubpull.fields import JSONB class Request(models.Model): """A web request. """ user = models.ForeignKey(User, null=True, blank=True, related_name='requests') method = models.CharField(max_length=20) path = models.TextField() def __unicode__(self): return "%s %s" % (self.method, self.path) OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE") class UpdateLog(models.Model): """Store a change to a single row in a table. """ table = models.CharField(max_length=200) type = models.CharField(max_length=1, choices=OPERATION_TYPE.items()) when = models.DateTimeField(auto_now_add=True) request = models.ForeignKey(Request, null=True, blank=True, related_name='changes') old = JSONB(null=True, blank=True) new = JSONB(null=True, blank=True) def save(self, **kw): raise ValidationError("Instances of this class cannot be using Django") def __unicode__(self): return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
<commit_before>""" Models. """ from django.contrib.auth.models import User from django.core.exceptions import ValidationError from django.db import models from pubsubpull.fields import JSONB class Request(models.Model): """A web request. """ user = models.ForeignKey(User, null=True, blank=True, related_name='requests') method = models.CharField(max_length=20) path = models.TextField() OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE") class UpdateLog(models.Model): """Store a change to a single row in a table. """ table = models.CharField(max_length=200) type = models.CharField(max_length=1, choices=OPERATION_TYPE.items()) when = models.DateTimeField(auto_now_add=True) request = models.ForeignKey(Request, null=True, blank=True, related_name='changes') old = JSONB(null=True, blank=True) new = JSONB(null=True, blank=True) def save(self, **kw): raise ValidationError("Instances of this class cannot be using Django") def __unicode__(self): return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when) <commit_msg>Add more useful display of the request data.<commit_after>""" Models. """ from django.contrib.auth.models import User from django.core.exceptions import ValidationError from django.db import models from pubsubpull.fields import JSONB class Request(models.Model): """A web request. """ user = models.ForeignKey(User, null=True, blank=True, related_name='requests') method = models.CharField(max_length=20) path = models.TextField() def __unicode__(self): return "%s %s" % (self.method, self.path) OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE") class UpdateLog(models.Model): """Store a change to a single row in a table. """ table = models.CharField(max_length=200) type = models.CharField(max_length=1, choices=OPERATION_TYPE.items()) when = models.DateTimeField(auto_now_add=True) request = models.ForeignKey(Request, null=True, blank=True, related_name='changes') old = JSONB(null=True, blank=True) new = JSONB(null=True, blank=True) def save(self, **kw): raise ValidationError("Instances of this class cannot be using Django") def __unicode__(self): return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
5b8241ad808bd11971d0d684bafd6f9019e58397
tests/contrib/flask/tests.py
tests/contrib/flask/tests.py
import logging from flask import Flask from raven.base import Client from raven.contrib.flask import Sentry from unittest2 import TestCase class TempStoreClient(Client): def __init__(self, *args, **kwargs): self.events = [] super(TempStoreClient, self).__init__(*args, **kwargs) def send(self, **kwargs): self.events.append(kwargs) def create_app(): app = Flask(__name__) @app.route('/an-error/', methods=['GET', 'POST']) def an_error(): raise ValueError('hello world') return app class FlaskTest(TestCase): def setUp(self): self.app = create_app() self.client = self.app.test_client() def test_error_handler(self): client = TempStoreClient() sentry = Sentry(self.app, client=client) response = self.client.get('/an-error/?foo=bar') self.assertEquals(response.status_code, 500) self.assertEquals(len(client.events), 1) event = client.events.pop(0) self.assertEquals(event['class_name'], 'ValueError') self.assertEquals(event['level'], logging.ERROR) self.assertEquals(event['message'], 'hello world') self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
import logging from flask import Flask from raven.base import Client from raven.contrib.flask import Sentry from unittest2 import TestCase class TempStoreClient(Client): def __init__(self, *args, **kwargs): self.events = [] super(TempStoreClient, self).__init__(*args, **kwargs) def send(self, **kwargs): self.events.append(kwargs) def create_app(): app = Flask(__name__) @app.route('/an-error/', methods=['GET', 'POST']) def an_error(): raise ValueError('hello world') return app class FlaskTest(TestCase): def setUp(self): self.app = create_app() self.client = self.app.test_client() def test_error_handler(self): client = TempStoreClient() sentry = Sentry(self.app, client=client) response = self.client.get('/an-error/?foo=bar') self.assertEquals(response.status_code, 500) self.assertEquals(len(client.events), 1) event = client.events.pop(0) self.assertEquals(event['class_name'], 'ValueError') self.assertEquals(event['level'], logging.ERROR) self.assertEquals(event['message'], 'hello world') self.assertEquals(event['url'], 'http://localhost/an-error/?foo=bar') self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
Add url test for Flask
Add url test for Flask
Python
bsd-3-clause
nikolas/raven-python,Photonomie/raven-python,jmagnusson/raven-python,inspirehep/raven-python,danriti/raven-python,lopter/raven-python-old,nikolas/raven-python,johansteffner/raven-python,johansteffner/raven-python,daikeren/opbeat_python,daikeren/opbeat_python,someonehan/raven-python,inspirehep/raven-python,jmagnusson/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,jbarbuto/raven-python,johansteffner/raven-python,danriti/raven-python,icereval/raven-python,dirtycoder/opbeat_python,jmagnusson/raven-python,someonehan/raven-python,patrys/opbeat_python,nikolas/raven-python,mitsuhiko/raven,openlabs/raven,beniwohli/apm-agent-python,tarkatronic/opbeat_python,Photonomie/raven-python,inspirehep/raven-python,recht/raven-python,lepture/raven-python,getsentry/raven-python,nikolas/raven-python,beniwohli/apm-agent-python,jbarbuto/raven-python,arthurlogilab/raven-python,getsentry/raven-python,percipient/raven-python,ewdurbin/raven-python,icereval/raven-python,jmp0xf/raven-python,tarkatronic/opbeat_python,someonehan/raven-python,ronaldevers/raven-python,ewdurbin/raven-python,dirtycoder/opbeat_python,jmp0xf/raven-python,akheron/raven-python,akheron/raven-python,akalipetis/raven-python,tarkatronic/opbeat_python,Goldmund-Wyldebeast-Wunderliebe/raven-python,Photonomie/raven-python,beniwohli/apm-agent-python,beniwohli/apm-agent-python,lepture/raven-python,recht/raven-python,jbarbuto/raven-python,inspirehep/raven-python,daikeren/opbeat_python,smarkets/raven-python,smarkets/raven-python,akheron/raven-python,smarkets/raven-python,mitsuhiko/raven,jmp0xf/raven-python,ticosax/opbeat_python,patrys/opbeat_python,dirtycoder/opbeat_python,percipient/raven-python,ronaldevers/raven-python,patrys/opbeat_python,recht/raven-python,dbravender/raven-python,hzy/raven-python,ticosax/opbeat_python,patrys/opbeat_python,dbravender/raven-python,arthurlogilab/raven-python,smarkets/raven-python,akalipetis/raven-python,icereval/raven-python,arthurlogilab/raven-python,ticosax/opbeat_python,lepture/raven-python,percipient/raven-python,getsentry/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,collective/mr.poe,hzy/raven-python,jbarbuto/raven-python,akalipetis/raven-python,ronaldevers/raven-python,hzy/raven-python,danriti/raven-python,ewdurbin/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,alex/raven,arthurlogilab/raven-python,dbravender/raven-python,icereval/raven-python
import logging from flask import Flask from raven.base import Client from raven.contrib.flask import Sentry from unittest2 import TestCase class TempStoreClient(Client): def __init__(self, *args, **kwargs): self.events = [] super(TempStoreClient, self).__init__(*args, **kwargs) def send(self, **kwargs): self.events.append(kwargs) def create_app(): app = Flask(__name__) @app.route('/an-error/', methods=['GET', 'POST']) def an_error(): raise ValueError('hello world') return app class FlaskTest(TestCase): def setUp(self): self.app = create_app() self.client = self.app.test_client() def test_error_handler(self): client = TempStoreClient() sentry = Sentry(self.app, client=client) response = self.client.get('/an-error/?foo=bar') self.assertEquals(response.status_code, 500) self.assertEquals(len(client.events), 1) event = client.events.pop(0) self.assertEquals(event['class_name'], 'ValueError') self.assertEquals(event['level'], logging.ERROR) self.assertEquals(event['message'], 'hello world') self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error') Add url test for Flask
import logging from flask import Flask from raven.base import Client from raven.contrib.flask import Sentry from unittest2 import TestCase class TempStoreClient(Client): def __init__(self, *args, **kwargs): self.events = [] super(TempStoreClient, self).__init__(*args, **kwargs) def send(self, **kwargs): self.events.append(kwargs) def create_app(): app = Flask(__name__) @app.route('/an-error/', methods=['GET', 'POST']) def an_error(): raise ValueError('hello world') return app class FlaskTest(TestCase): def setUp(self): self.app = create_app() self.client = self.app.test_client() def test_error_handler(self): client = TempStoreClient() sentry = Sentry(self.app, client=client) response = self.client.get('/an-error/?foo=bar') self.assertEquals(response.status_code, 500) self.assertEquals(len(client.events), 1) event = client.events.pop(0) self.assertEquals(event['class_name'], 'ValueError') self.assertEquals(event['level'], logging.ERROR) self.assertEquals(event['message'], 'hello world') self.assertEquals(event['url'], 'http://localhost/an-error/?foo=bar') self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
<commit_before>import logging from flask import Flask from raven.base import Client from raven.contrib.flask import Sentry from unittest2 import TestCase class TempStoreClient(Client): def __init__(self, *args, **kwargs): self.events = [] super(TempStoreClient, self).__init__(*args, **kwargs) def send(self, **kwargs): self.events.append(kwargs) def create_app(): app = Flask(__name__) @app.route('/an-error/', methods=['GET', 'POST']) def an_error(): raise ValueError('hello world') return app class FlaskTest(TestCase): def setUp(self): self.app = create_app() self.client = self.app.test_client() def test_error_handler(self): client = TempStoreClient() sentry = Sentry(self.app, client=client) response = self.client.get('/an-error/?foo=bar') self.assertEquals(response.status_code, 500) self.assertEquals(len(client.events), 1) event = client.events.pop(0) self.assertEquals(event['class_name'], 'ValueError') self.assertEquals(event['level'], logging.ERROR) self.assertEquals(event['message'], 'hello world') self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error') <commit_msg>Add url test for Flask<commit_after>
import logging from flask import Flask from raven.base import Client from raven.contrib.flask import Sentry from unittest2 import TestCase class TempStoreClient(Client): def __init__(self, *args, **kwargs): self.events = [] super(TempStoreClient, self).__init__(*args, **kwargs) def send(self, **kwargs): self.events.append(kwargs) def create_app(): app = Flask(__name__) @app.route('/an-error/', methods=['GET', 'POST']) def an_error(): raise ValueError('hello world') return app class FlaskTest(TestCase): def setUp(self): self.app = create_app() self.client = self.app.test_client() def test_error_handler(self): client = TempStoreClient() sentry = Sentry(self.app, client=client) response = self.client.get('/an-error/?foo=bar') self.assertEquals(response.status_code, 500) self.assertEquals(len(client.events), 1) event = client.events.pop(0) self.assertEquals(event['class_name'], 'ValueError') self.assertEquals(event['level'], logging.ERROR) self.assertEquals(event['message'], 'hello world') self.assertEquals(event['url'], 'http://localhost/an-error/?foo=bar') self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
import logging from flask import Flask from raven.base import Client from raven.contrib.flask import Sentry from unittest2 import TestCase class TempStoreClient(Client): def __init__(self, *args, **kwargs): self.events = [] super(TempStoreClient, self).__init__(*args, **kwargs) def send(self, **kwargs): self.events.append(kwargs) def create_app(): app = Flask(__name__) @app.route('/an-error/', methods=['GET', 'POST']) def an_error(): raise ValueError('hello world') return app class FlaskTest(TestCase): def setUp(self): self.app = create_app() self.client = self.app.test_client() def test_error_handler(self): client = TempStoreClient() sentry = Sentry(self.app, client=client) response = self.client.get('/an-error/?foo=bar') self.assertEquals(response.status_code, 500) self.assertEquals(len(client.events), 1) event = client.events.pop(0) self.assertEquals(event['class_name'], 'ValueError') self.assertEquals(event['level'], logging.ERROR) self.assertEquals(event['message'], 'hello world') self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error') Add url test for Flaskimport logging from flask import Flask from raven.base import Client from raven.contrib.flask import Sentry from unittest2 import TestCase class TempStoreClient(Client): def __init__(self, *args, **kwargs): self.events = [] super(TempStoreClient, self).__init__(*args, **kwargs) def send(self, **kwargs): self.events.append(kwargs) def create_app(): app = Flask(__name__) @app.route('/an-error/', methods=['GET', 'POST']) def an_error(): raise ValueError('hello world') return app class FlaskTest(TestCase): def setUp(self): self.app = create_app() self.client = self.app.test_client() def test_error_handler(self): client = TempStoreClient() sentry = Sentry(self.app, client=client) response = self.client.get('/an-error/?foo=bar') self.assertEquals(response.status_code, 500) self.assertEquals(len(client.events), 1) event = client.events.pop(0) self.assertEquals(event['class_name'], 'ValueError') self.assertEquals(event['level'], logging.ERROR) self.assertEquals(event['message'], 'hello world') self.assertEquals(event['url'], 'http://localhost/an-error/?foo=bar') self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
<commit_before>import logging from flask import Flask from raven.base import Client from raven.contrib.flask import Sentry from unittest2 import TestCase class TempStoreClient(Client): def __init__(self, *args, **kwargs): self.events = [] super(TempStoreClient, self).__init__(*args, **kwargs) def send(self, **kwargs): self.events.append(kwargs) def create_app(): app = Flask(__name__) @app.route('/an-error/', methods=['GET', 'POST']) def an_error(): raise ValueError('hello world') return app class FlaskTest(TestCase): def setUp(self): self.app = create_app() self.client = self.app.test_client() def test_error_handler(self): client = TempStoreClient() sentry = Sentry(self.app, client=client) response = self.client.get('/an-error/?foo=bar') self.assertEquals(response.status_code, 500) self.assertEquals(len(client.events), 1) event = client.events.pop(0) self.assertEquals(event['class_name'], 'ValueError') self.assertEquals(event['level'], logging.ERROR) self.assertEquals(event['message'], 'hello world') self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error') <commit_msg>Add url test for Flask<commit_after>import logging from flask import Flask from raven.base import Client from raven.contrib.flask import Sentry from unittest2 import TestCase class TempStoreClient(Client): def __init__(self, *args, **kwargs): self.events = [] super(TempStoreClient, self).__init__(*args, **kwargs) def send(self, **kwargs): self.events.append(kwargs) def create_app(): app = Flask(__name__) @app.route('/an-error/', methods=['GET', 'POST']) def an_error(): raise ValueError('hello world') return app class FlaskTest(TestCase): def setUp(self): self.app = create_app() self.client = self.app.test_client() def test_error_handler(self): client = TempStoreClient() sentry = Sentry(self.app, client=client) response = self.client.get('/an-error/?foo=bar') self.assertEquals(response.status_code, 500) self.assertEquals(len(client.events), 1) event = client.events.pop(0) self.assertEquals(event['class_name'], 'ValueError') self.assertEquals(event['level'], logging.ERROR) self.assertEquals(event['message'], 'hello world') self.assertEquals(event['url'], 'http://localhost/an-error/?foo=bar') self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
4ccc5ea6cf25adb029f5e08cc0675e2b8415abdf
LayerView.py
LayerView.py
from UM.View.View import View from UM.View.Renderer import Renderer from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator from UM.Resources import Resources class LayerView(View): def __init__(self): super().__init__() self._material = None def beginRendering(self): scene = self.getController().getScene() renderer = self.getRenderer() if not self._material: self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'color.frag')) self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0]) for node in DepthFirstIterator(scene.getRoot()): if not node.render(renderer): if node.getMeshData() and node.isVisible(): try: layerData = node.getMeshData().layerData except AttributeError: continue renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLineLoop) def endRendering(self): pass
from UM.View.View import View from UM.View.Renderer import Renderer from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator from UM.Resources import Resources class LayerView(View): def __init__(self): super().__init__() self._material = None def beginRendering(self): scene = self.getController().getScene() renderer = self.getRenderer() if not self._material: self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'vertexcolor.frag')) self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0]) for node in DepthFirstIterator(scene.getRoot()): if not node.render(renderer): if node.getMeshData() and node.isVisible(): try: layerData = node.getMeshData().layerData except AttributeError: continue renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLines) def endRendering(self): pass
Support colours for rendering the layer view
Support colours for rendering the layer view
Python
agpl-3.0
markwal/Cura,DeskboxBrazil/Cura,ad1217/Cura,Curahelper/Cura,senttech/Cura,derekhe/Cura,fxtentacle/Cura,ynotstartups/Wanhao,ad1217/Cura,markwal/Cura,fxtentacle/Cura,Curahelper/Cura,quillford/Cura,hmflash/Cura,ynotstartups/Wanhao,fieldOfView/Cura,hmflash/Cura,totalretribution/Cura,quillford/Cura,lo0ol/Ultimaker-Cura,lo0ol/Ultimaker-Cura,senttech/Cura,fieldOfView/Cura,bq/Ultimaker-Cura,derekhe/Cura,DeskboxBrazil/Cura,bq/Ultimaker-Cura,totalretribution/Cura
from UM.View.View import View from UM.View.Renderer import Renderer from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator from UM.Resources import Resources class LayerView(View): def __init__(self): super().__init__() self._material = None def beginRendering(self): scene = self.getController().getScene() renderer = self.getRenderer() if not self._material: self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'color.frag')) self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0]) for node in DepthFirstIterator(scene.getRoot()): if not node.render(renderer): if node.getMeshData() and node.isVisible(): try: layerData = node.getMeshData().layerData except AttributeError: continue renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLineLoop) def endRendering(self): pass Support colours for rendering the layer view
from UM.View.View import View from UM.View.Renderer import Renderer from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator from UM.Resources import Resources class LayerView(View): def __init__(self): super().__init__() self._material = None def beginRendering(self): scene = self.getController().getScene() renderer = self.getRenderer() if not self._material: self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'vertexcolor.frag')) self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0]) for node in DepthFirstIterator(scene.getRoot()): if not node.render(renderer): if node.getMeshData() and node.isVisible(): try: layerData = node.getMeshData().layerData except AttributeError: continue renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLines) def endRendering(self): pass
<commit_before>from UM.View.View import View from UM.View.Renderer import Renderer from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator from UM.Resources import Resources class LayerView(View): def __init__(self): super().__init__() self._material = None def beginRendering(self): scene = self.getController().getScene() renderer = self.getRenderer() if not self._material: self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'color.frag')) self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0]) for node in DepthFirstIterator(scene.getRoot()): if not node.render(renderer): if node.getMeshData() and node.isVisible(): try: layerData = node.getMeshData().layerData except AttributeError: continue renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLineLoop) def endRendering(self): pass <commit_msg>Support colours for rendering the layer view<commit_after>
from UM.View.View import View from UM.View.Renderer import Renderer from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator from UM.Resources import Resources class LayerView(View): def __init__(self): super().__init__() self._material = None def beginRendering(self): scene = self.getController().getScene() renderer = self.getRenderer() if not self._material: self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'vertexcolor.frag')) self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0]) for node in DepthFirstIterator(scene.getRoot()): if not node.render(renderer): if node.getMeshData() and node.isVisible(): try: layerData = node.getMeshData().layerData except AttributeError: continue renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLines) def endRendering(self): pass
from UM.View.View import View from UM.View.Renderer import Renderer from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator from UM.Resources import Resources class LayerView(View): def __init__(self): super().__init__() self._material = None def beginRendering(self): scene = self.getController().getScene() renderer = self.getRenderer() if not self._material: self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'color.frag')) self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0]) for node in DepthFirstIterator(scene.getRoot()): if not node.render(renderer): if node.getMeshData() and node.isVisible(): try: layerData = node.getMeshData().layerData except AttributeError: continue renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLineLoop) def endRendering(self): pass Support colours for rendering the layer viewfrom UM.View.View import View from UM.View.Renderer import Renderer from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator from UM.Resources import Resources class LayerView(View): def __init__(self): super().__init__() self._material = None def beginRendering(self): scene = self.getController().getScene() renderer = self.getRenderer() if not self._material: self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'vertexcolor.frag')) self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0]) for node in DepthFirstIterator(scene.getRoot()): if not node.render(renderer): if node.getMeshData() and node.isVisible(): try: layerData = node.getMeshData().layerData except AttributeError: continue renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLines) def endRendering(self): pass
<commit_before>from UM.View.View import View from UM.View.Renderer import Renderer from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator from UM.Resources import Resources class LayerView(View): def __init__(self): super().__init__() self._material = None def beginRendering(self): scene = self.getController().getScene() renderer = self.getRenderer() if not self._material: self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'color.frag')) self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0]) for node in DepthFirstIterator(scene.getRoot()): if not node.render(renderer): if node.getMeshData() and node.isVisible(): try: layerData = node.getMeshData().layerData except AttributeError: continue renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLineLoop) def endRendering(self): pass <commit_msg>Support colours for rendering the layer view<commit_after>from UM.View.View import View from UM.View.Renderer import Renderer from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator from UM.Resources import Resources class LayerView(View): def __init__(self): super().__init__() self._material = None def beginRendering(self): scene = self.getController().getScene() renderer = self.getRenderer() if not self._material: self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'vertexcolor.frag')) self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0]) for node in DepthFirstIterator(scene.getRoot()): if not node.render(renderer): if node.getMeshData() and node.isVisible(): try: layerData = node.getMeshData().layerData except AttributeError: continue renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLines) def endRendering(self): pass
a3a5d2d6b76a4e903fea232b746b2df8b208ec9e
km3pipe/tests/test_plot.py
km3pipe/tests/test_plot.py
# Filename: test_plot.py # pylint: disable=locally-disabled,C0111,R0904,C0103 import numpy as np from km3pipe.testing import TestCase from km3pipe.plot import bincenters __author__ = "Moritz Lotze" __copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration." __credits__ = [] __license__ = "MIT" __maintainer__ = "Moritz Lotze" __email__ = "mlotze@km3net.de" __status__ = "Development" class TestBins(TestCase): def test_binlims(self): bins = np.linspace(0, 20, 21) assert bincenters(bins).shape[0] == bins.shape[0] - 1
# Filename: test_plot.py # pylint: disable=locally-disabled,C0111,R0904,C0103 import numpy as np from km3pipe.testing import TestCase, patch from km3pipe.plot import bincenters, meshgrid, automeshgrid, diag __author__ = "Moritz Lotze" __copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration." __credits__ = [] __license__ = "MIT" __maintainer__ = "Moritz Lotze" __email__ = "mlotze@km3net.de" __status__ = "Development" class TestBins(TestCase): def test_binlims(self): bins = np.linspace(0, 20, 21) assert bincenters(bins).shape[0] == bins.shape[0] - 1 class TestMeshStuff(TestCase): def test_meshgrid(self): xx, yy = meshgrid(-1, 1, 0.8) assert np.allclose([[-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6]], xx) assert np.allclose([[-1.0, -1.0, -1.0], [-0.2, -0.2, -0.2], [0.6, 0.6, 0.6]], yy) def test_meshgrid_with_y_specs(self): xx, yy = meshgrid(-1, 1, 0.8, -10, 10, 8) assert np.allclose([[-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6]], xx) assert np.allclose([[-10, -10, -10], [-2, -2, -2], [6, 6, 6]], yy) class TestDiag(TestCase): def test_call(self): diag()
Add tests for plot functions
Add tests for plot functions
Python
mit
tamasgal/km3pipe,tamasgal/km3pipe
# Filename: test_plot.py # pylint: disable=locally-disabled,C0111,R0904,C0103 import numpy as np from km3pipe.testing import TestCase from km3pipe.plot import bincenters __author__ = "Moritz Lotze" __copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration." __credits__ = [] __license__ = "MIT" __maintainer__ = "Moritz Lotze" __email__ = "mlotze@km3net.de" __status__ = "Development" class TestBins(TestCase): def test_binlims(self): bins = np.linspace(0, 20, 21) assert bincenters(bins).shape[0] == bins.shape[0] - 1 Add tests for plot functions
# Filename: test_plot.py # pylint: disable=locally-disabled,C0111,R0904,C0103 import numpy as np from km3pipe.testing import TestCase, patch from km3pipe.plot import bincenters, meshgrid, automeshgrid, diag __author__ = "Moritz Lotze" __copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration." __credits__ = [] __license__ = "MIT" __maintainer__ = "Moritz Lotze" __email__ = "mlotze@km3net.de" __status__ = "Development" class TestBins(TestCase): def test_binlims(self): bins = np.linspace(0, 20, 21) assert bincenters(bins).shape[0] == bins.shape[0] - 1 class TestMeshStuff(TestCase): def test_meshgrid(self): xx, yy = meshgrid(-1, 1, 0.8) assert np.allclose([[-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6]], xx) assert np.allclose([[-1.0, -1.0, -1.0], [-0.2, -0.2, -0.2], [0.6, 0.6, 0.6]], yy) def test_meshgrid_with_y_specs(self): xx, yy = meshgrid(-1, 1, 0.8, -10, 10, 8) assert np.allclose([[-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6]], xx) assert np.allclose([[-10, -10, -10], [-2, -2, -2], [6, 6, 6]], yy) class TestDiag(TestCase): def test_call(self): diag()
<commit_before># Filename: test_plot.py # pylint: disable=locally-disabled,C0111,R0904,C0103 import numpy as np from km3pipe.testing import TestCase from km3pipe.plot import bincenters __author__ = "Moritz Lotze" __copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration." __credits__ = [] __license__ = "MIT" __maintainer__ = "Moritz Lotze" __email__ = "mlotze@km3net.de" __status__ = "Development" class TestBins(TestCase): def test_binlims(self): bins = np.linspace(0, 20, 21) assert bincenters(bins).shape[0] == bins.shape[0] - 1 <commit_msg>Add tests for plot functions<commit_after>
# Filename: test_plot.py # pylint: disable=locally-disabled,C0111,R0904,C0103 import numpy as np from km3pipe.testing import TestCase, patch from km3pipe.plot import bincenters, meshgrid, automeshgrid, diag __author__ = "Moritz Lotze" __copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration." __credits__ = [] __license__ = "MIT" __maintainer__ = "Moritz Lotze" __email__ = "mlotze@km3net.de" __status__ = "Development" class TestBins(TestCase): def test_binlims(self): bins = np.linspace(0, 20, 21) assert bincenters(bins).shape[0] == bins.shape[0] - 1 class TestMeshStuff(TestCase): def test_meshgrid(self): xx, yy = meshgrid(-1, 1, 0.8) assert np.allclose([[-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6]], xx) assert np.allclose([[-1.0, -1.0, -1.0], [-0.2, -0.2, -0.2], [0.6, 0.6, 0.6]], yy) def test_meshgrid_with_y_specs(self): xx, yy = meshgrid(-1, 1, 0.8, -10, 10, 8) assert np.allclose([[-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6]], xx) assert np.allclose([[-10, -10, -10], [-2, -2, -2], [6, 6, 6]], yy) class TestDiag(TestCase): def test_call(self): diag()
# Filename: test_plot.py # pylint: disable=locally-disabled,C0111,R0904,C0103 import numpy as np from km3pipe.testing import TestCase from km3pipe.plot import bincenters __author__ = "Moritz Lotze" __copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration." __credits__ = [] __license__ = "MIT" __maintainer__ = "Moritz Lotze" __email__ = "mlotze@km3net.de" __status__ = "Development" class TestBins(TestCase): def test_binlims(self): bins = np.linspace(0, 20, 21) assert bincenters(bins).shape[0] == bins.shape[0] - 1 Add tests for plot functions# Filename: test_plot.py # pylint: disable=locally-disabled,C0111,R0904,C0103 import numpy as np from km3pipe.testing import TestCase, patch from km3pipe.plot import bincenters, meshgrid, automeshgrid, diag __author__ = "Moritz Lotze" __copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration." __credits__ = [] __license__ = "MIT" __maintainer__ = "Moritz Lotze" __email__ = "mlotze@km3net.de" __status__ = "Development" class TestBins(TestCase): def test_binlims(self): bins = np.linspace(0, 20, 21) assert bincenters(bins).shape[0] == bins.shape[0] - 1 class TestMeshStuff(TestCase): def test_meshgrid(self): xx, yy = meshgrid(-1, 1, 0.8) assert np.allclose([[-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6]], xx) assert np.allclose([[-1.0, -1.0, -1.0], [-0.2, -0.2, -0.2], [0.6, 0.6, 0.6]], yy) def test_meshgrid_with_y_specs(self): xx, yy = meshgrid(-1, 1, 0.8, -10, 10, 8) assert np.allclose([[-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6]], xx) assert np.allclose([[-10, -10, -10], [-2, -2, -2], [6, 6, 6]], yy) class TestDiag(TestCase): def test_call(self): diag()
<commit_before># Filename: test_plot.py # pylint: disable=locally-disabled,C0111,R0904,C0103 import numpy as np from km3pipe.testing import TestCase from km3pipe.plot import bincenters __author__ = "Moritz Lotze" __copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration." __credits__ = [] __license__ = "MIT" __maintainer__ = "Moritz Lotze" __email__ = "mlotze@km3net.de" __status__ = "Development" class TestBins(TestCase): def test_binlims(self): bins = np.linspace(0, 20, 21) assert bincenters(bins).shape[0] == bins.shape[0] - 1 <commit_msg>Add tests for plot functions<commit_after># Filename: test_plot.py # pylint: disable=locally-disabled,C0111,R0904,C0103 import numpy as np from km3pipe.testing import TestCase, patch from km3pipe.plot import bincenters, meshgrid, automeshgrid, diag __author__ = "Moritz Lotze" __copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration." __credits__ = [] __license__ = "MIT" __maintainer__ = "Moritz Lotze" __email__ = "mlotze@km3net.de" __status__ = "Development" class TestBins(TestCase): def test_binlims(self): bins = np.linspace(0, 20, 21) assert bincenters(bins).shape[0] == bins.shape[0] - 1 class TestMeshStuff(TestCase): def test_meshgrid(self): xx, yy = meshgrid(-1, 1, 0.8) assert np.allclose([[-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6]], xx) assert np.allclose([[-1.0, -1.0, -1.0], [-0.2, -0.2, -0.2], [0.6, 0.6, 0.6]], yy) def test_meshgrid_with_y_specs(self): xx, yy = meshgrid(-1, 1, 0.8, -10, 10, 8) assert np.allclose([[-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6], [-1.0, -0.2, 0.6]], xx) assert np.allclose([[-10, -10, -10], [-2, -2, -2], [6, 6, 6]], yy) class TestDiag(TestCase): def test_call(self): diag()
ef4c9f6a2e6fc1db01d93d937d24e444b0bb0ede
tests/memory_profiling.py
tests/memory_profiling.py
""" Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector. """ import inspect import sys import time import memory_profiler import vector_test try: from pvectorc import pvector except ImportError: print("No C implementation of PVector available, terminating") sys.exit() PROFILING_DURATION = 2.0 def run_function(fn): stop = time.time() + PROFILING_DURATION while time.time() < stop: fn(pvector) def detect_memory_leak(samples): # Skip the first half to get rid of the build up period and the last since it seems # a little less precise samples = samples[int(len(samples)/2):len(samples)-1] return not samples.count(samples[0]) > len(samples) - 2 def profile_tests(): test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction) if fn[0].startswith('test_')] for name, fn in test_functions: # There are a couple of tests that are not run for the C implementation, skip those fn_args = inspect.getargspec(fn)[0] if 'pvector' in fn_args: print('Executing %s' % name) result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1) assert not detect_memory_leak(result), (name, result) if __name__ == "__main__": profile_tests()
""" Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector. """ import inspect import sys import time import memory_profiler import vector_test try: from pvectorc import pvector except ImportError: print("No C implementation of PVector available, terminating") sys.exit() PROFILING_DURATION = 2.0 def run_function(fn): stop = time.time() + PROFILING_DURATION while time.time() < stop: fn(pvector) def detect_memory_leak(samples): # Skip the first samples to get rid of the build up period and the last sample since it seems # a little less precise rising = 0 for i in range(5, len(samples)-1): if samples[i] < samples[i+1]: rising += 1 return (rising / float(len(samples) - 6)) > 0.2 def profile_tests(): test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction) if fn[0].startswith('test_')] for name, fn in test_functions: # There are a couple of tests that are not run for the C implementation, skip those fn_args = inspect.getargspec(fn)[0] if 'pvector' in fn_args: print('Executing %s' % name) result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1) assert not detect_memory_leak(result), (name, result) if __name__ == "__main__": profile_tests()
Improve memory error detection for less false positives
Improve memory error detection for less false positives
Python
mit
tobgu/pyrsistent,jkbjh/pyrsistent,Futrell/pyrsistent,tobgu/pyrsistent,jml/pyrsistent,jml/pyrsistent,tobgu/pyrsistent,jkbjh/pyrsistent,Futrell/pyrsistent,jkbjh/pyrsistent,Futrell/pyrsistent,jml/pyrsistent
""" Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector. """ import inspect import sys import time import memory_profiler import vector_test try: from pvectorc import pvector except ImportError: print("No C implementation of PVector available, terminating") sys.exit() PROFILING_DURATION = 2.0 def run_function(fn): stop = time.time() + PROFILING_DURATION while time.time() < stop: fn(pvector) def detect_memory_leak(samples): # Skip the first half to get rid of the build up period and the last since it seems # a little less precise samples = samples[int(len(samples)/2):len(samples)-1] return not samples.count(samples[0]) > len(samples) - 2 def profile_tests(): test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction) if fn[0].startswith('test_')] for name, fn in test_functions: # There are a couple of tests that are not run for the C implementation, skip those fn_args = inspect.getargspec(fn)[0] if 'pvector' in fn_args: print('Executing %s' % name) result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1) assert not detect_memory_leak(result), (name, result) if __name__ == "__main__": profile_tests()Improve memory error detection for less false positives
""" Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector. """ import inspect import sys import time import memory_profiler import vector_test try: from pvectorc import pvector except ImportError: print("No C implementation of PVector available, terminating") sys.exit() PROFILING_DURATION = 2.0 def run_function(fn): stop = time.time() + PROFILING_DURATION while time.time() < stop: fn(pvector) def detect_memory_leak(samples): # Skip the first samples to get rid of the build up period and the last sample since it seems # a little less precise rising = 0 for i in range(5, len(samples)-1): if samples[i] < samples[i+1]: rising += 1 return (rising / float(len(samples) - 6)) > 0.2 def profile_tests(): test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction) if fn[0].startswith('test_')] for name, fn in test_functions: # There are a couple of tests that are not run for the C implementation, skip those fn_args = inspect.getargspec(fn)[0] if 'pvector' in fn_args: print('Executing %s' % name) result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1) assert not detect_memory_leak(result), (name, result) if __name__ == "__main__": profile_tests()
<commit_before>""" Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector. """ import inspect import sys import time import memory_profiler import vector_test try: from pvectorc import pvector except ImportError: print("No C implementation of PVector available, terminating") sys.exit() PROFILING_DURATION = 2.0 def run_function(fn): stop = time.time() + PROFILING_DURATION while time.time() < stop: fn(pvector) def detect_memory_leak(samples): # Skip the first half to get rid of the build up period and the last since it seems # a little less precise samples = samples[int(len(samples)/2):len(samples)-1] return not samples.count(samples[0]) > len(samples) - 2 def profile_tests(): test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction) if fn[0].startswith('test_')] for name, fn in test_functions: # There are a couple of tests that are not run for the C implementation, skip those fn_args = inspect.getargspec(fn)[0] if 'pvector' in fn_args: print('Executing %s' % name) result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1) assert not detect_memory_leak(result), (name, result) if __name__ == "__main__": profile_tests()<commit_msg>Improve memory error detection for less false positives<commit_after>
""" Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector. """ import inspect import sys import time import memory_profiler import vector_test try: from pvectorc import pvector except ImportError: print("No C implementation of PVector available, terminating") sys.exit() PROFILING_DURATION = 2.0 def run_function(fn): stop = time.time() + PROFILING_DURATION while time.time() < stop: fn(pvector) def detect_memory_leak(samples): # Skip the first samples to get rid of the build up period and the last sample since it seems # a little less precise rising = 0 for i in range(5, len(samples)-1): if samples[i] < samples[i+1]: rising += 1 return (rising / float(len(samples) - 6)) > 0.2 def profile_tests(): test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction) if fn[0].startswith('test_')] for name, fn in test_functions: # There are a couple of tests that are not run for the C implementation, skip those fn_args = inspect.getargspec(fn)[0] if 'pvector' in fn_args: print('Executing %s' % name) result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1) assert not detect_memory_leak(result), (name, result) if __name__ == "__main__": profile_tests()
""" Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector. """ import inspect import sys import time import memory_profiler import vector_test try: from pvectorc import pvector except ImportError: print("No C implementation of PVector available, terminating") sys.exit() PROFILING_DURATION = 2.0 def run_function(fn): stop = time.time() + PROFILING_DURATION while time.time() < stop: fn(pvector) def detect_memory_leak(samples): # Skip the first half to get rid of the build up period and the last since it seems # a little less precise samples = samples[int(len(samples)/2):len(samples)-1] return not samples.count(samples[0]) > len(samples) - 2 def profile_tests(): test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction) if fn[0].startswith('test_')] for name, fn in test_functions: # There are a couple of tests that are not run for the C implementation, skip those fn_args = inspect.getargspec(fn)[0] if 'pvector' in fn_args: print('Executing %s' % name) result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1) assert not detect_memory_leak(result), (name, result) if __name__ == "__main__": profile_tests()Improve memory error detection for less false positives""" Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector. """ import inspect import sys import time import memory_profiler import vector_test try: from pvectorc import pvector except ImportError: print("No C implementation of PVector available, terminating") sys.exit() PROFILING_DURATION = 2.0 def run_function(fn): stop = time.time() + PROFILING_DURATION while time.time() < stop: fn(pvector) def detect_memory_leak(samples): # Skip the first samples to get rid of the build up period and the last sample since it seems # a little less precise rising = 0 for i in range(5, len(samples)-1): if samples[i] < samples[i+1]: rising += 1 return (rising / float(len(samples) - 6)) > 0.2 def profile_tests(): test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction) if fn[0].startswith('test_')] for name, fn in test_functions: # There are a couple of tests that are not run for the C implementation, skip those fn_args = inspect.getargspec(fn)[0] if 'pvector' in fn_args: print('Executing %s' % name) result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1) assert not detect_memory_leak(result), (name, result) if __name__ == "__main__": profile_tests()
<commit_before>""" Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector. """ import inspect import sys import time import memory_profiler import vector_test try: from pvectorc import pvector except ImportError: print("No C implementation of PVector available, terminating") sys.exit() PROFILING_DURATION = 2.0 def run_function(fn): stop = time.time() + PROFILING_DURATION while time.time() < stop: fn(pvector) def detect_memory_leak(samples): # Skip the first half to get rid of the build up period and the last since it seems # a little less precise samples = samples[int(len(samples)/2):len(samples)-1] return not samples.count(samples[0]) > len(samples) - 2 def profile_tests(): test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction) if fn[0].startswith('test_')] for name, fn in test_functions: # There are a couple of tests that are not run for the C implementation, skip those fn_args = inspect.getargspec(fn)[0] if 'pvector' in fn_args: print('Executing %s' % name) result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1) assert not detect_memory_leak(result), (name, result) if __name__ == "__main__": profile_tests()<commit_msg>Improve memory error detection for less false positives<commit_after>""" Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector. """ import inspect import sys import time import memory_profiler import vector_test try: from pvectorc import pvector except ImportError: print("No C implementation of PVector available, terminating") sys.exit() PROFILING_DURATION = 2.0 def run_function(fn): stop = time.time() + PROFILING_DURATION while time.time() < stop: fn(pvector) def detect_memory_leak(samples): # Skip the first samples to get rid of the build up period and the last sample since it seems # a little less precise rising = 0 for i in range(5, len(samples)-1): if samples[i] < samples[i+1]: rising += 1 return (rising / float(len(samples) - 6)) > 0.2 def profile_tests(): test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction) if fn[0].startswith('test_')] for name, fn in test_functions: # There are a couple of tests that are not run for the C implementation, skip those fn_args = inspect.getargspec(fn)[0] if 'pvector' in fn_args: print('Executing %s' % name) result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1) assert not detect_memory_leak(result), (name, result) if __name__ == "__main__": profile_tests()
ee2db892b4dafa33115779166773e248c17a1b43
kyoto/tests/test_client.py
kyoto/tests/test_client.py
import unittest import kyoto.server import kyoto.tests.dummy import kyoto.client class ServiceTestCase(unittest.TestCase): def setUp(self): self.address = ('localhost', 1337) self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy]) self.server.start() self.service = kyoto.client.Service(self.address, ":dummy") def test_invalid_service_name_type(self): with self.assertRaises(ValueError): service = kyoto.client.Service(self.address, "dummy") def test_sync_request(self): response = self.service.call(":echo", ["hello"]) self.assertEqual(response, "hello?") def test_async_request(self): response = self.service.cast(":echo", ["hello"]) self.assertEqual(response, None) def tearDown(self): self.server.stop()
import unittest import kyoto.server import kyoto.tests.dummy import kyoto.client class ServiceTestCase(unittest.TestCase): def setUp(self): self.address = ('localhost', 1337) self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy]) self.server.start() self.service = kyoto.client.Service(self.address, ":dummy") def test_invalid_module_name_type(self): with self.assertRaises(ValueError): service = kyoto.client.Service(self.address, "dummy") service = kyoto.client.Service(self.address, ":dummy") def test_sync_request(self): response = self.service.call(":echo", ["hello"]) self.assertEqual(response, "hello?") def test_async_request(self): response = self.service.cast(":echo", ["hello"]) self.assertEqual(response, None) def tearDown(self): self.server.stop()
Add valid module name test case
Add valid module name test case
Python
mit
kyoto-project/kyoto
import unittest import kyoto.server import kyoto.tests.dummy import kyoto.client class ServiceTestCase(unittest.TestCase): def setUp(self): self.address = ('localhost', 1337) self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy]) self.server.start() self.service = kyoto.client.Service(self.address, ":dummy") def test_invalid_service_name_type(self): with self.assertRaises(ValueError): service = kyoto.client.Service(self.address, "dummy") def test_sync_request(self): response = self.service.call(":echo", ["hello"]) self.assertEqual(response, "hello?") def test_async_request(self): response = self.service.cast(":echo", ["hello"]) self.assertEqual(response, None) def tearDown(self): self.server.stop() Add valid module name test case
import unittest import kyoto.server import kyoto.tests.dummy import kyoto.client class ServiceTestCase(unittest.TestCase): def setUp(self): self.address = ('localhost', 1337) self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy]) self.server.start() self.service = kyoto.client.Service(self.address, ":dummy") def test_invalid_module_name_type(self): with self.assertRaises(ValueError): service = kyoto.client.Service(self.address, "dummy") service = kyoto.client.Service(self.address, ":dummy") def test_sync_request(self): response = self.service.call(":echo", ["hello"]) self.assertEqual(response, "hello?") def test_async_request(self): response = self.service.cast(":echo", ["hello"]) self.assertEqual(response, None) def tearDown(self): self.server.stop()
<commit_before>import unittest import kyoto.server import kyoto.tests.dummy import kyoto.client class ServiceTestCase(unittest.TestCase): def setUp(self): self.address = ('localhost', 1337) self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy]) self.server.start() self.service = kyoto.client.Service(self.address, ":dummy") def test_invalid_service_name_type(self): with self.assertRaises(ValueError): service = kyoto.client.Service(self.address, "dummy") def test_sync_request(self): response = self.service.call(":echo", ["hello"]) self.assertEqual(response, "hello?") def test_async_request(self): response = self.service.cast(":echo", ["hello"]) self.assertEqual(response, None) def tearDown(self): self.server.stop() <commit_msg>Add valid module name test case<commit_after>
import unittest import kyoto.server import kyoto.tests.dummy import kyoto.client class ServiceTestCase(unittest.TestCase): def setUp(self): self.address = ('localhost', 1337) self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy]) self.server.start() self.service = kyoto.client.Service(self.address, ":dummy") def test_invalid_module_name_type(self): with self.assertRaises(ValueError): service = kyoto.client.Service(self.address, "dummy") service = kyoto.client.Service(self.address, ":dummy") def test_sync_request(self): response = self.service.call(":echo", ["hello"]) self.assertEqual(response, "hello?") def test_async_request(self): response = self.service.cast(":echo", ["hello"]) self.assertEqual(response, None) def tearDown(self): self.server.stop()
import unittest import kyoto.server import kyoto.tests.dummy import kyoto.client class ServiceTestCase(unittest.TestCase): def setUp(self): self.address = ('localhost', 1337) self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy]) self.server.start() self.service = kyoto.client.Service(self.address, ":dummy") def test_invalid_service_name_type(self): with self.assertRaises(ValueError): service = kyoto.client.Service(self.address, "dummy") def test_sync_request(self): response = self.service.call(":echo", ["hello"]) self.assertEqual(response, "hello?") def test_async_request(self): response = self.service.cast(":echo", ["hello"]) self.assertEqual(response, None) def tearDown(self): self.server.stop() Add valid module name test caseimport unittest import kyoto.server import kyoto.tests.dummy import kyoto.client class ServiceTestCase(unittest.TestCase): def setUp(self): self.address = ('localhost', 1337) self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy]) self.server.start() self.service = kyoto.client.Service(self.address, ":dummy") def test_invalid_module_name_type(self): with self.assertRaises(ValueError): service = kyoto.client.Service(self.address, "dummy") service = kyoto.client.Service(self.address, ":dummy") def test_sync_request(self): response = self.service.call(":echo", ["hello"]) self.assertEqual(response, "hello?") def test_async_request(self): response = self.service.cast(":echo", ["hello"]) self.assertEqual(response, None) def tearDown(self): self.server.stop()
<commit_before>import unittest import kyoto.server import kyoto.tests.dummy import kyoto.client class ServiceTestCase(unittest.TestCase): def setUp(self): self.address = ('localhost', 1337) self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy]) self.server.start() self.service = kyoto.client.Service(self.address, ":dummy") def test_invalid_service_name_type(self): with self.assertRaises(ValueError): service = kyoto.client.Service(self.address, "dummy") def test_sync_request(self): response = self.service.call(":echo", ["hello"]) self.assertEqual(response, "hello?") def test_async_request(self): response = self.service.cast(":echo", ["hello"]) self.assertEqual(response, None) def tearDown(self): self.server.stop() <commit_msg>Add valid module name test case<commit_after>import unittest import kyoto.server import kyoto.tests.dummy import kyoto.client class ServiceTestCase(unittest.TestCase): def setUp(self): self.address = ('localhost', 1337) self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy]) self.server.start() self.service = kyoto.client.Service(self.address, ":dummy") def test_invalid_module_name_type(self): with self.assertRaises(ValueError): service = kyoto.client.Service(self.address, "dummy") service = kyoto.client.Service(self.address, ":dummy") def test_sync_request(self): response = self.service.call(":echo", ["hello"]) self.assertEqual(response, "hello?") def test_async_request(self): response = self.service.cast(":echo", ["hello"]) self.assertEqual(response, None) def tearDown(self): self.server.stop()
84304d8c04f59421a76b7c070eb9bdcf58a72567
callbackLoader.py
callbackLoader.py
# -*- coding: latin-1 -*- ''' Created on 16.10.2012 @author: Teemu Pkknen ''' import imp import sys import os from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError def get_callback_module( name ): scriptDir = os.path.dirname(os.path.realpath(__file__)) # Already loaded? try: return sys.modules[name] except KeyError: pass fp = pathname = description = None try: fp, pathname, description = imp.find_module(name, [os.getcwdu(), scriptDir]) return imp.load_module(name, fp, pathname, description) except: return None finally: if fp: fp.close()
# -*- coding: latin-1 -*- ''' Created on 16.10.2012 @author: Teemu Pkknen ''' import imp import sys import os import ntpath from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError def path_leaf(path): head, tail = ntpath.split(path) return tail or ntpath.basename(head) def get_callback_module( name ): scriptDir = os.path.dirname(os.path.realpath(__file__)) callback_module_dir = scriptDir + '/' + ntpath.dirname( name ) callback_module_name = path_leaf( name ) # Already loaded? try: return sys.modules[name] except KeyError: pass fp = pathname = description = None try: fp, pathname, description = imp.find_module(callback_module_name, [callback_module_dir, os.getcwdu(), scriptDir]) return imp.load_module(name, fp, pathname, description) except: return None finally: if fp: fp.close()
Make callback loader take into account directory names in loadable module name
Make callback loader take into account directory names in loadable module name
Python
mit
fire-uta/ir-simulation,fire-uta/ir-simulation
# -*- coding: latin-1 -*- ''' Created on 16.10.2012 @author: Teemu Pkknen ''' import imp import sys import os from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError def get_callback_module( name ): scriptDir = os.path.dirname(os.path.realpath(__file__)) # Already loaded? try: return sys.modules[name] except KeyError: pass fp = pathname = description = None try: fp, pathname, description = imp.find_module(name, [os.getcwdu(), scriptDir]) return imp.load_module(name, fp, pathname, description) except: return None finally: if fp: fp.close() Make callback loader take into account directory names in loadable module name
# -*- coding: latin-1 -*- ''' Created on 16.10.2012 @author: Teemu Pkknen ''' import imp import sys import os import ntpath from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError def path_leaf(path): head, tail = ntpath.split(path) return tail or ntpath.basename(head) def get_callback_module( name ): scriptDir = os.path.dirname(os.path.realpath(__file__)) callback_module_dir = scriptDir + '/' + ntpath.dirname( name ) callback_module_name = path_leaf( name ) # Already loaded? try: return sys.modules[name] except KeyError: pass fp = pathname = description = None try: fp, pathname, description = imp.find_module(callback_module_name, [callback_module_dir, os.getcwdu(), scriptDir]) return imp.load_module(name, fp, pathname, description) except: return None finally: if fp: fp.close()
<commit_before># -*- coding: latin-1 -*- ''' Created on 16.10.2012 @author: Teemu Pkknen ''' import imp import sys import os from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError def get_callback_module( name ): scriptDir = os.path.dirname(os.path.realpath(__file__)) # Already loaded? try: return sys.modules[name] except KeyError: pass fp = pathname = description = None try: fp, pathname, description = imp.find_module(name, [os.getcwdu(), scriptDir]) return imp.load_module(name, fp, pathname, description) except: return None finally: if fp: fp.close() <commit_msg>Make callback loader take into account directory names in loadable module name<commit_after>
# -*- coding: latin-1 -*- ''' Created on 16.10.2012 @author: Teemu Pkknen ''' import imp import sys import os import ntpath from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError def path_leaf(path): head, tail = ntpath.split(path) return tail or ntpath.basename(head) def get_callback_module( name ): scriptDir = os.path.dirname(os.path.realpath(__file__)) callback_module_dir = scriptDir + '/' + ntpath.dirname( name ) callback_module_name = path_leaf( name ) # Already loaded? try: return sys.modules[name] except KeyError: pass fp = pathname = description = None try: fp, pathname, description = imp.find_module(callback_module_name, [callback_module_dir, os.getcwdu(), scriptDir]) return imp.load_module(name, fp, pathname, description) except: return None finally: if fp: fp.close()
# -*- coding: latin-1 -*- ''' Created on 16.10.2012 @author: Teemu Pkknen ''' import imp import sys import os from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError def get_callback_module( name ): scriptDir = os.path.dirname(os.path.realpath(__file__)) # Already loaded? try: return sys.modules[name] except KeyError: pass fp = pathname = description = None try: fp, pathname, description = imp.find_module(name, [os.getcwdu(), scriptDir]) return imp.load_module(name, fp, pathname, description) except: return None finally: if fp: fp.close() Make callback loader take into account directory names in loadable module name# -*- coding: latin-1 -*- ''' Created on 16.10.2012 @author: Teemu Pkknen ''' import imp import sys import os import ntpath from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError def path_leaf(path): head, tail = ntpath.split(path) return tail or ntpath.basename(head) def get_callback_module( name ): scriptDir = os.path.dirname(os.path.realpath(__file__)) callback_module_dir = scriptDir + '/' + ntpath.dirname( name ) callback_module_name = path_leaf( name ) # Already loaded? try: return sys.modules[name] except KeyError: pass fp = pathname = description = None try: fp, pathname, description = imp.find_module(callback_module_name, [callback_module_dir, os.getcwdu(), scriptDir]) return imp.load_module(name, fp, pathname, description) except: return None finally: if fp: fp.close()
<commit_before># -*- coding: latin-1 -*- ''' Created on 16.10.2012 @author: Teemu Pkknen ''' import imp import sys import os from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError def get_callback_module( name ): scriptDir = os.path.dirname(os.path.realpath(__file__)) # Already loaded? try: return sys.modules[name] except KeyError: pass fp = pathname = description = None try: fp, pathname, description = imp.find_module(name, [os.getcwdu(), scriptDir]) return imp.load_module(name, fp, pathname, description) except: return None finally: if fp: fp.close() <commit_msg>Make callback loader take into account directory names in loadable module name<commit_after># -*- coding: latin-1 -*- ''' Created on 16.10.2012 @author: Teemu Pkknen ''' import imp import sys import os import ntpath from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError def path_leaf(path): head, tail = ntpath.split(path) return tail or ntpath.basename(head) def get_callback_module( name ): scriptDir = os.path.dirname(os.path.realpath(__file__)) callback_module_dir = scriptDir + '/' + ntpath.dirname( name ) callback_module_name = path_leaf( name ) # Already loaded? try: return sys.modules[name] except KeyError: pass fp = pathname = description = None try: fp, pathname, description = imp.find_module(callback_module_name, [callback_module_dir, os.getcwdu(), scriptDir]) return imp.load_module(name, fp, pathname, description) except: return None finally: if fp: fp.close()
4a509970cb48b64046f88193efc141344437b151
tests/test_list_struct.py
tests/test_list_struct.py
import pytest from hypothesis import given from hypothesis.strategies import lists, integers, floats, one_of, composite from datatyping.datatyping import validate def test_empty(): assert validate([], []) is None @given(li=lists(integers())) def test_plain(li): assert validate([int], li) is None @given(lst=lists(floats(), min_size=1)) def test_plain_type_error(lst): with pytest.raises(TypeError): validate([int], lst) @given(lst=one_of(lists(integers(), min_size=5), lists(integers(), max_size=3))) def test_list_lengths(lst): with pytest.raises(ValueError): validate([int, int, int, str], lst) @given(lst=lists(lists(integers()))) def test_nested(lst): assert validate([[int]], lst) with pytest.raises(TypeError): validate([int], lst) @composite def heavy_nested_data(draw): return [draw(lists(integers)), draw(floats()), lists(lists(floats()))] @given(lst=heavy_nested_data()) def test_heavy_nested(lst): assert validate([[int], float, [[float]]], lst) is None with pytest.raises(TypeError): assert validate([[str], int, int], lst) with pytest.raises(ValueError): validate([[[float]]], lst)
import pytest from hypothesis import given from hypothesis.strategies import lists, integers, floats, one_of, composite from datatyping.datatyping import validate def test_empty(): assert validate([], []) is None @given(li=lists(integers())) def test_plain(li): assert validate([int], li) is None @given(lst=lists(floats(), min_size=1)) def test_plain_type_error(lst): with pytest.raises(TypeError): validate([int], lst) @given(lst=one_of(lists(integers(), min_size=5), lists(integers(), max_size=3))) def test_list_lengths(lst): with pytest.raises(ValueError): validate([int, int, int, str], lst) @given(lst=lists(lists(integers(), min_size=1), min_size=1)) def test_nested(lst): assert validate([[int]], lst) is None with pytest.raises(TypeError): validate([int], lst) @composite def heavy_nested_data(draw): return [draw(lists(integers(), min_size=1, max_size=3)), draw(floats()), draw(lists(lists(floats(), min_size=1, max_size=3), min_size=1, max_size=3))] @given(lst=heavy_nested_data()) def test_heavy_nested(lst): assert validate([[int], float, [[float]]], lst) is None with pytest.raises(TypeError): validate([[str], int, int], lst) with pytest.raises(TypeError): validate([[[float]]], lst)
Fix up mistakes in tests
Fix up mistakes in tests
Python
mit
Zaab1t/datatyping
import pytest from hypothesis import given from hypothesis.strategies import lists, integers, floats, one_of, composite from datatyping.datatyping import validate def test_empty(): assert validate([], []) is None @given(li=lists(integers())) def test_plain(li): assert validate([int], li) is None @given(lst=lists(floats(), min_size=1)) def test_plain_type_error(lst): with pytest.raises(TypeError): validate([int], lst) @given(lst=one_of(lists(integers(), min_size=5), lists(integers(), max_size=3))) def test_list_lengths(lst): with pytest.raises(ValueError): validate([int, int, int, str], lst) @given(lst=lists(lists(integers()))) def test_nested(lst): assert validate([[int]], lst) with pytest.raises(TypeError): validate([int], lst) @composite def heavy_nested_data(draw): return [draw(lists(integers)), draw(floats()), lists(lists(floats()))] @given(lst=heavy_nested_data()) def test_heavy_nested(lst): assert validate([[int], float, [[float]]], lst) is None with pytest.raises(TypeError): assert validate([[str], int, int], lst) with pytest.raises(ValueError): validate([[[float]]], lst) Fix up mistakes in tests
import pytest from hypothesis import given from hypothesis.strategies import lists, integers, floats, one_of, composite from datatyping.datatyping import validate def test_empty(): assert validate([], []) is None @given(li=lists(integers())) def test_plain(li): assert validate([int], li) is None @given(lst=lists(floats(), min_size=1)) def test_plain_type_error(lst): with pytest.raises(TypeError): validate([int], lst) @given(lst=one_of(lists(integers(), min_size=5), lists(integers(), max_size=3))) def test_list_lengths(lst): with pytest.raises(ValueError): validate([int, int, int, str], lst) @given(lst=lists(lists(integers(), min_size=1), min_size=1)) def test_nested(lst): assert validate([[int]], lst) is None with pytest.raises(TypeError): validate([int], lst) @composite def heavy_nested_data(draw): return [draw(lists(integers(), min_size=1, max_size=3)), draw(floats()), draw(lists(lists(floats(), min_size=1, max_size=3), min_size=1, max_size=3))] @given(lst=heavy_nested_data()) def test_heavy_nested(lst): assert validate([[int], float, [[float]]], lst) is None with pytest.raises(TypeError): validate([[str], int, int], lst) with pytest.raises(TypeError): validate([[[float]]], lst)
<commit_before>import pytest from hypothesis import given from hypothesis.strategies import lists, integers, floats, one_of, composite from datatyping.datatyping import validate def test_empty(): assert validate([], []) is None @given(li=lists(integers())) def test_plain(li): assert validate([int], li) is None @given(lst=lists(floats(), min_size=1)) def test_plain_type_error(lst): with pytest.raises(TypeError): validate([int], lst) @given(lst=one_of(lists(integers(), min_size=5), lists(integers(), max_size=3))) def test_list_lengths(lst): with pytest.raises(ValueError): validate([int, int, int, str], lst) @given(lst=lists(lists(integers()))) def test_nested(lst): assert validate([[int]], lst) with pytest.raises(TypeError): validate([int], lst) @composite def heavy_nested_data(draw): return [draw(lists(integers)), draw(floats()), lists(lists(floats()))] @given(lst=heavy_nested_data()) def test_heavy_nested(lst): assert validate([[int], float, [[float]]], lst) is None with pytest.raises(TypeError): assert validate([[str], int, int], lst) with pytest.raises(ValueError): validate([[[float]]], lst) <commit_msg>Fix up mistakes in tests<commit_after>
import pytest from hypothesis import given from hypothesis.strategies import lists, integers, floats, one_of, composite from datatyping.datatyping import validate def test_empty(): assert validate([], []) is None @given(li=lists(integers())) def test_plain(li): assert validate([int], li) is None @given(lst=lists(floats(), min_size=1)) def test_plain_type_error(lst): with pytest.raises(TypeError): validate([int], lst) @given(lst=one_of(lists(integers(), min_size=5), lists(integers(), max_size=3))) def test_list_lengths(lst): with pytest.raises(ValueError): validate([int, int, int, str], lst) @given(lst=lists(lists(integers(), min_size=1), min_size=1)) def test_nested(lst): assert validate([[int]], lst) is None with pytest.raises(TypeError): validate([int], lst) @composite def heavy_nested_data(draw): return [draw(lists(integers(), min_size=1, max_size=3)), draw(floats()), draw(lists(lists(floats(), min_size=1, max_size=3), min_size=1, max_size=3))] @given(lst=heavy_nested_data()) def test_heavy_nested(lst): assert validate([[int], float, [[float]]], lst) is None with pytest.raises(TypeError): validate([[str], int, int], lst) with pytest.raises(TypeError): validate([[[float]]], lst)
import pytest from hypothesis import given from hypothesis.strategies import lists, integers, floats, one_of, composite from datatyping.datatyping import validate def test_empty(): assert validate([], []) is None @given(li=lists(integers())) def test_plain(li): assert validate([int], li) is None @given(lst=lists(floats(), min_size=1)) def test_plain_type_error(lst): with pytest.raises(TypeError): validate([int], lst) @given(lst=one_of(lists(integers(), min_size=5), lists(integers(), max_size=3))) def test_list_lengths(lst): with pytest.raises(ValueError): validate([int, int, int, str], lst) @given(lst=lists(lists(integers()))) def test_nested(lst): assert validate([[int]], lst) with pytest.raises(TypeError): validate([int], lst) @composite def heavy_nested_data(draw): return [draw(lists(integers)), draw(floats()), lists(lists(floats()))] @given(lst=heavy_nested_data()) def test_heavy_nested(lst): assert validate([[int], float, [[float]]], lst) is None with pytest.raises(TypeError): assert validate([[str], int, int], lst) with pytest.raises(ValueError): validate([[[float]]], lst) Fix up mistakes in testsimport pytest from hypothesis import given from hypothesis.strategies import lists, integers, floats, one_of, composite from datatyping.datatyping import validate def test_empty(): assert validate([], []) is None @given(li=lists(integers())) def test_plain(li): assert validate([int], li) is None @given(lst=lists(floats(), min_size=1)) def test_plain_type_error(lst): with pytest.raises(TypeError): validate([int], lst) @given(lst=one_of(lists(integers(), min_size=5), lists(integers(), max_size=3))) def test_list_lengths(lst): with pytest.raises(ValueError): validate([int, int, int, str], lst) @given(lst=lists(lists(integers(), min_size=1), min_size=1)) def test_nested(lst): assert validate([[int]], lst) is None with pytest.raises(TypeError): validate([int], lst) @composite def heavy_nested_data(draw): return [draw(lists(integers(), min_size=1, max_size=3)), draw(floats()), draw(lists(lists(floats(), min_size=1, max_size=3), min_size=1, max_size=3))] @given(lst=heavy_nested_data()) def test_heavy_nested(lst): assert validate([[int], float, [[float]]], lst) is None with pytest.raises(TypeError): validate([[str], int, int], lst) with pytest.raises(TypeError): validate([[[float]]], lst)
<commit_before>import pytest from hypothesis import given from hypothesis.strategies import lists, integers, floats, one_of, composite from datatyping.datatyping import validate def test_empty(): assert validate([], []) is None @given(li=lists(integers())) def test_plain(li): assert validate([int], li) is None @given(lst=lists(floats(), min_size=1)) def test_plain_type_error(lst): with pytest.raises(TypeError): validate([int], lst) @given(lst=one_of(lists(integers(), min_size=5), lists(integers(), max_size=3))) def test_list_lengths(lst): with pytest.raises(ValueError): validate([int, int, int, str], lst) @given(lst=lists(lists(integers()))) def test_nested(lst): assert validate([[int]], lst) with pytest.raises(TypeError): validate([int], lst) @composite def heavy_nested_data(draw): return [draw(lists(integers)), draw(floats()), lists(lists(floats()))] @given(lst=heavy_nested_data()) def test_heavy_nested(lst): assert validate([[int], float, [[float]]], lst) is None with pytest.raises(TypeError): assert validate([[str], int, int], lst) with pytest.raises(ValueError): validate([[[float]]], lst) <commit_msg>Fix up mistakes in tests<commit_after>import pytest from hypothesis import given from hypothesis.strategies import lists, integers, floats, one_of, composite from datatyping.datatyping import validate def test_empty(): assert validate([], []) is None @given(li=lists(integers())) def test_plain(li): assert validate([int], li) is None @given(lst=lists(floats(), min_size=1)) def test_plain_type_error(lst): with pytest.raises(TypeError): validate([int], lst) @given(lst=one_of(lists(integers(), min_size=5), lists(integers(), max_size=3))) def test_list_lengths(lst): with pytest.raises(ValueError): validate([int, int, int, str], lst) @given(lst=lists(lists(integers(), min_size=1), min_size=1)) def test_nested(lst): assert validate([[int]], lst) is None with pytest.raises(TypeError): validate([int], lst) @composite def heavy_nested_data(draw): return [draw(lists(integers(), min_size=1, max_size=3)), draw(floats()), draw(lists(lists(floats(), min_size=1, max_size=3), min_size=1, max_size=3))] @given(lst=heavy_nested_data()) def test_heavy_nested(lst): assert validate([[int], float, [[float]]], lst) is None with pytest.raises(TypeError): validate([[str], int, int], lst) with pytest.raises(TypeError): validate([[[float]]], lst)
3ce9f6d8537c6b6d0ec5a5e09c5f1f6b7b34699c
troposphere/eventschemas.py
troposphere/eventschemas.py
# Copyright (c) 2012-2019, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 14.1.0 from troposphere import Tags from . import AWSObject class Discoverer(AWSObject): resource_type = "AWS::EventSchemas::Discoverer" props = { "Description": (str, False), "SourceArn": (str, True), "Tags": (Tags, False), } class Registry(AWSObject): resource_type = "AWS::EventSchemas::Registry" props = { "Description": (str, False), "RegistryName": (str, False), "Tags": (Tags, False), } class RegistryPolicy(AWSObject): resource_type = "AWS::EventSchemas::RegistryPolicy" props = { "Policy": (dict, True), "RegistryName": (str, True), "RevisionId": (str, False), } class Schema(AWSObject): resource_type = "AWS::EventSchemas::Schema" props = { "Content": (str, True), "Description": (str, False), "RegistryName": (str, True), "SchemaName": (str, False), "Tags": (Tags, False), "Type": (str, True), }
# Copyright (c) 2012-2021, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 41.0.0 from troposphere import Tags from . import AWSObject from .validators import boolean class Discoverer(AWSObject): resource_type = "AWS::EventSchemas::Discoverer" props = { "CrossAccount": (boolean, False), "Description": (str, False), "SourceArn": (str, True), "Tags": (Tags, False), } class Registry(AWSObject): resource_type = "AWS::EventSchemas::Registry" props = { "Description": (str, False), "RegistryName": (str, False), "Tags": (Tags, False), } class RegistryPolicy(AWSObject): resource_type = "AWS::EventSchemas::RegistryPolicy" props = { "Policy": (dict, True), "RegistryName": (str, True), "RevisionId": (str, False), } class Schema(AWSObject): resource_type = "AWS::EventSchemas::Schema" props = { "Content": (str, True), "Description": (str, False), "RegistryName": (str, True), "SchemaName": (str, False), "Tags": (Tags, False), "Type": (str, True), }
Update EventSchemas per 2021-09-02 changes
Update EventSchemas per 2021-09-02 changes
Python
bsd-2-clause
cloudtools/troposphere,cloudtools/troposphere
# Copyright (c) 2012-2019, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 14.1.0 from troposphere import Tags from . import AWSObject class Discoverer(AWSObject): resource_type = "AWS::EventSchemas::Discoverer" props = { "Description": (str, False), "SourceArn": (str, True), "Tags": (Tags, False), } class Registry(AWSObject): resource_type = "AWS::EventSchemas::Registry" props = { "Description": (str, False), "RegistryName": (str, False), "Tags": (Tags, False), } class RegistryPolicy(AWSObject): resource_type = "AWS::EventSchemas::RegistryPolicy" props = { "Policy": (dict, True), "RegistryName": (str, True), "RevisionId": (str, False), } class Schema(AWSObject): resource_type = "AWS::EventSchemas::Schema" props = { "Content": (str, True), "Description": (str, False), "RegistryName": (str, True), "SchemaName": (str, False), "Tags": (Tags, False), "Type": (str, True), } Update EventSchemas per 2021-09-02 changes
# Copyright (c) 2012-2021, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 41.0.0 from troposphere import Tags from . import AWSObject from .validators import boolean class Discoverer(AWSObject): resource_type = "AWS::EventSchemas::Discoverer" props = { "CrossAccount": (boolean, False), "Description": (str, False), "SourceArn": (str, True), "Tags": (Tags, False), } class Registry(AWSObject): resource_type = "AWS::EventSchemas::Registry" props = { "Description": (str, False), "RegistryName": (str, False), "Tags": (Tags, False), } class RegistryPolicy(AWSObject): resource_type = "AWS::EventSchemas::RegistryPolicy" props = { "Policy": (dict, True), "RegistryName": (str, True), "RevisionId": (str, False), } class Schema(AWSObject): resource_type = "AWS::EventSchemas::Schema" props = { "Content": (str, True), "Description": (str, False), "RegistryName": (str, True), "SchemaName": (str, False), "Tags": (Tags, False), "Type": (str, True), }
<commit_before># Copyright (c) 2012-2019, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 14.1.0 from troposphere import Tags from . import AWSObject class Discoverer(AWSObject): resource_type = "AWS::EventSchemas::Discoverer" props = { "Description": (str, False), "SourceArn": (str, True), "Tags": (Tags, False), } class Registry(AWSObject): resource_type = "AWS::EventSchemas::Registry" props = { "Description": (str, False), "RegistryName": (str, False), "Tags": (Tags, False), } class RegistryPolicy(AWSObject): resource_type = "AWS::EventSchemas::RegistryPolicy" props = { "Policy": (dict, True), "RegistryName": (str, True), "RevisionId": (str, False), } class Schema(AWSObject): resource_type = "AWS::EventSchemas::Schema" props = { "Content": (str, True), "Description": (str, False), "RegistryName": (str, True), "SchemaName": (str, False), "Tags": (Tags, False), "Type": (str, True), } <commit_msg>Update EventSchemas per 2021-09-02 changes<commit_after>
# Copyright (c) 2012-2021, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 41.0.0 from troposphere import Tags from . import AWSObject from .validators import boolean class Discoverer(AWSObject): resource_type = "AWS::EventSchemas::Discoverer" props = { "CrossAccount": (boolean, False), "Description": (str, False), "SourceArn": (str, True), "Tags": (Tags, False), } class Registry(AWSObject): resource_type = "AWS::EventSchemas::Registry" props = { "Description": (str, False), "RegistryName": (str, False), "Tags": (Tags, False), } class RegistryPolicy(AWSObject): resource_type = "AWS::EventSchemas::RegistryPolicy" props = { "Policy": (dict, True), "RegistryName": (str, True), "RevisionId": (str, False), } class Schema(AWSObject): resource_type = "AWS::EventSchemas::Schema" props = { "Content": (str, True), "Description": (str, False), "RegistryName": (str, True), "SchemaName": (str, False), "Tags": (Tags, False), "Type": (str, True), }
# Copyright (c) 2012-2019, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 14.1.0 from troposphere import Tags from . import AWSObject class Discoverer(AWSObject): resource_type = "AWS::EventSchemas::Discoverer" props = { "Description": (str, False), "SourceArn": (str, True), "Tags": (Tags, False), } class Registry(AWSObject): resource_type = "AWS::EventSchemas::Registry" props = { "Description": (str, False), "RegistryName": (str, False), "Tags": (Tags, False), } class RegistryPolicy(AWSObject): resource_type = "AWS::EventSchemas::RegistryPolicy" props = { "Policy": (dict, True), "RegistryName": (str, True), "RevisionId": (str, False), } class Schema(AWSObject): resource_type = "AWS::EventSchemas::Schema" props = { "Content": (str, True), "Description": (str, False), "RegistryName": (str, True), "SchemaName": (str, False), "Tags": (Tags, False), "Type": (str, True), } Update EventSchemas per 2021-09-02 changes# Copyright (c) 2012-2021, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 41.0.0 from troposphere import Tags from . import AWSObject from .validators import boolean class Discoverer(AWSObject): resource_type = "AWS::EventSchemas::Discoverer" props = { "CrossAccount": (boolean, False), "Description": (str, False), "SourceArn": (str, True), "Tags": (Tags, False), } class Registry(AWSObject): resource_type = "AWS::EventSchemas::Registry" props = { "Description": (str, False), "RegistryName": (str, False), "Tags": (Tags, False), } class RegistryPolicy(AWSObject): resource_type = "AWS::EventSchemas::RegistryPolicy" props = { "Policy": (dict, True), "RegistryName": (str, True), "RevisionId": (str, False), } class Schema(AWSObject): resource_type = "AWS::EventSchemas::Schema" props = { "Content": (str, True), "Description": (str, False), "RegistryName": (str, True), "SchemaName": (str, False), "Tags": (Tags, False), "Type": (str, True), }
<commit_before># Copyright (c) 2012-2019, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 14.1.0 from troposphere import Tags from . import AWSObject class Discoverer(AWSObject): resource_type = "AWS::EventSchemas::Discoverer" props = { "Description": (str, False), "SourceArn": (str, True), "Tags": (Tags, False), } class Registry(AWSObject): resource_type = "AWS::EventSchemas::Registry" props = { "Description": (str, False), "RegistryName": (str, False), "Tags": (Tags, False), } class RegistryPolicy(AWSObject): resource_type = "AWS::EventSchemas::RegistryPolicy" props = { "Policy": (dict, True), "RegistryName": (str, True), "RevisionId": (str, False), } class Schema(AWSObject): resource_type = "AWS::EventSchemas::Schema" props = { "Content": (str, True), "Description": (str, False), "RegistryName": (str, True), "SchemaName": (str, False), "Tags": (Tags, False), "Type": (str, True), } <commit_msg>Update EventSchemas per 2021-09-02 changes<commit_after># Copyright (c) 2012-2021, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 41.0.0 from troposphere import Tags from . import AWSObject from .validators import boolean class Discoverer(AWSObject): resource_type = "AWS::EventSchemas::Discoverer" props = { "CrossAccount": (boolean, False), "Description": (str, False), "SourceArn": (str, True), "Tags": (Tags, False), } class Registry(AWSObject): resource_type = "AWS::EventSchemas::Registry" props = { "Description": (str, False), "RegistryName": (str, False), "Tags": (Tags, False), } class RegistryPolicy(AWSObject): resource_type = "AWS::EventSchemas::RegistryPolicy" props = { "Policy": (dict, True), "RegistryName": (str, True), "RevisionId": (str, False), } class Schema(AWSObject): resource_type = "AWS::EventSchemas::Schema" props = { "Content": (str, True), "Description": (str, False), "RegistryName": (str, True), "SchemaName": (str, False), "Tags": (Tags, False), "Type": (str, True), }
76ed79593a832c1cf85615d21b31f18f2c7adebf
yanico/session/__init__.py
yanico/session/__init__.py
# Copyright 2015-2016 Masayuki Yamamoto # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handle nicovideo.jp user_session.""" import pkg_resources class LoaderNotFoundError(Exception): """Session loader is not found.""" class UserSessionNotFoundError(Exception): """Profile exists, but user_session is not found.""" def load(ltype, profile): for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype): load_func = entry.load() return load_func(profile) raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
# Copyright 2015-2016 Masayuki Yamamoto # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handle nicovideo.jp user_session.""" import pkg_resources class LoaderNotFoundError(Exception): """Session loader is not found.""" class UserSessionNotFoundError(Exception): """Profile exists, but user_session is not found.""" def load(ltype, profile): """Return nicovideo.jp user session string. Args: ltype (str): loader type profile (str): file path for profile Returns: str: user session Raises: LoaderNotFoundError Error from loader """ for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype): load_func = entry.load() return load_func(profile) raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
Add docstring into load function
Add docstring into load function Follow to Google style.
Python
apache-2.0
ma8ma/yanico
# Copyright 2015-2016 Masayuki Yamamoto # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handle nicovideo.jp user_session.""" import pkg_resources class LoaderNotFoundError(Exception): """Session loader is not found.""" class UserSessionNotFoundError(Exception): """Profile exists, but user_session is not found.""" def load(ltype, profile): for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype): load_func = entry.load() return load_func(profile) raise LoaderNotFoundError('{} loader is not found.'.format(ltype)) Add docstring into load function Follow to Google style.
# Copyright 2015-2016 Masayuki Yamamoto # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handle nicovideo.jp user_session.""" import pkg_resources class LoaderNotFoundError(Exception): """Session loader is not found.""" class UserSessionNotFoundError(Exception): """Profile exists, but user_session is not found.""" def load(ltype, profile): """Return nicovideo.jp user session string. Args: ltype (str): loader type profile (str): file path for profile Returns: str: user session Raises: LoaderNotFoundError Error from loader """ for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype): load_func = entry.load() return load_func(profile) raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
<commit_before># Copyright 2015-2016 Masayuki Yamamoto # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handle nicovideo.jp user_session.""" import pkg_resources class LoaderNotFoundError(Exception): """Session loader is not found.""" class UserSessionNotFoundError(Exception): """Profile exists, but user_session is not found.""" def load(ltype, profile): for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype): load_func = entry.load() return load_func(profile) raise LoaderNotFoundError('{} loader is not found.'.format(ltype)) <commit_msg>Add docstring into load function Follow to Google style.<commit_after>
# Copyright 2015-2016 Masayuki Yamamoto # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handle nicovideo.jp user_session.""" import pkg_resources class LoaderNotFoundError(Exception): """Session loader is not found.""" class UserSessionNotFoundError(Exception): """Profile exists, but user_session is not found.""" def load(ltype, profile): """Return nicovideo.jp user session string. Args: ltype (str): loader type profile (str): file path for profile Returns: str: user session Raises: LoaderNotFoundError Error from loader """ for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype): load_func = entry.load() return load_func(profile) raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
# Copyright 2015-2016 Masayuki Yamamoto # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handle nicovideo.jp user_session.""" import pkg_resources class LoaderNotFoundError(Exception): """Session loader is not found.""" class UserSessionNotFoundError(Exception): """Profile exists, but user_session is not found.""" def load(ltype, profile): for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype): load_func = entry.load() return load_func(profile) raise LoaderNotFoundError('{} loader is not found.'.format(ltype)) Add docstring into load function Follow to Google style.# Copyright 2015-2016 Masayuki Yamamoto # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handle nicovideo.jp user_session.""" import pkg_resources class LoaderNotFoundError(Exception): """Session loader is not found.""" class UserSessionNotFoundError(Exception): """Profile exists, but user_session is not found.""" def load(ltype, profile): """Return nicovideo.jp user session string. Args: ltype (str): loader type profile (str): file path for profile Returns: str: user session Raises: LoaderNotFoundError Error from loader """ for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype): load_func = entry.load() return load_func(profile) raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
<commit_before># Copyright 2015-2016 Masayuki Yamamoto # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handle nicovideo.jp user_session.""" import pkg_resources class LoaderNotFoundError(Exception): """Session loader is not found.""" class UserSessionNotFoundError(Exception): """Profile exists, but user_session is not found.""" def load(ltype, profile): for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype): load_func = entry.load() return load_func(profile) raise LoaderNotFoundError('{} loader is not found.'.format(ltype)) <commit_msg>Add docstring into load function Follow to Google style.<commit_after># Copyright 2015-2016 Masayuki Yamamoto # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Handle nicovideo.jp user_session.""" import pkg_resources class LoaderNotFoundError(Exception): """Session loader is not found.""" class UserSessionNotFoundError(Exception): """Profile exists, but user_session is not found.""" def load(ltype, profile): """Return nicovideo.jp user session string. Args: ltype (str): loader type profile (str): file path for profile Returns: str: user session Raises: LoaderNotFoundError Error from loader """ for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype): load_func = entry.load() return load_func(profile) raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
086e2bb85d0076c55dff886154664dc7179561fa
utils/summary_downloader.py
utils/summary_downloader.py
#!/usr/bin/env python # -*- coding: utf-8 -*- class SummaryDownloader(): # base url for official schedule json page SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule" # url template for official json gamefeed page JSON_GAME_FEED_URL_TEMPLATE = ( "http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
#!/usr/bin/env python # -*- coding: utf-8 -*- from dateutil.parser import parse from dateutil.relativedelta import DAILY from dateutil.rrule import rrule class SummaryDownloader(): # base url for official schedule json page SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule" # url template for official json gamefeed page JSON_GAME_FEED_URL_TEMPLATE = ( "http://statsapi.web.nhl.com/api/v1/game/%s/feed/live") MAX_DOWNLOAD_WORKERS = 8 def __init__(self, tgt_dir, date, to_date='', threads=0): self.date = parse(date) if to_date: self.to_date = parse(to_date) else: self.to_date = self.date # preparing list of dates to download summary data for self.game_dates = list( rrule(DAILY, dtstart=self.date, until=self.to_date)) print(self.game_dates) if __name__ == '__main__': date = "1997/04/20" d = SummaryDownloader(r"d:\tmp", date)
Add constructor to summary downloader
Add constructor to summary downloader
Python
mit
leaffan/pynhldb
#!/usr/bin/env python # -*- coding: utf-8 -*- class SummaryDownloader(): # base url for official schedule json page SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule" # url template for official json gamefeed page JSON_GAME_FEED_URL_TEMPLATE = ( "http://statsapi.web.nhl.com/api/v1/game/%s/feed/live") Add constructor to summary downloader
#!/usr/bin/env python # -*- coding: utf-8 -*- from dateutil.parser import parse from dateutil.relativedelta import DAILY from dateutil.rrule import rrule class SummaryDownloader(): # base url for official schedule json page SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule" # url template for official json gamefeed page JSON_GAME_FEED_URL_TEMPLATE = ( "http://statsapi.web.nhl.com/api/v1/game/%s/feed/live") MAX_DOWNLOAD_WORKERS = 8 def __init__(self, tgt_dir, date, to_date='', threads=0): self.date = parse(date) if to_date: self.to_date = parse(to_date) else: self.to_date = self.date # preparing list of dates to download summary data for self.game_dates = list( rrule(DAILY, dtstart=self.date, until=self.to_date)) print(self.game_dates) if __name__ == '__main__': date = "1997/04/20" d = SummaryDownloader(r"d:\tmp", date)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- class SummaryDownloader(): # base url for official schedule json page SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule" # url template for official json gamefeed page JSON_GAME_FEED_URL_TEMPLATE = ( "http://statsapi.web.nhl.com/api/v1/game/%s/feed/live") <commit_msg>Add constructor to summary downloader<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from dateutil.parser import parse from dateutil.relativedelta import DAILY from dateutil.rrule import rrule class SummaryDownloader(): # base url for official schedule json page SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule" # url template for official json gamefeed page JSON_GAME_FEED_URL_TEMPLATE = ( "http://statsapi.web.nhl.com/api/v1/game/%s/feed/live") MAX_DOWNLOAD_WORKERS = 8 def __init__(self, tgt_dir, date, to_date='', threads=0): self.date = parse(date) if to_date: self.to_date = parse(to_date) else: self.to_date = self.date # preparing list of dates to download summary data for self.game_dates = list( rrule(DAILY, dtstart=self.date, until=self.to_date)) print(self.game_dates) if __name__ == '__main__': date = "1997/04/20" d = SummaryDownloader(r"d:\tmp", date)
#!/usr/bin/env python # -*- coding: utf-8 -*- class SummaryDownloader(): # base url for official schedule json page SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule" # url template for official json gamefeed page JSON_GAME_FEED_URL_TEMPLATE = ( "http://statsapi.web.nhl.com/api/v1/game/%s/feed/live") Add constructor to summary downloader#!/usr/bin/env python # -*- coding: utf-8 -*- from dateutil.parser import parse from dateutil.relativedelta import DAILY from dateutil.rrule import rrule class SummaryDownloader(): # base url for official schedule json page SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule" # url template for official json gamefeed page JSON_GAME_FEED_URL_TEMPLATE = ( "http://statsapi.web.nhl.com/api/v1/game/%s/feed/live") MAX_DOWNLOAD_WORKERS = 8 def __init__(self, tgt_dir, date, to_date='', threads=0): self.date = parse(date) if to_date: self.to_date = parse(to_date) else: self.to_date = self.date # preparing list of dates to download summary data for self.game_dates = list( rrule(DAILY, dtstart=self.date, until=self.to_date)) print(self.game_dates) if __name__ == '__main__': date = "1997/04/20" d = SummaryDownloader(r"d:\tmp", date)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- class SummaryDownloader(): # base url for official schedule json page SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule" # url template for official json gamefeed page JSON_GAME_FEED_URL_TEMPLATE = ( "http://statsapi.web.nhl.com/api/v1/game/%s/feed/live") <commit_msg>Add constructor to summary downloader<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from dateutil.parser import parse from dateutil.relativedelta import DAILY from dateutil.rrule import rrule class SummaryDownloader(): # base url for official schedule json page SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule" # url template for official json gamefeed page JSON_GAME_FEED_URL_TEMPLATE = ( "http://statsapi.web.nhl.com/api/v1/game/%s/feed/live") MAX_DOWNLOAD_WORKERS = 8 def __init__(self, tgt_dir, date, to_date='', threads=0): self.date = parse(date) if to_date: self.to_date = parse(to_date) else: self.to_date = self.date # preparing list of dates to download summary data for self.game_dates = list( rrule(DAILY, dtstart=self.date, until=self.to_date)) print(self.game_dates) if __name__ == '__main__': date = "1997/04/20" d = SummaryDownloader(r"d:\tmp", date)
10a241938d5469f9da3d7d8a695963ac7cff87b2
website/mosaic/settings_gondor.py
website/mosaic/settings_gondor.py
import os import urlparse from .settings import * DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS if "GONDOR_DATABASE_URL" in os.environ: urlparse.uses_netloc.append("postgres") url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"]) DATABASES = { "default": { "ENGINE": { "postgres": "django.db.backends.postgresql_psycopg2" }[url.scheme], "NAME": url.path[1:], "USER": url.username, "PASSWORD": url.password, "HOST": url.hostname, "PORT": url.port } } MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media") STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static") MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/" FILE_UPLOAD_PERMISSIONS = 0640 SECRET_KEY = '8f89aej9fa89898__fd9er38948934fjsfoijdaf.'
import os import urlparse from .settings import * DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS if "GONDOR_DATABASE_URL" in os.environ: urlparse.uses_netloc.append("postgres") url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"]) DATABASES = { "default": { "ENGINE": { "postgres": "django.db.backends.postgresql_psycopg2" }[url.scheme], "NAME": url.path[1:], "USER": url.username, "PASSWORD": url.password, "HOST": url.hostname, "PORT": url.port } } MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media") STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static") MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/" FILE_UPLOAD_PERMISSIONS = 0640 SECRET_KEY = os.environ.get('DJANGO_SECRET_KEY')
Remove secret key from the gondor settings file
Remove secret key from the gondor settings file
Python
mit
sema/django-2012,sema/django-2012
import os import urlparse from .settings import * DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS if "GONDOR_DATABASE_URL" in os.environ: urlparse.uses_netloc.append("postgres") url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"]) DATABASES = { "default": { "ENGINE": { "postgres": "django.db.backends.postgresql_psycopg2" }[url.scheme], "NAME": url.path[1:], "USER": url.username, "PASSWORD": url.password, "HOST": url.hostname, "PORT": url.port } } MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media") STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static") MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/" FILE_UPLOAD_PERMISSIONS = 0640 SECRET_KEY = '8f89aej9fa89898__fd9er38948934fjsfoijdaf.'Remove secret key from the gondor settings file
import os import urlparse from .settings import * DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS if "GONDOR_DATABASE_URL" in os.environ: urlparse.uses_netloc.append("postgres") url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"]) DATABASES = { "default": { "ENGINE": { "postgres": "django.db.backends.postgresql_psycopg2" }[url.scheme], "NAME": url.path[1:], "USER": url.username, "PASSWORD": url.password, "HOST": url.hostname, "PORT": url.port } } MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media") STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static") MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/" FILE_UPLOAD_PERMISSIONS = 0640 SECRET_KEY = os.environ.get('DJANGO_SECRET_KEY')
<commit_before>import os import urlparse from .settings import * DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS if "GONDOR_DATABASE_URL" in os.environ: urlparse.uses_netloc.append("postgres") url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"]) DATABASES = { "default": { "ENGINE": { "postgres": "django.db.backends.postgresql_psycopg2" }[url.scheme], "NAME": url.path[1:], "USER": url.username, "PASSWORD": url.password, "HOST": url.hostname, "PORT": url.port } } MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media") STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static") MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/" FILE_UPLOAD_PERMISSIONS = 0640 SECRET_KEY = '8f89aej9fa89898__fd9er38948934fjsfoijdaf.'<commit_msg>Remove secret key from the gondor settings file<commit_after>
import os import urlparse from .settings import * DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS if "GONDOR_DATABASE_URL" in os.environ: urlparse.uses_netloc.append("postgres") url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"]) DATABASES = { "default": { "ENGINE": { "postgres": "django.db.backends.postgresql_psycopg2" }[url.scheme], "NAME": url.path[1:], "USER": url.username, "PASSWORD": url.password, "HOST": url.hostname, "PORT": url.port } } MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media") STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static") MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/" FILE_UPLOAD_PERMISSIONS = 0640 SECRET_KEY = os.environ.get('DJANGO_SECRET_KEY')
import os import urlparse from .settings import * DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS if "GONDOR_DATABASE_URL" in os.environ: urlparse.uses_netloc.append("postgres") url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"]) DATABASES = { "default": { "ENGINE": { "postgres": "django.db.backends.postgresql_psycopg2" }[url.scheme], "NAME": url.path[1:], "USER": url.username, "PASSWORD": url.password, "HOST": url.hostname, "PORT": url.port } } MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media") STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static") MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/" FILE_UPLOAD_PERMISSIONS = 0640 SECRET_KEY = '8f89aej9fa89898__fd9er38948934fjsfoijdaf.'Remove secret key from the gondor settings fileimport os import urlparse from .settings import * DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS if "GONDOR_DATABASE_URL" in os.environ: urlparse.uses_netloc.append("postgres") url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"]) DATABASES = { "default": { "ENGINE": { "postgres": "django.db.backends.postgresql_psycopg2" }[url.scheme], "NAME": url.path[1:], "USER": url.username, "PASSWORD": url.password, "HOST": url.hostname, "PORT": url.port } } MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media") STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static") MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/" FILE_UPLOAD_PERMISSIONS = 0640 SECRET_KEY = os.environ.get('DJANGO_SECRET_KEY')
<commit_before>import os import urlparse from .settings import * DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS if "GONDOR_DATABASE_URL" in os.environ: urlparse.uses_netloc.append("postgres") url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"]) DATABASES = { "default": { "ENGINE": { "postgres": "django.db.backends.postgresql_psycopg2" }[url.scheme], "NAME": url.path[1:], "USER": url.username, "PASSWORD": url.password, "HOST": url.hostname, "PORT": url.port } } MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media") STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static") MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/" FILE_UPLOAD_PERMISSIONS = 0640 SECRET_KEY = '8f89aej9fa89898__fd9er38948934fjsfoijdaf.'<commit_msg>Remove secret key from the gondor settings file<commit_after>import os import urlparse from .settings import * DEBUG = True TEMPLATE_DEBUG = DEBUG ADMINS = ( # ('Your Name', 'your_email@example.com'), ) MANAGERS = ADMINS if "GONDOR_DATABASE_URL" in os.environ: urlparse.uses_netloc.append("postgres") url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"]) DATABASES = { "default": { "ENGINE": { "postgres": "django.db.backends.postgresql_psycopg2" }[url.scheme], "NAME": url.path[1:], "USER": url.username, "PASSWORD": url.password, "HOST": url.hostname, "PORT": url.port } } MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media") STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static") MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/" FILE_UPLOAD_PERMISSIONS = 0640 SECRET_KEY = os.environ.get('DJANGO_SECRET_KEY')
e8e7bb5b7f063cc48b761fc17ef8f2264a17a2ce
dthm4kaiako/config/__init__.py
dthm4kaiako/config/__init__.py
"""Configuration for Django system.""" __version__ = "0.17.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
"""Configuration for Django system.""" __version__ = "0.17.3" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
Increment version number to 0.17.3
Increment version number to 0.17.3
Python
mit
uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers
"""Configuration for Django system.""" __version__ = "0.17.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) Increment version number to 0.17.3
"""Configuration for Django system.""" __version__ = "0.17.3" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
<commit_before>"""Configuration for Django system.""" __version__ = "0.17.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) <commit_msg>Increment version number to 0.17.3<commit_after>
"""Configuration for Django system.""" __version__ = "0.17.3" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
"""Configuration for Django system.""" __version__ = "0.17.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) Increment version number to 0.17.3"""Configuration for Django system.""" __version__ = "0.17.3" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
<commit_before>"""Configuration for Django system.""" __version__ = "0.17.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) <commit_msg>Increment version number to 0.17.3<commit_after>"""Configuration for Django system.""" __version__ = "0.17.3" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
50992031229ea903418935613cd5e1e561b04c91
Control/PID.py
Control/PID.py
class PID: def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0): self.Kp = Kp # Proporiional gain self.Ki = Ki # Integral gain self.Kd = Kd # Derivative gain self.integral = 0 def incrementTime(self, error, dt): self.integral = self.integral + error*self.Ki*dt def computeCorrection(self, error, derror): correction = self.Kp*(error) + self.Kd*derror + self.integral return correction
class PID: def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0): self.Kp = Kp # Proportional gain self.Ki = Ki # Integral gain self.Kd = Kd # Derivative gain self.integral = 0 def incrementTime(self, error, dt): self.integral = self.integral + error*self.Ki*dt def computeCorrection(self, error, derror): correction = self.Kp*(error) + self.Kd*derror + self.integral return correction
Correct typing error and arrange indentation
Correct typing error and arrange indentation
Python
mit
baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite
class PID: def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0): self.Kp = Kp # Proporiional gain self.Ki = Ki # Integral gain self.Kd = Kd # Derivative gain self.integral = 0 def incrementTime(self, error, dt): self.integral = self.integral + error*self.Ki*dt def computeCorrection(self, error, derror): correction = self.Kp*(error) + self.Kd*derror + self.integral return correction Correct typing error and arrange indentation
class PID: def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0): self.Kp = Kp # Proportional gain self.Ki = Ki # Integral gain self.Kd = Kd # Derivative gain self.integral = 0 def incrementTime(self, error, dt): self.integral = self.integral + error*self.Ki*dt def computeCorrection(self, error, derror): correction = self.Kp*(error) + self.Kd*derror + self.integral return correction
<commit_before>class PID: def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0): self.Kp = Kp # Proporiional gain self.Ki = Ki # Integral gain self.Kd = Kd # Derivative gain self.integral = 0 def incrementTime(self, error, dt): self.integral = self.integral + error*self.Ki*dt def computeCorrection(self, error, derror): correction = self.Kp*(error) + self.Kd*derror + self.integral return correction <commit_msg>Correct typing error and arrange indentation<commit_after>
class PID: def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0): self.Kp = Kp # Proportional gain self.Ki = Ki # Integral gain self.Kd = Kd # Derivative gain self.integral = 0 def incrementTime(self, error, dt): self.integral = self.integral + error*self.Ki*dt def computeCorrection(self, error, derror): correction = self.Kp*(error) + self.Kd*derror + self.integral return correction
class PID: def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0): self.Kp = Kp # Proporiional gain self.Ki = Ki # Integral gain self.Kd = Kd # Derivative gain self.integral = 0 def incrementTime(self, error, dt): self.integral = self.integral + error*self.Ki*dt def computeCorrection(self, error, derror): correction = self.Kp*(error) + self.Kd*derror + self.integral return correction Correct typing error and arrange indentationclass PID: def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0): self.Kp = Kp # Proportional gain self.Ki = Ki # Integral gain self.Kd = Kd # Derivative gain self.integral = 0 def incrementTime(self, error, dt): self.integral = self.integral + error*self.Ki*dt def computeCorrection(self, error, derror): correction = self.Kp*(error) + self.Kd*derror + self.integral return correction
<commit_before>class PID: def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0): self.Kp = Kp # Proporiional gain self.Ki = Ki # Integral gain self.Kd = Kd # Derivative gain self.integral = 0 def incrementTime(self, error, dt): self.integral = self.integral + error*self.Ki*dt def computeCorrection(self, error, derror): correction = self.Kp*(error) + self.Kd*derror + self.integral return correction <commit_msg>Correct typing error and arrange indentation<commit_after>class PID: def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0): self.Kp = Kp # Proportional gain self.Ki = Ki # Integral gain self.Kd = Kd # Derivative gain self.integral = 0 def incrementTime(self, error, dt): self.integral = self.integral + error*self.Ki*dt def computeCorrection(self, error, derror): correction = self.Kp*(error) + self.Kd*derror + self.integral return correction
7d5a259460b4e8b8325fa55793ed4456425bda78
xd/tool/log.py
xd/tool/log.py
import logging class ConsoleFormatter(logging.Formatter): """A logging formatter for use when logging to console. Log message above logging.INFO will be prefixed with the levelname, fx.: ERROR: this is wrong And logging.DEBUG messages will be prefixed with name of the logger, which should normally be the module name, fx.: xd.tool.shell: chdir /home/user/my-project To achive this (logger name being the module name), all modules should setup logging this way: import logging log = logging.getLogger(__name__) """ def __init__(self): """Initialize the formatter.""" logging.Formatter.__init__(self) return def format(self, record): """Format the specified record as text.""" record.message = record.getMessage() fmt = "" if record.levelno > logging.INFO: fmt += "%(levelname)s: " if record.levelno == logging.DEBUG: fmt += "%(name)s: " fmt += "%(message)s" s = fmt % record.__dict__ if record.exc_info: if not record.exc_text: record.exc_text = self.formatException(record.exc_info) if record.exc_text: if s[-1:] != "\n": s += "\n\n" s = s + record.exc_text + "\n" return s def init(): """Initialize logging module for logging to console. The root_logger will be setup and initialized to output print out logging.INFO level and above. """ console_formatter = ConsoleFormatter() console_logger = logging.StreamHandler() console_logger.setFormatter(console_formatter) root_logger = logging.getLogger() root_logger.addHandler(console_logger) root_logger.setLevel(logging.INFO)
import logging class ConsoleFormatter(logging.Formatter): """A logging formatter for use when logging to console. Log message above logging.INFO will be prefixed with the levelname, fx.: ERROR: this is wrong And logging.DEBUG messages will be prefixed with name of the logger, which should normally be the module name, fx.: xd.tool.shell: chdir /home/user/my-project To achive this (logger name being the module name), all modules should setup logging this way: import logging log = logging.getLogger(__name__) """ def __init__(self): """Initialize the formatter.""" logging.Formatter.__init__(self) return def formatMessage(self, record): """Format the specified record message as text.""" fmt = "" if record.levelno > logging.INFO: fmt += "%(levelname)s: " if record.levelno == logging.DEBUG: fmt += "%(name)s: " fmt += "%(message)s" return fmt % record.__dict__ def init(): """Initialize logging module for logging to console. The root_logger will be setup and initialized to output print out logging.INFO level and above. """ console_formatter = ConsoleFormatter() console_logger = logging.StreamHandler() console_logger.setFormatter(console_formatter) root_logger = logging.getLogger() root_logger.addHandler(console_logger) root_logger.setLevel(logging.INFO)
Refactor ConsoleFormatter.format() method to formatMessage()
Refactor ConsoleFormatter.format() method to formatMessage() Re-use the standard Formatter.format() method, and only override the Formatter.formatMessage() method. Signed-off-by: Esben Haabendal <da90c138e4a9573086862393cde34fa33d74f6e5@haabendal.dk>
Python
mit
esben/xd-tool,XD-embedded/xd-tool,esben/xd-tool,XD-embedded/xd-tool
import logging class ConsoleFormatter(logging.Formatter): """A logging formatter for use when logging to console. Log message above logging.INFO will be prefixed with the levelname, fx.: ERROR: this is wrong And logging.DEBUG messages will be prefixed with name of the logger, which should normally be the module name, fx.: xd.tool.shell: chdir /home/user/my-project To achive this (logger name being the module name), all modules should setup logging this way: import logging log = logging.getLogger(__name__) """ def __init__(self): """Initialize the formatter.""" logging.Formatter.__init__(self) return def format(self, record): """Format the specified record as text.""" record.message = record.getMessage() fmt = "" if record.levelno > logging.INFO: fmt += "%(levelname)s: " if record.levelno == logging.DEBUG: fmt += "%(name)s: " fmt += "%(message)s" s = fmt % record.__dict__ if record.exc_info: if not record.exc_text: record.exc_text = self.formatException(record.exc_info) if record.exc_text: if s[-1:] != "\n": s += "\n\n" s = s + record.exc_text + "\n" return s def init(): """Initialize logging module for logging to console. The root_logger will be setup and initialized to output print out logging.INFO level and above. """ console_formatter = ConsoleFormatter() console_logger = logging.StreamHandler() console_logger.setFormatter(console_formatter) root_logger = logging.getLogger() root_logger.addHandler(console_logger) root_logger.setLevel(logging.INFO) Refactor ConsoleFormatter.format() method to formatMessage() Re-use the standard Formatter.format() method, and only override the Formatter.formatMessage() method. Signed-off-by: Esben Haabendal <da90c138e4a9573086862393cde34fa33d74f6e5@haabendal.dk>
import logging class ConsoleFormatter(logging.Formatter): """A logging formatter for use when logging to console. Log message above logging.INFO will be prefixed with the levelname, fx.: ERROR: this is wrong And logging.DEBUG messages will be prefixed with name of the logger, which should normally be the module name, fx.: xd.tool.shell: chdir /home/user/my-project To achive this (logger name being the module name), all modules should setup logging this way: import logging log = logging.getLogger(__name__) """ def __init__(self): """Initialize the formatter.""" logging.Formatter.__init__(self) return def formatMessage(self, record): """Format the specified record message as text.""" fmt = "" if record.levelno > logging.INFO: fmt += "%(levelname)s: " if record.levelno == logging.DEBUG: fmt += "%(name)s: " fmt += "%(message)s" return fmt % record.__dict__ def init(): """Initialize logging module for logging to console. The root_logger will be setup and initialized to output print out logging.INFO level and above. """ console_formatter = ConsoleFormatter() console_logger = logging.StreamHandler() console_logger.setFormatter(console_formatter) root_logger = logging.getLogger() root_logger.addHandler(console_logger) root_logger.setLevel(logging.INFO)
<commit_before>import logging class ConsoleFormatter(logging.Formatter): """A logging formatter for use when logging to console. Log message above logging.INFO will be prefixed with the levelname, fx.: ERROR: this is wrong And logging.DEBUG messages will be prefixed with name of the logger, which should normally be the module name, fx.: xd.tool.shell: chdir /home/user/my-project To achive this (logger name being the module name), all modules should setup logging this way: import logging log = logging.getLogger(__name__) """ def __init__(self): """Initialize the formatter.""" logging.Formatter.__init__(self) return def format(self, record): """Format the specified record as text.""" record.message = record.getMessage() fmt = "" if record.levelno > logging.INFO: fmt += "%(levelname)s: " if record.levelno == logging.DEBUG: fmt += "%(name)s: " fmt += "%(message)s" s = fmt % record.__dict__ if record.exc_info: if not record.exc_text: record.exc_text = self.formatException(record.exc_info) if record.exc_text: if s[-1:] != "\n": s += "\n\n" s = s + record.exc_text + "\n" return s def init(): """Initialize logging module for logging to console. The root_logger will be setup and initialized to output print out logging.INFO level and above. """ console_formatter = ConsoleFormatter() console_logger = logging.StreamHandler() console_logger.setFormatter(console_formatter) root_logger = logging.getLogger() root_logger.addHandler(console_logger) root_logger.setLevel(logging.INFO) <commit_msg>Refactor ConsoleFormatter.format() method to formatMessage() Re-use the standard Formatter.format() method, and only override the Formatter.formatMessage() method. Signed-off-by: Esben Haabendal <da90c138e4a9573086862393cde34fa33d74f6e5@haabendal.dk><commit_after>
import logging class ConsoleFormatter(logging.Formatter): """A logging formatter for use when logging to console. Log message above logging.INFO will be prefixed with the levelname, fx.: ERROR: this is wrong And logging.DEBUG messages will be prefixed with name of the logger, which should normally be the module name, fx.: xd.tool.shell: chdir /home/user/my-project To achive this (logger name being the module name), all modules should setup logging this way: import logging log = logging.getLogger(__name__) """ def __init__(self): """Initialize the formatter.""" logging.Formatter.__init__(self) return def formatMessage(self, record): """Format the specified record message as text.""" fmt = "" if record.levelno > logging.INFO: fmt += "%(levelname)s: " if record.levelno == logging.DEBUG: fmt += "%(name)s: " fmt += "%(message)s" return fmt % record.__dict__ def init(): """Initialize logging module for logging to console. The root_logger will be setup and initialized to output print out logging.INFO level and above. """ console_formatter = ConsoleFormatter() console_logger = logging.StreamHandler() console_logger.setFormatter(console_formatter) root_logger = logging.getLogger() root_logger.addHandler(console_logger) root_logger.setLevel(logging.INFO)
import logging class ConsoleFormatter(logging.Formatter): """A logging formatter for use when logging to console. Log message above logging.INFO will be prefixed with the levelname, fx.: ERROR: this is wrong And logging.DEBUG messages will be prefixed with name of the logger, which should normally be the module name, fx.: xd.tool.shell: chdir /home/user/my-project To achive this (logger name being the module name), all modules should setup logging this way: import logging log = logging.getLogger(__name__) """ def __init__(self): """Initialize the formatter.""" logging.Formatter.__init__(self) return def format(self, record): """Format the specified record as text.""" record.message = record.getMessage() fmt = "" if record.levelno > logging.INFO: fmt += "%(levelname)s: " if record.levelno == logging.DEBUG: fmt += "%(name)s: " fmt += "%(message)s" s = fmt % record.__dict__ if record.exc_info: if not record.exc_text: record.exc_text = self.formatException(record.exc_info) if record.exc_text: if s[-1:] != "\n": s += "\n\n" s = s + record.exc_text + "\n" return s def init(): """Initialize logging module for logging to console. The root_logger will be setup and initialized to output print out logging.INFO level and above. """ console_formatter = ConsoleFormatter() console_logger = logging.StreamHandler() console_logger.setFormatter(console_formatter) root_logger = logging.getLogger() root_logger.addHandler(console_logger) root_logger.setLevel(logging.INFO) Refactor ConsoleFormatter.format() method to formatMessage() Re-use the standard Formatter.format() method, and only override the Formatter.formatMessage() method. Signed-off-by: Esben Haabendal <da90c138e4a9573086862393cde34fa33d74f6e5@haabendal.dk>import logging class ConsoleFormatter(logging.Formatter): """A logging formatter for use when logging to console. Log message above logging.INFO will be prefixed with the levelname, fx.: ERROR: this is wrong And logging.DEBUG messages will be prefixed with name of the logger, which should normally be the module name, fx.: xd.tool.shell: chdir /home/user/my-project To achive this (logger name being the module name), all modules should setup logging this way: import logging log = logging.getLogger(__name__) """ def __init__(self): """Initialize the formatter.""" logging.Formatter.__init__(self) return def formatMessage(self, record): """Format the specified record message as text.""" fmt = "" if record.levelno > logging.INFO: fmt += "%(levelname)s: " if record.levelno == logging.DEBUG: fmt += "%(name)s: " fmt += "%(message)s" return fmt % record.__dict__ def init(): """Initialize logging module for logging to console. The root_logger will be setup and initialized to output print out logging.INFO level and above. """ console_formatter = ConsoleFormatter() console_logger = logging.StreamHandler() console_logger.setFormatter(console_formatter) root_logger = logging.getLogger() root_logger.addHandler(console_logger) root_logger.setLevel(logging.INFO)
<commit_before>import logging class ConsoleFormatter(logging.Formatter): """A logging formatter for use when logging to console. Log message above logging.INFO will be prefixed with the levelname, fx.: ERROR: this is wrong And logging.DEBUG messages will be prefixed with name of the logger, which should normally be the module name, fx.: xd.tool.shell: chdir /home/user/my-project To achive this (logger name being the module name), all modules should setup logging this way: import logging log = logging.getLogger(__name__) """ def __init__(self): """Initialize the formatter.""" logging.Formatter.__init__(self) return def format(self, record): """Format the specified record as text.""" record.message = record.getMessage() fmt = "" if record.levelno > logging.INFO: fmt += "%(levelname)s: " if record.levelno == logging.DEBUG: fmt += "%(name)s: " fmt += "%(message)s" s = fmt % record.__dict__ if record.exc_info: if not record.exc_text: record.exc_text = self.formatException(record.exc_info) if record.exc_text: if s[-1:] != "\n": s += "\n\n" s = s + record.exc_text + "\n" return s def init(): """Initialize logging module for logging to console. The root_logger will be setup and initialized to output print out logging.INFO level and above. """ console_formatter = ConsoleFormatter() console_logger = logging.StreamHandler() console_logger.setFormatter(console_formatter) root_logger = logging.getLogger() root_logger.addHandler(console_logger) root_logger.setLevel(logging.INFO) <commit_msg>Refactor ConsoleFormatter.format() method to formatMessage() Re-use the standard Formatter.format() method, and only override the Formatter.formatMessage() method. Signed-off-by: Esben Haabendal <da90c138e4a9573086862393cde34fa33d74f6e5@haabendal.dk><commit_after>import logging class ConsoleFormatter(logging.Formatter): """A logging formatter for use when logging to console. Log message above logging.INFO will be prefixed with the levelname, fx.: ERROR: this is wrong And logging.DEBUG messages will be prefixed with name of the logger, which should normally be the module name, fx.: xd.tool.shell: chdir /home/user/my-project To achive this (logger name being the module name), all modules should setup logging this way: import logging log = logging.getLogger(__name__) """ def __init__(self): """Initialize the formatter.""" logging.Formatter.__init__(self) return def formatMessage(self, record): """Format the specified record message as text.""" fmt = "" if record.levelno > logging.INFO: fmt += "%(levelname)s: " if record.levelno == logging.DEBUG: fmt += "%(name)s: " fmt += "%(message)s" return fmt % record.__dict__ def init(): """Initialize logging module for logging to console. The root_logger will be setup and initialized to output print out logging.INFO level and above. """ console_formatter = ConsoleFormatter() console_logger = logging.StreamHandler() console_logger.setFormatter(console_formatter) root_logger = logging.getLogger() root_logger.addHandler(console_logger) root_logger.setLevel(logging.INFO)
5162275b9b6136f2b97d195384bb9979a0d79bfc
script/lib/config.py
script/lib/config.py
#!/usr/bin/env python import platform import sys BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '9f5271d31e0f32eac5a20ef6f543e3f1d43ad645' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
#!/usr/bin/env python import platform import sys BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '56984fa0e4c3c745652510f342c0fb2724d846c2' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
Upgrade libchromiumcontent for dbus headers
Upgrade libchromiumcontent for dbus headers
Python
mit
ianscrivener/electron,chriskdon/electron,yalexx/electron,subblue/electron,nekuz0r/electron,systembugtj/electron,trankmichael/electron,posix4e/electron,bitemyapp/electron,beni55/electron,mrwizard82d1/electron,Faiz7412/electron,rajatsingla28/electron,tomashanacek/electron,kokdemo/electron,darwin/electron,vipulroxx/electron,nicobot/electron,tincan24/electron,jcblw/electron,neutrous/electron,iftekeriba/electron,arturts/electron,robinvandernoord/electron,matiasinsaurralde/electron,pandoraui/electron,vaginessa/electron,jaanus/electron,jlord/electron,faizalpribadi/electron,BionicClick/electron,Faiz7412/electron,ianscrivener/electron,sky7sea/electron,brave/electron,jtburke/electron,jaanus/electron,hokein/atom-shell,jhen0409/electron,evgenyzinoviev/electron,faizalpribadi/electron,yan-foto/electron,cos2004/electron,natgolov/electron,astoilkov/electron,cqqccqc/electron,stevemao/electron,dahal/electron,edulan/electron,greyhwndz/electron,wolfflow/electron,simongregory/electron,DivyaKMenon/electron,GoooIce/electron,IonicaBizauKitchen/electron,leftstick/electron,MaxWhere/electron,thomsonreuters/electron,shaundunne/electron,bitemyapp/electron,seanchas116/electron,aichingm/electron,stevekinney/electron,bwiggs/electron,leethomas/electron,eric-seekas/electron,John-Lin/electron,vipulroxx/electron,synaptek/electron,neutrous/electron,jonatasfreitasv/electron,egoist/electron,rreimann/electron,brave/muon,thingsinjars/electron,hokein/atom-shell,ankitaggarwal011/electron,digideskio/electron,adcentury/electron,kcrt/electron,Ivshti/electron,mjaniszew/electron,natgolov/electron,yan-foto/electron,fomojola/electron,RobertJGabriel/electron,JussMee15/electron,jaanus/electron,Faiz7412/electron,jtburke/electron,chriskdon/electron,jjz/electron,thomsonreuters/electron,vaginessa/electron,astoilkov/electron,aecca/electron,miniak/electron,nicholasess/electron,seanchas116/electron,GoooIce/electron,brave/electron,digideskio/electron,mattotodd/electron,Floato/electron,wan-qy/electron,nicobot/electron,maxogden/atom-shell,wan-qy/electron,farmisen/electron,posix4e/electron,thingsinjars/electron,evgenyzinoviev/electron,shockone/electron,leolujuyi/electron,fabien-d/electron,minggo/electron,brave/muon,wan-qy/electron,jhen0409/electron,sircharleswatson/electron,destan/electron,brenca/electron,rhencke/electron,shockone/electron,simongregory/electron,adcentury/electron,fomojola/electron,bwiggs/electron,sshiting/electron,tinydew4/electron,icattlecoder/electron,evgenyzinoviev/electron,xiruibing/electron,bright-sparks/electron,dongjoon-hyun/electron,zhakui/electron,seanchas116/electron,mattotodd/electron,jjz/electron,miniak/electron,benweissmann/electron,Faiz7412/electron,dkfiresky/electron,jannishuebl/electron,sky7sea/electron,JesselJohn/electron,joneit/electron,soulteary/electron,micalan/electron,vaginessa/electron,setzer777/electron,JesselJohn/electron,JesselJohn/electron,adamjgray/electron,shennushi/electron,medixdev/electron,takashi/electron,bbondy/electron,kenmozi/electron,BionicClick/electron,yalexx/electron,baiwyc119/electron,twolfson/electron,aliib/electron,renaesop/electron,fffej/electron,tinydew4/electron,nagyistoce/electron-atom-shell,simongregory/electron,meowlab/electron,gabrielPeart/electron,robinvandernoord/electron,nicholasess/electron,yan-foto/electron,arusakov/electron,shiftkey/electron,roadev/electron,RIAEvangelist/electron,christian-bromann/electron,egoist/electron,trankmichael/electron,d-salas/electron,bobwol/electron,zhakui/electron,d-salas/electron,kikong/electron,minggo/electron,gamedevsam/electron,yalexx/electron,baiwyc119/electron,gabriel/electron,fritx/electron,voidbridge/electron,jsutcodes/electron,Jacobichou/electron,BionicClick/electron,arturts/electron,jcblw/electron,yalexx/electron,JussMee15/electron,thompsonemerson/electron,deepak1556/atom-shell,robinvandernoord/electron,aichingm/electron,shockone/electron,egoist/electron,RIAEvangelist/electron,nekuz0r/electron,gabrielPeart/electron,nagyistoce/electron-atom-shell,maxogden/atom-shell,cqqccqc/electron,MaxGraey/electron,webmechanicx/electron,eric-seekas/electron,DivyaKMenon/electron,electron/electron,joaomoreno/atom-shell,mubassirhayat/electron,destan/electron,arusakov/electron,vipulroxx/electron,baiwyc119/electron,pirafrank/electron,greyhwndz/electron,gabriel/electron,renaesop/electron,RobertJGabriel/electron,setzer777/electron,egoist/electron,thomsonreuters/electron,kokdemo/electron,matiasinsaurralde/electron,Gerhut/electron,tincan24/electron,rsvip/electron,bpasero/electron,maxogden/atom-shell,eriser/electron,joneit/electron,iftekeriba/electron,takashi/electron,lzpfmh/electron,mhkeller/electron,jacksondc/electron,synaptek/electron,DivyaKMenon/electron,jiaz/electron,takashi/electron,icattlecoder/electron,fritx/electron,tincan24/electron,trankmichael/electron,vaginessa/electron,the-ress/electron,aliib/electron,vHanda/electron,rreimann/electron,simongregory/electron,bruce/electron,maxogden/atom-shell,cos2004/electron,beni55/electron,bitemyapp/electron,preco21/electron,ianscrivener/electron,aliib/electron,jjz/electron,systembugtj/electron,roadev/electron,arturts/electron,MaxWhere/electron,coderhaoxin/electron,jonatasfreitasv/electron,bpasero/electron,gstack/infinium-shell,saronwei/electron,subblue/electron,MaxWhere/electron,shaundunne/electron,brenca/electron,dongjoon-hyun/electron,joneit/electron,sky7sea/electron,digideskio/electron,vaginessa/electron,digideskio/electron,jlhbaseball15/electron,greyhwndz/electron,fireball-x/atom-shell,shaundunne/electron,arusakov/electron,yalexx/electron,trigrass2/electron,kostia/electron,bruce/electron,subblue/electron,medixdev/electron,edulan/electron,rhencke/electron,wolfflow/electron,bbondy/electron,chriskdon/electron,stevemao/electron,deed02392/electron,evgenyzinoviev/electron,dkfiresky/electron,leftstick/electron,mattdesl/electron,kenmozi/electron,John-Lin/electron,egoist/electron,aliib/electron,gerhardberger/electron,beni55/electron,beni55/electron,gamedevsam/electron,darwin/electron,kenmozi/electron,stevekinney/electron,xiruibing/electron,LadyNaggaga/electron,medixdev/electron,kostia/electron,mrwizard82d1/electron,pandoraui/electron,saronwei/electron,dahal/electron,gbn972/electron,leolujuyi/electron,nicobot/electron,shaundunne/electron,davazp/electron,voidbridge/electron,rhencke/electron,rajatsingla28/electron,zhakui/electron,sircharleswatson/electron,anko/electron,ervinb/electron,fomojola/electron,darwin/electron,DivyaKMenon/electron,MaxGraey/electron,tonyganch/electron,gstack/infinium-shell,egoist/electron,roadev/electron,jannishuebl/electron,edulan/electron,jsutcodes/electron,noikiy/electron,chrisswk/electron,simonfork/electron,thomsonreuters/electron,mattdesl/electron,stevemao/electron,IonicaBizauKitchen/electron,nicholasess/electron,minggo/electron,rprichard/electron,sshiting/electron,felixrieseberg/electron,bwiggs/electron,christian-bromann/electron,micalan/electron,Neron-X5/electron,michaelchiche/electron,leethomas/electron,tomashanacek/electron,benweissmann/electron,rajatsingla28/electron,d-salas/electron,kokdemo/electron,shiftkey/electron,noikiy/electron,ianscrivener/electron,sky7sea/electron,sky7sea/electron,chrisswk/electron,smczk/electron,the-ress/electron,RobertJGabriel/electron,sshiting/electron,twolfson/electron,cos2004/electron,kcrt/electron,Zagorakiss/electron,simonfork/electron,tinydew4/electron,jcblw/electron,cos2004/electron,gerhardberger/electron,carsonmcdonald/electron,kcrt/electron,timruffles/electron,bwiggs/electron,jacksondc/electron,kazupon/electron,pirafrank/electron,preco21/electron,nicobot/electron,miniak/electron,shennushi/electron,fabien-d/electron,Jacobichou/electron,fffej/electron,ankitaggarwal011/electron,arturts/electron,chriskdon/electron,mattdesl/electron,setzer777/electron,micalan/electron,nicholasess/electron,Gerhut/electron,edulan/electron,Zagorakiss/electron,Jonekee/electron,Rokt33r/electron,jtburke/electron,pandoraui/electron,joaomoreno/atom-shell,jacksondc/electron,vHanda/electron,soulteary/electron,d-salas/electron,rsvip/electron,xfstudio/electron,fomojola/electron,etiktin/electron,eriser/electron,darwin/electron,rsvip/electron,vipulroxx/electron,jsutcodes/electron,tinydew4/electron,preco21/electron,MaxGraey/electron,gabriel/electron,saronwei/electron,brave/muon,pombredanne/electron,jaanus/electron,Zagorakiss/electron,aaron-goshine/electron,pirafrank/electron,soulteary/electron,roadev/electron,davazp/electron,Rokt33r/electron,coderhaoxin/electron,systembugtj/electron,ervinb/electron,faizalpribadi/electron,astoilkov/electron,thompsonemerson/electron,carsonmcdonald/electron,rhencke/electron,miniak/electron,carsonmcdonald/electron,bright-sparks/electron,aichingm/electron,takashi/electron,jlord/electron,natgolov/electron,leethomas/electron,LadyNaggaga/electron,lzpfmh/electron,tincan24/electron,jsutcodes/electron,preco21/electron,kenmozi/electron,Ivshti/electron,jiaz/electron,christian-bromann/electron,zhakui/electron,fomojola/electron,Andrey-Pavlov/electron,dahal/electron,JussMee15/electron,jlhbaseball15/electron,zhakui/electron,trankmichael/electron,lrlna/electron,fomojola/electron,setzer777/electron,adamjgray/electron,mjaniszew/electron,jtburke/electron,noikiy/electron,smczk/electron,RobertJGabriel/electron,adamjgray/electron,wolfflow/electron,joneit/electron,bruce/electron,Evercoder/electron,Jacobichou/electron,thingsinjars/electron,coderhaoxin/electron,the-ress/electron,hokein/atom-shell,electron/electron,eriser/electron,jonatasfreitasv/electron,ervinb/electron,bruce/electron,jannishuebl/electron,cqqccqc/electron,roadev/electron,tomashanacek/electron,fritx/electron,synaptek/electron,mubassirhayat/electron,jannishuebl/electron,evgenyzinoviev/electron,mattdesl/electron,DivyaKMenon/electron,gamedevsam/electron,biblerule/UMCTelnetHub,eric-seekas/electron,michaelchiche/electron,michaelchiche/electron,sshiting/electron,Evercoder/electron,destan/electron,gbn972/electron,LadyNaggaga/electron,icattlecoder/electron,greyhwndz/electron,roadev/electron,oiledCode/electron,gabriel/electron,kostia/electron,jannishuebl/electron,RobertJGabriel/electron,gerhardberger/electron,Jacobichou/electron,aliib/electron,simonfork/electron,coderhaoxin/electron,posix4e/electron,kazupon/electron,jacksondc/electron,stevemao/electron,thompsonemerson/electron,nicobot/electron,John-Lin/electron,etiktin/electron,cos2004/electron,hokein/atom-shell,biblerule/UMCTelnetHub,iftekeriba/electron,eriser/electron,medixdev/electron,kcrt/electron,shaundunne/electron,Neron-X5/electron,vHanda/electron,arusakov/electron,aaron-goshine/electron,stevekinney/electron,howmuchcomputer/electron,deed02392/electron,nicobot/electron,soulteary/electron,jhen0409/electron,Ivshti/electron,jlhbaseball15/electron,xfstudio/electron,fireball-x/atom-shell,JussMee15/electron,Gerhut/electron,mubassirhayat/electron,sircharleswatson/electron,preco21/electron,gamedevsam/electron,nicholasess/electron,bpasero/electron,JesselJohn/electron,jhen0409/electron,jhen0409/electron,astoilkov/electron,fireball-x/atom-shell,gerhardberger/electron,dkfiresky/electron,bbondy/electron,GoooIce/electron,digideskio/electron,meowlab/electron,John-Lin/electron,icattlecoder/electron,tonyganch/electron,sky7sea/electron,LadyNaggaga/electron,minggo/electron,tinydew4/electron,mrwizard82d1/electron,dongjoon-hyun/electron,carsonmcdonald/electron,Andrey-Pavlov/electron,leftstick/electron,SufianHassan/electron,deed02392/electron,smczk/electron,gstack/infinium-shell,Neron-X5/electron,anko/electron,bobwol/electron,aecca/electron,leethomas/electron,subblue/electron,pandoraui/electron,davazp/electron,adcentury/electron,stevekinney/electron,jiaz/electron,minggo/electron,lzpfmh/electron,Rokt33r/electron,bobwol/electron,bbondy/electron,timruffles/electron,robinvandernoord/electron,aaron-goshine/electron,mirrh/electron,deepak1556/atom-shell,pirafrank/electron,Neron-X5/electron,oiledCode/electron,abhishekgahlot/electron,voidbridge/electron,lzpfmh/electron,rreimann/electron,davazp/electron,BionicClick/electron,micalan/electron,rajatsingla28/electron,xiruibing/electron,destan/electron,arusakov/electron,tomashanacek/electron,greyhwndz/electron,fireball-x/atom-shell,pandoraui/electron,biblerule/UMCTelnetHub,shaundunne/electron,eric-seekas/electron,timruffles/electron,darwin/electron,gabrielPeart/electron,minggo/electron,Andrey-Pavlov/electron,takashi/electron,nekuz0r/electron,mirrh/electron,eriser/electron,rprichard/electron,xfstudio/electron,bwiggs/electron,Floato/electron,nekuz0r/electron,shiftkey/electron,subblue/electron,edulan/electron,wan-qy/electron,matiasinsaurralde/electron,oiledCode/electron,abhishekgahlot/electron,MaxWhere/electron,howmuchcomputer/electron,jcblw/electron,the-ress/electron,sshiting/electron,tomashanacek/electron,trigrass2/electron,brave/electron,benweissmann/electron,maxogden/atom-shell,noikiy/electron,medixdev/electron,brave/muon,smczk/electron,kikong/electron,aichingm/electron,carsonmcdonald/electron,twolfson/electron,natgolov/electron,kokdemo/electron,RIAEvangelist/electron,arusakov/electron,abhishekgahlot/electron,seanchas116/electron,xfstudio/electron,shockone/electron,mattdesl/electron,gabrielPeart/electron,micalan/electron,jlord/electron,aecca/electron,Andrey-Pavlov/electron,setzer777/electron,bruce/electron,pombredanne/electron,chrisswk/electron,dongjoon-hyun/electron,simongregory/electron,bitemyapp/electron,mjaniszew/electron,evgenyzinoviev/electron,ankitaggarwal011/electron,jacksondc/electron,baiwyc119/electron,tylergibson/electron,Zagorakiss/electron,mubassirhayat/electron,electron/electron,tylergibson/electron,chrisswk/electron,gbn972/electron,zhakui/electron,cqqccqc/electron,shennushi/electron,jtburke/electron,deepak1556/atom-shell,tylergibson/electron,jiaz/electron,stevemao/electron,fabien-d/electron,rsvip/electron,bbondy/electron,stevemao/electron,shockone/electron,rsvip/electron,GoooIce/electron,leolujuyi/electron,thomsonreuters/electron,GoooIce/electron,iftekeriba/electron,DivyaKMenon/electron,bobwol/electron,wan-qy/electron,systembugtj/electron,wan-qy/electron,Gerhut/electron,biblerule/UMCTelnetHub,adcentury/electron,pirafrank/electron,trankmichael/electron,shennushi/electron,nagyistoce/electron-atom-shell,faizalpribadi/electron,tylergibson/electron,chrisswk/electron,gbn972/electron,joneit/electron,kikong/electron,gbn972/electron,mirrh/electron,leolujuyi/electron,MaxWhere/electron,IonicaBizauKitchen/electron,abhishekgahlot/electron,thomsonreuters/electron,joaomoreno/atom-shell,vHanda/electron,greyhwndz/electron,eric-seekas/electron,aaron-goshine/electron,mattotodd/electron,deed02392/electron,twolfson/electron,beni55/electron,kikong/electron,jlord/electron,saronwei/electron,ianscrivener/electron,Rokt33r/electron,Jacobichou/electron,preco21/electron,aecca/electron,miniak/electron,SufianHassan/electron,rreimann/electron,iftekeriba/electron,sircharleswatson/electron,trigrass2/electron,cos2004/electron,anko/electron,Floato/electron,mirrh/electron,mjaniszew/electron,JussMee15/electron,thingsinjars/electron,michaelchiche/electron,tinydew4/electron,mattotodd/electron,nekuz0r/electron,systembugtj/electron,mjaniszew/electron,aecca/electron,smczk/electron,felixrieseberg/electron,anko/electron,MaxGraey/electron,dongjoon-hyun/electron,wolfflow/electron,felixrieseberg/electron,xiruibing/electron,farmisen/electron,tylergibson/electron,brenca/electron,rajatsingla28/electron,adamjgray/electron,mhkeller/electron,mhkeller/electron,eriser/electron,JesselJohn/electron,mjaniszew/electron,jjz/electron,christian-bromann/electron,pandoraui/electron,gbn972/electron,matiasinsaurralde/electron,mirrh/electron,noikiy/electron,LadyNaggaga/electron,carsonmcdonald/electron,jjz/electron,fffej/electron,JussMee15/electron,Zagorakiss/electron,tomashanacek/electron,fabien-d/electron,jiaz/electron,gamedevsam/electron,saronwei/electron,voidbridge/electron,michaelchiche/electron,adcentury/electron,icattlecoder/electron,shennushi/electron,pombredanne/electron,ankitaggarwal011/electron,jlhbaseball15/electron,John-Lin/electron,Evercoder/electron,biblerule/UMCTelnetHub,electron/electron,ervinb/electron,jlhbaseball15/electron,tincan24/electron,thompsonemerson/electron,xiruibing/electron,electron/electron,bpasero/electron,simonfork/electron,adcentury/electron,tylergibson/electron,bpasero/electron,simonfork/electron,etiktin/electron,bobwol/electron,stevekinney/electron,aaron-goshine/electron,electron/electron,wolfflow/electron,bpasero/electron,ervinb/electron,rhencke/electron,seanchas116/electron,synaptek/electron,iftekeriba/electron,Gerhut/electron,bobwol/electron,mhkeller/electron,shiftkey/electron,kikong/electron,tincan24/electron,kenmozi/electron,fffej/electron,pombredanne/electron,Evercoder/electron,systembugtj/electron,Floato/electron,vipulroxx/electron,electron/electron,oiledCode/electron,fritx/electron,joneit/electron,neutrous/electron,mubassirhayat/electron,kcrt/electron,bright-sparks/electron,etiktin/electron,jonatasfreitasv/electron,aaron-goshine/electron,trigrass2/electron,eric-seekas/electron,Rokt33r/electron,micalan/electron,leolujuyi/electron,brenca/electron,JesselJohn/electron,leftstick/electron,bwiggs/electron,jaanus/electron,LadyNaggaga/electron,webmechanicx/electron,bitemyapp/electron,SufianHassan/electron,davazp/electron,Gerhut/electron,renaesop/electron,mhkeller/electron,synaptek/electron,chriskdon/electron,jaanus/electron,John-Lin/electron,mirrh/electron,trigrass2/electron,oiledCode/electron,aecca/electron,jjz/electron,xfstudio/electron,Jonekee/electron,deepak1556/atom-shell,trigrass2/electron,gerhardberger/electron,SufianHassan/electron,shiftkey/electron,xfstudio/electron,aichingm/electron,dahal/electron,Zagorakiss/electron,hokein/atom-shell,vipulroxx/electron,jsutcodes/electron,thingsinjars/electron,voidbridge/electron,mattotodd/electron,adamjgray/electron,howmuchcomputer/electron,soulteary/electron,beni55/electron,Ivshti/electron,baiwyc119/electron,mrwizard82d1/electron,ianscrivener/electron,kazupon/electron,pirafrank/electron,timruffles/electron,faizalpribadi/electron,abhishekgahlot/electron,farmisen/electron,Jacobichou/electron,gerhardberger/electron,lrlna/electron,vHanda/electron,gabrielPeart/electron,cqqccqc/electron,sircharleswatson/electron,mrwizard82d1/electron,brenca/electron,pombredanne/electron,icattlecoder/electron,simonfork/electron,deepak1556/atom-shell,fritx/electron,tonyganch/electron,RIAEvangelist/electron,brave/electron,trankmichael/electron,chriskdon/electron,digideskio/electron,gabriel/electron,mattdesl/electron,thingsinjars/electron,SufianHassan/electron,rajatsingla28/electron,dahal/electron,miniak/electron,gstack/infinium-shell,kazupon/electron,farmisen/electron,nekuz0r/electron,fffej/electron,vaginessa/electron,arturts/electron,faizalpribadi/electron,jiaz/electron,jcblw/electron,nagyistoce/electron-atom-shell,biblerule/UMCTelnetHub,nagyistoce/electron-atom-shell,thompsonemerson/electron,GoooIce/electron,lrlna/electron,bbondy/electron,rprichard/electron,MaxWhere/electron,cqqccqc/electron,webmechanicx/electron,posix4e/electron,posix4e/electron,yan-foto/electron,jonatasfreitasv/electron,twolfson/electron,tonyganch/electron,ankitaggarwal011/electron,neutrous/electron,benweissmann/electron,jlord/electron,mattotodd/electron,leftstick/electron,dongjoon-hyun/electron,fireball-x/atom-shell,d-salas/electron,howmuchcomputer/electron,kcrt/electron,IonicaBizauKitchen/electron,Andrey-Pavlov/electron,lzpfmh/electron,gabriel/electron,Evercoder/electron,kokdemo/electron,fffej/electron,howmuchcomputer/electron,aliib/electron,shockone/electron,gabrielPeart/electron,wolfflow/electron,jsutcodes/electron,aichingm/electron,bruce/electron,natgolov/electron,brave/electron,the-ress/electron,joaomoreno/atom-shell,anko/electron,matiasinsaurralde/electron,renaesop/electron,lzpfmh/electron,oiledCode/electron,rprichard/electron,bright-sparks/electron,lrlna/electron,BionicClick/electron,dahal/electron,Neron-X5/electron,adamjgray/electron,subblue/electron,edulan/electron,stevekinney/electron,rhencke/electron,robinvandernoord/electron,coderhaoxin/electron,ervinb/electron,pombredanne/electron,twolfson/electron,astoilkov/electron,jcblw/electron,the-ress/electron,farmisen/electron,leethomas/electron,leethomas/electron,Floato/electron,matiasinsaurralde/electron,felixrieseberg/electron,sircharleswatson/electron,voidbridge/electron,gerhardberger/electron,anko/electron,d-salas/electron,arturts/electron,Jonekee/electron,SufianHassan/electron,dkfiresky/electron,coderhaoxin/electron,mhkeller/electron,neutrous/electron,IonicaBizauKitchen/electron,lrlna/electron,seanchas116/electron,takashi/electron,the-ress/electron,etiktin/electron,Andrey-Pavlov/electron,posix4e/electron,fabien-d/electron,IonicaBizauKitchen/electron,bpasero/electron,noikiy/electron,farmisen/electron,kostia/electron,deed02392/electron,saronwei/electron,yalexx/electron,tonyganch/electron,MaxGraey/electron,christian-bromann/electron,sshiting/electron,webmechanicx/electron,gamedevsam/electron,destan/electron,felixrieseberg/electron,benweissmann/electron,Rokt33r/electron,ankitaggarwal011/electron,dkfiresky/electron,deed02392/electron,RIAEvangelist/electron,neutrous/electron,Jonekee/electron,brave/muon,Floato/electron,timruffles/electron,meowlab/electron,kenmozi/electron,howmuchcomputer/electron,kostia/electron,xiruibing/electron,Faiz7412/electron,robinvandernoord/electron,renaesop/electron,Neron-X5/electron,webmechanicx/electron,Ivshti/electron,meowlab/electron,brave/electron,yan-foto/electron,nicholasess/electron,Jonekee/electron,meowlab/electron,soulteary/electron,tonyganch/electron,smczk/electron,natgolov/electron,jacksondc/electron,lrlna/electron,leftstick/electron,jlhbaseball15/electron,kostia/electron,abhishekgahlot/electron,dkfiresky/electron,mrwizard82d1/electron,medixdev/electron,synaptek/electron,fritx/electron,rreimann/electron,joaomoreno/atom-shell,benweissmann/electron,RobertJGabriel/electron,Evercoder/electron,jonatasfreitasv/electron,bitemyapp/electron,Jonekee/electron,joaomoreno/atom-shell,christian-bromann/electron,kokdemo/electron,etiktin/electron,renaesop/electron,kazupon/electron,simongregory/electron,brave/muon,felixrieseberg/electron,jhen0409/electron,BionicClick/electron,vHanda/electron,RIAEvangelist/electron,michaelchiche/electron,destan/electron,davazp/electron,brenca/electron,astoilkov/electron,meowlab/electron,setzer777/electron,baiwyc119/electron,thompsonemerson/electron,yan-foto/electron,shiftkey/electron,kazupon/electron,jannishuebl/electron,webmechanicx/electron,jtburke/electron,bright-sparks/electron,bright-sparks/electron,rreimann/electron,shennushi/electron,gstack/infinium-shell,leolujuyi/electron
#!/usr/bin/env python import platform import sys BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '9f5271d31e0f32eac5a20ef6f543e3f1d43ad645' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform] Upgrade libchromiumcontent for dbus headers
#!/usr/bin/env python import platform import sys BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '56984fa0e4c3c745652510f342c0fb2724d846c2' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
<commit_before>#!/usr/bin/env python import platform import sys BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '9f5271d31e0f32eac5a20ef6f543e3f1d43ad645' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform] <commit_msg>Upgrade libchromiumcontent for dbus headers<commit_after>
#!/usr/bin/env python import platform import sys BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '56984fa0e4c3c745652510f342c0fb2724d846c2' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
#!/usr/bin/env python import platform import sys BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '9f5271d31e0f32eac5a20ef6f543e3f1d43ad645' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform] Upgrade libchromiumcontent for dbus headers#!/usr/bin/env python import platform import sys BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '56984fa0e4c3c745652510f342c0fb2724d846c2' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
<commit_before>#!/usr/bin/env python import platform import sys BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '9f5271d31e0f32eac5a20ef6f543e3f1d43ad645' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform] <commit_msg>Upgrade libchromiumcontent for dbus headers<commit_after>#!/usr/bin/env python import platform import sys BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '56984fa0e4c3c745652510f342c0fb2724d846c2' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
5a2c03b9369ccd00cc8c5c7bca4b2fc40bb18a7f
passpie/credential.py
passpie/credential.py
import re def split_fullname(fullname): rgx = re.compile(r"(?P<login>.*)@(?P<name>.*)") try: name = rgx.match(fullname).group("name") login = rgx.match(fullname).group("login") except AttributeError: raise ValueError("Not a valid name") return login if login else "_", name def make_fullname(login, name): return "{}@{}".format("_" if login is None else login, name)
import re def split_fullname(fullname): rgx = re.compile(r"(?P<login>.*)?@(?P<name>.*)") try: name = rgx.match(fullname).group("name") login = rgx.match(fullname).group("login") except AttributeError: raise ValueError("Not a valid name") return login if login else "_", name def make_fullname(login, name): return "{}@{}".format("_" if login is None else login, name)
Fix regex for spliting fullnames
Fix regex for spliting fullnames
Python
mit
marcwebbie/passpie,scorphus/passpie,marcwebbie/passpie,eiginn/passpie,scorphus/passpie,eiginn/passpie
import re def split_fullname(fullname): rgx = re.compile(r"(?P<login>.*)@(?P<name>.*)") try: name = rgx.match(fullname).group("name") login = rgx.match(fullname).group("login") except AttributeError: raise ValueError("Not a valid name") return login if login else "_", name def make_fullname(login, name): return "{}@{}".format("_" if login is None else login, name) Fix regex for spliting fullnames
import re def split_fullname(fullname): rgx = re.compile(r"(?P<login>.*)?@(?P<name>.*)") try: name = rgx.match(fullname).group("name") login = rgx.match(fullname).group("login") except AttributeError: raise ValueError("Not a valid name") return login if login else "_", name def make_fullname(login, name): return "{}@{}".format("_" if login is None else login, name)
<commit_before>import re def split_fullname(fullname): rgx = re.compile(r"(?P<login>.*)@(?P<name>.*)") try: name = rgx.match(fullname).group("name") login = rgx.match(fullname).group("login") except AttributeError: raise ValueError("Not a valid name") return login if login else "_", name def make_fullname(login, name): return "{}@{}".format("_" if login is None else login, name) <commit_msg>Fix regex for spliting fullnames<commit_after>
import re def split_fullname(fullname): rgx = re.compile(r"(?P<login>.*)?@(?P<name>.*)") try: name = rgx.match(fullname).group("name") login = rgx.match(fullname).group("login") except AttributeError: raise ValueError("Not a valid name") return login if login else "_", name def make_fullname(login, name): return "{}@{}".format("_" if login is None else login, name)
import re def split_fullname(fullname): rgx = re.compile(r"(?P<login>.*)@(?P<name>.*)") try: name = rgx.match(fullname).group("name") login = rgx.match(fullname).group("login") except AttributeError: raise ValueError("Not a valid name") return login if login else "_", name def make_fullname(login, name): return "{}@{}".format("_" if login is None else login, name) Fix regex for spliting fullnamesimport re def split_fullname(fullname): rgx = re.compile(r"(?P<login>.*)?@(?P<name>.*)") try: name = rgx.match(fullname).group("name") login = rgx.match(fullname).group("login") except AttributeError: raise ValueError("Not a valid name") return login if login else "_", name def make_fullname(login, name): return "{}@{}".format("_" if login is None else login, name)
<commit_before>import re def split_fullname(fullname): rgx = re.compile(r"(?P<login>.*)@(?P<name>.*)") try: name = rgx.match(fullname).group("name") login = rgx.match(fullname).group("login") except AttributeError: raise ValueError("Not a valid name") return login if login else "_", name def make_fullname(login, name): return "{}@{}".format("_" if login is None else login, name) <commit_msg>Fix regex for spliting fullnames<commit_after>import re def split_fullname(fullname): rgx = re.compile(r"(?P<login>.*)?@(?P<name>.*)") try: name = rgx.match(fullname).group("name") login = rgx.match(fullname).group("login") except AttributeError: raise ValueError("Not a valid name") return login if login else "_", name def make_fullname(login, name): return "{}@{}".format("_" if login is None else login, name)
335abda444cbd5651af0d9a298570144627c7022
passwordless/utils.py
passwordless/utils.py
import os import random import uuid from django.contrib.auth.hashers import make_password,is_password_usable from django.utils import timezone from datetime import timedelta WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt') def make_token(): """ Generate a random token suitable for activation/confirmation via email A hex-encoded random UUID has plenty of entropy to be secure enough for our needs. """ return uuid.uuid4().hex def expiration_date(): """ AuthToken objects expire 1 hour after creation by default """ return timezone.now() + timedelta(hours=1) def new_app_password(size=6): f = open(WORDLIST_FILE, 'r') words = [] for i in range(size): words.append(next(f).strip()) for num,line in enumerate(f): j = random.randrange(size+num) if j < size: words[j] = line.strip() return words
import os import random import uuid from django.utils import timezone from datetime import timedelta WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt') def make_token(): """ Generate a random token suitable for activation/confirmation via email A hex-encoded random UUID has plenty of entropy to be secure enough for our needs. """ return uuid.uuid4().hex def expiration_date(): """ AuthToken objects expire 1 hour after creation by default """ return timezone.now() + timedelta(hours=1) def new_app_password(size=6): f = open(WORDLIST_FILE, 'r') words = [] for i in range(size): words.append(next(f).strip()) for num,line in enumerate(f): j = random.randrange(size+num) if j < size: words[j] = line.strip() return ' '.join(words)
Return app passwords as string
Return app passwords as string
Python
mit
Kromey/fbxnano,Kromey/akwriters,Kromey/fbxnano,Kromey/fbxnano,Kromey/akwriters,Kromey/fbxnano,Kromey/akwriters,Kromey/akwriters
import os import random import uuid from django.contrib.auth.hashers import make_password,is_password_usable from django.utils import timezone from datetime import timedelta WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt') def make_token(): """ Generate a random token suitable for activation/confirmation via email A hex-encoded random UUID has plenty of entropy to be secure enough for our needs. """ return uuid.uuid4().hex def expiration_date(): """ AuthToken objects expire 1 hour after creation by default """ return timezone.now() + timedelta(hours=1) def new_app_password(size=6): f = open(WORDLIST_FILE, 'r') words = [] for i in range(size): words.append(next(f).strip()) for num,line in enumerate(f): j = random.randrange(size+num) if j < size: words[j] = line.strip() return words Return app passwords as string
import os import random import uuid from django.utils import timezone from datetime import timedelta WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt') def make_token(): """ Generate a random token suitable for activation/confirmation via email A hex-encoded random UUID has plenty of entropy to be secure enough for our needs. """ return uuid.uuid4().hex def expiration_date(): """ AuthToken objects expire 1 hour after creation by default """ return timezone.now() + timedelta(hours=1) def new_app_password(size=6): f = open(WORDLIST_FILE, 'r') words = [] for i in range(size): words.append(next(f).strip()) for num,line in enumerate(f): j = random.randrange(size+num) if j < size: words[j] = line.strip() return ' '.join(words)
<commit_before>import os import random import uuid from django.contrib.auth.hashers import make_password,is_password_usable from django.utils import timezone from datetime import timedelta WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt') def make_token(): """ Generate a random token suitable for activation/confirmation via email A hex-encoded random UUID has plenty of entropy to be secure enough for our needs. """ return uuid.uuid4().hex def expiration_date(): """ AuthToken objects expire 1 hour after creation by default """ return timezone.now() + timedelta(hours=1) def new_app_password(size=6): f = open(WORDLIST_FILE, 'r') words = [] for i in range(size): words.append(next(f).strip()) for num,line in enumerate(f): j = random.randrange(size+num) if j < size: words[j] = line.strip() return words <commit_msg>Return app passwords as string<commit_after>
import os import random import uuid from django.utils import timezone from datetime import timedelta WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt') def make_token(): """ Generate a random token suitable for activation/confirmation via email A hex-encoded random UUID has plenty of entropy to be secure enough for our needs. """ return uuid.uuid4().hex def expiration_date(): """ AuthToken objects expire 1 hour after creation by default """ return timezone.now() + timedelta(hours=1) def new_app_password(size=6): f = open(WORDLIST_FILE, 'r') words = [] for i in range(size): words.append(next(f).strip()) for num,line in enumerate(f): j = random.randrange(size+num) if j < size: words[j] = line.strip() return ' '.join(words)
import os import random import uuid from django.contrib.auth.hashers import make_password,is_password_usable from django.utils import timezone from datetime import timedelta WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt') def make_token(): """ Generate a random token suitable for activation/confirmation via email A hex-encoded random UUID has plenty of entropy to be secure enough for our needs. """ return uuid.uuid4().hex def expiration_date(): """ AuthToken objects expire 1 hour after creation by default """ return timezone.now() + timedelta(hours=1) def new_app_password(size=6): f = open(WORDLIST_FILE, 'r') words = [] for i in range(size): words.append(next(f).strip()) for num,line in enumerate(f): j = random.randrange(size+num) if j < size: words[j] = line.strip() return words Return app passwords as stringimport os import random import uuid from django.utils import timezone from datetime import timedelta WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt') def make_token(): """ Generate a random token suitable for activation/confirmation via email A hex-encoded random UUID has plenty of entropy to be secure enough for our needs. """ return uuid.uuid4().hex def expiration_date(): """ AuthToken objects expire 1 hour after creation by default """ return timezone.now() + timedelta(hours=1) def new_app_password(size=6): f = open(WORDLIST_FILE, 'r') words = [] for i in range(size): words.append(next(f).strip()) for num,line in enumerate(f): j = random.randrange(size+num) if j < size: words[j] = line.strip() return ' '.join(words)
<commit_before>import os import random import uuid from django.contrib.auth.hashers import make_password,is_password_usable from django.utils import timezone from datetime import timedelta WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt') def make_token(): """ Generate a random token suitable for activation/confirmation via email A hex-encoded random UUID has plenty of entropy to be secure enough for our needs. """ return uuid.uuid4().hex def expiration_date(): """ AuthToken objects expire 1 hour after creation by default """ return timezone.now() + timedelta(hours=1) def new_app_password(size=6): f = open(WORDLIST_FILE, 'r') words = [] for i in range(size): words.append(next(f).strip()) for num,line in enumerate(f): j = random.randrange(size+num) if j < size: words[j] = line.strip() return words <commit_msg>Return app passwords as string<commit_after>import os import random import uuid from django.utils import timezone from datetime import timedelta WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt') def make_token(): """ Generate a random token suitable for activation/confirmation via email A hex-encoded random UUID has plenty of entropy to be secure enough for our needs. """ return uuid.uuid4().hex def expiration_date(): """ AuthToken objects expire 1 hour after creation by default """ return timezone.now() + timedelta(hours=1) def new_app_password(size=6): f = open(WORDLIST_FILE, 'r') words = [] for i in range(size): words.append(next(f).strip()) for num,line in enumerate(f): j = random.randrange(size+num) if j < size: words[j] = line.strip() return ' '.join(words)
551d86f64e1dadf54a4c63b633af6523dd5cdc05
urbansim/utils/logutil.py
urbansim/utils/logutil.py
import contextlib import logging @contextlib.contextmanager def log_start_finish(msg, logger, level=logging.DEBUG): """ A context manager to log messages with "start: " and "finish: " prefixes before and after a block. Parameters ---------- msg : str Will be prefixed with "start: " and "finish: ". logger : logging.Logger level : int, optional Level at which to log, passed to ``logger.log``. """ logger.log(level, 'start: ' + msg) yield logger.log(level, 'finish: ' + msg)
import contextlib import logging US_LOG_FMT = ('%(asctime)s|%(levelname)s|%(name)s|' '%(funcName)s|%(filename)s|%(lineno)s|%(message)s') US_LOG_DATE_FMT = '%Y-%m-%d %H:%M:%S' US_FMT = logging.Formatter(fmt=US_LOG_FMT, datefmt=US_LOG_DATE_FMT) @contextlib.contextmanager def log_start_finish(msg, logger, level=logging.DEBUG): """ A context manager to log messages with "start: " and "finish: " prefixes before and after a block. Parameters ---------- msg : str Will be prefixed with "start: " and "finish: ". logger : logging.Logger level : int, optional Level at which to log, passed to ``logger.log``. """ logger.log(level, 'start: ' + msg) yield logger.log(level, 'finish: ' + msg) def set_log_level(level): """ Set the logging level for urbansim. Parameters ---------- level : int A supporting logging level. Use logging constants like logging.DEBUG. """ logging.getLogger('urbansim').setLevel(level) def log_to_stream(level=None): """ Send log messages to the console. """ handler = logging.StreamHandler() handler.setFormatter(US_FMT) if level is not None: handler.setLevel(level) logger = logging.getLogger() logger.addHandler(handler) def log_to_file(filename, level=None): """ Send log output to the given file. Parameters ---------- filename : str level : int, optional Optional logging level for the file handler. """ handler = logging.FileHandler(filename) handler.setFormatter(US_FMT) if level is not None: handler.setLevel(level) logger = logging.getLogger() logger.addHandler(handler)
Add utilities for controlling urbansim logging.
Add utilities for controlling urbansim logging.
Python
bsd-3-clause
UDST/urbansim,waddell/urbansim,waddell/urbansim,ual/urbansim,UDST/urbansim,VladimirTyrin/urbansim,waddell/urbansim,ual/urbansim,AZMAG/urbansim,SANDAG/urbansim,synthicity/urbansim,SANDAG/urbansim,bricegnichols/urbansim,synthicity/urbansim,SANDAG/urbansim,AZMAG/urbansim,synthicity/urbansim,apdjustino/urbansim,ual/urbansim,bricegnichols/urbansim,AZMAG/urbansim,apdjustino/urbansim,apdjustino/urbansim,ual/urbansim,SANDAG/urbansim,synthicity/urbansim,bricegnichols/urbansim,VladimirTyrin/urbansim,AZMAG/urbansim,VladimirTyrin/urbansim,waddell/urbansim,VladimirTyrin/urbansim,bricegnichols/urbansim,UDST/urbansim,apdjustino/urbansim,UDST/urbansim
import contextlib import logging @contextlib.contextmanager def log_start_finish(msg, logger, level=logging.DEBUG): """ A context manager to log messages with "start: " and "finish: " prefixes before and after a block. Parameters ---------- msg : str Will be prefixed with "start: " and "finish: ". logger : logging.Logger level : int, optional Level at which to log, passed to ``logger.log``. """ logger.log(level, 'start: ' + msg) yield logger.log(level, 'finish: ' + msg) Add utilities for controlling urbansim logging.
import contextlib import logging US_LOG_FMT = ('%(asctime)s|%(levelname)s|%(name)s|' '%(funcName)s|%(filename)s|%(lineno)s|%(message)s') US_LOG_DATE_FMT = '%Y-%m-%d %H:%M:%S' US_FMT = logging.Formatter(fmt=US_LOG_FMT, datefmt=US_LOG_DATE_FMT) @contextlib.contextmanager def log_start_finish(msg, logger, level=logging.DEBUG): """ A context manager to log messages with "start: " and "finish: " prefixes before and after a block. Parameters ---------- msg : str Will be prefixed with "start: " and "finish: ". logger : logging.Logger level : int, optional Level at which to log, passed to ``logger.log``. """ logger.log(level, 'start: ' + msg) yield logger.log(level, 'finish: ' + msg) def set_log_level(level): """ Set the logging level for urbansim. Parameters ---------- level : int A supporting logging level. Use logging constants like logging.DEBUG. """ logging.getLogger('urbansim').setLevel(level) def log_to_stream(level=None): """ Send log messages to the console. """ handler = logging.StreamHandler() handler.setFormatter(US_FMT) if level is not None: handler.setLevel(level) logger = logging.getLogger() logger.addHandler(handler) def log_to_file(filename, level=None): """ Send log output to the given file. Parameters ---------- filename : str level : int, optional Optional logging level for the file handler. """ handler = logging.FileHandler(filename) handler.setFormatter(US_FMT) if level is not None: handler.setLevel(level) logger = logging.getLogger() logger.addHandler(handler)
<commit_before>import contextlib import logging @contextlib.contextmanager def log_start_finish(msg, logger, level=logging.DEBUG): """ A context manager to log messages with "start: " and "finish: " prefixes before and after a block. Parameters ---------- msg : str Will be prefixed with "start: " and "finish: ". logger : logging.Logger level : int, optional Level at which to log, passed to ``logger.log``. """ logger.log(level, 'start: ' + msg) yield logger.log(level, 'finish: ' + msg) <commit_msg>Add utilities for controlling urbansim logging.<commit_after>
import contextlib import logging US_LOG_FMT = ('%(asctime)s|%(levelname)s|%(name)s|' '%(funcName)s|%(filename)s|%(lineno)s|%(message)s') US_LOG_DATE_FMT = '%Y-%m-%d %H:%M:%S' US_FMT = logging.Formatter(fmt=US_LOG_FMT, datefmt=US_LOG_DATE_FMT) @contextlib.contextmanager def log_start_finish(msg, logger, level=logging.DEBUG): """ A context manager to log messages with "start: " and "finish: " prefixes before and after a block. Parameters ---------- msg : str Will be prefixed with "start: " and "finish: ". logger : logging.Logger level : int, optional Level at which to log, passed to ``logger.log``. """ logger.log(level, 'start: ' + msg) yield logger.log(level, 'finish: ' + msg) def set_log_level(level): """ Set the logging level for urbansim. Parameters ---------- level : int A supporting logging level. Use logging constants like logging.DEBUG. """ logging.getLogger('urbansim').setLevel(level) def log_to_stream(level=None): """ Send log messages to the console. """ handler = logging.StreamHandler() handler.setFormatter(US_FMT) if level is not None: handler.setLevel(level) logger = logging.getLogger() logger.addHandler(handler) def log_to_file(filename, level=None): """ Send log output to the given file. Parameters ---------- filename : str level : int, optional Optional logging level for the file handler. """ handler = logging.FileHandler(filename) handler.setFormatter(US_FMT) if level is not None: handler.setLevel(level) logger = logging.getLogger() logger.addHandler(handler)
import contextlib import logging @contextlib.contextmanager def log_start_finish(msg, logger, level=logging.DEBUG): """ A context manager to log messages with "start: " and "finish: " prefixes before and after a block. Parameters ---------- msg : str Will be prefixed with "start: " and "finish: ". logger : logging.Logger level : int, optional Level at which to log, passed to ``logger.log``. """ logger.log(level, 'start: ' + msg) yield logger.log(level, 'finish: ' + msg) Add utilities for controlling urbansim logging.import contextlib import logging US_LOG_FMT = ('%(asctime)s|%(levelname)s|%(name)s|' '%(funcName)s|%(filename)s|%(lineno)s|%(message)s') US_LOG_DATE_FMT = '%Y-%m-%d %H:%M:%S' US_FMT = logging.Formatter(fmt=US_LOG_FMT, datefmt=US_LOG_DATE_FMT) @contextlib.contextmanager def log_start_finish(msg, logger, level=logging.DEBUG): """ A context manager to log messages with "start: " and "finish: " prefixes before and after a block. Parameters ---------- msg : str Will be prefixed with "start: " and "finish: ". logger : logging.Logger level : int, optional Level at which to log, passed to ``logger.log``. """ logger.log(level, 'start: ' + msg) yield logger.log(level, 'finish: ' + msg) def set_log_level(level): """ Set the logging level for urbansim. Parameters ---------- level : int A supporting logging level. Use logging constants like logging.DEBUG. """ logging.getLogger('urbansim').setLevel(level) def log_to_stream(level=None): """ Send log messages to the console. """ handler = logging.StreamHandler() handler.setFormatter(US_FMT) if level is not None: handler.setLevel(level) logger = logging.getLogger() logger.addHandler(handler) def log_to_file(filename, level=None): """ Send log output to the given file. Parameters ---------- filename : str level : int, optional Optional logging level for the file handler. """ handler = logging.FileHandler(filename) handler.setFormatter(US_FMT) if level is not None: handler.setLevel(level) logger = logging.getLogger() logger.addHandler(handler)
<commit_before>import contextlib import logging @contextlib.contextmanager def log_start_finish(msg, logger, level=logging.DEBUG): """ A context manager to log messages with "start: " and "finish: " prefixes before and after a block. Parameters ---------- msg : str Will be prefixed with "start: " and "finish: ". logger : logging.Logger level : int, optional Level at which to log, passed to ``logger.log``. """ logger.log(level, 'start: ' + msg) yield logger.log(level, 'finish: ' + msg) <commit_msg>Add utilities for controlling urbansim logging.<commit_after>import contextlib import logging US_LOG_FMT = ('%(asctime)s|%(levelname)s|%(name)s|' '%(funcName)s|%(filename)s|%(lineno)s|%(message)s') US_LOG_DATE_FMT = '%Y-%m-%d %H:%M:%S' US_FMT = logging.Formatter(fmt=US_LOG_FMT, datefmt=US_LOG_DATE_FMT) @contextlib.contextmanager def log_start_finish(msg, logger, level=logging.DEBUG): """ A context manager to log messages with "start: " and "finish: " prefixes before and after a block. Parameters ---------- msg : str Will be prefixed with "start: " and "finish: ". logger : logging.Logger level : int, optional Level at which to log, passed to ``logger.log``. """ logger.log(level, 'start: ' + msg) yield logger.log(level, 'finish: ' + msg) def set_log_level(level): """ Set the logging level for urbansim. Parameters ---------- level : int A supporting logging level. Use logging constants like logging.DEBUG. """ logging.getLogger('urbansim').setLevel(level) def log_to_stream(level=None): """ Send log messages to the console. """ handler = logging.StreamHandler() handler.setFormatter(US_FMT) if level is not None: handler.setLevel(level) logger = logging.getLogger() logger.addHandler(handler) def log_to_file(filename, level=None): """ Send log output to the given file. Parameters ---------- filename : str level : int, optional Optional logging level for the file handler. """ handler = logging.FileHandler(filename) handler.setFormatter(US_FMT) if level is not None: handler.setLevel(level) logger = logging.getLogger() logger.addHandler(handler)
57b375d7bab3b88137b2ef5d6b0c38056b758a48
Mscthesis/IO/municipios_parser.py
Mscthesis/IO/municipios_parser.py
""" municipios_parser ----------------- Module which contains the process of parsing data of municipios. TODO ---- """ import pandas as pd from pythonUtils.ProcessTools import Processer class Municipios_Parser(Processer): """This class is the one which controls the parsing process of municipios information. """ indices = None files = '' def __init__(self, logfile, bool_inform=False): "Instantiation of the class remembering it is a subclass of Processer." self.proc_name = "Municipios parser" self.proc_desc = "Parser the standarize data from file" self.subproc_desc = [] self.t_expended_subproc = [] self.logfile = logfile def parse(self, filepath): "Parse the data from the file given." data = pd.read_csv(filepath, sep=';', index_col=0) typevars = {} typevars['pop_vars'] = ['Poblacion', "Superficie", "Densidad"] typevars['loc_vars'] = ["longitud", "latitud"] return data, typevars
""" municipios_parser ----------------- Module which contains the process of parsing data of municipios. TODO ---- """ import pandas as pd from pythonUtils.ProcessTools import Processer class Municipios_Parser(Processer): """This class is the one which controls the parsing process of municipios information. """ indices = None files = '' def __init__(self, logfile, bool_inform=False): "Instantiation of the class remembering it is a subclass of Processer." self.proc_name = "Municipios parser" self.proc_desc = "Parser the standarize data from file" self.subproc_desc = [] self.t_expended_subproc = [] self.logfile = logfile def parse(self, filepath): "Parse the data from the file given." data = pd.read_csv(filepath, sep=';', index_col=0) typevars = {} typevars['feat_vars'] = ['Poblacion', "Superficie", "Densidad"] typevars['loc_vars'] = ["longitud", "latitud"] return data, typevars
Change in the typ output.
Change in the typ output.
Python
mit
tgquintela/Mscthesis
""" municipios_parser ----------------- Module which contains the process of parsing data of municipios. TODO ---- """ import pandas as pd from pythonUtils.ProcessTools import Processer class Municipios_Parser(Processer): """This class is the one which controls the parsing process of municipios information. """ indices = None files = '' def __init__(self, logfile, bool_inform=False): "Instantiation of the class remembering it is a subclass of Processer." self.proc_name = "Municipios parser" self.proc_desc = "Parser the standarize data from file" self.subproc_desc = [] self.t_expended_subproc = [] self.logfile = logfile def parse(self, filepath): "Parse the data from the file given." data = pd.read_csv(filepath, sep=';', index_col=0) typevars = {} typevars['pop_vars'] = ['Poblacion', "Superficie", "Densidad"] typevars['loc_vars'] = ["longitud", "latitud"] return data, typevars Change in the typ output.
""" municipios_parser ----------------- Module which contains the process of parsing data of municipios. TODO ---- """ import pandas as pd from pythonUtils.ProcessTools import Processer class Municipios_Parser(Processer): """This class is the one which controls the parsing process of municipios information. """ indices = None files = '' def __init__(self, logfile, bool_inform=False): "Instantiation of the class remembering it is a subclass of Processer." self.proc_name = "Municipios parser" self.proc_desc = "Parser the standarize data from file" self.subproc_desc = [] self.t_expended_subproc = [] self.logfile = logfile def parse(self, filepath): "Parse the data from the file given." data = pd.read_csv(filepath, sep=';', index_col=0) typevars = {} typevars['feat_vars'] = ['Poblacion', "Superficie", "Densidad"] typevars['loc_vars'] = ["longitud", "latitud"] return data, typevars
<commit_before> """ municipios_parser ----------------- Module which contains the process of parsing data of municipios. TODO ---- """ import pandas as pd from pythonUtils.ProcessTools import Processer class Municipios_Parser(Processer): """This class is the one which controls the parsing process of municipios information. """ indices = None files = '' def __init__(self, logfile, bool_inform=False): "Instantiation of the class remembering it is a subclass of Processer." self.proc_name = "Municipios parser" self.proc_desc = "Parser the standarize data from file" self.subproc_desc = [] self.t_expended_subproc = [] self.logfile = logfile def parse(self, filepath): "Parse the data from the file given." data = pd.read_csv(filepath, sep=';', index_col=0) typevars = {} typevars['pop_vars'] = ['Poblacion', "Superficie", "Densidad"] typevars['loc_vars'] = ["longitud", "latitud"] return data, typevars <commit_msg>Change in the typ output.<commit_after>
""" municipios_parser ----------------- Module which contains the process of parsing data of municipios. TODO ---- """ import pandas as pd from pythonUtils.ProcessTools import Processer class Municipios_Parser(Processer): """This class is the one which controls the parsing process of municipios information. """ indices = None files = '' def __init__(self, logfile, bool_inform=False): "Instantiation of the class remembering it is a subclass of Processer." self.proc_name = "Municipios parser" self.proc_desc = "Parser the standarize data from file" self.subproc_desc = [] self.t_expended_subproc = [] self.logfile = logfile def parse(self, filepath): "Parse the data from the file given." data = pd.read_csv(filepath, sep=';', index_col=0) typevars = {} typevars['feat_vars'] = ['Poblacion', "Superficie", "Densidad"] typevars['loc_vars'] = ["longitud", "latitud"] return data, typevars
""" municipios_parser ----------------- Module which contains the process of parsing data of municipios. TODO ---- """ import pandas as pd from pythonUtils.ProcessTools import Processer class Municipios_Parser(Processer): """This class is the one which controls the parsing process of municipios information. """ indices = None files = '' def __init__(self, logfile, bool_inform=False): "Instantiation of the class remembering it is a subclass of Processer." self.proc_name = "Municipios parser" self.proc_desc = "Parser the standarize data from file" self.subproc_desc = [] self.t_expended_subproc = [] self.logfile = logfile def parse(self, filepath): "Parse the data from the file given." data = pd.read_csv(filepath, sep=';', index_col=0) typevars = {} typevars['pop_vars'] = ['Poblacion', "Superficie", "Densidad"] typevars['loc_vars'] = ["longitud", "latitud"] return data, typevars Change in the typ output. """ municipios_parser ----------------- Module which contains the process of parsing data of municipios. TODO ---- """ import pandas as pd from pythonUtils.ProcessTools import Processer class Municipios_Parser(Processer): """This class is the one which controls the parsing process of municipios information. """ indices = None files = '' def __init__(self, logfile, bool_inform=False): "Instantiation of the class remembering it is a subclass of Processer." self.proc_name = "Municipios parser" self.proc_desc = "Parser the standarize data from file" self.subproc_desc = [] self.t_expended_subproc = [] self.logfile = logfile def parse(self, filepath): "Parse the data from the file given." data = pd.read_csv(filepath, sep=';', index_col=0) typevars = {} typevars['feat_vars'] = ['Poblacion', "Superficie", "Densidad"] typevars['loc_vars'] = ["longitud", "latitud"] return data, typevars
<commit_before> """ municipios_parser ----------------- Module which contains the process of parsing data of municipios. TODO ---- """ import pandas as pd from pythonUtils.ProcessTools import Processer class Municipios_Parser(Processer): """This class is the one which controls the parsing process of municipios information. """ indices = None files = '' def __init__(self, logfile, bool_inform=False): "Instantiation of the class remembering it is a subclass of Processer." self.proc_name = "Municipios parser" self.proc_desc = "Parser the standarize data from file" self.subproc_desc = [] self.t_expended_subproc = [] self.logfile = logfile def parse(self, filepath): "Parse the data from the file given." data = pd.read_csv(filepath, sep=';', index_col=0) typevars = {} typevars['pop_vars'] = ['Poblacion', "Superficie", "Densidad"] typevars['loc_vars'] = ["longitud", "latitud"] return data, typevars <commit_msg>Change in the typ output.<commit_after> """ municipios_parser ----------------- Module which contains the process of parsing data of municipios. TODO ---- """ import pandas as pd from pythonUtils.ProcessTools import Processer class Municipios_Parser(Processer): """This class is the one which controls the parsing process of municipios information. """ indices = None files = '' def __init__(self, logfile, bool_inform=False): "Instantiation of the class remembering it is a subclass of Processer." self.proc_name = "Municipios parser" self.proc_desc = "Parser the standarize data from file" self.subproc_desc = [] self.t_expended_subproc = [] self.logfile = logfile def parse(self, filepath): "Parse the data from the file given." data = pd.read_csv(filepath, sep=';', index_col=0) typevars = {} typevars['feat_vars'] = ['Poblacion', "Superficie", "Densidad"] typevars['loc_vars'] = ["longitud", "latitud"] return data, typevars
4854015a61f0b582065b0d5561df231314abcce1
django_redux_generator/management/commands/redux_generator.py
django_redux_generator/management/commands/redux_generator.py
from django.core.management.base import BaseCommand from django.template.loader import render_to_string class Command(BaseCommand): help = 'Generate redux boilerplate' def add_arguments(self, parser): parser.add_argument('action_name', type=str) parser.add_argument('--thunk', action='store_true', dest='thunk', default=False, help='Generate a redux thunk') def handle(self, *args, **options): if options['thunk']: template_name = 'django_redux_generator/thunk_fetch.js' else: template_name = 'django_redux_generator/action_creator.js' print(render_to_string(template_name, { 'action_name': options['action_name'], }))
from django.core.management.base import BaseCommand from django.template.loader import render_to_string class Command(BaseCommand): help = 'Generate redux boilerplate' def add_arguments(self, parser): parser.add_argument('action_name', type=str) parser.add_argument('--thunk', action='store_true', dest='thunk', default=False, help='Generate a redux thunk') def handle(self, *args, **options): if options['thunk']: template_name = 'django_redux_generator/thunk_fetch.js' else: template_name = 'django_redux_generator/action_creator.js' return render_to_string(template_name, { 'action_name': options['action_name'], })
Return the output rather than print
Return the output rather than print
Python
mit
rapilabs/django-redux-generator,rapilabs/django-redux-generator
from django.core.management.base import BaseCommand from django.template.loader import render_to_string class Command(BaseCommand): help = 'Generate redux boilerplate' def add_arguments(self, parser): parser.add_argument('action_name', type=str) parser.add_argument('--thunk', action='store_true', dest='thunk', default=False, help='Generate a redux thunk') def handle(self, *args, **options): if options['thunk']: template_name = 'django_redux_generator/thunk_fetch.js' else: template_name = 'django_redux_generator/action_creator.js' print(render_to_string(template_name, { 'action_name': options['action_name'], })) Return the output rather than print
from django.core.management.base import BaseCommand from django.template.loader import render_to_string class Command(BaseCommand): help = 'Generate redux boilerplate' def add_arguments(self, parser): parser.add_argument('action_name', type=str) parser.add_argument('--thunk', action='store_true', dest='thunk', default=False, help='Generate a redux thunk') def handle(self, *args, **options): if options['thunk']: template_name = 'django_redux_generator/thunk_fetch.js' else: template_name = 'django_redux_generator/action_creator.js' return render_to_string(template_name, { 'action_name': options['action_name'], })
<commit_before>from django.core.management.base import BaseCommand from django.template.loader import render_to_string class Command(BaseCommand): help = 'Generate redux boilerplate' def add_arguments(self, parser): parser.add_argument('action_name', type=str) parser.add_argument('--thunk', action='store_true', dest='thunk', default=False, help='Generate a redux thunk') def handle(self, *args, **options): if options['thunk']: template_name = 'django_redux_generator/thunk_fetch.js' else: template_name = 'django_redux_generator/action_creator.js' print(render_to_string(template_name, { 'action_name': options['action_name'], })) <commit_msg>Return the output rather than print<commit_after>
from django.core.management.base import BaseCommand from django.template.loader import render_to_string class Command(BaseCommand): help = 'Generate redux boilerplate' def add_arguments(self, parser): parser.add_argument('action_name', type=str) parser.add_argument('--thunk', action='store_true', dest='thunk', default=False, help='Generate a redux thunk') def handle(self, *args, **options): if options['thunk']: template_name = 'django_redux_generator/thunk_fetch.js' else: template_name = 'django_redux_generator/action_creator.js' return render_to_string(template_name, { 'action_name': options['action_name'], })
from django.core.management.base import BaseCommand from django.template.loader import render_to_string class Command(BaseCommand): help = 'Generate redux boilerplate' def add_arguments(self, parser): parser.add_argument('action_name', type=str) parser.add_argument('--thunk', action='store_true', dest='thunk', default=False, help='Generate a redux thunk') def handle(self, *args, **options): if options['thunk']: template_name = 'django_redux_generator/thunk_fetch.js' else: template_name = 'django_redux_generator/action_creator.js' print(render_to_string(template_name, { 'action_name': options['action_name'], })) Return the output rather than printfrom django.core.management.base import BaseCommand from django.template.loader import render_to_string class Command(BaseCommand): help = 'Generate redux boilerplate' def add_arguments(self, parser): parser.add_argument('action_name', type=str) parser.add_argument('--thunk', action='store_true', dest='thunk', default=False, help='Generate a redux thunk') def handle(self, *args, **options): if options['thunk']: template_name = 'django_redux_generator/thunk_fetch.js' else: template_name = 'django_redux_generator/action_creator.js' return render_to_string(template_name, { 'action_name': options['action_name'], })
<commit_before>from django.core.management.base import BaseCommand from django.template.loader import render_to_string class Command(BaseCommand): help = 'Generate redux boilerplate' def add_arguments(self, parser): parser.add_argument('action_name', type=str) parser.add_argument('--thunk', action='store_true', dest='thunk', default=False, help='Generate a redux thunk') def handle(self, *args, **options): if options['thunk']: template_name = 'django_redux_generator/thunk_fetch.js' else: template_name = 'django_redux_generator/action_creator.js' print(render_to_string(template_name, { 'action_name': options['action_name'], })) <commit_msg>Return the output rather than print<commit_after>from django.core.management.base import BaseCommand from django.template.loader import render_to_string class Command(BaseCommand): help = 'Generate redux boilerplate' def add_arguments(self, parser): parser.add_argument('action_name', type=str) parser.add_argument('--thunk', action='store_true', dest='thunk', default=False, help='Generate a redux thunk') def handle(self, *args, **options): if options['thunk']: template_name = 'django_redux_generator/thunk_fetch.js' else: template_name = 'django_redux_generator/action_creator.js' return render_to_string(template_name, { 'action_name': options['action_name'], })
d8ba1531b2e0faa71c57e8970af471ec2caa4a18
en-2014-06-21-unit-testing-with-unittest-mock-patch/chdir2.py
en-2014-06-21-unit-testing-with-unittest-mock-patch/chdir2.py
""" chdir2 ~~~~~~ An alternative implementation of :func:`chdir.chdir`. :copyright: © 2014 by Petr Zemek <s3rvac@gmail.com> :license: BSD, see LICENSE for more details """ import os class chdir2(): """An alternative implementation of :func:`chdir.chdir`.""" def __init__(self, dir): self.dir = dir def __enter__(self): self.orig_cwd = os.getcwd() os.chdir(self.dir) def __exit__(self, *exc_info): os.chdir(self.orig_cwd)
""" chdir2 ~~~~~~ An alternative implementation of :func:`chdir.chdir()`. :copyright: © 2014 by Petr Zemek <s3rvac@gmail.com> :license: BSD, see LICENSE for more details """ import os class chdir2(): """An alternative implementation of :func:`chdir.chdir()`.""" def __init__(self, dir): self.dir = dir def __enter__(self): self.orig_cwd = os.getcwd() os.chdir(self.dir) def __exit__(self, *exc_info): os.chdir(self.orig_cwd)
Add missing parentheses after 'chdir'.
blog/en-2014-06-21: Add missing parentheses after 'chdir'. It is a function, so we better add parentheses to make this clearer.
Python
bsd-3-clause
s3rvac/blog,s3rvac/blog,s3rvac/blog,s3rvac/blog
""" chdir2 ~~~~~~ An alternative implementation of :func:`chdir.chdir`. :copyright: © 2014 by Petr Zemek <s3rvac@gmail.com> :license: BSD, see LICENSE for more details """ import os class chdir2(): """An alternative implementation of :func:`chdir.chdir`.""" def __init__(self, dir): self.dir = dir def __enter__(self): self.orig_cwd = os.getcwd() os.chdir(self.dir) def __exit__(self, *exc_info): os.chdir(self.orig_cwd) blog/en-2014-06-21: Add missing parentheses after 'chdir'. It is a function, so we better add parentheses to make this clearer.
""" chdir2 ~~~~~~ An alternative implementation of :func:`chdir.chdir()`. :copyright: © 2014 by Petr Zemek <s3rvac@gmail.com> :license: BSD, see LICENSE for more details """ import os class chdir2(): """An alternative implementation of :func:`chdir.chdir()`.""" def __init__(self, dir): self.dir = dir def __enter__(self): self.orig_cwd = os.getcwd() os.chdir(self.dir) def __exit__(self, *exc_info): os.chdir(self.orig_cwd)
<commit_before>""" chdir2 ~~~~~~ An alternative implementation of :func:`chdir.chdir`. :copyright: © 2014 by Petr Zemek <s3rvac@gmail.com> :license: BSD, see LICENSE for more details """ import os class chdir2(): """An alternative implementation of :func:`chdir.chdir`.""" def __init__(self, dir): self.dir = dir def __enter__(self): self.orig_cwd = os.getcwd() os.chdir(self.dir) def __exit__(self, *exc_info): os.chdir(self.orig_cwd) <commit_msg>blog/en-2014-06-21: Add missing parentheses after 'chdir'. It is a function, so we better add parentheses to make this clearer.<commit_after>
""" chdir2 ~~~~~~ An alternative implementation of :func:`chdir.chdir()`. :copyright: © 2014 by Petr Zemek <s3rvac@gmail.com> :license: BSD, see LICENSE for more details """ import os class chdir2(): """An alternative implementation of :func:`chdir.chdir()`.""" def __init__(self, dir): self.dir = dir def __enter__(self): self.orig_cwd = os.getcwd() os.chdir(self.dir) def __exit__(self, *exc_info): os.chdir(self.orig_cwd)
""" chdir2 ~~~~~~ An alternative implementation of :func:`chdir.chdir`. :copyright: © 2014 by Petr Zemek <s3rvac@gmail.com> :license: BSD, see LICENSE for more details """ import os class chdir2(): """An alternative implementation of :func:`chdir.chdir`.""" def __init__(self, dir): self.dir = dir def __enter__(self): self.orig_cwd = os.getcwd() os.chdir(self.dir) def __exit__(self, *exc_info): os.chdir(self.orig_cwd) blog/en-2014-06-21: Add missing parentheses after 'chdir'. It is a function, so we better add parentheses to make this clearer.""" chdir2 ~~~~~~ An alternative implementation of :func:`chdir.chdir()`. :copyright: © 2014 by Petr Zemek <s3rvac@gmail.com> :license: BSD, see LICENSE for more details """ import os class chdir2(): """An alternative implementation of :func:`chdir.chdir()`.""" def __init__(self, dir): self.dir = dir def __enter__(self): self.orig_cwd = os.getcwd() os.chdir(self.dir) def __exit__(self, *exc_info): os.chdir(self.orig_cwd)
<commit_before>""" chdir2 ~~~~~~ An alternative implementation of :func:`chdir.chdir`. :copyright: © 2014 by Petr Zemek <s3rvac@gmail.com> :license: BSD, see LICENSE for more details """ import os class chdir2(): """An alternative implementation of :func:`chdir.chdir`.""" def __init__(self, dir): self.dir = dir def __enter__(self): self.orig_cwd = os.getcwd() os.chdir(self.dir) def __exit__(self, *exc_info): os.chdir(self.orig_cwd) <commit_msg>blog/en-2014-06-21: Add missing parentheses after 'chdir'. It is a function, so we better add parentheses to make this clearer.<commit_after>""" chdir2 ~~~~~~ An alternative implementation of :func:`chdir.chdir()`. :copyright: © 2014 by Petr Zemek <s3rvac@gmail.com> :license: BSD, see LICENSE for more details """ import os class chdir2(): """An alternative implementation of :func:`chdir.chdir()`.""" def __init__(self, dir): self.dir = dir def __enter__(self): self.orig_cwd = os.getcwd() os.chdir(self.dir) def __exit__(self, *exc_info): os.chdir(self.orig_cwd)
f59919efefb78fffff564ec17c55f6df644e8d7e
server/lib/python/cartodb_services/cartodb_services/here/__init__.py
server/lib/python/cartodb_services/cartodb_services/here/__init__.py
from cartodb_services.here.geocoder import HereMapsGeocoder from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder from cartodb_services.here.routing import HereMapsRoutingIsoline
from cartodb_services.here.geocoder import HereMapsGeocoder, HereMapsGeocoderV7 from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder, HereMapsBulkGeocoderV7 from cartodb_services.here.service_factory import get_geocoder, get_bulk_geocoder, get_routing_isoline from cartodb_services.here.routing import HereMapsRoutingIsoline
Add new imports for apikey parameter support
Add new imports for apikey parameter support
Python
bsd-3-clause
CartoDB/dataservices-api,CartoDB/dataservices-api,CartoDB/dataservices-api,CartoDB/dataservices-api,CartoDB/geocoder-api,CartoDB/geocoder-api,CartoDB/geocoder-api,CartoDB/geocoder-api
from cartodb_services.here.geocoder import HereMapsGeocoder from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder from cartodb_services.here.routing import HereMapsRoutingIsoline Add new imports for apikey parameter support
from cartodb_services.here.geocoder import HereMapsGeocoder, HereMapsGeocoderV7 from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder, HereMapsBulkGeocoderV7 from cartodb_services.here.service_factory import get_geocoder, get_bulk_geocoder, get_routing_isoline from cartodb_services.here.routing import HereMapsRoutingIsoline
<commit_before>from cartodb_services.here.geocoder import HereMapsGeocoder from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder from cartodb_services.here.routing import HereMapsRoutingIsoline <commit_msg>Add new imports for apikey parameter support<commit_after>
from cartodb_services.here.geocoder import HereMapsGeocoder, HereMapsGeocoderV7 from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder, HereMapsBulkGeocoderV7 from cartodb_services.here.service_factory import get_geocoder, get_bulk_geocoder, get_routing_isoline from cartodb_services.here.routing import HereMapsRoutingIsoline
from cartodb_services.here.geocoder import HereMapsGeocoder from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder from cartodb_services.here.routing import HereMapsRoutingIsoline Add new imports for apikey parameter supportfrom cartodb_services.here.geocoder import HereMapsGeocoder, HereMapsGeocoderV7 from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder, HereMapsBulkGeocoderV7 from cartodb_services.here.service_factory import get_geocoder, get_bulk_geocoder, get_routing_isoline from cartodb_services.here.routing import HereMapsRoutingIsoline
<commit_before>from cartodb_services.here.geocoder import HereMapsGeocoder from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder from cartodb_services.here.routing import HereMapsRoutingIsoline <commit_msg>Add new imports for apikey parameter support<commit_after>from cartodb_services.here.geocoder import HereMapsGeocoder, HereMapsGeocoderV7 from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder, HereMapsBulkGeocoderV7 from cartodb_services.here.service_factory import get_geocoder, get_bulk_geocoder, get_routing_isoline from cartodb_services.here.routing import HereMapsRoutingIsoline
a6405ccfc7f53f601088206c216c5167fd86359f
symposion/teams/backends.py
symposion/teams/backends.py
from django.db.models import Q from .models import Team class TeamPermissionsBackend(object): def authenticate(self, username=None, password=None): return None def get_team_permissions(self, user_obj, obj=None): """ Returns a set of permission strings that this user has through his/her team memberships. """ if user_obj.is_anonymous() or obj is not None: return set() if not hasattr(user_obj, "_team_perm_cache"): memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="manager") | Q(memberships__state="member"), ) perms = memberships.values_list( "permissions__content_type__app_label", "permissions__codename" ).order_by() user_obj._team_perm_cache = set(["%s.%s" % (ct, name) for ct, name in perms]) return user_obj._team_perm_cache def has_perm(self, user_obj, perm, obj=None): if not user_obj.is_active: return False return perm in self.get_team_permissions(user_obj, obj)
from django.db.models import Q from .models import Team class TeamPermissionsBackend(object): def authenticate(self, username=None, password=None): return None def get_team_permissions(self, user_obj, obj=None): """ Returns a set of permission strings that this user has through his/her team memberships. """ if user_obj.is_anonymous() or obj is not None: return set() if not hasattr(user_obj, "_team_perm_cache"): # Member permissions memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="member"), ) perms = memberships.values_list( "permissions__content_type__app_label", "permissions__codename" ).order_by() permissions = ["%s.%s" % (ct, name) for ct, name in perms] # Manager permissions memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="manager"), ) perms = memberships.values_list( "manager_permissions__content_type__app_label", "manager_permissions__codename" ).order_by() permissions += ["%s.%s" % (ct, name) for ct, name in perms] user_obj._team_perm_cache = set(permissions) return user_obj._team_perm_cache def has_perm(self, user_obj, perm, obj=None): if not user_obj.is_active: return False return perm in self.get_team_permissions(user_obj, obj)
Fix team permissions backend not pulling out manager_permissions
Fix team permissions backend not pulling out manager_permissions Something like request.user.has_perm('reviews.can_manage_%s' % proposal.kind.section.slug) Will aways return false as the backend does a lookup of team membership (member or manager) but only grabs the 'permissions' and not the 'manager_permissions' field
Python
bsd-3-clause
pyconau2017/symposion,pyconau2017/symposion
from django.db.models import Q from .models import Team class TeamPermissionsBackend(object): def authenticate(self, username=None, password=None): return None def get_team_permissions(self, user_obj, obj=None): """ Returns a set of permission strings that this user has through his/her team memberships. """ if user_obj.is_anonymous() or obj is not None: return set() if not hasattr(user_obj, "_team_perm_cache"): memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="manager") | Q(memberships__state="member"), ) perms = memberships.values_list( "permissions__content_type__app_label", "permissions__codename" ).order_by() user_obj._team_perm_cache = set(["%s.%s" % (ct, name) for ct, name in perms]) return user_obj._team_perm_cache def has_perm(self, user_obj, perm, obj=None): if not user_obj.is_active: return False return perm in self.get_team_permissions(user_obj, obj) Fix team permissions backend not pulling out manager_permissions Something like request.user.has_perm('reviews.can_manage_%s' % proposal.kind.section.slug) Will aways return false as the backend does a lookup of team membership (member or manager) but only grabs the 'permissions' and not the 'manager_permissions' field
from django.db.models import Q from .models import Team class TeamPermissionsBackend(object): def authenticate(self, username=None, password=None): return None def get_team_permissions(self, user_obj, obj=None): """ Returns a set of permission strings that this user has through his/her team memberships. """ if user_obj.is_anonymous() or obj is not None: return set() if not hasattr(user_obj, "_team_perm_cache"): # Member permissions memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="member"), ) perms = memberships.values_list( "permissions__content_type__app_label", "permissions__codename" ).order_by() permissions = ["%s.%s" % (ct, name) for ct, name in perms] # Manager permissions memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="manager"), ) perms = memberships.values_list( "manager_permissions__content_type__app_label", "manager_permissions__codename" ).order_by() permissions += ["%s.%s" % (ct, name) for ct, name in perms] user_obj._team_perm_cache = set(permissions) return user_obj._team_perm_cache def has_perm(self, user_obj, perm, obj=None): if not user_obj.is_active: return False return perm in self.get_team_permissions(user_obj, obj)
<commit_before>from django.db.models import Q from .models import Team class TeamPermissionsBackend(object): def authenticate(self, username=None, password=None): return None def get_team_permissions(self, user_obj, obj=None): """ Returns a set of permission strings that this user has through his/her team memberships. """ if user_obj.is_anonymous() or obj is not None: return set() if not hasattr(user_obj, "_team_perm_cache"): memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="manager") | Q(memberships__state="member"), ) perms = memberships.values_list( "permissions__content_type__app_label", "permissions__codename" ).order_by() user_obj._team_perm_cache = set(["%s.%s" % (ct, name) for ct, name in perms]) return user_obj._team_perm_cache def has_perm(self, user_obj, perm, obj=None): if not user_obj.is_active: return False return perm in self.get_team_permissions(user_obj, obj) <commit_msg>Fix team permissions backend not pulling out manager_permissions Something like request.user.has_perm('reviews.can_manage_%s' % proposal.kind.section.slug) Will aways return false as the backend does a lookup of team membership (member or manager) but only grabs the 'permissions' and not the 'manager_permissions' field<commit_after>
from django.db.models import Q from .models import Team class TeamPermissionsBackend(object): def authenticate(self, username=None, password=None): return None def get_team_permissions(self, user_obj, obj=None): """ Returns a set of permission strings that this user has through his/her team memberships. """ if user_obj.is_anonymous() or obj is not None: return set() if not hasattr(user_obj, "_team_perm_cache"): # Member permissions memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="member"), ) perms = memberships.values_list( "permissions__content_type__app_label", "permissions__codename" ).order_by() permissions = ["%s.%s" % (ct, name) for ct, name in perms] # Manager permissions memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="manager"), ) perms = memberships.values_list( "manager_permissions__content_type__app_label", "manager_permissions__codename" ).order_by() permissions += ["%s.%s" % (ct, name) for ct, name in perms] user_obj._team_perm_cache = set(permissions) return user_obj._team_perm_cache def has_perm(self, user_obj, perm, obj=None): if not user_obj.is_active: return False return perm in self.get_team_permissions(user_obj, obj)
from django.db.models import Q from .models import Team class TeamPermissionsBackend(object): def authenticate(self, username=None, password=None): return None def get_team_permissions(self, user_obj, obj=None): """ Returns a set of permission strings that this user has through his/her team memberships. """ if user_obj.is_anonymous() or obj is not None: return set() if not hasattr(user_obj, "_team_perm_cache"): memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="manager") | Q(memberships__state="member"), ) perms = memberships.values_list( "permissions__content_type__app_label", "permissions__codename" ).order_by() user_obj._team_perm_cache = set(["%s.%s" % (ct, name) for ct, name in perms]) return user_obj._team_perm_cache def has_perm(self, user_obj, perm, obj=None): if not user_obj.is_active: return False return perm in self.get_team_permissions(user_obj, obj) Fix team permissions backend not pulling out manager_permissions Something like request.user.has_perm('reviews.can_manage_%s' % proposal.kind.section.slug) Will aways return false as the backend does a lookup of team membership (member or manager) but only grabs the 'permissions' and not the 'manager_permissions' fieldfrom django.db.models import Q from .models import Team class TeamPermissionsBackend(object): def authenticate(self, username=None, password=None): return None def get_team_permissions(self, user_obj, obj=None): """ Returns a set of permission strings that this user has through his/her team memberships. """ if user_obj.is_anonymous() or obj is not None: return set() if not hasattr(user_obj, "_team_perm_cache"): # Member permissions memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="member"), ) perms = memberships.values_list( "permissions__content_type__app_label", "permissions__codename" ).order_by() permissions = ["%s.%s" % (ct, name) for ct, name in perms] # Manager permissions memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="manager"), ) perms = memberships.values_list( "manager_permissions__content_type__app_label", "manager_permissions__codename" ).order_by() permissions += ["%s.%s" % (ct, name) for ct, name in perms] user_obj._team_perm_cache = set(permissions) return user_obj._team_perm_cache def has_perm(self, user_obj, perm, obj=None): if not user_obj.is_active: return False return perm in self.get_team_permissions(user_obj, obj)
<commit_before>from django.db.models import Q from .models import Team class TeamPermissionsBackend(object): def authenticate(self, username=None, password=None): return None def get_team_permissions(self, user_obj, obj=None): """ Returns a set of permission strings that this user has through his/her team memberships. """ if user_obj.is_anonymous() or obj is not None: return set() if not hasattr(user_obj, "_team_perm_cache"): memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="manager") | Q(memberships__state="member"), ) perms = memberships.values_list( "permissions__content_type__app_label", "permissions__codename" ).order_by() user_obj._team_perm_cache = set(["%s.%s" % (ct, name) for ct, name in perms]) return user_obj._team_perm_cache def has_perm(self, user_obj, perm, obj=None): if not user_obj.is_active: return False return perm in self.get_team_permissions(user_obj, obj) <commit_msg>Fix team permissions backend not pulling out manager_permissions Something like request.user.has_perm('reviews.can_manage_%s' % proposal.kind.section.slug) Will aways return false as the backend does a lookup of team membership (member or manager) but only grabs the 'permissions' and not the 'manager_permissions' field<commit_after>from django.db.models import Q from .models import Team class TeamPermissionsBackend(object): def authenticate(self, username=None, password=None): return None def get_team_permissions(self, user_obj, obj=None): """ Returns a set of permission strings that this user has through his/her team memberships. """ if user_obj.is_anonymous() or obj is not None: return set() if not hasattr(user_obj, "_team_perm_cache"): # Member permissions memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="member"), ) perms = memberships.values_list( "permissions__content_type__app_label", "permissions__codename" ).order_by() permissions = ["%s.%s" % (ct, name) for ct, name in perms] # Manager permissions memberships = Team.objects.filter( Q(memberships__user=user_obj), Q(memberships__state="manager"), ) perms = memberships.values_list( "manager_permissions__content_type__app_label", "manager_permissions__codename" ).order_by() permissions += ["%s.%s" % (ct, name) for ct, name in perms] user_obj._team_perm_cache = set(permissions) return user_obj._team_perm_cache def has_perm(self, user_obj, perm, obj=None): if not user_obj.is_active: return False return perm in self.get_team_permissions(user_obj, obj)
23072e882edb6da55cb12ef0591a786235249670
ome/__main__.py
ome/__main__.py
# ome - Object Message Expressions # Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved. import sys from .command import command_args from .error import OmeError from .terminal import stderr def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) if command_args.verbose: print('ome: using target {}'.format(target.name)) print('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: if command_args.verbose: print('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main()
# ome - Object Message Expressions # Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved. import sys from .command import command_args from .error import OmeError from .terminal import stderr def print_verbose(*args, **kwargs): if command_args.verbose: print(*args, **kwargs) def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) print_verbose('ome: using target {}'.format(target.name)) print_verbose('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: print_verbose('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main()
Use print_verbose for conditional printing.
Use print_verbose for conditional printing.
Python
mit
shaurz/ome,shaurz/ome
# ome - Object Message Expressions # Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved. import sys from .command import command_args from .error import OmeError from .terminal import stderr def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) if command_args.verbose: print('ome: using target {}'.format(target.name)) print('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: if command_args.verbose: print('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main() Use print_verbose for conditional printing.
# ome - Object Message Expressions # Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved. import sys from .command import command_args from .error import OmeError from .terminal import stderr def print_verbose(*args, **kwargs): if command_args.verbose: print(*args, **kwargs) def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) print_verbose('ome: using target {}'.format(target.name)) print_verbose('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: print_verbose('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main()
<commit_before># ome - Object Message Expressions # Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved. import sys from .command import command_args from .error import OmeError from .terminal import stderr def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) if command_args.verbose: print('ome: using target {}'.format(target.name)) print('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: if command_args.verbose: print('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main() <commit_msg>Use print_verbose for conditional printing.<commit_after>
# ome - Object Message Expressions # Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved. import sys from .command import command_args from .error import OmeError from .terminal import stderr def print_verbose(*args, **kwargs): if command_args.verbose: print(*args, **kwargs) def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) print_verbose('ome: using target {}'.format(target.name)) print_verbose('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: print_verbose('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main()
# ome - Object Message Expressions # Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved. import sys from .command import command_args from .error import OmeError from .terminal import stderr def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) if command_args.verbose: print('ome: using target {}'.format(target.name)) print('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: if command_args.verbose: print('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main() Use print_verbose for conditional printing.# ome - Object Message Expressions # Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved. import sys from .command import command_args from .error import OmeError from .terminal import stderr def print_verbose(*args, **kwargs): if command_args.verbose: print(*args, **kwargs) def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) print_verbose('ome: using target {}'.format(target.name)) print_verbose('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: print_verbose('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main()
<commit_before># ome - Object Message Expressions # Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved. import sys from .command import command_args from .error import OmeError from .terminal import stderr def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) if command_args.verbose: print('ome: using target {}'.format(target.name)) print('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: if command_args.verbose: print('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main() <commit_msg>Use print_verbose for conditional printing.<commit_after># ome - Object Message Expressions # Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved. import sys from .command import command_args from .error import OmeError from .terminal import stderr def print_verbose(*args, **kwargs): if command_args.verbose: print(*args, **kwargs) def main(): stderr.reset() try: from . import compiler target = compiler.get_target(command_args.target) build_options = compiler.BuildOptions(target) backend = compiler.get_backend(target, command_args.backend) print_verbose('ome: using target {}'.format(target.name)) print_verbose('ome: using backend {} {}'.format(backend.name, backend.version)) for filename in command_args.filename: print_verbose('ome: compiling {}'.format(filename)) if command_args.print_code: print(compiler.compile_file(filename, target).decode(target.encoding)) else: build_options.make_executable(filename, backend) except OmeError as error: error.write_ansi(stderr) stderr.reset() sys.exit(1) if __name__ == '__main__': if sys.version_info[0] < 3: sys.exit('ome: error: please use python 3.x') main()
36d2b9843160d9c3d439bc36c0188840fcdfa8b5
examples/rmg/minimal_sensitivity/input.py
examples/rmg/minimal_sensitivity/input.py
# Data sources database( thermoLibraries = ['primaryThermoLibrary'], reactionLibraries = [], seedMechanisms = [], kineticsDepositories = ['training'], kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'], kineticsEstimator = 'rate rules', ) # Constraints on generated species generatedSpeciesConstraints( maximumRadicalElectrons = 2, ) # List of species species( label='ethane', reactive=True, structure=SMILES("CC"), ) # Reaction systems simpleReactor( temperature=(1350,'K'), pressure=(1.0,'bar'), initialMoleFractions={ "ethane": 1.0, }, terminationConversion={ 'ethane': 0.9, }, terminationTime=(1e6,'s'), sensitivity=['ethane'], sensitivityThreshold=0.01, ) simulator( atol=1e-16, rtol=1e-8, sens_atol=1e-6, sens_rtol=1e-4, ) model( toleranceKeepInEdge=0.0, toleranceMoveToCore=0.1, toleranceInterruptSimulation=0.1, maximumEdgeSpecies=100000 ) options( units='si', saveRestartPeriod=None, saveSimulationProfiles=True, generateOutputHTML=False, generatePlots=False, )
# Data sources database( thermoLibraries = ['primaryThermoLibrary'], reactionLibraries = [], seedMechanisms = [], kineticsDepositories = ['training'], kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'], kineticsEstimator = 'rate rules', ) # Constraints on generated species generatedSpeciesConstraints( maximumRadicalElectrons = 2, ) # List of species species( label='ethane', reactive=True, structure=SMILES("CC"), ) # Reaction systems simpleReactor( temperature=(1350,'K'), pressure=(1.0,'bar'), initialMoleFractions={ "ethane": 1.0, }, terminationConversion={ 'ethane': 0.9, }, terminationTime=(1e6,'s'), sensitivity=['ethane'], sensitivityThreshold=0.01, ) simulator( atol=1e-16, rtol=1e-8, sens_atol=1e-6, sens_rtol=1e-4, ) model( toleranceKeepInEdge=0.0, toleranceMoveToCore=0.1, toleranceInterruptSimulation=0.1, maximumEdgeSpecies=100000 ) options( units='si', saveRestartPeriod=None, saveSimulationProfiles=False, generateOutputHTML=False, generatePlots=False, )
Change saveSimulationProfiles to False in minimal_sensitivity
Change saveSimulationProfiles to False in minimal_sensitivity just to test a diff parameter in this job
Python
mit
chatelak/RMG-Py,nyee/RMG-Py,chatelak/RMG-Py,pierrelb/RMG-Py,nickvandewiele/RMG-Py,pierrelb/RMG-Py,nyee/RMG-Py,nickvandewiele/RMG-Py
# Data sources database( thermoLibraries = ['primaryThermoLibrary'], reactionLibraries = [], seedMechanisms = [], kineticsDepositories = ['training'], kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'], kineticsEstimator = 'rate rules', ) # Constraints on generated species generatedSpeciesConstraints( maximumRadicalElectrons = 2, ) # List of species species( label='ethane', reactive=True, structure=SMILES("CC"), ) # Reaction systems simpleReactor( temperature=(1350,'K'), pressure=(1.0,'bar'), initialMoleFractions={ "ethane": 1.0, }, terminationConversion={ 'ethane': 0.9, }, terminationTime=(1e6,'s'), sensitivity=['ethane'], sensitivityThreshold=0.01, ) simulator( atol=1e-16, rtol=1e-8, sens_atol=1e-6, sens_rtol=1e-4, ) model( toleranceKeepInEdge=0.0, toleranceMoveToCore=0.1, toleranceInterruptSimulation=0.1, maximumEdgeSpecies=100000 ) options( units='si', saveRestartPeriod=None, saveSimulationProfiles=True, generateOutputHTML=False, generatePlots=False, ) Change saveSimulationProfiles to False in minimal_sensitivity just to test a diff parameter in this job
# Data sources database( thermoLibraries = ['primaryThermoLibrary'], reactionLibraries = [], seedMechanisms = [], kineticsDepositories = ['training'], kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'], kineticsEstimator = 'rate rules', ) # Constraints on generated species generatedSpeciesConstraints( maximumRadicalElectrons = 2, ) # List of species species( label='ethane', reactive=True, structure=SMILES("CC"), ) # Reaction systems simpleReactor( temperature=(1350,'K'), pressure=(1.0,'bar'), initialMoleFractions={ "ethane": 1.0, }, terminationConversion={ 'ethane': 0.9, }, terminationTime=(1e6,'s'), sensitivity=['ethane'], sensitivityThreshold=0.01, ) simulator( atol=1e-16, rtol=1e-8, sens_atol=1e-6, sens_rtol=1e-4, ) model( toleranceKeepInEdge=0.0, toleranceMoveToCore=0.1, toleranceInterruptSimulation=0.1, maximumEdgeSpecies=100000 ) options( units='si', saveRestartPeriod=None, saveSimulationProfiles=False, generateOutputHTML=False, generatePlots=False, )
<commit_before># Data sources database( thermoLibraries = ['primaryThermoLibrary'], reactionLibraries = [], seedMechanisms = [], kineticsDepositories = ['training'], kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'], kineticsEstimator = 'rate rules', ) # Constraints on generated species generatedSpeciesConstraints( maximumRadicalElectrons = 2, ) # List of species species( label='ethane', reactive=True, structure=SMILES("CC"), ) # Reaction systems simpleReactor( temperature=(1350,'K'), pressure=(1.0,'bar'), initialMoleFractions={ "ethane": 1.0, }, terminationConversion={ 'ethane': 0.9, }, terminationTime=(1e6,'s'), sensitivity=['ethane'], sensitivityThreshold=0.01, ) simulator( atol=1e-16, rtol=1e-8, sens_atol=1e-6, sens_rtol=1e-4, ) model( toleranceKeepInEdge=0.0, toleranceMoveToCore=0.1, toleranceInterruptSimulation=0.1, maximumEdgeSpecies=100000 ) options( units='si', saveRestartPeriod=None, saveSimulationProfiles=True, generateOutputHTML=False, generatePlots=False, ) <commit_msg>Change saveSimulationProfiles to False in minimal_sensitivity just to test a diff parameter in this job<commit_after>
# Data sources database( thermoLibraries = ['primaryThermoLibrary'], reactionLibraries = [], seedMechanisms = [], kineticsDepositories = ['training'], kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'], kineticsEstimator = 'rate rules', ) # Constraints on generated species generatedSpeciesConstraints( maximumRadicalElectrons = 2, ) # List of species species( label='ethane', reactive=True, structure=SMILES("CC"), ) # Reaction systems simpleReactor( temperature=(1350,'K'), pressure=(1.0,'bar'), initialMoleFractions={ "ethane": 1.0, }, terminationConversion={ 'ethane': 0.9, }, terminationTime=(1e6,'s'), sensitivity=['ethane'], sensitivityThreshold=0.01, ) simulator( atol=1e-16, rtol=1e-8, sens_atol=1e-6, sens_rtol=1e-4, ) model( toleranceKeepInEdge=0.0, toleranceMoveToCore=0.1, toleranceInterruptSimulation=0.1, maximumEdgeSpecies=100000 ) options( units='si', saveRestartPeriod=None, saveSimulationProfiles=False, generateOutputHTML=False, generatePlots=False, )
# Data sources database( thermoLibraries = ['primaryThermoLibrary'], reactionLibraries = [], seedMechanisms = [], kineticsDepositories = ['training'], kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'], kineticsEstimator = 'rate rules', ) # Constraints on generated species generatedSpeciesConstraints( maximumRadicalElectrons = 2, ) # List of species species( label='ethane', reactive=True, structure=SMILES("CC"), ) # Reaction systems simpleReactor( temperature=(1350,'K'), pressure=(1.0,'bar'), initialMoleFractions={ "ethane": 1.0, }, terminationConversion={ 'ethane': 0.9, }, terminationTime=(1e6,'s'), sensitivity=['ethane'], sensitivityThreshold=0.01, ) simulator( atol=1e-16, rtol=1e-8, sens_atol=1e-6, sens_rtol=1e-4, ) model( toleranceKeepInEdge=0.0, toleranceMoveToCore=0.1, toleranceInterruptSimulation=0.1, maximumEdgeSpecies=100000 ) options( units='si', saveRestartPeriod=None, saveSimulationProfiles=True, generateOutputHTML=False, generatePlots=False, ) Change saveSimulationProfiles to False in minimal_sensitivity just to test a diff parameter in this job# Data sources database( thermoLibraries = ['primaryThermoLibrary'], reactionLibraries = [], seedMechanisms = [], kineticsDepositories = ['training'], kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'], kineticsEstimator = 'rate rules', ) # Constraints on generated species generatedSpeciesConstraints( maximumRadicalElectrons = 2, ) # List of species species( label='ethane', reactive=True, structure=SMILES("CC"), ) # Reaction systems simpleReactor( temperature=(1350,'K'), pressure=(1.0,'bar'), initialMoleFractions={ "ethane": 1.0, }, terminationConversion={ 'ethane': 0.9, }, terminationTime=(1e6,'s'), sensitivity=['ethane'], sensitivityThreshold=0.01, ) simulator( atol=1e-16, rtol=1e-8, sens_atol=1e-6, sens_rtol=1e-4, ) model( toleranceKeepInEdge=0.0, toleranceMoveToCore=0.1, toleranceInterruptSimulation=0.1, maximumEdgeSpecies=100000 ) options( units='si', saveRestartPeriod=None, saveSimulationProfiles=False, generateOutputHTML=False, generatePlots=False, )
<commit_before># Data sources database( thermoLibraries = ['primaryThermoLibrary'], reactionLibraries = [], seedMechanisms = [], kineticsDepositories = ['training'], kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'], kineticsEstimator = 'rate rules', ) # Constraints on generated species generatedSpeciesConstraints( maximumRadicalElectrons = 2, ) # List of species species( label='ethane', reactive=True, structure=SMILES("CC"), ) # Reaction systems simpleReactor( temperature=(1350,'K'), pressure=(1.0,'bar'), initialMoleFractions={ "ethane": 1.0, }, terminationConversion={ 'ethane': 0.9, }, terminationTime=(1e6,'s'), sensitivity=['ethane'], sensitivityThreshold=0.01, ) simulator( atol=1e-16, rtol=1e-8, sens_atol=1e-6, sens_rtol=1e-4, ) model( toleranceKeepInEdge=0.0, toleranceMoveToCore=0.1, toleranceInterruptSimulation=0.1, maximumEdgeSpecies=100000 ) options( units='si', saveRestartPeriod=None, saveSimulationProfiles=True, generateOutputHTML=False, generatePlots=False, ) <commit_msg>Change saveSimulationProfiles to False in minimal_sensitivity just to test a diff parameter in this job<commit_after># Data sources database( thermoLibraries = ['primaryThermoLibrary'], reactionLibraries = [], seedMechanisms = [], kineticsDepositories = ['training'], kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'], kineticsEstimator = 'rate rules', ) # Constraints on generated species generatedSpeciesConstraints( maximumRadicalElectrons = 2, ) # List of species species( label='ethane', reactive=True, structure=SMILES("CC"), ) # Reaction systems simpleReactor( temperature=(1350,'K'), pressure=(1.0,'bar'), initialMoleFractions={ "ethane": 1.0, }, terminationConversion={ 'ethane': 0.9, }, terminationTime=(1e6,'s'), sensitivity=['ethane'], sensitivityThreshold=0.01, ) simulator( atol=1e-16, rtol=1e-8, sens_atol=1e-6, sens_rtol=1e-4, ) model( toleranceKeepInEdge=0.0, toleranceMoveToCore=0.1, toleranceInterruptSimulation=0.1, maximumEdgeSpecies=100000 ) options( units='si', saveRestartPeriod=None, saveSimulationProfiles=False, generateOutputHTML=False, generatePlots=False, )
1bf4116bbd449769d209c4ff98b609b72bd312aa
api/views.py
api/views.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib.auth.models import User from rest_framework import viewsets from core.models import Timesheet, Task, Entry from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer, EntrySerializer) class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer class TimesheetViewSet(viewsets.ModelViewSet): queryset = Timesheet.objects.all() serializer_class = TimesheetSerializer filter_fields = ('id',) class TaskViewSet(viewsets.ModelViewSet): queryset = Task.objects.all() serializer_class = TaskSerializer filter_fields = ('id', 'timesheet',) class EntryViewSet(viewsets.ModelViewSet): queryset = Entry.objects.all() serializer_class = EntrySerializer filter_fields = ('id', 'user', 'task', 'task__timesheet',)
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib.auth.models import User from rest_framework import viewsets import django_filters from core.models import Timesheet, Task, Entry from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer, EntrySerializer) class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer class TimesheetViewSet(viewsets.ModelViewSet): queryset = Timesheet.objects.all() serializer_class = TimesheetSerializer filter_fields = ('id',) class TaskViewSet(viewsets.ModelViewSet): queryset = Task.objects.all() serializer_class = TaskSerializer filter_fields = ('id', 'timesheet',) class EntryFilter(django_filters.rest_framework.FilterSet): min_date = django_filters.DateFilter(name="date", lookup_expr="gte") max_date = django_filters.DateFilter(name="date", lookup_expr="lte") class Meta: model = Entry fields = ('id', 'date', 'user', 'task', 'task__timesheet',) class EntryViewSet(viewsets.ModelViewSet): queryset = Entry.objects.all() serializer_class = EntrySerializer filter_class = EntryFilter
Add date min-max filtering to API
Add date min-max filtering to API
Python
bsd-2-clause
Leahelisabeth/timestrap,muhleder/timestrap,cdubz/timestrap,Leahelisabeth/timestrap,muhleder/timestrap,overshard/timestrap,overshard/timestrap,Leahelisabeth/timestrap,muhleder/timestrap,cdubz/timestrap,Leahelisabeth/timestrap,overshard/timestrap,cdubz/timestrap
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib.auth.models import User from rest_framework import viewsets from core.models import Timesheet, Task, Entry from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer, EntrySerializer) class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer class TimesheetViewSet(viewsets.ModelViewSet): queryset = Timesheet.objects.all() serializer_class = TimesheetSerializer filter_fields = ('id',) class TaskViewSet(viewsets.ModelViewSet): queryset = Task.objects.all() serializer_class = TaskSerializer filter_fields = ('id', 'timesheet',) class EntryViewSet(viewsets.ModelViewSet): queryset = Entry.objects.all() serializer_class = EntrySerializer filter_fields = ('id', 'user', 'task', 'task__timesheet',) Add date min-max filtering to API
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib.auth.models import User from rest_framework import viewsets import django_filters from core.models import Timesheet, Task, Entry from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer, EntrySerializer) class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer class TimesheetViewSet(viewsets.ModelViewSet): queryset = Timesheet.objects.all() serializer_class = TimesheetSerializer filter_fields = ('id',) class TaskViewSet(viewsets.ModelViewSet): queryset = Task.objects.all() serializer_class = TaskSerializer filter_fields = ('id', 'timesheet',) class EntryFilter(django_filters.rest_framework.FilterSet): min_date = django_filters.DateFilter(name="date", lookup_expr="gte") max_date = django_filters.DateFilter(name="date", lookup_expr="lte") class Meta: model = Entry fields = ('id', 'date', 'user', 'task', 'task__timesheet',) class EntryViewSet(viewsets.ModelViewSet): queryset = Entry.objects.all() serializer_class = EntrySerializer filter_class = EntryFilter
<commit_before># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib.auth.models import User from rest_framework import viewsets from core.models import Timesheet, Task, Entry from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer, EntrySerializer) class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer class TimesheetViewSet(viewsets.ModelViewSet): queryset = Timesheet.objects.all() serializer_class = TimesheetSerializer filter_fields = ('id',) class TaskViewSet(viewsets.ModelViewSet): queryset = Task.objects.all() serializer_class = TaskSerializer filter_fields = ('id', 'timesheet',) class EntryViewSet(viewsets.ModelViewSet): queryset = Entry.objects.all() serializer_class = EntrySerializer filter_fields = ('id', 'user', 'task', 'task__timesheet',) <commit_msg>Add date min-max filtering to API<commit_after>
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib.auth.models import User from rest_framework import viewsets import django_filters from core.models import Timesheet, Task, Entry from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer, EntrySerializer) class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer class TimesheetViewSet(viewsets.ModelViewSet): queryset = Timesheet.objects.all() serializer_class = TimesheetSerializer filter_fields = ('id',) class TaskViewSet(viewsets.ModelViewSet): queryset = Task.objects.all() serializer_class = TaskSerializer filter_fields = ('id', 'timesheet',) class EntryFilter(django_filters.rest_framework.FilterSet): min_date = django_filters.DateFilter(name="date", lookup_expr="gte") max_date = django_filters.DateFilter(name="date", lookup_expr="lte") class Meta: model = Entry fields = ('id', 'date', 'user', 'task', 'task__timesheet',) class EntryViewSet(viewsets.ModelViewSet): queryset = Entry.objects.all() serializer_class = EntrySerializer filter_class = EntryFilter
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib.auth.models import User from rest_framework import viewsets from core.models import Timesheet, Task, Entry from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer, EntrySerializer) class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer class TimesheetViewSet(viewsets.ModelViewSet): queryset = Timesheet.objects.all() serializer_class = TimesheetSerializer filter_fields = ('id',) class TaskViewSet(viewsets.ModelViewSet): queryset = Task.objects.all() serializer_class = TaskSerializer filter_fields = ('id', 'timesheet',) class EntryViewSet(viewsets.ModelViewSet): queryset = Entry.objects.all() serializer_class = EntrySerializer filter_fields = ('id', 'user', 'task', 'task__timesheet',) Add date min-max filtering to API# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib.auth.models import User from rest_framework import viewsets import django_filters from core.models import Timesheet, Task, Entry from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer, EntrySerializer) class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer class TimesheetViewSet(viewsets.ModelViewSet): queryset = Timesheet.objects.all() serializer_class = TimesheetSerializer filter_fields = ('id',) class TaskViewSet(viewsets.ModelViewSet): queryset = Task.objects.all() serializer_class = TaskSerializer filter_fields = ('id', 'timesheet',) class EntryFilter(django_filters.rest_framework.FilterSet): min_date = django_filters.DateFilter(name="date", lookup_expr="gte") max_date = django_filters.DateFilter(name="date", lookup_expr="lte") class Meta: model = Entry fields = ('id', 'date', 'user', 'task', 'task__timesheet',) class EntryViewSet(viewsets.ModelViewSet): queryset = Entry.objects.all() serializer_class = EntrySerializer filter_class = EntryFilter
<commit_before># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib.auth.models import User from rest_framework import viewsets from core.models import Timesheet, Task, Entry from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer, EntrySerializer) class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer class TimesheetViewSet(viewsets.ModelViewSet): queryset = Timesheet.objects.all() serializer_class = TimesheetSerializer filter_fields = ('id',) class TaskViewSet(viewsets.ModelViewSet): queryset = Task.objects.all() serializer_class = TaskSerializer filter_fields = ('id', 'timesheet',) class EntryViewSet(viewsets.ModelViewSet): queryset = Entry.objects.all() serializer_class = EntrySerializer filter_fields = ('id', 'user', 'task', 'task__timesheet',) <commit_msg>Add date min-max filtering to API<commit_after># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib.auth.models import User from rest_framework import viewsets import django_filters from core.models import Timesheet, Task, Entry from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer, EntrySerializer) class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer class TimesheetViewSet(viewsets.ModelViewSet): queryset = Timesheet.objects.all() serializer_class = TimesheetSerializer filter_fields = ('id',) class TaskViewSet(viewsets.ModelViewSet): queryset = Task.objects.all() serializer_class = TaskSerializer filter_fields = ('id', 'timesheet',) class EntryFilter(django_filters.rest_framework.FilterSet): min_date = django_filters.DateFilter(name="date", lookup_expr="gte") max_date = django_filters.DateFilter(name="date", lookup_expr="lte") class Meta: model = Entry fields = ('id', 'date', 'user', 'task', 'task__timesheet',) class EntryViewSet(viewsets.ModelViewSet): queryset = Entry.objects.all() serializer_class = EntrySerializer filter_class = EntryFilter
b5e11827929f37da8d18616f1fb3fc2d62591515
djangocms_spa/decorators.py
djangocms_spa/decorators.py
from functools import wraps from django.conf import settings from django.core.cache import cache from django.template.response import ContentNotRenderedError from django.utils.decorators import available_attrs def cache_view(view_func): @wraps(view_func, assigned=available_attrs(view_func)) def _wrapped_view_func(view, *args, **kwargs): cache_key = view.request.get_full_path() cached_response = cache.get(cache_key) if cached_response and not view.request.user.is_authenticated(): return cached_response response = view_func(view, *args, **kwargs) if response.status_code == 200 and not view.request.user.is_authenticated(): try: set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) except ContentNotRenderedError: response.add_post_render_callback( lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) ) return response return _wrapped_view_func def set_cache_after_rendering(cache_key, response, timeout): cache.set(cache_key, response, timeout)
from functools import wraps from django.conf import settings from django.core.cache import cache from django.template.response import ContentNotRenderedError from django.utils.decorators import available_attrs def cache_view(view_func): @wraps(view_func, assigned=available_attrs(view_func)) def _wrapped_view_func(view, *args, **kwargs): request = view.request language_code = request.LANGUAGE_CODE cache_key = '{path}:{lang}'.format(path=request.get_full_path(), lang=language_code) cached_response = cache.get(cache_key) if cached_response and not request.user.is_authenticated(): return cached_response response = view_func(view, *args, **kwargs) if response.status_code == 200 and not request.user.is_authenticated(): try: set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) except ContentNotRenderedError: response.add_post_render_callback( lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) ) return response return _wrapped_view_func def set_cache_after_rendering(cache_key, response, timeout): cache.set(cache_key, response, timeout)
Add language code to cache key explicitly
[language_activation] Add language code to cache key explicitly
Python
mit
dreipol/djangocms-spa,dreipol/djangocms-spa
from functools import wraps from django.conf import settings from django.core.cache import cache from django.template.response import ContentNotRenderedError from django.utils.decorators import available_attrs def cache_view(view_func): @wraps(view_func, assigned=available_attrs(view_func)) def _wrapped_view_func(view, *args, **kwargs): cache_key = view.request.get_full_path() cached_response = cache.get(cache_key) if cached_response and not view.request.user.is_authenticated(): return cached_response response = view_func(view, *args, **kwargs) if response.status_code == 200 and not view.request.user.is_authenticated(): try: set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) except ContentNotRenderedError: response.add_post_render_callback( lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) ) return response return _wrapped_view_func def set_cache_after_rendering(cache_key, response, timeout): cache.set(cache_key, response, timeout) [language_activation] Add language code to cache key explicitly
from functools import wraps from django.conf import settings from django.core.cache import cache from django.template.response import ContentNotRenderedError from django.utils.decorators import available_attrs def cache_view(view_func): @wraps(view_func, assigned=available_attrs(view_func)) def _wrapped_view_func(view, *args, **kwargs): request = view.request language_code = request.LANGUAGE_CODE cache_key = '{path}:{lang}'.format(path=request.get_full_path(), lang=language_code) cached_response = cache.get(cache_key) if cached_response and not request.user.is_authenticated(): return cached_response response = view_func(view, *args, **kwargs) if response.status_code == 200 and not request.user.is_authenticated(): try: set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) except ContentNotRenderedError: response.add_post_render_callback( lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) ) return response return _wrapped_view_func def set_cache_after_rendering(cache_key, response, timeout): cache.set(cache_key, response, timeout)
<commit_before>from functools import wraps from django.conf import settings from django.core.cache import cache from django.template.response import ContentNotRenderedError from django.utils.decorators import available_attrs def cache_view(view_func): @wraps(view_func, assigned=available_attrs(view_func)) def _wrapped_view_func(view, *args, **kwargs): cache_key = view.request.get_full_path() cached_response = cache.get(cache_key) if cached_response and not view.request.user.is_authenticated(): return cached_response response = view_func(view, *args, **kwargs) if response.status_code == 200 and not view.request.user.is_authenticated(): try: set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) except ContentNotRenderedError: response.add_post_render_callback( lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) ) return response return _wrapped_view_func def set_cache_after_rendering(cache_key, response, timeout): cache.set(cache_key, response, timeout) <commit_msg>[language_activation] Add language code to cache key explicitly<commit_after>
from functools import wraps from django.conf import settings from django.core.cache import cache from django.template.response import ContentNotRenderedError from django.utils.decorators import available_attrs def cache_view(view_func): @wraps(view_func, assigned=available_attrs(view_func)) def _wrapped_view_func(view, *args, **kwargs): request = view.request language_code = request.LANGUAGE_CODE cache_key = '{path}:{lang}'.format(path=request.get_full_path(), lang=language_code) cached_response = cache.get(cache_key) if cached_response and not request.user.is_authenticated(): return cached_response response = view_func(view, *args, **kwargs) if response.status_code == 200 and not request.user.is_authenticated(): try: set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) except ContentNotRenderedError: response.add_post_render_callback( lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) ) return response return _wrapped_view_func def set_cache_after_rendering(cache_key, response, timeout): cache.set(cache_key, response, timeout)
from functools import wraps from django.conf import settings from django.core.cache import cache from django.template.response import ContentNotRenderedError from django.utils.decorators import available_attrs def cache_view(view_func): @wraps(view_func, assigned=available_attrs(view_func)) def _wrapped_view_func(view, *args, **kwargs): cache_key = view.request.get_full_path() cached_response = cache.get(cache_key) if cached_response and not view.request.user.is_authenticated(): return cached_response response = view_func(view, *args, **kwargs) if response.status_code == 200 and not view.request.user.is_authenticated(): try: set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) except ContentNotRenderedError: response.add_post_render_callback( lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) ) return response return _wrapped_view_func def set_cache_after_rendering(cache_key, response, timeout): cache.set(cache_key, response, timeout) [language_activation] Add language code to cache key explicitlyfrom functools import wraps from django.conf import settings from django.core.cache import cache from django.template.response import ContentNotRenderedError from django.utils.decorators import available_attrs def cache_view(view_func): @wraps(view_func, assigned=available_attrs(view_func)) def _wrapped_view_func(view, *args, **kwargs): request = view.request language_code = request.LANGUAGE_CODE cache_key = '{path}:{lang}'.format(path=request.get_full_path(), lang=language_code) cached_response = cache.get(cache_key) if cached_response and not request.user.is_authenticated(): return cached_response response = view_func(view, *args, **kwargs) if response.status_code == 200 and not request.user.is_authenticated(): try: set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) except ContentNotRenderedError: response.add_post_render_callback( lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) ) return response return _wrapped_view_func def set_cache_after_rendering(cache_key, response, timeout): cache.set(cache_key, response, timeout)
<commit_before>from functools import wraps from django.conf import settings from django.core.cache import cache from django.template.response import ContentNotRenderedError from django.utils.decorators import available_attrs def cache_view(view_func): @wraps(view_func, assigned=available_attrs(view_func)) def _wrapped_view_func(view, *args, **kwargs): cache_key = view.request.get_full_path() cached_response = cache.get(cache_key) if cached_response and not view.request.user.is_authenticated(): return cached_response response = view_func(view, *args, **kwargs) if response.status_code == 200 and not view.request.user.is_authenticated(): try: set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) except ContentNotRenderedError: response.add_post_render_callback( lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) ) return response return _wrapped_view_func def set_cache_after_rendering(cache_key, response, timeout): cache.set(cache_key, response, timeout) <commit_msg>[language_activation] Add language code to cache key explicitly<commit_after>from functools import wraps from django.conf import settings from django.core.cache import cache from django.template.response import ContentNotRenderedError from django.utils.decorators import available_attrs def cache_view(view_func): @wraps(view_func, assigned=available_attrs(view_func)) def _wrapped_view_func(view, *args, **kwargs): request = view.request language_code = request.LANGUAGE_CODE cache_key = '{path}:{lang}'.format(path=request.get_full_path(), lang=language_code) cached_response = cache.get(cache_key) if cached_response and not request.user.is_authenticated(): return cached_response response = view_func(view, *args, **kwargs) if response.status_code == 200 and not request.user.is_authenticated(): try: set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) except ContentNotRenderedError: response.add_post_render_callback( lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT) ) return response return _wrapped_view_func def set_cache_after_rendering(cache_key, response, timeout): cache.set(cache_key, response, timeout)
e0aab62f2a693ca20a81c9e55c4220f379ac9eb1
socialregistration/templatetags/socialregistration_tags.py
socialregistration/templatetags/socialregistration_tags.py
from django import template register = template.Library() @register.tag def social_csrf_token(): """ Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration work with both Django v1.2 and Django < v1.2 """ return CsrfNode() class CsrfNode(template.Node): def render(self, context): try: from django.template.defaulttags import CsrfTokenNode return CsrfTokenNode().render(context) except ImportError: return u''
from django import template register = template.Library() @register.tag def social_csrf_token(parser, token): """ Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration work with both Django v1.2 and Django < v1.2 """ return CsrfNode() class CsrfNode(template.Node): def render(self, context): try: from django.template.defaulttags import CsrfTokenNode return CsrfTokenNode().render(context) except ImportError: return u''
Add necessary arguments to the social_csrf_token tag.
Add necessary arguments to the social_csrf_token tag.
Python
mit
praekelt/django-socialregistration,aditweb/django-socialregistration,minlex/django-socialregistration,kapt/django-socialregistration,coxmediagroup/django-socialregistration,aditweb/django-socialregistration,flashingpumpkin/django-socialregistration,mark-adams/django-socialregistration,mark-adams/django-socialregistration,minlex/django-socialregistration,brodie/django-socialregistration,minlex/django-socialregistration,lgapontes/django-socialregistration,bopo/django-socialregistration,0101/django-socialregistration,aditweb/django-socialregistration,lgapontes/django-socialregistration,amakhnach/django-socialregistration,bopo/django-socialregistration,flashingpumpkin/django-socialregistration,coxmediagroup/django-socialregistration,itmustbejj/django-socialregistration,lgapontes/django-socialregistration,Soovox/django-socialregistration,bopo/django-socialregistration,mark-adams/django-socialregistration,kapt/django-socialregistration,brodie/django-socialregistration
from django import template register = template.Library() @register.tag def social_csrf_token(): """ Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration work with both Django v1.2 and Django < v1.2 """ return CsrfNode() class CsrfNode(template.Node): def render(self, context): try: from django.template.defaulttags import CsrfTokenNode return CsrfTokenNode().render(context) except ImportError: return u''Add necessary arguments to the social_csrf_token tag.
from django import template register = template.Library() @register.tag def social_csrf_token(parser, token): """ Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration work with both Django v1.2 and Django < v1.2 """ return CsrfNode() class CsrfNode(template.Node): def render(self, context): try: from django.template.defaulttags import CsrfTokenNode return CsrfTokenNode().render(context) except ImportError: return u''
<commit_before>from django import template register = template.Library() @register.tag def social_csrf_token(): """ Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration work with both Django v1.2 and Django < v1.2 """ return CsrfNode() class CsrfNode(template.Node): def render(self, context): try: from django.template.defaulttags import CsrfTokenNode return CsrfTokenNode().render(context) except ImportError: return u''<commit_msg>Add necessary arguments to the social_csrf_token tag.<commit_after>
from django import template register = template.Library() @register.tag def social_csrf_token(parser, token): """ Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration work with both Django v1.2 and Django < v1.2 """ return CsrfNode() class CsrfNode(template.Node): def render(self, context): try: from django.template.defaulttags import CsrfTokenNode return CsrfTokenNode().render(context) except ImportError: return u''
from django import template register = template.Library() @register.tag def social_csrf_token(): """ Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration work with both Django v1.2 and Django < v1.2 """ return CsrfNode() class CsrfNode(template.Node): def render(self, context): try: from django.template.defaulttags import CsrfTokenNode return CsrfTokenNode().render(context) except ImportError: return u''Add necessary arguments to the social_csrf_token tag.from django import template register = template.Library() @register.tag def social_csrf_token(parser, token): """ Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration work with both Django v1.2 and Django < v1.2 """ return CsrfNode() class CsrfNode(template.Node): def render(self, context): try: from django.template.defaulttags import CsrfTokenNode return CsrfTokenNode().render(context) except ImportError: return u''
<commit_before>from django import template register = template.Library() @register.tag def social_csrf_token(): """ Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration work with both Django v1.2 and Django < v1.2 """ return CsrfNode() class CsrfNode(template.Node): def render(self, context): try: from django.template.defaulttags import CsrfTokenNode return CsrfTokenNode().render(context) except ImportError: return u''<commit_msg>Add necessary arguments to the social_csrf_token tag.<commit_after>from django import template register = template.Library() @register.tag def social_csrf_token(parser, token): """ Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration work with both Django v1.2 and Django < v1.2 """ return CsrfNode() class CsrfNode(template.Node): def render(self, context): try: from django.template.defaulttags import CsrfTokenNode return CsrfTokenNode().render(context) except ImportError: return u''
7ea8420e9653765d960938340124b8c2274c69fc
Site/Settings.py
Site/Settings.py
import os settings = {} domain = 'localhost:8080' framework = 'flask' debug = False clearOnLoad = False emailOnRegister = False registrationDisabled = False cacheDirectory = "./Site/Cache" siteDown = False database = 'osf20120530' # Mongo cookieDomain = '.openscienceframework.org' # Beaker static = '%s/static' % os.path.dirname(os.path.abspath(__file__)) local = True
import os settings = {} domain = 'localhost:8080' framework = 'flask' debug = False clearOnLoad = False emailOnRegister = False registrationDisabled = False cacheDirectory = "./Site/Cache" siteDown = False database = 'osf20120530' # Mongo cookieDomain = '.openscienceframework.org' # Beaker static = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'static') local = True
Improve OS compatibility for settings
Improve OS compatibility for settings
Python
apache-2.0
cslzchen/osf.io,mattclark/osf.io,kwierman/osf.io,SSJohns/osf.io,caseyrollins/osf.io,HarryRybacki/osf.io,monikagrabowska/osf.io,HarryRybacki/osf.io,Nesiehr/osf.io,himanshuo/osf.io,SSJohns/osf.io,jnayak1/osf.io,billyhunt/osf.io,fabianvf/osf.io,CenterForOpenScience/osf.io,TomHeatwole/osf.io,chennan47/osf.io,HarryRybacki/osf.io,jinluyuan/osf.io,CenterForOpenScience/osf.io,brandonPurvis/osf.io,jnayak1/osf.io,cosenal/osf.io,samanehsan/osf.io,Nesiehr/osf.io,GageGaskins/osf.io,bdyetton/prettychart,jinluyuan/osf.io,arpitar/osf.io,jmcarp/osf.io,kch8qx/osf.io,Nesiehr/osf.io,monikagrabowska/osf.io,Johnetordoff/osf.io,cwisecarver/osf.io,reinaH/osf.io,jeffreyliu3230/osf.io,RomanZWang/osf.io,barbour-em/osf.io,lamdnhan/osf.io,revanthkolli/osf.io,amyshi188/osf.io,danielneis/osf.io,ckc6cz/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,njantrania/osf.io,GaryKriebel/osf.io,doublebits/osf.io,barbour-em/osf.io,revanthkolli/osf.io,aaxelb/osf.io,mluo613/osf.io,brianjgeiger/osf.io,kushG/osf.io,icereval/osf.io,jolene-esposito/osf.io,jolene-esposito/osf.io,wearpants/osf.io,felliott/osf.io,mluo613/osf.io,zachjanicki/osf.io,arpitar/osf.io,mattclark/osf.io,brandonPurvis/osf.io,rdhyee/osf.io,jeffreyliu3230/osf.io,cslzchen/osf.io,ticklemepierce/osf.io,zachjanicki/osf.io,kch8qx/osf.io,rdhyee/osf.io,AndrewSallans/osf.io,danielneis/osf.io,KAsante95/osf.io,KAsante95/osf.io,GageGaskins/osf.io,kch8qx/osf.io,pattisdr/osf.io,barbour-em/osf.io,amyshi188/osf.io,asanfilippo7/osf.io,alexschiller/osf.io,KAsante95/osf.io,jmcarp/osf.io,ticklemepierce/osf.io,lyndsysimon/osf.io,leb2dg/osf.io,Ghalko/osf.io,fabianvf/osf.io,samchrisinger/osf.io,GageGaskins/osf.io,wearpants/osf.io,sbt9uc/osf.io,AndrewSallans/osf.io,felliott/osf.io,samanehsan/osf.io,cosenal/osf.io,bdyetton/prettychart,bdyetton/prettychart,binoculars/osf.io,mluo613/osf.io,MerlinZhang/osf.io,caneruguz/osf.io,mfraezz/osf.io,amyshi188/osf.io,crcresearch/osf.io,dplorimer/osf,kushG/osf.io,Johnetordoff/osf.io,leb2dg/osf.io,TomBaxter/osf.io,alexschiller/osf.io,acshi/osf.io,arpitar/osf.io,GageGaskins/osf.io,amyshi188/osf.io,erinspace/osf.io,haoyuchen1992/osf.io,lamdnhan/osf.io,himanshuo/osf.io,Ghalko/osf.io,caneruguz/osf.io,HalcyonChimera/osf.io,felliott/osf.io,zachjanicki/osf.io,billyhunt/osf.io,caseyrygt/osf.io,danielneis/osf.io,mattclark/osf.io,cldershem/osf.io,billyhunt/osf.io,monikagrabowska/osf.io,zkraime/osf.io,acshi/osf.io,mfraezz/osf.io,leb2dg/osf.io,dplorimer/osf,sloria/osf.io,binoculars/osf.io,sbt9uc/osf.io,leb2dg/osf.io,laurenrevere/osf.io,acshi/osf.io,dplorimer/osf,acshi/osf.io,crcresearch/osf.io,jnayak1/osf.io,chrisseto/osf.io,alexschiller/osf.io,SSJohns/osf.io,baylee-d/osf.io,aaxelb/osf.io,sbt9uc/osf.io,Nesiehr/osf.io,kushG/osf.io,saradbowman/osf.io,samchrisinger/osf.io,cldershem/osf.io,zamattiac/osf.io,hmoco/osf.io,caseyrollins/osf.io,GaryKriebel/osf.io,TomBaxter/osf.io,adlius/osf.io,hmoco/osf.io,petermalcolm/osf.io,jnayak1/osf.io,ZobairAlijan/osf.io,barbour-em/osf.io,RomanZWang/osf.io,abought/osf.io,felliott/osf.io,chrisseto/osf.io,erinspace/osf.io,mluke93/osf.io,wearpants/osf.io,cldershem/osf.io,Ghalko/osf.io,zamattiac/osf.io,cwisecarver/osf.io,mfraezz/osf.io,caneruguz/osf.io,ticklemepierce/osf.io,GaryKriebel/osf.io,monikagrabowska/osf.io,emetsger/osf.io,brianjgeiger/osf.io,icereval/osf.io,danielneis/osf.io,cldershem/osf.io,laurenrevere/osf.io,GaryKriebel/osf.io,monikagrabowska/osf.io,DanielSBrown/osf.io,doublebits/osf.io,chrisseto/osf.io,zkraime/osf.io,brianjgeiger/osf.io,brandonPurvis/osf.io,kushG/osf.io,lyndsysimon/osf.io,petermalcolm/osf.io,samchrisinger/osf.io,caseyrollins/osf.io,himanshuo/osf.io,aaxelb/osf.io,jmcarp/osf.io,erinspace/osf.io,ticklemepierce/osf.io,RomanZWang/osf.io,billyhunt/osf.io,kch8qx/osf.io,reinaH/osf.io,jolene-esposito/osf.io,mfraezz/osf.io,emetsger/osf.io,doublebits/osf.io,samanehsan/osf.io,RomanZWang/osf.io,haoyuchen1992/osf.io,himanshuo/osf.io,dplorimer/osf,cosenal/osf.io,njantrania/osf.io,chennan47/osf.io,billyhunt/osf.io,MerlinZhang/osf.io,jeffreyliu3230/osf.io,HarryRybacki/osf.io,cslzchen/osf.io,caseyrygt/osf.io,baylee-d/osf.io,ZobairAlijan/osf.io,TomHeatwole/osf.io,GageGaskins/osf.io,njantrania/osf.io,kwierman/osf.io,asanfilippo7/osf.io,DanielSBrown/osf.io,petermalcolm/osf.io,asanfilippo7/osf.io,alexschiller/osf.io,ckc6cz/osf.io,alexschiller/osf.io,adlius/osf.io,kwierman/osf.io,chennan47/osf.io,mluke93/osf.io,revanthkolli/osf.io,njantrania/osf.io,petermalcolm/osf.io,samchrisinger/osf.io,rdhyee/osf.io,sbt9uc/osf.io,ckc6cz/osf.io,hmoco/osf.io,TomHeatwole/osf.io,caneruguz/osf.io,mluke93/osf.io,caseyrygt/osf.io,saradbowman/osf.io,cwisecarver/osf.io,lamdnhan/osf.io,hmoco/osf.io,zachjanicki/osf.io,MerlinZhang/osf.io,wearpants/osf.io,abought/osf.io,Johnetordoff/osf.io,ZobairAlijan/osf.io,reinaH/osf.io,abought/osf.io,cslzchen/osf.io,zamattiac/osf.io,abought/osf.io,samanehsan/osf.io,SSJohns/osf.io,CenterForOpenScience/osf.io,adlius/osf.io,jinluyuan/osf.io,arpitar/osf.io,jolene-esposito/osf.io,zamattiac/osf.io,fabianvf/osf.io,aaxelb/osf.io,doublebits/osf.io,haoyuchen1992/osf.io,KAsante95/osf.io,caseyrygt/osf.io,brandonPurvis/osf.io,icereval/osf.io,sloria/osf.io,HalcyonChimera/osf.io,ckc6cz/osf.io,lamdnhan/osf.io,lyndsysimon/osf.io,DanielSBrown/osf.io,rdhyee/osf.io,lyndsysimon/osf.io,bdyetton/prettychart,zkraime/osf.io,emetsger/osf.io,doublebits/osf.io,mluo613/osf.io,kch8qx/osf.io,baylee-d/osf.io,pattisdr/osf.io,laurenrevere/osf.io,jmcarp/osf.io,HalcyonChimera/osf.io,sloria/osf.io,ZobairAlijan/osf.io,crcresearch/osf.io,KAsante95/osf.io,HalcyonChimera/osf.io,zkraime/osf.io,mluke93/osf.io,kwierman/osf.io,chrisseto/osf.io,Ghalko/osf.io,jeffreyliu3230/osf.io,fabianvf/osf.io,haoyuchen1992/osf.io,RomanZWang/osf.io,MerlinZhang/osf.io,TomHeatwole/osf.io,binoculars/osf.io,asanfilippo7/osf.io,pattisdr/osf.io,emetsger/osf.io,cosenal/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,mluo613/osf.io,revanthkolli/osf.io,cwisecarver/osf.io,jinluyuan/osf.io,reinaH/osf.io,DanielSBrown/osf.io,adlius/osf.io,acshi/osf.io,TomBaxter/osf.io
import os settings = {} domain = 'localhost:8080' framework = 'flask' debug = False clearOnLoad = False emailOnRegister = False registrationDisabled = False cacheDirectory = "./Site/Cache" siteDown = False database = 'osf20120530' # Mongo cookieDomain = '.openscienceframework.org' # Beaker static = '%s/static' % os.path.dirname(os.path.abspath(__file__)) local = TrueImprove OS compatibility for settings
import os settings = {} domain = 'localhost:8080' framework = 'flask' debug = False clearOnLoad = False emailOnRegister = False registrationDisabled = False cacheDirectory = "./Site/Cache" siteDown = False database = 'osf20120530' # Mongo cookieDomain = '.openscienceframework.org' # Beaker static = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'static') local = True
<commit_before>import os settings = {} domain = 'localhost:8080' framework = 'flask' debug = False clearOnLoad = False emailOnRegister = False registrationDisabled = False cacheDirectory = "./Site/Cache" siteDown = False database = 'osf20120530' # Mongo cookieDomain = '.openscienceframework.org' # Beaker static = '%s/static' % os.path.dirname(os.path.abspath(__file__)) local = True<commit_msg>Improve OS compatibility for settings<commit_after>
import os settings = {} domain = 'localhost:8080' framework = 'flask' debug = False clearOnLoad = False emailOnRegister = False registrationDisabled = False cacheDirectory = "./Site/Cache" siteDown = False database = 'osf20120530' # Mongo cookieDomain = '.openscienceframework.org' # Beaker static = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'static') local = True
import os settings = {} domain = 'localhost:8080' framework = 'flask' debug = False clearOnLoad = False emailOnRegister = False registrationDisabled = False cacheDirectory = "./Site/Cache" siteDown = False database = 'osf20120530' # Mongo cookieDomain = '.openscienceframework.org' # Beaker static = '%s/static' % os.path.dirname(os.path.abspath(__file__)) local = TrueImprove OS compatibility for settingsimport os settings = {} domain = 'localhost:8080' framework = 'flask' debug = False clearOnLoad = False emailOnRegister = False registrationDisabled = False cacheDirectory = "./Site/Cache" siteDown = False database = 'osf20120530' # Mongo cookieDomain = '.openscienceframework.org' # Beaker static = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'static') local = True
<commit_before>import os settings = {} domain = 'localhost:8080' framework = 'flask' debug = False clearOnLoad = False emailOnRegister = False registrationDisabled = False cacheDirectory = "./Site/Cache" siteDown = False database = 'osf20120530' # Mongo cookieDomain = '.openscienceframework.org' # Beaker static = '%s/static' % os.path.dirname(os.path.abspath(__file__)) local = True<commit_msg>Improve OS compatibility for settings<commit_after>import os settings = {} domain = 'localhost:8080' framework = 'flask' debug = False clearOnLoad = False emailOnRegister = False registrationDisabled = False cacheDirectory = "./Site/Cache" siteDown = False database = 'osf20120530' # Mongo cookieDomain = '.openscienceframework.org' # Beaker static = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'static') local = True
09ec6e4611a763e823a5e3d15fb233a0132fd06b
imagersite/imagersite/tests.py
imagersite/imagersite/tests.py
from django.test import Client, TestCase from django.contrib.auth.models import User HOME = '/' REGISTER = '/accounts/register/' LOGIN = '/login' LOGOUT = '/logout' class UnauthenticatedUser(TestCase): """Create unauth user for testing.""" def setUp(self): """Setup unauth user.""" client = Client() self.home = client.get(HOME) self.login = client.get(LOGIN) self.logout = client.get(LOGOUT) self.register = client.get(REGISTER) def test_no_user_in_db(self): """No user i db.""" self.assertFalse(User.objects.count()) def test_homepage(self): """Test homepage can be reached.""" self.assertEqual(self.home.status_code, 200) def test_login(self): """Test login cna be reached.""" self.assertEqual(self.login.status_code, 200) def test_logout(self): """Test logout can be reached.""" self.assertEqual(self.logout.status_code, 200) def test_register(self): """Test register can be reached.""" self.assertEqual(self.register.status_code, 200)
"""Tests for project level urls and views.""" from __future__ import unicode_literals from django.contrib.staticfiles import finders from django.test import Client, TestCase from django.contrib.auth.models import User HOME = '/' REGISTER = '/accounts/register/' LOGIN = '/login' LOGOUT = '/logout' DEFAULT_IMAGE = finders.find('static/imagersite/images/default-image.jpg') class UnauthenticatedUser(TestCase): """Create unauth user for testing.""" def setUp(self): """Setup unauth user.""" client = Client() self.home = client.get(HOME) self.login = client.get(LOGIN) self.logout = client.get(LOGOUT) self.register = client.get(REGISTER) def test_no_user_in_db(self): """No user i db.""" self.assertFalse(User.objects.count()) def test_homepage(self): """Test homepage can be reached.""" self.assertEqual(self.home.status_code, 200) def test_login(self): """Test login cna be reached.""" self.assertEqual(self.login.status_code, 200) def test_logout(self): """Test logout can be reached.""" self.assertEqual(self.logout.status_code, 200) def test_register(self): """Test register can be reached.""" self.assertEqual(self.register.status_code, 200) def test_default_image(self): """Test default image shows up.""" img_path = self.home.context['image'] self.assertEqual(img_path, DEFAULT_IMAGE)
Add passing test for default image
Add passing test for default image
Python
mit
SeleniumK/django-imager,SeleniumK/django-imager,SeleniumK/django-imager
from django.test import Client, TestCase from django.contrib.auth.models import User HOME = '/' REGISTER = '/accounts/register/' LOGIN = '/login' LOGOUT = '/logout' class UnauthenticatedUser(TestCase): """Create unauth user for testing.""" def setUp(self): """Setup unauth user.""" client = Client() self.home = client.get(HOME) self.login = client.get(LOGIN) self.logout = client.get(LOGOUT) self.register = client.get(REGISTER) def test_no_user_in_db(self): """No user i db.""" self.assertFalse(User.objects.count()) def test_homepage(self): """Test homepage can be reached.""" self.assertEqual(self.home.status_code, 200) def test_login(self): """Test login cna be reached.""" self.assertEqual(self.login.status_code, 200) def test_logout(self): """Test logout can be reached.""" self.assertEqual(self.logout.status_code, 200) def test_register(self): """Test register can be reached.""" self.assertEqual(self.register.status_code, 200) Add passing test for default image
"""Tests for project level urls and views.""" from __future__ import unicode_literals from django.contrib.staticfiles import finders from django.test import Client, TestCase from django.contrib.auth.models import User HOME = '/' REGISTER = '/accounts/register/' LOGIN = '/login' LOGOUT = '/logout' DEFAULT_IMAGE = finders.find('static/imagersite/images/default-image.jpg') class UnauthenticatedUser(TestCase): """Create unauth user for testing.""" def setUp(self): """Setup unauth user.""" client = Client() self.home = client.get(HOME) self.login = client.get(LOGIN) self.logout = client.get(LOGOUT) self.register = client.get(REGISTER) def test_no_user_in_db(self): """No user i db.""" self.assertFalse(User.objects.count()) def test_homepage(self): """Test homepage can be reached.""" self.assertEqual(self.home.status_code, 200) def test_login(self): """Test login cna be reached.""" self.assertEqual(self.login.status_code, 200) def test_logout(self): """Test logout can be reached.""" self.assertEqual(self.logout.status_code, 200) def test_register(self): """Test register can be reached.""" self.assertEqual(self.register.status_code, 200) def test_default_image(self): """Test default image shows up.""" img_path = self.home.context['image'] self.assertEqual(img_path, DEFAULT_IMAGE)
<commit_before>from django.test import Client, TestCase from django.contrib.auth.models import User HOME = '/' REGISTER = '/accounts/register/' LOGIN = '/login' LOGOUT = '/logout' class UnauthenticatedUser(TestCase): """Create unauth user for testing.""" def setUp(self): """Setup unauth user.""" client = Client() self.home = client.get(HOME) self.login = client.get(LOGIN) self.logout = client.get(LOGOUT) self.register = client.get(REGISTER) def test_no_user_in_db(self): """No user i db.""" self.assertFalse(User.objects.count()) def test_homepage(self): """Test homepage can be reached.""" self.assertEqual(self.home.status_code, 200) def test_login(self): """Test login cna be reached.""" self.assertEqual(self.login.status_code, 200) def test_logout(self): """Test logout can be reached.""" self.assertEqual(self.logout.status_code, 200) def test_register(self): """Test register can be reached.""" self.assertEqual(self.register.status_code, 200) <commit_msg>Add passing test for default image<commit_after>
"""Tests for project level urls and views.""" from __future__ import unicode_literals from django.contrib.staticfiles import finders from django.test import Client, TestCase from django.contrib.auth.models import User HOME = '/' REGISTER = '/accounts/register/' LOGIN = '/login' LOGOUT = '/logout' DEFAULT_IMAGE = finders.find('static/imagersite/images/default-image.jpg') class UnauthenticatedUser(TestCase): """Create unauth user for testing.""" def setUp(self): """Setup unauth user.""" client = Client() self.home = client.get(HOME) self.login = client.get(LOGIN) self.logout = client.get(LOGOUT) self.register = client.get(REGISTER) def test_no_user_in_db(self): """No user i db.""" self.assertFalse(User.objects.count()) def test_homepage(self): """Test homepage can be reached.""" self.assertEqual(self.home.status_code, 200) def test_login(self): """Test login cna be reached.""" self.assertEqual(self.login.status_code, 200) def test_logout(self): """Test logout can be reached.""" self.assertEqual(self.logout.status_code, 200) def test_register(self): """Test register can be reached.""" self.assertEqual(self.register.status_code, 200) def test_default_image(self): """Test default image shows up.""" img_path = self.home.context['image'] self.assertEqual(img_path, DEFAULT_IMAGE)
from django.test import Client, TestCase from django.contrib.auth.models import User HOME = '/' REGISTER = '/accounts/register/' LOGIN = '/login' LOGOUT = '/logout' class UnauthenticatedUser(TestCase): """Create unauth user for testing.""" def setUp(self): """Setup unauth user.""" client = Client() self.home = client.get(HOME) self.login = client.get(LOGIN) self.logout = client.get(LOGOUT) self.register = client.get(REGISTER) def test_no_user_in_db(self): """No user i db.""" self.assertFalse(User.objects.count()) def test_homepage(self): """Test homepage can be reached.""" self.assertEqual(self.home.status_code, 200) def test_login(self): """Test login cna be reached.""" self.assertEqual(self.login.status_code, 200) def test_logout(self): """Test logout can be reached.""" self.assertEqual(self.logout.status_code, 200) def test_register(self): """Test register can be reached.""" self.assertEqual(self.register.status_code, 200) Add passing test for default image"""Tests for project level urls and views.""" from __future__ import unicode_literals from django.contrib.staticfiles import finders from django.test import Client, TestCase from django.contrib.auth.models import User HOME = '/' REGISTER = '/accounts/register/' LOGIN = '/login' LOGOUT = '/logout' DEFAULT_IMAGE = finders.find('static/imagersite/images/default-image.jpg') class UnauthenticatedUser(TestCase): """Create unauth user for testing.""" def setUp(self): """Setup unauth user.""" client = Client() self.home = client.get(HOME) self.login = client.get(LOGIN) self.logout = client.get(LOGOUT) self.register = client.get(REGISTER) def test_no_user_in_db(self): """No user i db.""" self.assertFalse(User.objects.count()) def test_homepage(self): """Test homepage can be reached.""" self.assertEqual(self.home.status_code, 200) def test_login(self): """Test login cna be reached.""" self.assertEqual(self.login.status_code, 200) def test_logout(self): """Test logout can be reached.""" self.assertEqual(self.logout.status_code, 200) def test_register(self): """Test register can be reached.""" self.assertEqual(self.register.status_code, 200) def test_default_image(self): """Test default image shows up.""" img_path = self.home.context['image'] self.assertEqual(img_path, DEFAULT_IMAGE)
<commit_before>from django.test import Client, TestCase from django.contrib.auth.models import User HOME = '/' REGISTER = '/accounts/register/' LOGIN = '/login' LOGOUT = '/logout' class UnauthenticatedUser(TestCase): """Create unauth user for testing.""" def setUp(self): """Setup unauth user.""" client = Client() self.home = client.get(HOME) self.login = client.get(LOGIN) self.logout = client.get(LOGOUT) self.register = client.get(REGISTER) def test_no_user_in_db(self): """No user i db.""" self.assertFalse(User.objects.count()) def test_homepage(self): """Test homepage can be reached.""" self.assertEqual(self.home.status_code, 200) def test_login(self): """Test login cna be reached.""" self.assertEqual(self.login.status_code, 200) def test_logout(self): """Test logout can be reached.""" self.assertEqual(self.logout.status_code, 200) def test_register(self): """Test register can be reached.""" self.assertEqual(self.register.status_code, 200) <commit_msg>Add passing test for default image<commit_after>"""Tests for project level urls and views.""" from __future__ import unicode_literals from django.contrib.staticfiles import finders from django.test import Client, TestCase from django.contrib.auth.models import User HOME = '/' REGISTER = '/accounts/register/' LOGIN = '/login' LOGOUT = '/logout' DEFAULT_IMAGE = finders.find('static/imagersite/images/default-image.jpg') class UnauthenticatedUser(TestCase): """Create unauth user for testing.""" def setUp(self): """Setup unauth user.""" client = Client() self.home = client.get(HOME) self.login = client.get(LOGIN) self.logout = client.get(LOGOUT) self.register = client.get(REGISTER) def test_no_user_in_db(self): """No user i db.""" self.assertFalse(User.objects.count()) def test_homepage(self): """Test homepage can be reached.""" self.assertEqual(self.home.status_code, 200) def test_login(self): """Test login cna be reached.""" self.assertEqual(self.login.status_code, 200) def test_logout(self): """Test logout can be reached.""" self.assertEqual(self.logout.status_code, 200) def test_register(self): """Test register can be reached.""" self.assertEqual(self.register.status_code, 200) def test_default_image(self): """Test default image shows up.""" img_path = self.home.context['image'] self.assertEqual(img_path, DEFAULT_IMAGE)
24c24ab8a6c662079c397a8d91228b3b8d45f033
testing/test_integration.py
testing/test_integration.py
import sys from setuptools_scm.utils import do def test_pyproject_support(tmpdir, monkeypatch): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.ensure("package", dir=42) pkg.join("pyproject.toml").write( """[tool.setuptools_scm] fallback_version = "12.34" """ ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "12.34"
import sys import pytest from setuptools_scm.utils import do @pytest.fixture def wd(wd): wd("git init") wd("git config user.email test@example.com") wd('git config user.name "a test"') wd.add_command = "git add ." wd.commit_command = "git commit -m test-{reason}" return wd def test_pyproject_support(tmpdir, monkeypatch): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.ensure("package", dir=42) pkg.join("pyproject.toml").write( """[tool.setuptools_scm] fallback_version = "12.34" """ ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "12.34" @pytest.mark.xfail def test_pyproject_support_with_git(tmpdir, monkeypatch, wd): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.join("wd") pkg.join("pyproject.toml").write( """[tool.setuptools_scm]""" ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "0.1.dev0"
Test pyproject.toml integration using git
Test pyproject.toml integration using git See #374.
Python
mit
pypa/setuptools_scm,pypa/setuptools_scm,RonnyPfannschmidt/setuptools_scm,RonnyPfannschmidt/setuptools_scm
import sys from setuptools_scm.utils import do def test_pyproject_support(tmpdir, monkeypatch): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.ensure("package", dir=42) pkg.join("pyproject.toml").write( """[tool.setuptools_scm] fallback_version = "12.34" """ ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "12.34" Test pyproject.toml integration using git See #374.
import sys import pytest from setuptools_scm.utils import do @pytest.fixture def wd(wd): wd("git init") wd("git config user.email test@example.com") wd('git config user.name "a test"') wd.add_command = "git add ." wd.commit_command = "git commit -m test-{reason}" return wd def test_pyproject_support(tmpdir, monkeypatch): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.ensure("package", dir=42) pkg.join("pyproject.toml").write( """[tool.setuptools_scm] fallback_version = "12.34" """ ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "12.34" @pytest.mark.xfail def test_pyproject_support_with_git(tmpdir, monkeypatch, wd): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.join("wd") pkg.join("pyproject.toml").write( """[tool.setuptools_scm]""" ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "0.1.dev0"
<commit_before>import sys from setuptools_scm.utils import do def test_pyproject_support(tmpdir, monkeypatch): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.ensure("package", dir=42) pkg.join("pyproject.toml").write( """[tool.setuptools_scm] fallback_version = "12.34" """ ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "12.34" <commit_msg>Test pyproject.toml integration using git See #374.<commit_after>
import sys import pytest from setuptools_scm.utils import do @pytest.fixture def wd(wd): wd("git init") wd("git config user.email test@example.com") wd('git config user.name "a test"') wd.add_command = "git add ." wd.commit_command = "git commit -m test-{reason}" return wd def test_pyproject_support(tmpdir, monkeypatch): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.ensure("package", dir=42) pkg.join("pyproject.toml").write( """[tool.setuptools_scm] fallback_version = "12.34" """ ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "12.34" @pytest.mark.xfail def test_pyproject_support_with_git(tmpdir, monkeypatch, wd): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.join("wd") pkg.join("pyproject.toml").write( """[tool.setuptools_scm]""" ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "0.1.dev0"
import sys from setuptools_scm.utils import do def test_pyproject_support(tmpdir, monkeypatch): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.ensure("package", dir=42) pkg.join("pyproject.toml").write( """[tool.setuptools_scm] fallback_version = "12.34" """ ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "12.34" Test pyproject.toml integration using git See #374.import sys import pytest from setuptools_scm.utils import do @pytest.fixture def wd(wd): wd("git init") wd("git config user.email test@example.com") wd('git config user.name "a test"') wd.add_command = "git add ." wd.commit_command = "git commit -m test-{reason}" return wd def test_pyproject_support(tmpdir, monkeypatch): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.ensure("package", dir=42) pkg.join("pyproject.toml").write( """[tool.setuptools_scm] fallback_version = "12.34" """ ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "12.34" @pytest.mark.xfail def test_pyproject_support_with_git(tmpdir, monkeypatch, wd): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.join("wd") pkg.join("pyproject.toml").write( """[tool.setuptools_scm]""" ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "0.1.dev0"
<commit_before>import sys from setuptools_scm.utils import do def test_pyproject_support(tmpdir, monkeypatch): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.ensure("package", dir=42) pkg.join("pyproject.toml").write( """[tool.setuptools_scm] fallback_version = "12.34" """ ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "12.34" <commit_msg>Test pyproject.toml integration using git See #374.<commit_after>import sys import pytest from setuptools_scm.utils import do @pytest.fixture def wd(wd): wd("git init") wd("git config user.email test@example.com") wd('git config user.name "a test"') wd.add_command = "git add ." wd.commit_command = "git commit -m test-{reason}" return wd def test_pyproject_support(tmpdir, monkeypatch): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.ensure("package", dir=42) pkg.join("pyproject.toml").write( """[tool.setuptools_scm] fallback_version = "12.34" """ ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "12.34" @pytest.mark.xfail def test_pyproject_support_with_git(tmpdir, monkeypatch, wd): monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") pkg = tmpdir.join("wd") pkg.join("pyproject.toml").write( """[tool.setuptools_scm]""" ) pkg.join("setup.py").write("__import__('setuptools').setup()") res = do((sys.executable, "setup.py", "--version"), pkg) assert res == "0.1.dev0"
81936bfbac9254fcd90d294c299ad635504cd93c
police_api/service.py
police_api/service.py
import logging import requests from .exceptions import APIError from .version import __version__ logger = logging.getLogger(__name__) class BaseService(object): def __init__(self, api, **config): self.api = api self.config = { 'base_url': 'http://data.police.uk/api/', 'user_agent': 'police-api-client-python/%s' % __version__, } self.config.update(config) def raise_for_status(self, request): try: request.raise_for_status() except requests.models.HTTPError as e: raise APIError(e) def request(self, verb, method, **kwargs): verb = verb.upper() request_kwargs = { 'headers': { 'User-Agent': self.config['user_agent'], }, 'timeout': self.config.get('timeout', 30), } if 'username' in self.config: request_kwargs['auth'] = (self.config.get('username', ''), self.config.get('password', '')) if verb == 'GET': request_kwargs['params'] = kwargs else: request_kwargs['data'] = kwargs url = self.config['base_url'] + method logger.debug('%s %s' % (verb, url)) r = requests.request(verb, url, **request_kwargs) self.raise_for_status(r) return r.json()
import logging import requests from .exceptions import APIError from .version import __version__ logger = logging.getLogger(__name__) class BaseService(object): def __init__(self, api, **config): self.api = api self.config = { 'base_url': 'http://data.police.uk/api/', 'user_agent': 'police-api-client-python/%s' % __version__, } self.config.update(config) def raise_for_status(self, request): try: request.raise_for_status() except requests.models.HTTPError as e: raise APIError(e) def _make_request(self, verb, url, params={}): request_kwargs = { 'headers': { 'User-Agent': self.config['user_agent'], }, 'timeout': self.config.get('timeout', 30), } if 'username' in self.config: request_kwargs['auth'] = (self.config.get('username', ''), self.config.get('password', '')) if verb == 'GET': request_kwargs['params'] = params else: request_kwargs['data'] = params logger.debug('%s %s' % (verb, url)) r = requests.request(verb, url, **request_kwargs) self.raise_for_status(r) return r.json() def request(self, verb, method, **kwargs): url = self.config['base_url'] + method return self._make_request(verb.upper(), url, kwargs)
Refactor request mechanics into an internal method on BaseService
Refactor request mechanics into an internal method on BaseService
Python
mit
rkhleics/police-api-client-python
import logging import requests from .exceptions import APIError from .version import __version__ logger = logging.getLogger(__name__) class BaseService(object): def __init__(self, api, **config): self.api = api self.config = { 'base_url': 'http://data.police.uk/api/', 'user_agent': 'police-api-client-python/%s' % __version__, } self.config.update(config) def raise_for_status(self, request): try: request.raise_for_status() except requests.models.HTTPError as e: raise APIError(e) def request(self, verb, method, **kwargs): verb = verb.upper() request_kwargs = { 'headers': { 'User-Agent': self.config['user_agent'], }, 'timeout': self.config.get('timeout', 30), } if 'username' in self.config: request_kwargs['auth'] = (self.config.get('username', ''), self.config.get('password', '')) if verb == 'GET': request_kwargs['params'] = kwargs else: request_kwargs['data'] = kwargs url = self.config['base_url'] + method logger.debug('%s %s' % (verb, url)) r = requests.request(verb, url, **request_kwargs) self.raise_for_status(r) return r.json() Refactor request mechanics into an internal method on BaseService
import logging import requests from .exceptions import APIError from .version import __version__ logger = logging.getLogger(__name__) class BaseService(object): def __init__(self, api, **config): self.api = api self.config = { 'base_url': 'http://data.police.uk/api/', 'user_agent': 'police-api-client-python/%s' % __version__, } self.config.update(config) def raise_for_status(self, request): try: request.raise_for_status() except requests.models.HTTPError as e: raise APIError(e) def _make_request(self, verb, url, params={}): request_kwargs = { 'headers': { 'User-Agent': self.config['user_agent'], }, 'timeout': self.config.get('timeout', 30), } if 'username' in self.config: request_kwargs['auth'] = (self.config.get('username', ''), self.config.get('password', '')) if verb == 'GET': request_kwargs['params'] = params else: request_kwargs['data'] = params logger.debug('%s %s' % (verb, url)) r = requests.request(verb, url, **request_kwargs) self.raise_for_status(r) return r.json() def request(self, verb, method, **kwargs): url = self.config['base_url'] + method return self._make_request(verb.upper(), url, kwargs)
<commit_before>import logging import requests from .exceptions import APIError from .version import __version__ logger = logging.getLogger(__name__) class BaseService(object): def __init__(self, api, **config): self.api = api self.config = { 'base_url': 'http://data.police.uk/api/', 'user_agent': 'police-api-client-python/%s' % __version__, } self.config.update(config) def raise_for_status(self, request): try: request.raise_for_status() except requests.models.HTTPError as e: raise APIError(e) def request(self, verb, method, **kwargs): verb = verb.upper() request_kwargs = { 'headers': { 'User-Agent': self.config['user_agent'], }, 'timeout': self.config.get('timeout', 30), } if 'username' in self.config: request_kwargs['auth'] = (self.config.get('username', ''), self.config.get('password', '')) if verb == 'GET': request_kwargs['params'] = kwargs else: request_kwargs['data'] = kwargs url = self.config['base_url'] + method logger.debug('%s %s' % (verb, url)) r = requests.request(verb, url, **request_kwargs) self.raise_for_status(r) return r.json() <commit_msg>Refactor request mechanics into an internal method on BaseService<commit_after>
import logging import requests from .exceptions import APIError from .version import __version__ logger = logging.getLogger(__name__) class BaseService(object): def __init__(self, api, **config): self.api = api self.config = { 'base_url': 'http://data.police.uk/api/', 'user_agent': 'police-api-client-python/%s' % __version__, } self.config.update(config) def raise_for_status(self, request): try: request.raise_for_status() except requests.models.HTTPError as e: raise APIError(e) def _make_request(self, verb, url, params={}): request_kwargs = { 'headers': { 'User-Agent': self.config['user_agent'], }, 'timeout': self.config.get('timeout', 30), } if 'username' in self.config: request_kwargs['auth'] = (self.config.get('username', ''), self.config.get('password', '')) if verb == 'GET': request_kwargs['params'] = params else: request_kwargs['data'] = params logger.debug('%s %s' % (verb, url)) r = requests.request(verb, url, **request_kwargs) self.raise_for_status(r) return r.json() def request(self, verb, method, **kwargs): url = self.config['base_url'] + method return self._make_request(verb.upper(), url, kwargs)
import logging import requests from .exceptions import APIError from .version import __version__ logger = logging.getLogger(__name__) class BaseService(object): def __init__(self, api, **config): self.api = api self.config = { 'base_url': 'http://data.police.uk/api/', 'user_agent': 'police-api-client-python/%s' % __version__, } self.config.update(config) def raise_for_status(self, request): try: request.raise_for_status() except requests.models.HTTPError as e: raise APIError(e) def request(self, verb, method, **kwargs): verb = verb.upper() request_kwargs = { 'headers': { 'User-Agent': self.config['user_agent'], }, 'timeout': self.config.get('timeout', 30), } if 'username' in self.config: request_kwargs['auth'] = (self.config.get('username', ''), self.config.get('password', '')) if verb == 'GET': request_kwargs['params'] = kwargs else: request_kwargs['data'] = kwargs url = self.config['base_url'] + method logger.debug('%s %s' % (verb, url)) r = requests.request(verb, url, **request_kwargs) self.raise_for_status(r) return r.json() Refactor request mechanics into an internal method on BaseServiceimport logging import requests from .exceptions import APIError from .version import __version__ logger = logging.getLogger(__name__) class BaseService(object): def __init__(self, api, **config): self.api = api self.config = { 'base_url': 'http://data.police.uk/api/', 'user_agent': 'police-api-client-python/%s' % __version__, } self.config.update(config) def raise_for_status(self, request): try: request.raise_for_status() except requests.models.HTTPError as e: raise APIError(e) def _make_request(self, verb, url, params={}): request_kwargs = { 'headers': { 'User-Agent': self.config['user_agent'], }, 'timeout': self.config.get('timeout', 30), } if 'username' in self.config: request_kwargs['auth'] = (self.config.get('username', ''), self.config.get('password', '')) if verb == 'GET': request_kwargs['params'] = params else: request_kwargs['data'] = params logger.debug('%s %s' % (verb, url)) r = requests.request(verb, url, **request_kwargs) self.raise_for_status(r) return r.json() def request(self, verb, method, **kwargs): url = self.config['base_url'] + method return self._make_request(verb.upper(), url, kwargs)
<commit_before>import logging import requests from .exceptions import APIError from .version import __version__ logger = logging.getLogger(__name__) class BaseService(object): def __init__(self, api, **config): self.api = api self.config = { 'base_url': 'http://data.police.uk/api/', 'user_agent': 'police-api-client-python/%s' % __version__, } self.config.update(config) def raise_for_status(self, request): try: request.raise_for_status() except requests.models.HTTPError as e: raise APIError(e) def request(self, verb, method, **kwargs): verb = verb.upper() request_kwargs = { 'headers': { 'User-Agent': self.config['user_agent'], }, 'timeout': self.config.get('timeout', 30), } if 'username' in self.config: request_kwargs['auth'] = (self.config.get('username', ''), self.config.get('password', '')) if verb == 'GET': request_kwargs['params'] = kwargs else: request_kwargs['data'] = kwargs url = self.config['base_url'] + method logger.debug('%s %s' % (verb, url)) r = requests.request(verb, url, **request_kwargs) self.raise_for_status(r) return r.json() <commit_msg>Refactor request mechanics into an internal method on BaseService<commit_after>import logging import requests from .exceptions import APIError from .version import __version__ logger = logging.getLogger(__name__) class BaseService(object): def __init__(self, api, **config): self.api = api self.config = { 'base_url': 'http://data.police.uk/api/', 'user_agent': 'police-api-client-python/%s' % __version__, } self.config.update(config) def raise_for_status(self, request): try: request.raise_for_status() except requests.models.HTTPError as e: raise APIError(e) def _make_request(self, verb, url, params={}): request_kwargs = { 'headers': { 'User-Agent': self.config['user_agent'], }, 'timeout': self.config.get('timeout', 30), } if 'username' in self.config: request_kwargs['auth'] = (self.config.get('username', ''), self.config.get('password', '')) if verb == 'GET': request_kwargs['params'] = params else: request_kwargs['data'] = params logger.debug('%s %s' % (verb, url)) r = requests.request(verb, url, **request_kwargs) self.raise_for_status(r) return r.json() def request(self, verb, method, **kwargs): url = self.config['base_url'] + method return self._make_request(verb.upper(), url, kwargs)
178c25714aaae056c115f1580f19d833486a54ac
datapipe/targets/objects.py
datapipe/targets/objects.py
from ..target import Target import hashlib import dill import joblib class PyTarget(Target): def __init__(self, name, obj=None): self._name = name self._obj = obj super(PyTarget, self).__init__() if not obj is None: self.set(obj) def identifier(self): return self._name def get(self): return self._obj def set(self, obj): self._obj = obj def is_damaged(self): stored = self.stored() if stored: if self._obj is None: self._obj = stored._obj return stored._obj is None else: return joblib.hash(self._obj) != joblib.hash(stored._obj) else: return self._obj is None
from ..target import Target import hashlib import dill import joblib class PyTarget(Target): def __init__(self, name, obj=None): self._name = name self._obj = obj super(PyTarget, self).__init__() if not obj is None: self.set(obj) def identifier(self): return self._name def get(self): return self._obj def set(self, obj): self._obj = obj self._memory['obj'] = dill.dumps(obj).encode('base64') def is_damaged(self): mem = self.stored() if mem: if self._obj is None: self._obj = dill.loads(mem['obj'].decode('base64')) return stored._obj is None else: return joblib.hash(self._obj) != joblib.hash(dill.loads(mem['obj'].decode('base64'))) else: return self._obj is None
Make PyTarget object work again
Make PyTarget object work again We now save a base64 encoded pickled version of the object.
Python
mit
ibab/datapipe
from ..target import Target import hashlib import dill import joblib class PyTarget(Target): def __init__(self, name, obj=None): self._name = name self._obj = obj super(PyTarget, self).__init__() if not obj is None: self.set(obj) def identifier(self): return self._name def get(self): return self._obj def set(self, obj): self._obj = obj def is_damaged(self): stored = self.stored() if stored: if self._obj is None: self._obj = stored._obj return stored._obj is None else: return joblib.hash(self._obj) != joblib.hash(stored._obj) else: return self._obj is None Make PyTarget object work again We now save a base64 encoded pickled version of the object.
from ..target import Target import hashlib import dill import joblib class PyTarget(Target): def __init__(self, name, obj=None): self._name = name self._obj = obj super(PyTarget, self).__init__() if not obj is None: self.set(obj) def identifier(self): return self._name def get(self): return self._obj def set(self, obj): self._obj = obj self._memory['obj'] = dill.dumps(obj).encode('base64') def is_damaged(self): mem = self.stored() if mem: if self._obj is None: self._obj = dill.loads(mem['obj'].decode('base64')) return stored._obj is None else: return joblib.hash(self._obj) != joblib.hash(dill.loads(mem['obj'].decode('base64'))) else: return self._obj is None
<commit_before>from ..target import Target import hashlib import dill import joblib class PyTarget(Target): def __init__(self, name, obj=None): self._name = name self._obj = obj super(PyTarget, self).__init__() if not obj is None: self.set(obj) def identifier(self): return self._name def get(self): return self._obj def set(self, obj): self._obj = obj def is_damaged(self): stored = self.stored() if stored: if self._obj is None: self._obj = stored._obj return stored._obj is None else: return joblib.hash(self._obj) != joblib.hash(stored._obj) else: return self._obj is None <commit_msg>Make PyTarget object work again We now save a base64 encoded pickled version of the object.<commit_after>
from ..target import Target import hashlib import dill import joblib class PyTarget(Target): def __init__(self, name, obj=None): self._name = name self._obj = obj super(PyTarget, self).__init__() if not obj is None: self.set(obj) def identifier(self): return self._name def get(self): return self._obj def set(self, obj): self._obj = obj self._memory['obj'] = dill.dumps(obj).encode('base64') def is_damaged(self): mem = self.stored() if mem: if self._obj is None: self._obj = dill.loads(mem['obj'].decode('base64')) return stored._obj is None else: return joblib.hash(self._obj) != joblib.hash(dill.loads(mem['obj'].decode('base64'))) else: return self._obj is None
from ..target import Target import hashlib import dill import joblib class PyTarget(Target): def __init__(self, name, obj=None): self._name = name self._obj = obj super(PyTarget, self).__init__() if not obj is None: self.set(obj) def identifier(self): return self._name def get(self): return self._obj def set(self, obj): self._obj = obj def is_damaged(self): stored = self.stored() if stored: if self._obj is None: self._obj = stored._obj return stored._obj is None else: return joblib.hash(self._obj) != joblib.hash(stored._obj) else: return self._obj is None Make PyTarget object work again We now save a base64 encoded pickled version of the object.from ..target import Target import hashlib import dill import joblib class PyTarget(Target): def __init__(self, name, obj=None): self._name = name self._obj = obj super(PyTarget, self).__init__() if not obj is None: self.set(obj) def identifier(self): return self._name def get(self): return self._obj def set(self, obj): self._obj = obj self._memory['obj'] = dill.dumps(obj).encode('base64') def is_damaged(self): mem = self.stored() if mem: if self._obj is None: self._obj = dill.loads(mem['obj'].decode('base64')) return stored._obj is None else: return joblib.hash(self._obj) != joblib.hash(dill.loads(mem['obj'].decode('base64'))) else: return self._obj is None
<commit_before>from ..target import Target import hashlib import dill import joblib class PyTarget(Target): def __init__(self, name, obj=None): self._name = name self._obj = obj super(PyTarget, self).__init__() if not obj is None: self.set(obj) def identifier(self): return self._name def get(self): return self._obj def set(self, obj): self._obj = obj def is_damaged(self): stored = self.stored() if stored: if self._obj is None: self._obj = stored._obj return stored._obj is None else: return joblib.hash(self._obj) != joblib.hash(stored._obj) else: return self._obj is None <commit_msg>Make PyTarget object work again We now save a base64 encoded pickled version of the object.<commit_after>from ..target import Target import hashlib import dill import joblib class PyTarget(Target): def __init__(self, name, obj=None): self._name = name self._obj = obj super(PyTarget, self).__init__() if not obj is None: self.set(obj) def identifier(self): return self._name def get(self): return self._obj def set(self, obj): self._obj = obj self._memory['obj'] = dill.dumps(obj).encode('base64') def is_damaged(self): mem = self.stored() if mem: if self._obj is None: self._obj = dill.loads(mem['obj'].decode('base64')) return stored._obj is None else: return joblib.hash(self._obj) != joblib.hash(dill.loads(mem['obj'].decode('base64'))) else: return self._obj is None
4d3f809ba5e1b5109e6f2e73d9c9630371660210
Bookie/bookie/lib/access.py
Bookie/bookie/lib/access.py
"""Handle auth and authz activities in bookie""" from pyramid.httpexceptions import HTTPForbidden class Authorize(object): """Context manager to check if the user is authorized use: with Authorize(some_key): # do work Will return NotAuthorized if it fails """ def __init__(self, submitted_key, config_key): """Create the context manager""" self.api_key = config_key self.check_key = submitted_key def __enter__(self): """Verify api key set in constructor""" if self.api_key != self.check_key: raise HTTPForbidden('Invalid Authorization') def __exit__(self, exc_type, exc_value, traceback): """No cleanup work to do after usage""" pass
"""Handle auth and authz activities in bookie""" import logging from pyramid.httpexceptions import HTTPForbidden LOG = logging.getLogger(__name__) class Authorize(object): """Context manager to check if the user is authorized use: with Authorize(some_key): # do work Will return NotAuthorized if it fails """ def __init__(self, submitted_key, config_key): """Create the context manager""" self.api_key = config_key self.check_key = submitted_key def __enter__(self): """Verify api key set in constructor""" if self.api_key != self.check_key: LOG.error('Invalid API Key! {0} v {1}'.format(self.api_key, self.check_key)) raise HTTPForbidden('Invalid Authorization') def __exit__(self, exc_type, exc_value, traceback): """No cleanup work to do after usage""" pass
Update to make sure we log an error with an invalid key
Update to make sure we log an error with an invalid key
Python
agpl-3.0
GreenLunar/Bookie,adamlincoln/Bookie,wangjun/Bookie,bookieio/Bookie,GreenLunar/Bookie,teodesson/Bookie,adamlincoln/Bookie,pombredanne/Bookie,bookieio/Bookie,teodesson/Bookie,adamlincoln/Bookie,GreenLunar/Bookie,wangjun/Bookie,wangjun/Bookie,skmezanul/Bookie,wangjun/Bookie,GreenLunar/Bookie,skmezanul/Bookie,adamlincoln/Bookie,pombredanne/Bookie,skmezanul/Bookie,bookieio/Bookie,charany1/Bookie,teodesson/Bookie,skmezanul/Bookie,charany1/Bookie,pombredanne/Bookie,charany1/Bookie,teodesson/Bookie,bookieio/Bookie
"""Handle auth and authz activities in bookie""" from pyramid.httpexceptions import HTTPForbidden class Authorize(object): """Context manager to check if the user is authorized use: with Authorize(some_key): # do work Will return NotAuthorized if it fails """ def __init__(self, submitted_key, config_key): """Create the context manager""" self.api_key = config_key self.check_key = submitted_key def __enter__(self): """Verify api key set in constructor""" if self.api_key != self.check_key: raise HTTPForbidden('Invalid Authorization') def __exit__(self, exc_type, exc_value, traceback): """No cleanup work to do after usage""" pass Update to make sure we log an error with an invalid key
"""Handle auth and authz activities in bookie""" import logging from pyramid.httpexceptions import HTTPForbidden LOG = logging.getLogger(__name__) class Authorize(object): """Context manager to check if the user is authorized use: with Authorize(some_key): # do work Will return NotAuthorized if it fails """ def __init__(self, submitted_key, config_key): """Create the context manager""" self.api_key = config_key self.check_key = submitted_key def __enter__(self): """Verify api key set in constructor""" if self.api_key != self.check_key: LOG.error('Invalid API Key! {0} v {1}'.format(self.api_key, self.check_key)) raise HTTPForbidden('Invalid Authorization') def __exit__(self, exc_type, exc_value, traceback): """No cleanup work to do after usage""" pass
<commit_before>"""Handle auth and authz activities in bookie""" from pyramid.httpexceptions import HTTPForbidden class Authorize(object): """Context manager to check if the user is authorized use: with Authorize(some_key): # do work Will return NotAuthorized if it fails """ def __init__(self, submitted_key, config_key): """Create the context manager""" self.api_key = config_key self.check_key = submitted_key def __enter__(self): """Verify api key set in constructor""" if self.api_key != self.check_key: raise HTTPForbidden('Invalid Authorization') def __exit__(self, exc_type, exc_value, traceback): """No cleanup work to do after usage""" pass <commit_msg>Update to make sure we log an error with an invalid key<commit_after>
"""Handle auth and authz activities in bookie""" import logging from pyramid.httpexceptions import HTTPForbidden LOG = logging.getLogger(__name__) class Authorize(object): """Context manager to check if the user is authorized use: with Authorize(some_key): # do work Will return NotAuthorized if it fails """ def __init__(self, submitted_key, config_key): """Create the context manager""" self.api_key = config_key self.check_key = submitted_key def __enter__(self): """Verify api key set in constructor""" if self.api_key != self.check_key: LOG.error('Invalid API Key! {0} v {1}'.format(self.api_key, self.check_key)) raise HTTPForbidden('Invalid Authorization') def __exit__(self, exc_type, exc_value, traceback): """No cleanup work to do after usage""" pass
"""Handle auth and authz activities in bookie""" from pyramid.httpexceptions import HTTPForbidden class Authorize(object): """Context manager to check if the user is authorized use: with Authorize(some_key): # do work Will return NotAuthorized if it fails """ def __init__(self, submitted_key, config_key): """Create the context manager""" self.api_key = config_key self.check_key = submitted_key def __enter__(self): """Verify api key set in constructor""" if self.api_key != self.check_key: raise HTTPForbidden('Invalid Authorization') def __exit__(self, exc_type, exc_value, traceback): """No cleanup work to do after usage""" pass Update to make sure we log an error with an invalid key"""Handle auth and authz activities in bookie""" import logging from pyramid.httpexceptions import HTTPForbidden LOG = logging.getLogger(__name__) class Authorize(object): """Context manager to check if the user is authorized use: with Authorize(some_key): # do work Will return NotAuthorized if it fails """ def __init__(self, submitted_key, config_key): """Create the context manager""" self.api_key = config_key self.check_key = submitted_key def __enter__(self): """Verify api key set in constructor""" if self.api_key != self.check_key: LOG.error('Invalid API Key! {0} v {1}'.format(self.api_key, self.check_key)) raise HTTPForbidden('Invalid Authorization') def __exit__(self, exc_type, exc_value, traceback): """No cleanup work to do after usage""" pass
<commit_before>"""Handle auth and authz activities in bookie""" from pyramid.httpexceptions import HTTPForbidden class Authorize(object): """Context manager to check if the user is authorized use: with Authorize(some_key): # do work Will return NotAuthorized if it fails """ def __init__(self, submitted_key, config_key): """Create the context manager""" self.api_key = config_key self.check_key = submitted_key def __enter__(self): """Verify api key set in constructor""" if self.api_key != self.check_key: raise HTTPForbidden('Invalid Authorization') def __exit__(self, exc_type, exc_value, traceback): """No cleanup work to do after usage""" pass <commit_msg>Update to make sure we log an error with an invalid key<commit_after>"""Handle auth and authz activities in bookie""" import logging from pyramid.httpexceptions import HTTPForbidden LOG = logging.getLogger(__name__) class Authorize(object): """Context manager to check if the user is authorized use: with Authorize(some_key): # do work Will return NotAuthorized if it fails """ def __init__(self, submitted_key, config_key): """Create the context manager""" self.api_key = config_key self.check_key = submitted_key def __enter__(self): """Verify api key set in constructor""" if self.api_key != self.check_key: LOG.error('Invalid API Key! {0} v {1}'.format(self.api_key, self.check_key)) raise HTTPForbidden('Invalid Authorization') def __exit__(self, exc_type, exc_value, traceback): """No cleanup work to do after usage""" pass
8b4ea06ae8c61f0745a13e4c0118d6f499a31738
app.py
app.py
from flask import Flask, jsonify, request from dotenv import load_dotenv, find_dotenv from twilio import twiml from Queue import Queue from threading import Thread from time import sleep load_dotenv(find_dotenv()) directions = ['forward', 'backward'] task_q = Queue() def send_rasp(task_q): while True: sleep(2) if not task_q.empty(): message = task_q.get() print(message) rasp_signal = Thread(target=send_rasp, args=(task_q, )) rasp_signal.setDaemon(True) rasp_signal.start() app = Flask(__name__) @app.route('/message', methods=['POST']) def roomba_command(): # twilio text message body = request.form['Body'] resp = handle_twilio_message(body) twilio_resp = twiml.Response() twilio_resp.message(resp) return str(twilio_resp) def handle_twilio_message(message): if message.lower() in directions: task_q.put(message.lower()) return 'Message sent' try: degree = float(message) except ValueError as e: return 'Invalid command' task_q.put(str(degree)) return 'Message sent' if __name__ == '__main__': app.run(debug=True)
from flask import Flask, jsonify, request from dotenv import load_dotenv, find_dotenv from twilio import twiml from Queue import Queue from threading import Thread from time import sleep load_dotenv(find_dotenv()) directions = ['forward', 'backward'] task_q = Queue() def send_rasp(task_q): while True: sleep(2) if task_q.empty(): continue message = task_q.get() print(message) rasp_signal = Thread(target=send_rasp, args=(task_q, )) rasp_signal.setDaemon(True) rasp_signal.start() app = Flask(__name__) @app.route('/message', methods=['POST']) def roomba_command(): # twilio text message body = request.form['Body'] resp = handle_twilio_message(body) twilio_resp = twiml.Response() twilio_resp.message(resp) return str(twilio_resp) def handle_twilio_message(message): if message.lower() in directions: task_q.put(message.lower()) return 'Message sent' try: degree = float(message) except ValueError as e: return 'Invalid command' task_q.put(str(degree)) return 'Message sent' if __name__ == '__main__': app.run(debug=True)
Use continue in task loop
Use continue in task loop
Python
mit
tforrest/twilio-plays-roomba-flask
from flask import Flask, jsonify, request from dotenv import load_dotenv, find_dotenv from twilio import twiml from Queue import Queue from threading import Thread from time import sleep load_dotenv(find_dotenv()) directions = ['forward', 'backward'] task_q = Queue() def send_rasp(task_q): while True: sleep(2) if not task_q.empty(): message = task_q.get() print(message) rasp_signal = Thread(target=send_rasp, args=(task_q, )) rasp_signal.setDaemon(True) rasp_signal.start() app = Flask(__name__) @app.route('/message', methods=['POST']) def roomba_command(): # twilio text message body = request.form['Body'] resp = handle_twilio_message(body) twilio_resp = twiml.Response() twilio_resp.message(resp) return str(twilio_resp) def handle_twilio_message(message): if message.lower() in directions: task_q.put(message.lower()) return 'Message sent' try: degree = float(message) except ValueError as e: return 'Invalid command' task_q.put(str(degree)) return 'Message sent' if __name__ == '__main__': app.run(debug=True)Use continue in task loop
from flask import Flask, jsonify, request from dotenv import load_dotenv, find_dotenv from twilio import twiml from Queue import Queue from threading import Thread from time import sleep load_dotenv(find_dotenv()) directions = ['forward', 'backward'] task_q = Queue() def send_rasp(task_q): while True: sleep(2) if task_q.empty(): continue message = task_q.get() print(message) rasp_signal = Thread(target=send_rasp, args=(task_q, )) rasp_signal.setDaemon(True) rasp_signal.start() app = Flask(__name__) @app.route('/message', methods=['POST']) def roomba_command(): # twilio text message body = request.form['Body'] resp = handle_twilio_message(body) twilio_resp = twiml.Response() twilio_resp.message(resp) return str(twilio_resp) def handle_twilio_message(message): if message.lower() in directions: task_q.put(message.lower()) return 'Message sent' try: degree = float(message) except ValueError as e: return 'Invalid command' task_q.put(str(degree)) return 'Message sent' if __name__ == '__main__': app.run(debug=True)
<commit_before>from flask import Flask, jsonify, request from dotenv import load_dotenv, find_dotenv from twilio import twiml from Queue import Queue from threading import Thread from time import sleep load_dotenv(find_dotenv()) directions = ['forward', 'backward'] task_q = Queue() def send_rasp(task_q): while True: sleep(2) if not task_q.empty(): message = task_q.get() print(message) rasp_signal = Thread(target=send_rasp, args=(task_q, )) rasp_signal.setDaemon(True) rasp_signal.start() app = Flask(__name__) @app.route('/message', methods=['POST']) def roomba_command(): # twilio text message body = request.form['Body'] resp = handle_twilio_message(body) twilio_resp = twiml.Response() twilio_resp.message(resp) return str(twilio_resp) def handle_twilio_message(message): if message.lower() in directions: task_q.put(message.lower()) return 'Message sent' try: degree = float(message) except ValueError as e: return 'Invalid command' task_q.put(str(degree)) return 'Message sent' if __name__ == '__main__': app.run(debug=True)<commit_msg>Use continue in task loop<commit_after>
from flask import Flask, jsonify, request from dotenv import load_dotenv, find_dotenv from twilio import twiml from Queue import Queue from threading import Thread from time import sleep load_dotenv(find_dotenv()) directions = ['forward', 'backward'] task_q = Queue() def send_rasp(task_q): while True: sleep(2) if task_q.empty(): continue message = task_q.get() print(message) rasp_signal = Thread(target=send_rasp, args=(task_q, )) rasp_signal.setDaemon(True) rasp_signal.start() app = Flask(__name__) @app.route('/message', methods=['POST']) def roomba_command(): # twilio text message body = request.form['Body'] resp = handle_twilio_message(body) twilio_resp = twiml.Response() twilio_resp.message(resp) return str(twilio_resp) def handle_twilio_message(message): if message.lower() in directions: task_q.put(message.lower()) return 'Message sent' try: degree = float(message) except ValueError as e: return 'Invalid command' task_q.put(str(degree)) return 'Message sent' if __name__ == '__main__': app.run(debug=True)
from flask import Flask, jsonify, request from dotenv import load_dotenv, find_dotenv from twilio import twiml from Queue import Queue from threading import Thread from time import sleep load_dotenv(find_dotenv()) directions = ['forward', 'backward'] task_q = Queue() def send_rasp(task_q): while True: sleep(2) if not task_q.empty(): message = task_q.get() print(message) rasp_signal = Thread(target=send_rasp, args=(task_q, )) rasp_signal.setDaemon(True) rasp_signal.start() app = Flask(__name__) @app.route('/message', methods=['POST']) def roomba_command(): # twilio text message body = request.form['Body'] resp = handle_twilio_message(body) twilio_resp = twiml.Response() twilio_resp.message(resp) return str(twilio_resp) def handle_twilio_message(message): if message.lower() in directions: task_q.put(message.lower()) return 'Message sent' try: degree = float(message) except ValueError as e: return 'Invalid command' task_q.put(str(degree)) return 'Message sent' if __name__ == '__main__': app.run(debug=True)Use continue in task loopfrom flask import Flask, jsonify, request from dotenv import load_dotenv, find_dotenv from twilio import twiml from Queue import Queue from threading import Thread from time import sleep load_dotenv(find_dotenv()) directions = ['forward', 'backward'] task_q = Queue() def send_rasp(task_q): while True: sleep(2) if task_q.empty(): continue message = task_q.get() print(message) rasp_signal = Thread(target=send_rasp, args=(task_q, )) rasp_signal.setDaemon(True) rasp_signal.start() app = Flask(__name__) @app.route('/message', methods=['POST']) def roomba_command(): # twilio text message body = request.form['Body'] resp = handle_twilio_message(body) twilio_resp = twiml.Response() twilio_resp.message(resp) return str(twilio_resp) def handle_twilio_message(message): if message.lower() in directions: task_q.put(message.lower()) return 'Message sent' try: degree = float(message) except ValueError as e: return 'Invalid command' task_q.put(str(degree)) return 'Message sent' if __name__ == '__main__': app.run(debug=True)
<commit_before>from flask import Flask, jsonify, request from dotenv import load_dotenv, find_dotenv from twilio import twiml from Queue import Queue from threading import Thread from time import sleep load_dotenv(find_dotenv()) directions = ['forward', 'backward'] task_q = Queue() def send_rasp(task_q): while True: sleep(2) if not task_q.empty(): message = task_q.get() print(message) rasp_signal = Thread(target=send_rasp, args=(task_q, )) rasp_signal.setDaemon(True) rasp_signal.start() app = Flask(__name__) @app.route('/message', methods=['POST']) def roomba_command(): # twilio text message body = request.form['Body'] resp = handle_twilio_message(body) twilio_resp = twiml.Response() twilio_resp.message(resp) return str(twilio_resp) def handle_twilio_message(message): if message.lower() in directions: task_q.put(message.lower()) return 'Message sent' try: degree = float(message) except ValueError as e: return 'Invalid command' task_q.put(str(degree)) return 'Message sent' if __name__ == '__main__': app.run(debug=True)<commit_msg>Use continue in task loop<commit_after>from flask import Flask, jsonify, request from dotenv import load_dotenv, find_dotenv from twilio import twiml from Queue import Queue from threading import Thread from time import sleep load_dotenv(find_dotenv()) directions = ['forward', 'backward'] task_q = Queue() def send_rasp(task_q): while True: sleep(2) if task_q.empty(): continue message = task_q.get() print(message) rasp_signal = Thread(target=send_rasp, args=(task_q, )) rasp_signal.setDaemon(True) rasp_signal.start() app = Flask(__name__) @app.route('/message', methods=['POST']) def roomba_command(): # twilio text message body = request.form['Body'] resp = handle_twilio_message(body) twilio_resp = twiml.Response() twilio_resp.message(resp) return str(twilio_resp) def handle_twilio_message(message): if message.lower() in directions: task_q.put(message.lower()) return 'Message sent' try: degree = float(message) except ValueError as e: return 'Invalid command' task_q.put(str(degree)) return 'Message sent' if __name__ == '__main__': app.run(debug=True)
0389fabdb0343b189b153cc909b05e88d3830b94
downstream_node/lib/node.py
downstream_node/lib/node.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from downstream_node.config import config from downstream_node.models import Challenges from heartbeat import Heartbeat from downstream_node.startup import db __all__ = ['create_token', 'delete_token', 'add_file', 'remove_file', 'gen_challenges', 'update_challenges'] def create_token(*args, **kwargs): raise NotImplementedError def delete_token(*args, **kwargs): raise NotImplementedError def add_file(*args, **kwargs): raise NotImplementedError def remove_file(*args, **kwargs): raise NotImplementedError def gen_challenges(filepath, root_seed): secret = getattr(config, 'HEARTBEAT_SECRET') hb = Heartbeat(filepath, secret=secret) hb.generate_challenges(1000, root_seed) for challenge in hb.challenges: chal = Challenges( filepath=filepath, root_seed=root_seed, block=challenge.block, seed=challenge.seed, response=challenge.response, ) db.session.add(chal) db.session.commit() def update_challenges(*args, **kwargs): raise NotImplementedError
#!/usr/bin/env python # -*- coding: utf-8 -*- from downstream_node.config import config from downstream_node.models import Challenges, Files from heartbeat import Heartbeat from downstream_node.startup import db __all__ = ['create_token', 'delete_token', 'add_file', 'remove_file', 'gen_challenges', 'update_challenges'] def create_token(*args, **kwargs): raise NotImplementedError def delete_token(*args, **kwargs): raise NotImplementedError def add_file(*args, **kwargs): raise NotImplementedError def remove_file(*args, **kwargs): raise NotImplementedError def gen_challenges(filepath, root_seed): secret = getattr(config, 'HEARTBEAT_SECRET') hb = Heartbeat(filepath, secret=secret) hb.generate_challenges(1000, root_seed) files = Files(name=filepath) db.session.add(files) for challenge in hb.challenges: chal = Challenges( filename=filepath, rootseed=root_seed, block=challenge.block, seed=challenge.seed, response=challenge.response, ) db.session.add(chal) db.session.commit() def update_challenges(*args, **kwargs): raise NotImplementedError
Fix for new column names
Fix for new column names
Python
mit
Storj/downstream-node,Storj/downstream-node
#!/usr/bin/env python # -*- coding: utf-8 -*- from downstream_node.config import config from downstream_node.models import Challenges from heartbeat import Heartbeat from downstream_node.startup import db __all__ = ['create_token', 'delete_token', 'add_file', 'remove_file', 'gen_challenges', 'update_challenges'] def create_token(*args, **kwargs): raise NotImplementedError def delete_token(*args, **kwargs): raise NotImplementedError def add_file(*args, **kwargs): raise NotImplementedError def remove_file(*args, **kwargs): raise NotImplementedError def gen_challenges(filepath, root_seed): secret = getattr(config, 'HEARTBEAT_SECRET') hb = Heartbeat(filepath, secret=secret) hb.generate_challenges(1000, root_seed) for challenge in hb.challenges: chal = Challenges( filepath=filepath, root_seed=root_seed, block=challenge.block, seed=challenge.seed, response=challenge.response, ) db.session.add(chal) db.session.commit() def update_challenges(*args, **kwargs): raise NotImplementedError Fix for new column names
#!/usr/bin/env python # -*- coding: utf-8 -*- from downstream_node.config import config from downstream_node.models import Challenges, Files from heartbeat import Heartbeat from downstream_node.startup import db __all__ = ['create_token', 'delete_token', 'add_file', 'remove_file', 'gen_challenges', 'update_challenges'] def create_token(*args, **kwargs): raise NotImplementedError def delete_token(*args, **kwargs): raise NotImplementedError def add_file(*args, **kwargs): raise NotImplementedError def remove_file(*args, **kwargs): raise NotImplementedError def gen_challenges(filepath, root_seed): secret = getattr(config, 'HEARTBEAT_SECRET') hb = Heartbeat(filepath, secret=secret) hb.generate_challenges(1000, root_seed) files = Files(name=filepath) db.session.add(files) for challenge in hb.challenges: chal = Challenges( filename=filepath, rootseed=root_seed, block=challenge.block, seed=challenge.seed, response=challenge.response, ) db.session.add(chal) db.session.commit() def update_challenges(*args, **kwargs): raise NotImplementedError
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from downstream_node.config import config from downstream_node.models import Challenges from heartbeat import Heartbeat from downstream_node.startup import db __all__ = ['create_token', 'delete_token', 'add_file', 'remove_file', 'gen_challenges', 'update_challenges'] def create_token(*args, **kwargs): raise NotImplementedError def delete_token(*args, **kwargs): raise NotImplementedError def add_file(*args, **kwargs): raise NotImplementedError def remove_file(*args, **kwargs): raise NotImplementedError def gen_challenges(filepath, root_seed): secret = getattr(config, 'HEARTBEAT_SECRET') hb = Heartbeat(filepath, secret=secret) hb.generate_challenges(1000, root_seed) for challenge in hb.challenges: chal = Challenges( filepath=filepath, root_seed=root_seed, block=challenge.block, seed=challenge.seed, response=challenge.response, ) db.session.add(chal) db.session.commit() def update_challenges(*args, **kwargs): raise NotImplementedError <commit_msg>Fix for new column names<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from downstream_node.config import config from downstream_node.models import Challenges, Files from heartbeat import Heartbeat from downstream_node.startup import db __all__ = ['create_token', 'delete_token', 'add_file', 'remove_file', 'gen_challenges', 'update_challenges'] def create_token(*args, **kwargs): raise NotImplementedError def delete_token(*args, **kwargs): raise NotImplementedError def add_file(*args, **kwargs): raise NotImplementedError def remove_file(*args, **kwargs): raise NotImplementedError def gen_challenges(filepath, root_seed): secret = getattr(config, 'HEARTBEAT_SECRET') hb = Heartbeat(filepath, secret=secret) hb.generate_challenges(1000, root_seed) files = Files(name=filepath) db.session.add(files) for challenge in hb.challenges: chal = Challenges( filename=filepath, rootseed=root_seed, block=challenge.block, seed=challenge.seed, response=challenge.response, ) db.session.add(chal) db.session.commit() def update_challenges(*args, **kwargs): raise NotImplementedError
#!/usr/bin/env python # -*- coding: utf-8 -*- from downstream_node.config import config from downstream_node.models import Challenges from heartbeat import Heartbeat from downstream_node.startup import db __all__ = ['create_token', 'delete_token', 'add_file', 'remove_file', 'gen_challenges', 'update_challenges'] def create_token(*args, **kwargs): raise NotImplementedError def delete_token(*args, **kwargs): raise NotImplementedError def add_file(*args, **kwargs): raise NotImplementedError def remove_file(*args, **kwargs): raise NotImplementedError def gen_challenges(filepath, root_seed): secret = getattr(config, 'HEARTBEAT_SECRET') hb = Heartbeat(filepath, secret=secret) hb.generate_challenges(1000, root_seed) for challenge in hb.challenges: chal = Challenges( filepath=filepath, root_seed=root_seed, block=challenge.block, seed=challenge.seed, response=challenge.response, ) db.session.add(chal) db.session.commit() def update_challenges(*args, **kwargs): raise NotImplementedError Fix for new column names#!/usr/bin/env python # -*- coding: utf-8 -*- from downstream_node.config import config from downstream_node.models import Challenges, Files from heartbeat import Heartbeat from downstream_node.startup import db __all__ = ['create_token', 'delete_token', 'add_file', 'remove_file', 'gen_challenges', 'update_challenges'] def create_token(*args, **kwargs): raise NotImplementedError def delete_token(*args, **kwargs): raise NotImplementedError def add_file(*args, **kwargs): raise NotImplementedError def remove_file(*args, **kwargs): raise NotImplementedError def gen_challenges(filepath, root_seed): secret = getattr(config, 'HEARTBEAT_SECRET') hb = Heartbeat(filepath, secret=secret) hb.generate_challenges(1000, root_seed) files = Files(name=filepath) db.session.add(files) for challenge in hb.challenges: chal = Challenges( filename=filepath, rootseed=root_seed, block=challenge.block, seed=challenge.seed, response=challenge.response, ) db.session.add(chal) db.session.commit() def update_challenges(*args, **kwargs): raise NotImplementedError
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from downstream_node.config import config from downstream_node.models import Challenges from heartbeat import Heartbeat from downstream_node.startup import db __all__ = ['create_token', 'delete_token', 'add_file', 'remove_file', 'gen_challenges', 'update_challenges'] def create_token(*args, **kwargs): raise NotImplementedError def delete_token(*args, **kwargs): raise NotImplementedError def add_file(*args, **kwargs): raise NotImplementedError def remove_file(*args, **kwargs): raise NotImplementedError def gen_challenges(filepath, root_seed): secret = getattr(config, 'HEARTBEAT_SECRET') hb = Heartbeat(filepath, secret=secret) hb.generate_challenges(1000, root_seed) for challenge in hb.challenges: chal = Challenges( filepath=filepath, root_seed=root_seed, block=challenge.block, seed=challenge.seed, response=challenge.response, ) db.session.add(chal) db.session.commit() def update_challenges(*args, **kwargs): raise NotImplementedError <commit_msg>Fix for new column names<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from downstream_node.config import config from downstream_node.models import Challenges, Files from heartbeat import Heartbeat from downstream_node.startup import db __all__ = ['create_token', 'delete_token', 'add_file', 'remove_file', 'gen_challenges', 'update_challenges'] def create_token(*args, **kwargs): raise NotImplementedError def delete_token(*args, **kwargs): raise NotImplementedError def add_file(*args, **kwargs): raise NotImplementedError def remove_file(*args, **kwargs): raise NotImplementedError def gen_challenges(filepath, root_seed): secret = getattr(config, 'HEARTBEAT_SECRET') hb = Heartbeat(filepath, secret=secret) hb.generate_challenges(1000, root_seed) files = Files(name=filepath) db.session.add(files) for challenge in hb.challenges: chal = Challenges( filename=filepath, rootseed=root_seed, block=challenge.block, seed=challenge.seed, response=challenge.response, ) db.session.add(chal) db.session.commit() def update_challenges(*args, **kwargs): raise NotImplementedError
efc1988d704a7a1231046dea8af65dcdba7897fd
py/fbx_write.py
py/fbx_write.py
# !/usr/bin/env python # Blender has moved to Python 3! import sys import os import bpy for sysarg in sys.argv: print(sysarg) py_args = sys.argv[sys.argv.index('--') + 1] py_args = py_args.split(' ') for arg in py_args: if (arg.startswith('basedir:')): basedir = arg.split('basedir:')[1] else: # can supply filename(s) with or without extension pdb_code = os.path.splitext(arg)[0] abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl') # This is the base directory, used for saving files molecule = bpy.ops.import_scene.x3d( filepath = abs_file_in ) abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx') bpy.ops.export_scene.fbx( filepath = abs_file_out )
# !/usr/bin/env python # Blender has moved to Python 3! import sys import os import bpy for sysarg in sys.argv: print(sysarg) py_args = sys.argv[sys.argv.index('--') + 1] py_args = py_args.split(' ') for arg in py_args: if (arg.startswith('basedir:')): basedir = arg.split('basedir:')[1] else: # can supply filename(s) with or without extension pdb_code = os.path.splitext(arg)[0] abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl') # This is the base directory, used for saving files molecule = bpy.ops.import_scene.x3d( filepath = abs_file_in ) abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx') bpy.ops.export_scene.fbx( filepath = abs_file_out ) bpy.ops.wm.quit_blender()
Quit Blender after writing FBX
Quit Blender after writing FBX
Python
mit
hackmcr15-code-a-la-mode/mol-vis-hack,hackmcr15-code-a-la-mode/mol-vis-hack
# !/usr/bin/env python # Blender has moved to Python 3! import sys import os import bpy for sysarg in sys.argv: print(sysarg) py_args = sys.argv[sys.argv.index('--') + 1] py_args = py_args.split(' ') for arg in py_args: if (arg.startswith('basedir:')): basedir = arg.split('basedir:')[1] else: # can supply filename(s) with or without extension pdb_code = os.path.splitext(arg)[0] abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl') # This is the base directory, used for saving files molecule = bpy.ops.import_scene.x3d( filepath = abs_file_in ) abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx') bpy.ops.export_scene.fbx( filepath = abs_file_out ) Quit Blender after writing FBX
# !/usr/bin/env python # Blender has moved to Python 3! import sys import os import bpy for sysarg in sys.argv: print(sysarg) py_args = sys.argv[sys.argv.index('--') + 1] py_args = py_args.split(' ') for arg in py_args: if (arg.startswith('basedir:')): basedir = arg.split('basedir:')[1] else: # can supply filename(s) with or without extension pdb_code = os.path.splitext(arg)[0] abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl') # This is the base directory, used for saving files molecule = bpy.ops.import_scene.x3d( filepath = abs_file_in ) abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx') bpy.ops.export_scene.fbx( filepath = abs_file_out ) bpy.ops.wm.quit_blender()
<commit_before># !/usr/bin/env python # Blender has moved to Python 3! import sys import os import bpy for sysarg in sys.argv: print(sysarg) py_args = sys.argv[sys.argv.index('--') + 1] py_args = py_args.split(' ') for arg in py_args: if (arg.startswith('basedir:')): basedir = arg.split('basedir:')[1] else: # can supply filename(s) with or without extension pdb_code = os.path.splitext(arg)[0] abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl') # This is the base directory, used for saving files molecule = bpy.ops.import_scene.x3d( filepath = abs_file_in ) abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx') bpy.ops.export_scene.fbx( filepath = abs_file_out ) <commit_msg>Quit Blender after writing FBX<commit_after>
# !/usr/bin/env python # Blender has moved to Python 3! import sys import os import bpy for sysarg in sys.argv: print(sysarg) py_args = sys.argv[sys.argv.index('--') + 1] py_args = py_args.split(' ') for arg in py_args: if (arg.startswith('basedir:')): basedir = arg.split('basedir:')[1] else: # can supply filename(s) with or without extension pdb_code = os.path.splitext(arg)[0] abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl') # This is the base directory, used for saving files molecule = bpy.ops.import_scene.x3d( filepath = abs_file_in ) abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx') bpy.ops.export_scene.fbx( filepath = abs_file_out ) bpy.ops.wm.quit_blender()
# !/usr/bin/env python # Blender has moved to Python 3! import sys import os import bpy for sysarg in sys.argv: print(sysarg) py_args = sys.argv[sys.argv.index('--') + 1] py_args = py_args.split(' ') for arg in py_args: if (arg.startswith('basedir:')): basedir = arg.split('basedir:')[1] else: # can supply filename(s) with or without extension pdb_code = os.path.splitext(arg)[0] abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl') # This is the base directory, used for saving files molecule = bpy.ops.import_scene.x3d( filepath = abs_file_in ) abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx') bpy.ops.export_scene.fbx( filepath = abs_file_out ) Quit Blender after writing FBX# !/usr/bin/env python # Blender has moved to Python 3! import sys import os import bpy for sysarg in sys.argv: print(sysarg) py_args = sys.argv[sys.argv.index('--') + 1] py_args = py_args.split(' ') for arg in py_args: if (arg.startswith('basedir:')): basedir = arg.split('basedir:')[1] else: # can supply filename(s) with or without extension pdb_code = os.path.splitext(arg)[0] abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl') # This is the base directory, used for saving files molecule = bpy.ops.import_scene.x3d( filepath = abs_file_in ) abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx') bpy.ops.export_scene.fbx( filepath = abs_file_out ) bpy.ops.wm.quit_blender()
<commit_before># !/usr/bin/env python # Blender has moved to Python 3! import sys import os import bpy for sysarg in sys.argv: print(sysarg) py_args = sys.argv[sys.argv.index('--') + 1] py_args = py_args.split(' ') for arg in py_args: if (arg.startswith('basedir:')): basedir = arg.split('basedir:')[1] else: # can supply filename(s) with or without extension pdb_code = os.path.splitext(arg)[0] abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl') # This is the base directory, used for saving files molecule = bpy.ops.import_scene.x3d( filepath = abs_file_in ) abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx') bpy.ops.export_scene.fbx( filepath = abs_file_out ) <commit_msg>Quit Blender after writing FBX<commit_after># !/usr/bin/env python # Blender has moved to Python 3! import sys import os import bpy for sysarg in sys.argv: print(sysarg) py_args = sys.argv[sys.argv.index('--') + 1] py_args = py_args.split(' ') for arg in py_args: if (arg.startswith('basedir:')): basedir = arg.split('basedir:')[1] else: # can supply filename(s) with or without extension pdb_code = os.path.splitext(arg)[0] abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl') # This is the base directory, used for saving files molecule = bpy.ops.import_scene.x3d( filepath = abs_file_in ) abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx') bpy.ops.export_scene.fbx( filepath = abs_file_out ) bpy.ops.wm.quit_blender()
1f3fce7cb415e739bdb745295807cceaf853d176
easy_thumbnails/__init__.py
easy_thumbnails/__init__.py
VERSION = (1, 0, 'alpha', 11) def get_version(join=' ', short=False): """ Return the version of this package as a string. The version number is built from a ``VERSION`` tuple, which should consist of integers, or trailing version information (such as 'alpha', 'beta' or 'final'). For example: >>> VERSION = (2, 0, 6) >>> get_version() '2.0.6' >>> VERSION = (1, 0, 'beta', 2) >>> get_version() '1.0 beta 2' Use the ``join`` argument to join the version elements by an alternate character to the default ``' '``. This is useful when building a distutils setup module:: from this_package import get_version setup( version=get_version(join='-'), # ... ) Use the ``short`` argument to get the version number without trailing version information. """ version = [] number = [] remainder = [] for i, bit in enumerate(VERSION): if isinstance(bit, int): number.append(str(bit)) else: remainder = [str(bit) for bit in VERSION[i:]] break if number: version.append('.'.join(number)) if not short: if remainder == ['alpha', 0]: version.append('pre-alpha') elif 'final' not in remainder: version.extend(remainder) return join.join(version)
VERSION = (1, 0, 'alpha', 12) def get_version(join=' ', short=False): """ Return the version of this package as a string. The version number is built from a ``VERSION`` tuple, which should consist of integers, or trailing version information (such as 'alpha', 'beta' or 'final'). For example: >>> VERSION = (2, 0, 6) >>> get_version() '2.0.6' >>> VERSION = (1, 0, 'beta', 2) >>> get_version() '1.0 beta 2' Use the ``join`` argument to join the version elements by an alternate character to the default ``' '``. This is useful when building a distutils setup module:: from this_package import get_version setup( version=get_version(join='-'), # ... ) Use the ``short`` argument to get the version number without trailing version information. """ version = [] number = [] remainder = [] for i, bit in enumerate(VERSION): if isinstance(bit, int): number.append(str(bit)) else: remainder = [str(bit) for bit in VERSION[i:]] break if number: version.append('.'.join(number)) if not short: if remainder == ['alpha', 0]: version.append('pre-alpha') elif 'final' not in remainder: version.extend(remainder) return join.join(version)
Bump version number for another release with ducktyping in it
Bump version number for another release with ducktyping in it
Python
bsd-3-clause
sandow-digital/easy-thumbnails-cropman,jrief/easy-thumbnails,Mactory/easy-thumbnails,emschorsch/easy-thumbnails,sandow-digital/easy-thumbnails-cropman,siovene/easy-thumbnails,jrief/easy-thumbnails,emschorsch/easy-thumbnails,SmileyChris/easy-thumbnails,jaddison/easy-thumbnails
VERSION = (1, 0, 'alpha', 11) def get_version(join=' ', short=False): """ Return the version of this package as a string. The version number is built from a ``VERSION`` tuple, which should consist of integers, or trailing version information (such as 'alpha', 'beta' or 'final'). For example: >>> VERSION = (2, 0, 6) >>> get_version() '2.0.6' >>> VERSION = (1, 0, 'beta', 2) >>> get_version() '1.0 beta 2' Use the ``join`` argument to join the version elements by an alternate character to the default ``' '``. This is useful when building a distutils setup module:: from this_package import get_version setup( version=get_version(join='-'), # ... ) Use the ``short`` argument to get the version number without trailing version information. """ version = [] number = [] remainder = [] for i, bit in enumerate(VERSION): if isinstance(bit, int): number.append(str(bit)) else: remainder = [str(bit) for bit in VERSION[i:]] break if number: version.append('.'.join(number)) if not short: if remainder == ['alpha', 0]: version.append('pre-alpha') elif 'final' not in remainder: version.extend(remainder) return join.join(version) Bump version number for another release with ducktyping in it
VERSION = (1, 0, 'alpha', 12) def get_version(join=' ', short=False): """ Return the version of this package as a string. The version number is built from a ``VERSION`` tuple, which should consist of integers, or trailing version information (such as 'alpha', 'beta' or 'final'). For example: >>> VERSION = (2, 0, 6) >>> get_version() '2.0.6' >>> VERSION = (1, 0, 'beta', 2) >>> get_version() '1.0 beta 2' Use the ``join`` argument to join the version elements by an alternate character to the default ``' '``. This is useful when building a distutils setup module:: from this_package import get_version setup( version=get_version(join='-'), # ... ) Use the ``short`` argument to get the version number without trailing version information. """ version = [] number = [] remainder = [] for i, bit in enumerate(VERSION): if isinstance(bit, int): number.append(str(bit)) else: remainder = [str(bit) for bit in VERSION[i:]] break if number: version.append('.'.join(number)) if not short: if remainder == ['alpha', 0]: version.append('pre-alpha') elif 'final' not in remainder: version.extend(remainder) return join.join(version)
<commit_before>VERSION = (1, 0, 'alpha', 11) def get_version(join=' ', short=False): """ Return the version of this package as a string. The version number is built from a ``VERSION`` tuple, which should consist of integers, or trailing version information (such as 'alpha', 'beta' or 'final'). For example: >>> VERSION = (2, 0, 6) >>> get_version() '2.0.6' >>> VERSION = (1, 0, 'beta', 2) >>> get_version() '1.0 beta 2' Use the ``join`` argument to join the version elements by an alternate character to the default ``' '``. This is useful when building a distutils setup module:: from this_package import get_version setup( version=get_version(join='-'), # ... ) Use the ``short`` argument to get the version number without trailing version information. """ version = [] number = [] remainder = [] for i, bit in enumerate(VERSION): if isinstance(bit, int): number.append(str(bit)) else: remainder = [str(bit) for bit in VERSION[i:]] break if number: version.append('.'.join(number)) if not short: if remainder == ['alpha', 0]: version.append('pre-alpha') elif 'final' not in remainder: version.extend(remainder) return join.join(version) <commit_msg>Bump version number for another release with ducktyping in it<commit_after>
VERSION = (1, 0, 'alpha', 12) def get_version(join=' ', short=False): """ Return the version of this package as a string. The version number is built from a ``VERSION`` tuple, which should consist of integers, or trailing version information (such as 'alpha', 'beta' or 'final'). For example: >>> VERSION = (2, 0, 6) >>> get_version() '2.0.6' >>> VERSION = (1, 0, 'beta', 2) >>> get_version() '1.0 beta 2' Use the ``join`` argument to join the version elements by an alternate character to the default ``' '``. This is useful when building a distutils setup module:: from this_package import get_version setup( version=get_version(join='-'), # ... ) Use the ``short`` argument to get the version number without trailing version information. """ version = [] number = [] remainder = [] for i, bit in enumerate(VERSION): if isinstance(bit, int): number.append(str(bit)) else: remainder = [str(bit) for bit in VERSION[i:]] break if number: version.append('.'.join(number)) if not short: if remainder == ['alpha', 0]: version.append('pre-alpha') elif 'final' not in remainder: version.extend(remainder) return join.join(version)
VERSION = (1, 0, 'alpha', 11) def get_version(join=' ', short=False): """ Return the version of this package as a string. The version number is built from a ``VERSION`` tuple, which should consist of integers, or trailing version information (such as 'alpha', 'beta' or 'final'). For example: >>> VERSION = (2, 0, 6) >>> get_version() '2.0.6' >>> VERSION = (1, 0, 'beta', 2) >>> get_version() '1.0 beta 2' Use the ``join`` argument to join the version elements by an alternate character to the default ``' '``. This is useful when building a distutils setup module:: from this_package import get_version setup( version=get_version(join='-'), # ... ) Use the ``short`` argument to get the version number without trailing version information. """ version = [] number = [] remainder = [] for i, bit in enumerate(VERSION): if isinstance(bit, int): number.append(str(bit)) else: remainder = [str(bit) for bit in VERSION[i:]] break if number: version.append('.'.join(number)) if not short: if remainder == ['alpha', 0]: version.append('pre-alpha') elif 'final' not in remainder: version.extend(remainder) return join.join(version) Bump version number for another release with ducktyping in itVERSION = (1, 0, 'alpha', 12) def get_version(join=' ', short=False): """ Return the version of this package as a string. The version number is built from a ``VERSION`` tuple, which should consist of integers, or trailing version information (such as 'alpha', 'beta' or 'final'). For example: >>> VERSION = (2, 0, 6) >>> get_version() '2.0.6' >>> VERSION = (1, 0, 'beta', 2) >>> get_version() '1.0 beta 2' Use the ``join`` argument to join the version elements by an alternate character to the default ``' '``. This is useful when building a distutils setup module:: from this_package import get_version setup( version=get_version(join='-'), # ... ) Use the ``short`` argument to get the version number without trailing version information. """ version = [] number = [] remainder = [] for i, bit in enumerate(VERSION): if isinstance(bit, int): number.append(str(bit)) else: remainder = [str(bit) for bit in VERSION[i:]] break if number: version.append('.'.join(number)) if not short: if remainder == ['alpha', 0]: version.append('pre-alpha') elif 'final' not in remainder: version.extend(remainder) return join.join(version)
<commit_before>VERSION = (1, 0, 'alpha', 11) def get_version(join=' ', short=False): """ Return the version of this package as a string. The version number is built from a ``VERSION`` tuple, which should consist of integers, or trailing version information (such as 'alpha', 'beta' or 'final'). For example: >>> VERSION = (2, 0, 6) >>> get_version() '2.0.6' >>> VERSION = (1, 0, 'beta', 2) >>> get_version() '1.0 beta 2' Use the ``join`` argument to join the version elements by an alternate character to the default ``' '``. This is useful when building a distutils setup module:: from this_package import get_version setup( version=get_version(join='-'), # ... ) Use the ``short`` argument to get the version number without trailing version information. """ version = [] number = [] remainder = [] for i, bit in enumerate(VERSION): if isinstance(bit, int): number.append(str(bit)) else: remainder = [str(bit) for bit in VERSION[i:]] break if number: version.append('.'.join(number)) if not short: if remainder == ['alpha', 0]: version.append('pre-alpha') elif 'final' not in remainder: version.extend(remainder) return join.join(version) <commit_msg>Bump version number for another release with ducktyping in it<commit_after>VERSION = (1, 0, 'alpha', 12) def get_version(join=' ', short=False): """ Return the version of this package as a string. The version number is built from a ``VERSION`` tuple, which should consist of integers, or trailing version information (such as 'alpha', 'beta' or 'final'). For example: >>> VERSION = (2, 0, 6) >>> get_version() '2.0.6' >>> VERSION = (1, 0, 'beta', 2) >>> get_version() '1.0 beta 2' Use the ``join`` argument to join the version elements by an alternate character to the default ``' '``. This is useful when building a distutils setup module:: from this_package import get_version setup( version=get_version(join='-'), # ... ) Use the ``short`` argument to get the version number without trailing version information. """ version = [] number = [] remainder = [] for i, bit in enumerate(VERSION): if isinstance(bit, int): number.append(str(bit)) else: remainder = [str(bit) for bit in VERSION[i:]] break if number: version.append('.'.join(number)) if not short: if remainder == ['alpha', 0]: version.append('pre-alpha') elif 'final' not in remainder: version.extend(remainder) return join.join(version)
aca031267748358c49eac96fe158ba0a2ec3a2e8
tota/drawers/json_replay.py
tota/drawers/json_replay.py
import json from os import path from tota.game import Drawer class JsonReplayDrawer(Drawer): def __init__(self, replay_dir): self.replay_dir = replay_dir def draw(self, game): """Draw the world with 'ascii'-art .""" things_data = [] tick_data = { 't': game.world.t, 'things': things_data, 'effects': [{ 'position': position, 'effect': effect, } for position, effect in game.world.effects.items()] } for thing in game.world.things.values(): thing_data = { 'id': id(thing), 'type': thing.__class__.__name__, 'position': thing.position, } if thing_data['type'] != 'Tree': thing_data.update({ 'life': thing.life, 'name': thing.name, 'team': thing.team, 'level': getattr(thing, 'level', None), 'xp': getattr(thing, 'xp', None), 'action': thing.last_action, 'target': thing.last_target, }) things_data.append(thing_data) tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t) with open(tick_path, 'w') as tick_file: json.dump(tick_data, tick_file, indent=2 if game.debug else None)
import json from os import path from tota.game import Drawer class JsonReplayDrawer(Drawer): def __init__(self, replay_dir): self.replay_dir = replay_dir def draw(self, game): """Draw the world with 'ascii'-art .""" things_data = [] tick_data = { 't': game.world.t, 'things': things_data, 'effects': [{ 'position': position, 'effect': effect, } for position, effect in game.world.effects.items()] } for thing in game.world.things.values(): thing_data = { 'id': id(thing), 'type': thing.__class__.__name__, 'position': thing.position, } if thing_data['type'] != 'Tree': thing_data.update({ 'life': thing.life, 'name': thing.name, 'team': thing.team, 'level': getattr(thing, 'level', None), 'xp': getattr(thing, 'xp', None), 'action': thing.last_action, 'target': thing.last_target, 'action_done': thing.last_action_done, }) things_data.append(thing_data) tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t) with open(tick_path, 'w') as tick_file: json.dump(tick_data, tick_file, indent=2 if game.debug else None)
Add done result to replay
Add done result to replay
Python
mit
dmoisset/tota,matuu/tota,fisadev/tota
import json from os import path from tota.game import Drawer class JsonReplayDrawer(Drawer): def __init__(self, replay_dir): self.replay_dir = replay_dir def draw(self, game): """Draw the world with 'ascii'-art .""" things_data = [] tick_data = { 't': game.world.t, 'things': things_data, 'effects': [{ 'position': position, 'effect': effect, } for position, effect in game.world.effects.items()] } for thing in game.world.things.values(): thing_data = { 'id': id(thing), 'type': thing.__class__.__name__, 'position': thing.position, } if thing_data['type'] != 'Tree': thing_data.update({ 'life': thing.life, 'name': thing.name, 'team': thing.team, 'level': getattr(thing, 'level', None), 'xp': getattr(thing, 'xp', None), 'action': thing.last_action, 'target': thing.last_target, }) things_data.append(thing_data) tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t) with open(tick_path, 'w') as tick_file: json.dump(tick_data, tick_file, indent=2 if game.debug else None) Add done result to replay
import json from os import path from tota.game import Drawer class JsonReplayDrawer(Drawer): def __init__(self, replay_dir): self.replay_dir = replay_dir def draw(self, game): """Draw the world with 'ascii'-art .""" things_data = [] tick_data = { 't': game.world.t, 'things': things_data, 'effects': [{ 'position': position, 'effect': effect, } for position, effect in game.world.effects.items()] } for thing in game.world.things.values(): thing_data = { 'id': id(thing), 'type': thing.__class__.__name__, 'position': thing.position, } if thing_data['type'] != 'Tree': thing_data.update({ 'life': thing.life, 'name': thing.name, 'team': thing.team, 'level': getattr(thing, 'level', None), 'xp': getattr(thing, 'xp', None), 'action': thing.last_action, 'target': thing.last_target, 'action_done': thing.last_action_done, }) things_data.append(thing_data) tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t) with open(tick_path, 'w') as tick_file: json.dump(tick_data, tick_file, indent=2 if game.debug else None)
<commit_before>import json from os import path from tota.game import Drawer class JsonReplayDrawer(Drawer): def __init__(self, replay_dir): self.replay_dir = replay_dir def draw(self, game): """Draw the world with 'ascii'-art .""" things_data = [] tick_data = { 't': game.world.t, 'things': things_data, 'effects': [{ 'position': position, 'effect': effect, } for position, effect in game.world.effects.items()] } for thing in game.world.things.values(): thing_data = { 'id': id(thing), 'type': thing.__class__.__name__, 'position': thing.position, } if thing_data['type'] != 'Tree': thing_data.update({ 'life': thing.life, 'name': thing.name, 'team': thing.team, 'level': getattr(thing, 'level', None), 'xp': getattr(thing, 'xp', None), 'action': thing.last_action, 'target': thing.last_target, }) things_data.append(thing_data) tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t) with open(tick_path, 'w') as tick_file: json.dump(tick_data, tick_file, indent=2 if game.debug else None) <commit_msg>Add done result to replay<commit_after>
import json from os import path from tota.game import Drawer class JsonReplayDrawer(Drawer): def __init__(self, replay_dir): self.replay_dir = replay_dir def draw(self, game): """Draw the world with 'ascii'-art .""" things_data = [] tick_data = { 't': game.world.t, 'things': things_data, 'effects': [{ 'position': position, 'effect': effect, } for position, effect in game.world.effects.items()] } for thing in game.world.things.values(): thing_data = { 'id': id(thing), 'type': thing.__class__.__name__, 'position': thing.position, } if thing_data['type'] != 'Tree': thing_data.update({ 'life': thing.life, 'name': thing.name, 'team': thing.team, 'level': getattr(thing, 'level', None), 'xp': getattr(thing, 'xp', None), 'action': thing.last_action, 'target': thing.last_target, 'action_done': thing.last_action_done, }) things_data.append(thing_data) tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t) with open(tick_path, 'w') as tick_file: json.dump(tick_data, tick_file, indent=2 if game.debug else None)
import json from os import path from tota.game import Drawer class JsonReplayDrawer(Drawer): def __init__(self, replay_dir): self.replay_dir = replay_dir def draw(self, game): """Draw the world with 'ascii'-art .""" things_data = [] tick_data = { 't': game.world.t, 'things': things_data, 'effects': [{ 'position': position, 'effect': effect, } for position, effect in game.world.effects.items()] } for thing in game.world.things.values(): thing_data = { 'id': id(thing), 'type': thing.__class__.__name__, 'position': thing.position, } if thing_data['type'] != 'Tree': thing_data.update({ 'life': thing.life, 'name': thing.name, 'team': thing.team, 'level': getattr(thing, 'level', None), 'xp': getattr(thing, 'xp', None), 'action': thing.last_action, 'target': thing.last_target, }) things_data.append(thing_data) tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t) with open(tick_path, 'w') as tick_file: json.dump(tick_data, tick_file, indent=2 if game.debug else None) Add done result to replayimport json from os import path from tota.game import Drawer class JsonReplayDrawer(Drawer): def __init__(self, replay_dir): self.replay_dir = replay_dir def draw(self, game): """Draw the world with 'ascii'-art .""" things_data = [] tick_data = { 't': game.world.t, 'things': things_data, 'effects': [{ 'position': position, 'effect': effect, } for position, effect in game.world.effects.items()] } for thing in game.world.things.values(): thing_data = { 'id': id(thing), 'type': thing.__class__.__name__, 'position': thing.position, } if thing_data['type'] != 'Tree': thing_data.update({ 'life': thing.life, 'name': thing.name, 'team': thing.team, 'level': getattr(thing, 'level', None), 'xp': getattr(thing, 'xp', None), 'action': thing.last_action, 'target': thing.last_target, 'action_done': thing.last_action_done, }) things_data.append(thing_data) tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t) with open(tick_path, 'w') as tick_file: json.dump(tick_data, tick_file, indent=2 if game.debug else None)
<commit_before>import json from os import path from tota.game import Drawer class JsonReplayDrawer(Drawer): def __init__(self, replay_dir): self.replay_dir = replay_dir def draw(self, game): """Draw the world with 'ascii'-art .""" things_data = [] tick_data = { 't': game.world.t, 'things': things_data, 'effects': [{ 'position': position, 'effect': effect, } for position, effect in game.world.effects.items()] } for thing in game.world.things.values(): thing_data = { 'id': id(thing), 'type': thing.__class__.__name__, 'position': thing.position, } if thing_data['type'] != 'Tree': thing_data.update({ 'life': thing.life, 'name': thing.name, 'team': thing.team, 'level': getattr(thing, 'level', None), 'xp': getattr(thing, 'xp', None), 'action': thing.last_action, 'target': thing.last_target, }) things_data.append(thing_data) tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t) with open(tick_path, 'w') as tick_file: json.dump(tick_data, tick_file, indent=2 if game.debug else None) <commit_msg>Add done result to replay<commit_after>import json from os import path from tota.game import Drawer class JsonReplayDrawer(Drawer): def __init__(self, replay_dir): self.replay_dir = replay_dir def draw(self, game): """Draw the world with 'ascii'-art .""" things_data = [] tick_data = { 't': game.world.t, 'things': things_data, 'effects': [{ 'position': position, 'effect': effect, } for position, effect in game.world.effects.items()] } for thing in game.world.things.values(): thing_data = { 'id': id(thing), 'type': thing.__class__.__name__, 'position': thing.position, } if thing_data['type'] != 'Tree': thing_data.update({ 'life': thing.life, 'name': thing.name, 'team': thing.team, 'level': getattr(thing, 'level', None), 'xp': getattr(thing, 'xp', None), 'action': thing.last_action, 'target': thing.last_target, 'action_done': thing.last_action_done, }) things_data.append(thing_data) tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t) with open(tick_path, 'w') as tick_file: json.dump(tick_data, tick_file, indent=2 if game.debug else None)
c3c703b411d05e6f2d52a0b3695b9dc22bc907d8
test/test_main.py
test/test_main.py
from mpf.main import main def test_main(): main()
import matplotlib matplotlib.use('Agg') # Not to use X server. For TravisCI. from mpf.main import main def test_main(): main()
Make matplotlib work with TravisCI
Make matplotlib work with TravisCI
Python
mit
Vayel/MPF,tartopum/MPF
from mpf.main import main def test_main(): main() Make matplotlib work with TravisCI
import matplotlib matplotlib.use('Agg') # Not to use X server. For TravisCI. from mpf.main import main def test_main(): main()
<commit_before>from mpf.main import main def test_main(): main() <commit_msg>Make matplotlib work with TravisCI<commit_after>
import matplotlib matplotlib.use('Agg') # Not to use X server. For TravisCI. from mpf.main import main def test_main(): main()
from mpf.main import main def test_main(): main() Make matplotlib work with TravisCIimport matplotlib matplotlib.use('Agg') # Not to use X server. For TravisCI. from mpf.main import main def test_main(): main()
<commit_before>from mpf.main import main def test_main(): main() <commit_msg>Make matplotlib work with TravisCI<commit_after>import matplotlib matplotlib.use('Agg') # Not to use X server. For TravisCI. from mpf.main import main def test_main(): main()
1a5aeabcdfae02125e167e8a221de4151819f5b5
test.py
test.py
import unittest from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen class RotorTestCase(unittest.TestCase): def test_rotor_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('E', rotor.encode('A')) def run_tests(): runner = unittest.TextTestRunner() suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase) runner.run(suite) if __name__ == '__main__': # pragma: no cover run_tests()
import unittest from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen class RotorTestCase(unittest.TestCase): def test_rotor_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('E', rotor.encode('A')) def test_rotor_reverse_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('U', rotor.encode_reverse('A')) def run_tests(): runner = unittest.TextTestRunner() suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase) runner.run(suite) if __name__ == '__main__': # pragma: no cover run_tests()
Test if default rotor encodes backwards properly
Test if default rotor encodes backwards properly
Python
mit
ranisalt/enigma
import unittest from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen class RotorTestCase(unittest.TestCase): def test_rotor_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('E', rotor.encode('A')) def run_tests(): runner = unittest.TextTestRunner() suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase) runner.run(suite) if __name__ == '__main__': # pragma: no cover run_tests()Test if default rotor encodes backwards properly
import unittest from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen class RotorTestCase(unittest.TestCase): def test_rotor_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('E', rotor.encode('A')) def test_rotor_reverse_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('U', rotor.encode_reverse('A')) def run_tests(): runner = unittest.TextTestRunner() suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase) runner.run(suite) if __name__ == '__main__': # pragma: no cover run_tests()
<commit_before>import unittest from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen class RotorTestCase(unittest.TestCase): def test_rotor_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('E', rotor.encode('A')) def run_tests(): runner = unittest.TextTestRunner() suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase) runner.run(suite) if __name__ == '__main__': # pragma: no cover run_tests()<commit_msg>Test if default rotor encodes backwards properly<commit_after>
import unittest from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen class RotorTestCase(unittest.TestCase): def test_rotor_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('E', rotor.encode('A')) def test_rotor_reverse_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('U', rotor.encode_reverse('A')) def run_tests(): runner = unittest.TextTestRunner() suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase) runner.run(suite) if __name__ == '__main__': # pragma: no cover run_tests()
import unittest from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen class RotorTestCase(unittest.TestCase): def test_rotor_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('E', rotor.encode('A')) def run_tests(): runner = unittest.TextTestRunner() suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase) runner.run(suite) if __name__ == '__main__': # pragma: no cover run_tests()Test if default rotor encodes backwards properlyimport unittest from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen class RotorTestCase(unittest.TestCase): def test_rotor_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('E', rotor.encode('A')) def test_rotor_reverse_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('U', rotor.encode_reverse('A')) def run_tests(): runner = unittest.TextTestRunner() suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase) runner.run(suite) if __name__ == '__main__': # pragma: no cover run_tests()
<commit_before>import unittest from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen class RotorTestCase(unittest.TestCase): def test_rotor_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('E', rotor.encode('A')) def run_tests(): runner = unittest.TextTestRunner() suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase) runner.run(suite) if __name__ == '__main__': # pragma: no cover run_tests()<commit_msg>Test if default rotor encodes backwards properly<commit_after>import unittest from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen class RotorTestCase(unittest.TestCase): def test_rotor_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('E', rotor.encode('A')) def test_rotor_reverse_encoding(self): rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q') self.assertEqual('U', rotor.encode_reverse('A')) def run_tests(): runner = unittest.TextTestRunner() suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase) runner.run(suite) if __name__ == '__main__': # pragma: no cover run_tests()
1a5cc5b69811db2ac63987ab329bd117e61f3f03
tests/__init__.py
tests/__init__.py
import os from functools import partial TESTS_DIR = os.path.dirname(os.path.abspath(__file__)) TEST_DATA_DIR = os.path.join(TESTS_DIR, 'test_data') AppEventTestXml = partial(open, os.path.join(TEST_DATA_DIR, 'app_event.xml'))
Make the test data accessible to the tests.
Make the test data accessible to the tests.
Python
bsd-3-clause
unt-libraries/django-premis-event-service,unt-libraries/django-premis-event-service,unt-libraries/django-premis-event-service
Make the test data accessible to the tests.
import os from functools import partial TESTS_DIR = os.path.dirname(os.path.abspath(__file__)) TEST_DATA_DIR = os.path.join(TESTS_DIR, 'test_data') AppEventTestXml = partial(open, os.path.join(TEST_DATA_DIR, 'app_event.xml'))
<commit_before><commit_msg>Make the test data accessible to the tests.<commit_after>
import os from functools import partial TESTS_DIR = os.path.dirname(os.path.abspath(__file__)) TEST_DATA_DIR = os.path.join(TESTS_DIR, 'test_data') AppEventTestXml = partial(open, os.path.join(TEST_DATA_DIR, 'app_event.xml'))
Make the test data accessible to the tests.import os from functools import partial TESTS_DIR = os.path.dirname(os.path.abspath(__file__)) TEST_DATA_DIR = os.path.join(TESTS_DIR, 'test_data') AppEventTestXml = partial(open, os.path.join(TEST_DATA_DIR, 'app_event.xml'))
<commit_before><commit_msg>Make the test data accessible to the tests.<commit_after>import os from functools import partial TESTS_DIR = os.path.dirname(os.path.abspath(__file__)) TEST_DATA_DIR = os.path.join(TESTS_DIR, 'test_data') AppEventTestXml = partial(open, os.path.join(TEST_DATA_DIR, 'app_event.xml'))
6160da958f4b8ecb1553c7bcca0b32bc1a5a1649
tests/conftest.py
tests/conftest.py
import os import shutil import tempfile import builtins import subprocess import pytest from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): shutil.rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) shutil.rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git']
import os import shutil import tempfile import builtins import subprocess import pytest import sys from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] def rmtree(dirname): """Remove a directory, even if it has read-only files (Windows). Git creates read-only files that must be removed on teardown. See https://stackoverflow.com/questions/2656322 for more info. Parameters ---------- dirname : str Directory to be removed """ try: shutil.rmtree(dirname) except PermissionError: if sys.platform == 'win32': subprocess.check_call(['del', '/F/S/Q', dirname], shell=True) else: raise
Make sure .git test directory is removed on Windows
Make sure .git test directory is removed on Windows
Python
bsd-3-clause
scopatz/rever,ergs/rever
import os import shutil import tempfile import builtins import subprocess import pytest from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): shutil.rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) shutil.rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] Make sure .git test directory is removed on Windows
import os import shutil import tempfile import builtins import subprocess import pytest import sys from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] def rmtree(dirname): """Remove a directory, even if it has read-only files (Windows). Git creates read-only files that must be removed on teardown. See https://stackoverflow.com/questions/2656322 for more info. Parameters ---------- dirname : str Directory to be removed """ try: shutil.rmtree(dirname) except PermissionError: if sys.platform == 'win32': subprocess.check_call(['del', '/F/S/Q', dirname], shell=True) else: raise
<commit_before>import os import shutil import tempfile import builtins import subprocess import pytest from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): shutil.rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) shutil.rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] <commit_msg>Make sure .git test directory is removed on Windows<commit_after>
import os import shutil import tempfile import builtins import subprocess import pytest import sys from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] def rmtree(dirname): """Remove a directory, even if it has read-only files (Windows). Git creates read-only files that must be removed on teardown. See https://stackoverflow.com/questions/2656322 for more info. Parameters ---------- dirname : str Directory to be removed """ try: shutil.rmtree(dirname) except PermissionError: if sys.platform == 'win32': subprocess.check_call(['del', '/F/S/Q', dirname], shell=True) else: raise
import os import shutil import tempfile import builtins import subprocess import pytest from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): shutil.rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) shutil.rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] Make sure .git test directory is removed on Windowsimport os import shutil import tempfile import builtins import subprocess import pytest import sys from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] def rmtree(dirname): """Remove a directory, even if it has read-only files (Windows). Git creates read-only files that must be removed on teardown. See https://stackoverflow.com/questions/2656322 for more info. Parameters ---------- dirname : str Directory to be removed """ try: shutil.rmtree(dirname) except PermissionError: if sys.platform == 'win32': subprocess.check_call(['del', '/F/S/Q', dirname], shell=True) else: raise
<commit_before>import os import shutil import tempfile import builtins import subprocess import pytest from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): shutil.rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) shutil.rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] <commit_msg>Make sure .git test directory is removed on Windows<commit_after>import os import shutil import tempfile import builtins import subprocess import pytest import sys from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] def rmtree(dirname): """Remove a directory, even if it has read-only files (Windows). Git creates read-only files that must be removed on teardown. See https://stackoverflow.com/questions/2656322 for more info. Parameters ---------- dirname : str Directory to be removed """ try: shutil.rmtree(dirname) except PermissionError: if sys.platform == 'win32': subprocess.check_call(['del', '/F/S/Q', dirname], shell=True) else: raise
c4e71b56e74ab8b81a670c690fef6942d4a412b4
ocds/storage/backends/fs.py
ocds/storage/backends/fs.py
import os import os.path import logging import datetime from .base import Storage from ocds.storage.errors import InvalidPath logger = logging.getLogger(__name__) class FSStorage(Storage): def __init__(self, base_path): self.base_path = base_path if not os.path.exists(self.base_path): logger.warn('Initial path not exists. Creating') try: os.makedirs(self.base_path) except (IOError, OSError) as e: logger.error("Couldn't create destination dir." "Error {}".format(e)) raise InvalidPath('Not destination folder') def _from_string(self, string): return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S') def _path_from_date(self, date): if isinstance(date, str): path =
import os import os.path import logging import datetime import simplejson as json from .base import Storage from ocds.export.helpers import encoder from ocds.storage.errors import InvalidPath join = os.path.join logger = logging.getLogger(__name__) class FSStorage(Storage): def __init__(self, base_path): self.base_path = base_path self.path_fmt = '%Y-%m-%d/%H/%M%/%S' if not os.path.exists(self.base_path): logger.warn('Initial path not exists. Creating') try: os.makedirs(self.base_path) except (IOError, OSError) as e: logger.error("Couldn't create destination dir." "Error {}".format(e)) raise InvalidPath('Not destination folder') def _walk(self): for path, _, files in os.walk(self.base_path): for f in files: yield join(path, f) def _write(self, obj): path = join(self.base_path, self._path_from_date(obj['date'])) file_path = join(path, '{}.json'.format(obj['id'])) with open(file_path, 'w') as out: out.write(encoder(obj)) def _load(self, key): with open(join(self.base_path, key)) as out: result = json.load(out) return result def _from_string(self, string): return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S') def _path_from_date(self, date): if isinstance(date, str): path = self._from_string(date).srtftime(self.path_fmt) if isinstance(date, datetime.date): path = date.strftime(self.path_fmt) return path def __contains__(self, key): try: fs = open(join(self.base_path, key)) result = True except (IOError, OSError): result = False finally: fs.close() return result def __iter__(self): for f in self._walk(): yield f def save(self, obj): self._write(obj) def get(self, key): return self._load(key)
Add basic file system storage
Add basic file system storage
Python
apache-2.0
yshalenyk/openprocurement.ocds.export,yshalenyk/ocds.storage,yshalenyk/ocds.export,yshalenyk/openprocurement.ocds.export
import os import os.path import logging import datetime from .base import Storage from ocds.storage.errors import InvalidPath logger = logging.getLogger(__name__) class FSStorage(Storage): def __init__(self, base_path): self.base_path = base_path if not os.path.exists(self.base_path): logger.warn('Initial path not exists. Creating') try: os.makedirs(self.base_path) except (IOError, OSError) as e: logger.error("Couldn't create destination dir." "Error {}".format(e)) raise InvalidPath('Not destination folder') def _from_string(self, string): return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S') def _path_from_date(self, date): if isinstance(date, str): path = Add basic file system storage
import os import os.path import logging import datetime import simplejson as json from .base import Storage from ocds.export.helpers import encoder from ocds.storage.errors import InvalidPath join = os.path.join logger = logging.getLogger(__name__) class FSStorage(Storage): def __init__(self, base_path): self.base_path = base_path self.path_fmt = '%Y-%m-%d/%H/%M%/%S' if not os.path.exists(self.base_path): logger.warn('Initial path not exists. Creating') try: os.makedirs(self.base_path) except (IOError, OSError) as e: logger.error("Couldn't create destination dir." "Error {}".format(e)) raise InvalidPath('Not destination folder') def _walk(self): for path, _, files in os.walk(self.base_path): for f in files: yield join(path, f) def _write(self, obj): path = join(self.base_path, self._path_from_date(obj['date'])) file_path = join(path, '{}.json'.format(obj['id'])) with open(file_path, 'w') as out: out.write(encoder(obj)) def _load(self, key): with open(join(self.base_path, key)) as out: result = json.load(out) return result def _from_string(self, string): return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S') def _path_from_date(self, date): if isinstance(date, str): path = self._from_string(date).srtftime(self.path_fmt) if isinstance(date, datetime.date): path = date.strftime(self.path_fmt) return path def __contains__(self, key): try: fs = open(join(self.base_path, key)) result = True except (IOError, OSError): result = False finally: fs.close() return result def __iter__(self): for f in self._walk(): yield f def save(self, obj): self._write(obj) def get(self, key): return self._load(key)
<commit_before>import os import os.path import logging import datetime from .base import Storage from ocds.storage.errors import InvalidPath logger = logging.getLogger(__name__) class FSStorage(Storage): def __init__(self, base_path): self.base_path = base_path if not os.path.exists(self.base_path): logger.warn('Initial path not exists. Creating') try: os.makedirs(self.base_path) except (IOError, OSError) as e: logger.error("Couldn't create destination dir." "Error {}".format(e)) raise InvalidPath('Not destination folder') def _from_string(self, string): return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S') def _path_from_date(self, date): if isinstance(date, str): path = <commit_msg>Add basic file system storage<commit_after>
import os import os.path import logging import datetime import simplejson as json from .base import Storage from ocds.export.helpers import encoder from ocds.storage.errors import InvalidPath join = os.path.join logger = logging.getLogger(__name__) class FSStorage(Storage): def __init__(self, base_path): self.base_path = base_path self.path_fmt = '%Y-%m-%d/%H/%M%/%S' if not os.path.exists(self.base_path): logger.warn('Initial path not exists. Creating') try: os.makedirs(self.base_path) except (IOError, OSError) as e: logger.error("Couldn't create destination dir." "Error {}".format(e)) raise InvalidPath('Not destination folder') def _walk(self): for path, _, files in os.walk(self.base_path): for f in files: yield join(path, f) def _write(self, obj): path = join(self.base_path, self._path_from_date(obj['date'])) file_path = join(path, '{}.json'.format(obj['id'])) with open(file_path, 'w') as out: out.write(encoder(obj)) def _load(self, key): with open(join(self.base_path, key)) as out: result = json.load(out) return result def _from_string(self, string): return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S') def _path_from_date(self, date): if isinstance(date, str): path = self._from_string(date).srtftime(self.path_fmt) if isinstance(date, datetime.date): path = date.strftime(self.path_fmt) return path def __contains__(self, key): try: fs = open(join(self.base_path, key)) result = True except (IOError, OSError): result = False finally: fs.close() return result def __iter__(self): for f in self._walk(): yield f def save(self, obj): self._write(obj) def get(self, key): return self._load(key)
import os import os.path import logging import datetime from .base import Storage from ocds.storage.errors import InvalidPath logger = logging.getLogger(__name__) class FSStorage(Storage): def __init__(self, base_path): self.base_path = base_path if not os.path.exists(self.base_path): logger.warn('Initial path not exists. Creating') try: os.makedirs(self.base_path) except (IOError, OSError) as e: logger.error("Couldn't create destination dir." "Error {}".format(e)) raise InvalidPath('Not destination folder') def _from_string(self, string): return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S') def _path_from_date(self, date): if isinstance(date, str): path = Add basic file system storageimport os import os.path import logging import datetime import simplejson as json from .base import Storage from ocds.export.helpers import encoder from ocds.storage.errors import InvalidPath join = os.path.join logger = logging.getLogger(__name__) class FSStorage(Storage): def __init__(self, base_path): self.base_path = base_path self.path_fmt = '%Y-%m-%d/%H/%M%/%S' if not os.path.exists(self.base_path): logger.warn('Initial path not exists. Creating') try: os.makedirs(self.base_path) except (IOError, OSError) as e: logger.error("Couldn't create destination dir." "Error {}".format(e)) raise InvalidPath('Not destination folder') def _walk(self): for path, _, files in os.walk(self.base_path): for f in files: yield join(path, f) def _write(self, obj): path = join(self.base_path, self._path_from_date(obj['date'])) file_path = join(path, '{}.json'.format(obj['id'])) with open(file_path, 'w') as out: out.write(encoder(obj)) def _load(self, key): with open(join(self.base_path, key)) as out: result = json.load(out) return result def _from_string(self, string): return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S') def _path_from_date(self, date): if isinstance(date, str): path = self._from_string(date).srtftime(self.path_fmt) if isinstance(date, datetime.date): path = date.strftime(self.path_fmt) return path def __contains__(self, key): try: fs = open(join(self.base_path, key)) result = True except (IOError, OSError): result = False finally: fs.close() return result def __iter__(self): for f in self._walk(): yield f def save(self, obj): self._write(obj) def get(self, key): return self._load(key)
<commit_before>import os import os.path import logging import datetime from .base import Storage from ocds.storage.errors import InvalidPath logger = logging.getLogger(__name__) class FSStorage(Storage): def __init__(self, base_path): self.base_path = base_path if not os.path.exists(self.base_path): logger.warn('Initial path not exists. Creating') try: os.makedirs(self.base_path) except (IOError, OSError) as e: logger.error("Couldn't create destination dir." "Error {}".format(e)) raise InvalidPath('Not destination folder') def _from_string(self, string): return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S') def _path_from_date(self, date): if isinstance(date, str): path = <commit_msg>Add basic file system storage<commit_after>import os import os.path import logging import datetime import simplejson as json from .base import Storage from ocds.export.helpers import encoder from ocds.storage.errors import InvalidPath join = os.path.join logger = logging.getLogger(__name__) class FSStorage(Storage): def __init__(self, base_path): self.base_path = base_path self.path_fmt = '%Y-%m-%d/%H/%M%/%S' if not os.path.exists(self.base_path): logger.warn('Initial path not exists. Creating') try: os.makedirs(self.base_path) except (IOError, OSError) as e: logger.error("Couldn't create destination dir." "Error {}".format(e)) raise InvalidPath('Not destination folder') def _walk(self): for path, _, files in os.walk(self.base_path): for f in files: yield join(path, f) def _write(self, obj): path = join(self.base_path, self._path_from_date(obj['date'])) file_path = join(path, '{}.json'.format(obj['id'])) with open(file_path, 'w') as out: out.write(encoder(obj)) def _load(self, key): with open(join(self.base_path, key)) as out: result = json.load(out) return result def _from_string(self, string): return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S') def _path_from_date(self, date): if isinstance(date, str): path = self._from_string(date).srtftime(self.path_fmt) if isinstance(date, datetime.date): path = date.strftime(self.path_fmt) return path def __contains__(self, key): try: fs = open(join(self.base_path, key)) result = True except (IOError, OSError): result = False finally: fs.close() return result def __iter__(self): for f in self._walk(): yield f def save(self, obj): self._write(obj) def get(self, key): return self._load(key)
2ccfb54f493bf0ffa07db910514a8429a2c51d73
changes/api/node_job_index.py
changes/api/node_job_index.py
from __future__ import absolute_import from sqlalchemy.orm import joinedload from changes.api.base import APIView from changes.api.serializer.models.job import JobWithBuildSerializer from changes.models import Job, JobStep, Node class NodeJobIndexAPIView(APIView): def get(self, node_id): node = Node.query.get(node_id) if node is None: return '', 404 jobs = Job.query.join( JobStep, JobStep.job_id == Job.id, ).options( joinedload(Job.build, innerjoin=True), ).filter( JobStep.node_id == node.id, ).order_by(Job.date_created.desc()) return self.paginate(jobs, serializers={ Job: JobWithBuildSerializer(), })
from __future__ import absolute_import from sqlalchemy.orm import joinedload from changes.api.base import APIView from changes.models import Build, Job, JobStep, Node class NodeJobIndexAPIView(APIView): def get(self, node_id): node = Node.query.get(node_id) if node is None: return '', 404 jobs = list(Job.query.join( JobStep, JobStep.job_id == Job.id, ).filter( JobStep.node_id == node.id, ).order_by(Job.date_created.desc())) build_list = list(Build.query.options( joinedload('project'), joinedload('author'), joinedload('source').joinedload('revision'), ).filter( Build.id.in_(j.build_id for j in jobs), )) build_map = dict( (b, d) for b, d in zip(build_list, self.serialize(build_list)) ) context = [] for job, data in zip(jobs, self.serialize(jobs)): print job, data data['build'] = build_map[job.build] context.append(data) return self.paginate(context, serialize=False)
Improve query patterns on node job list
Improve query patterns on node job list
Python
apache-2.0
wfxiang08/changes,wfxiang08/changes,bowlofstew/changes,wfxiang08/changes,dropbox/changes,dropbox/changes,dropbox/changes,bowlofstew/changes,bowlofstew/changes,wfxiang08/changes,dropbox/changes,bowlofstew/changes
from __future__ import absolute_import from sqlalchemy.orm import joinedload from changes.api.base import APIView from changes.api.serializer.models.job import JobWithBuildSerializer from changes.models import Job, JobStep, Node class NodeJobIndexAPIView(APIView): def get(self, node_id): node = Node.query.get(node_id) if node is None: return '', 404 jobs = Job.query.join( JobStep, JobStep.job_id == Job.id, ).options( joinedload(Job.build, innerjoin=True), ).filter( JobStep.node_id == node.id, ).order_by(Job.date_created.desc()) return self.paginate(jobs, serializers={ Job: JobWithBuildSerializer(), }) Improve query patterns on node job list
from __future__ import absolute_import from sqlalchemy.orm import joinedload from changes.api.base import APIView from changes.models import Build, Job, JobStep, Node class NodeJobIndexAPIView(APIView): def get(self, node_id): node = Node.query.get(node_id) if node is None: return '', 404 jobs = list(Job.query.join( JobStep, JobStep.job_id == Job.id, ).filter( JobStep.node_id == node.id, ).order_by(Job.date_created.desc())) build_list = list(Build.query.options( joinedload('project'), joinedload('author'), joinedload('source').joinedload('revision'), ).filter( Build.id.in_(j.build_id for j in jobs), )) build_map = dict( (b, d) for b, d in zip(build_list, self.serialize(build_list)) ) context = [] for job, data in zip(jobs, self.serialize(jobs)): print job, data data['build'] = build_map[job.build] context.append(data) return self.paginate(context, serialize=False)
<commit_before>from __future__ import absolute_import from sqlalchemy.orm import joinedload from changes.api.base import APIView from changes.api.serializer.models.job import JobWithBuildSerializer from changes.models import Job, JobStep, Node class NodeJobIndexAPIView(APIView): def get(self, node_id): node = Node.query.get(node_id) if node is None: return '', 404 jobs = Job.query.join( JobStep, JobStep.job_id == Job.id, ).options( joinedload(Job.build, innerjoin=True), ).filter( JobStep.node_id == node.id, ).order_by(Job.date_created.desc()) return self.paginate(jobs, serializers={ Job: JobWithBuildSerializer(), }) <commit_msg>Improve query patterns on node job list<commit_after>
from __future__ import absolute_import from sqlalchemy.orm import joinedload from changes.api.base import APIView from changes.models import Build, Job, JobStep, Node class NodeJobIndexAPIView(APIView): def get(self, node_id): node = Node.query.get(node_id) if node is None: return '', 404 jobs = list(Job.query.join( JobStep, JobStep.job_id == Job.id, ).filter( JobStep.node_id == node.id, ).order_by(Job.date_created.desc())) build_list = list(Build.query.options( joinedload('project'), joinedload('author'), joinedload('source').joinedload('revision'), ).filter( Build.id.in_(j.build_id for j in jobs), )) build_map = dict( (b, d) for b, d in zip(build_list, self.serialize(build_list)) ) context = [] for job, data in zip(jobs, self.serialize(jobs)): print job, data data['build'] = build_map[job.build] context.append(data) return self.paginate(context, serialize=False)
from __future__ import absolute_import from sqlalchemy.orm import joinedload from changes.api.base import APIView from changes.api.serializer.models.job import JobWithBuildSerializer from changes.models import Job, JobStep, Node class NodeJobIndexAPIView(APIView): def get(self, node_id): node = Node.query.get(node_id) if node is None: return '', 404 jobs = Job.query.join( JobStep, JobStep.job_id == Job.id, ).options( joinedload(Job.build, innerjoin=True), ).filter( JobStep.node_id == node.id, ).order_by(Job.date_created.desc()) return self.paginate(jobs, serializers={ Job: JobWithBuildSerializer(), }) Improve query patterns on node job listfrom __future__ import absolute_import from sqlalchemy.orm import joinedload from changes.api.base import APIView from changes.models import Build, Job, JobStep, Node class NodeJobIndexAPIView(APIView): def get(self, node_id): node = Node.query.get(node_id) if node is None: return '', 404 jobs = list(Job.query.join( JobStep, JobStep.job_id == Job.id, ).filter( JobStep.node_id == node.id, ).order_by(Job.date_created.desc())) build_list = list(Build.query.options( joinedload('project'), joinedload('author'), joinedload('source').joinedload('revision'), ).filter( Build.id.in_(j.build_id for j in jobs), )) build_map = dict( (b, d) for b, d in zip(build_list, self.serialize(build_list)) ) context = [] for job, data in zip(jobs, self.serialize(jobs)): print job, data data['build'] = build_map[job.build] context.append(data) return self.paginate(context, serialize=False)
<commit_before>from __future__ import absolute_import from sqlalchemy.orm import joinedload from changes.api.base import APIView from changes.api.serializer.models.job import JobWithBuildSerializer from changes.models import Job, JobStep, Node class NodeJobIndexAPIView(APIView): def get(self, node_id): node = Node.query.get(node_id) if node is None: return '', 404 jobs = Job.query.join( JobStep, JobStep.job_id == Job.id, ).options( joinedload(Job.build, innerjoin=True), ).filter( JobStep.node_id == node.id, ).order_by(Job.date_created.desc()) return self.paginate(jobs, serializers={ Job: JobWithBuildSerializer(), }) <commit_msg>Improve query patterns on node job list<commit_after>from __future__ import absolute_import from sqlalchemy.orm import joinedload from changes.api.base import APIView from changes.models import Build, Job, JobStep, Node class NodeJobIndexAPIView(APIView): def get(self, node_id): node = Node.query.get(node_id) if node is None: return '', 404 jobs = list(Job.query.join( JobStep, JobStep.job_id == Job.id, ).filter( JobStep.node_id == node.id, ).order_by(Job.date_created.desc())) build_list = list(Build.query.options( joinedload('project'), joinedload('author'), joinedload('source').joinedload('revision'), ).filter( Build.id.in_(j.build_id for j in jobs), )) build_map = dict( (b, d) for b, d in zip(build_list, self.serialize(build_list)) ) context = [] for job, data in zip(jobs, self.serialize(jobs)): print job, data data['build'] = build_map[job.build] context.append(data) return self.paginate(context, serialize=False)
c12f3e516eb28d306a103582495216253dd98e7e
feedreader/tasks/core.py
feedreader/tasks/core.py
from celery import Celery class Tasks(object): def __init__(self, debug=False): self.app = Celery() self.app.conf.update( CELERY_ACCEPT_CONTENT=['json'], CELERY_ALWAYS_EAGER=True, CELERY_ENABLE_UTC=True, CELERY_TASK_SERIALIZER='json', CELERY_RESULT_SERIALIZER='json', CELERY_TIMEZONE='America/Vancouver', ) if not debug: self.app.conf.update( BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//', CELERY_ALWAYS_EAGER=False, CELERY_RESULT_BACKEND='amqp', ) # register tasks with celery self.fetch_feed = self.app.task()(self.fetch_feed) # celery tasks def fetch_feed(self, url): return 'hola bro'
from celery import Celery class Tasks(object): def __init__(self, debug=False): self.app = Celery() self.app.conf.update( CELERY_ACCEPT_CONTENT=['json'], CELERY_ALWAYS_EAGER=True, CELERY_ENABLE_UTC=True, CELERY_TASK_SERIALIZER='json', CELERY_RESULT_SERIALIZER='json', CELERY_TIMEZONE='America/Vancouver', ) if not debug: self.app.conf.update( BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//', CELERY_ALWAYS_EAGER=False, CELERY_RESULT_BACKEND='amqp', ) # register tasks with celery self.fetch_feed = self.app.task()(self.fetch_feed) # celery tasks def fetch_feed(self, feed_url, last_modified=None, etag=None): """Fetch and parse the feed at the given URL. If the given URL is not a feed, this will attempt to find one. Raises SomeException if an error occurs. Returns dict containing: - feed_url: canonical url of the feed resource - feed: new instance of the Feed model, or None if the feed was unmodified - entries: list of new instances of the Entry model, or empty list if the feed was unmodified - last_modified: last modified date, if server provides one - etag: etag, if server provides one """ return # TODO
Add a stub task for fetch_feed
Add a stub task for fetch_feed
Python
mit
tdryer/feeder,tdryer/feeder
from celery import Celery class Tasks(object): def __init__(self, debug=False): self.app = Celery() self.app.conf.update( CELERY_ACCEPT_CONTENT=['json'], CELERY_ALWAYS_EAGER=True, CELERY_ENABLE_UTC=True, CELERY_TASK_SERIALIZER='json', CELERY_RESULT_SERIALIZER='json', CELERY_TIMEZONE='America/Vancouver', ) if not debug: self.app.conf.update( BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//', CELERY_ALWAYS_EAGER=False, CELERY_RESULT_BACKEND='amqp', ) # register tasks with celery self.fetch_feed = self.app.task()(self.fetch_feed) # celery tasks def fetch_feed(self, url): return 'hola bro' Add a stub task for fetch_feed
from celery import Celery class Tasks(object): def __init__(self, debug=False): self.app = Celery() self.app.conf.update( CELERY_ACCEPT_CONTENT=['json'], CELERY_ALWAYS_EAGER=True, CELERY_ENABLE_UTC=True, CELERY_TASK_SERIALIZER='json', CELERY_RESULT_SERIALIZER='json', CELERY_TIMEZONE='America/Vancouver', ) if not debug: self.app.conf.update( BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//', CELERY_ALWAYS_EAGER=False, CELERY_RESULT_BACKEND='amqp', ) # register tasks with celery self.fetch_feed = self.app.task()(self.fetch_feed) # celery tasks def fetch_feed(self, feed_url, last_modified=None, etag=None): """Fetch and parse the feed at the given URL. If the given URL is not a feed, this will attempt to find one. Raises SomeException if an error occurs. Returns dict containing: - feed_url: canonical url of the feed resource - feed: new instance of the Feed model, or None if the feed was unmodified - entries: list of new instances of the Entry model, or empty list if the feed was unmodified - last_modified: last modified date, if server provides one - etag: etag, if server provides one """ return # TODO
<commit_before>from celery import Celery class Tasks(object): def __init__(self, debug=False): self.app = Celery() self.app.conf.update( CELERY_ACCEPT_CONTENT=['json'], CELERY_ALWAYS_EAGER=True, CELERY_ENABLE_UTC=True, CELERY_TASK_SERIALIZER='json', CELERY_RESULT_SERIALIZER='json', CELERY_TIMEZONE='America/Vancouver', ) if not debug: self.app.conf.update( BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//', CELERY_ALWAYS_EAGER=False, CELERY_RESULT_BACKEND='amqp', ) # register tasks with celery self.fetch_feed = self.app.task()(self.fetch_feed) # celery tasks def fetch_feed(self, url): return 'hola bro' <commit_msg>Add a stub task for fetch_feed<commit_after>
from celery import Celery class Tasks(object): def __init__(self, debug=False): self.app = Celery() self.app.conf.update( CELERY_ACCEPT_CONTENT=['json'], CELERY_ALWAYS_EAGER=True, CELERY_ENABLE_UTC=True, CELERY_TASK_SERIALIZER='json', CELERY_RESULT_SERIALIZER='json', CELERY_TIMEZONE='America/Vancouver', ) if not debug: self.app.conf.update( BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//', CELERY_ALWAYS_EAGER=False, CELERY_RESULT_BACKEND='amqp', ) # register tasks with celery self.fetch_feed = self.app.task()(self.fetch_feed) # celery tasks def fetch_feed(self, feed_url, last_modified=None, etag=None): """Fetch and parse the feed at the given URL. If the given URL is not a feed, this will attempt to find one. Raises SomeException if an error occurs. Returns dict containing: - feed_url: canonical url of the feed resource - feed: new instance of the Feed model, or None if the feed was unmodified - entries: list of new instances of the Entry model, or empty list if the feed was unmodified - last_modified: last modified date, if server provides one - etag: etag, if server provides one """ return # TODO
from celery import Celery class Tasks(object): def __init__(self, debug=False): self.app = Celery() self.app.conf.update( CELERY_ACCEPT_CONTENT=['json'], CELERY_ALWAYS_EAGER=True, CELERY_ENABLE_UTC=True, CELERY_TASK_SERIALIZER='json', CELERY_RESULT_SERIALIZER='json', CELERY_TIMEZONE='America/Vancouver', ) if not debug: self.app.conf.update( BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//', CELERY_ALWAYS_EAGER=False, CELERY_RESULT_BACKEND='amqp', ) # register tasks with celery self.fetch_feed = self.app.task()(self.fetch_feed) # celery tasks def fetch_feed(self, url): return 'hola bro' Add a stub task for fetch_feedfrom celery import Celery class Tasks(object): def __init__(self, debug=False): self.app = Celery() self.app.conf.update( CELERY_ACCEPT_CONTENT=['json'], CELERY_ALWAYS_EAGER=True, CELERY_ENABLE_UTC=True, CELERY_TASK_SERIALIZER='json', CELERY_RESULT_SERIALIZER='json', CELERY_TIMEZONE='America/Vancouver', ) if not debug: self.app.conf.update( BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//', CELERY_ALWAYS_EAGER=False, CELERY_RESULT_BACKEND='amqp', ) # register tasks with celery self.fetch_feed = self.app.task()(self.fetch_feed) # celery tasks def fetch_feed(self, feed_url, last_modified=None, etag=None): """Fetch and parse the feed at the given URL. If the given URL is not a feed, this will attempt to find one. Raises SomeException if an error occurs. Returns dict containing: - feed_url: canonical url of the feed resource - feed: new instance of the Feed model, or None if the feed was unmodified - entries: list of new instances of the Entry model, or empty list if the feed was unmodified - last_modified: last modified date, if server provides one - etag: etag, if server provides one """ return # TODO
<commit_before>from celery import Celery class Tasks(object): def __init__(self, debug=False): self.app = Celery() self.app.conf.update( CELERY_ACCEPT_CONTENT=['json'], CELERY_ALWAYS_EAGER=True, CELERY_ENABLE_UTC=True, CELERY_TASK_SERIALIZER='json', CELERY_RESULT_SERIALIZER='json', CELERY_TIMEZONE='America/Vancouver', ) if not debug: self.app.conf.update( BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//', CELERY_ALWAYS_EAGER=False, CELERY_RESULT_BACKEND='amqp', ) # register tasks with celery self.fetch_feed = self.app.task()(self.fetch_feed) # celery tasks def fetch_feed(self, url): return 'hola bro' <commit_msg>Add a stub task for fetch_feed<commit_after>from celery import Celery class Tasks(object): def __init__(self, debug=False): self.app = Celery() self.app.conf.update( CELERY_ACCEPT_CONTENT=['json'], CELERY_ALWAYS_EAGER=True, CELERY_ENABLE_UTC=True, CELERY_TASK_SERIALIZER='json', CELERY_RESULT_SERIALIZER='json', CELERY_TIMEZONE='America/Vancouver', ) if not debug: self.app.conf.update( BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//', CELERY_ALWAYS_EAGER=False, CELERY_RESULT_BACKEND='amqp', ) # register tasks with celery self.fetch_feed = self.app.task()(self.fetch_feed) # celery tasks def fetch_feed(self, feed_url, last_modified=None, etag=None): """Fetch and parse the feed at the given URL. If the given URL is not a feed, this will attempt to find one. Raises SomeException if an error occurs. Returns dict containing: - feed_url: canonical url of the feed resource - feed: new instance of the Feed model, or None if the feed was unmodified - entries: list of new instances of the Entry model, or empty list if the feed was unmodified - last_modified: last modified date, if server provides one - etag: etag, if server provides one """ return # TODO
8beaab317d5da25edd093be42f57e35ac12408b8
feincms3/plugins/html.py
feincms3/plugins/html.py
""" Plugin providing a simple textarea where raw HTML, CSS and JS code can be entered. Most useful for people wanting to shoot themselves in the foot. """ from django.db import models from django.utils.html import mark_safe from django.utils.translation import ugettext_lazy as _ from content_editor.admin import ContentEditorInline __all__ = ("HTML", "HTMLInline", "render_html") class HTML(models.Model): """ Raw HTML plugin """ html = models.TextField( "HTML", help_text=_( "The content will be inserted directly into the page." " It is VERY important that the HTML snippet is well-formed!" ), ) class Meta: abstract = True verbose_name = "HTML" verbose_name_plural = "HTML" def __str__(self): return "" class HTMLInline(ContentEditorInline): """ Just available for consistency, absolutely no difference to a standard ``ContentEditorInline``. """ pass def render_html(plugin, **kwargs): """ Return the HTML code as safe string so that it is not escaped. Of course the contents are not guaranteed to be safe at all """ return mark_safe(plugin.html)
""" Plugin providing a simple textarea where raw HTML, CSS and JS code can be entered. Most useful for people wanting to shoot themselves in the foot. """ from django import forms from django.db import models from django.utils.html import mark_safe from django.utils.translation import ugettext_lazy as _ from content_editor.admin import ContentEditorInline __all__ = ("HTML", "HTMLInline", "render_html") class HTML(models.Model): """ Raw HTML plugin """ html = models.TextField( "HTML", help_text=_( "The content will be inserted directly into the page." " It is VERY important that the HTML snippet is well-formed!" ), ) class Meta: abstract = True verbose_name = "HTML" verbose_name_plural = "HTML" def __str__(self): return "" class HTMLInline(ContentEditorInline): """ Just available for consistency, absolutely no difference to a standard ``ContentEditorInline``. """ formfield_overrides = { models.TextField: { "widget": forms.Textarea( attrs={"rows": 3, "cols": 40, "class": "vLargeTextField"} ) } } def render_html(plugin, **kwargs): """ Return the HTML code as safe string so that it is not escaped. Of course the contents are not guaranteed to be safe at all """ return mark_safe(plugin.html)
Make the default HTML textarea smaller
Make the default HTML textarea smaller
Python
bsd-3-clause
matthiask/feincms3,matthiask/feincms3,matthiask/feincms3
""" Plugin providing a simple textarea where raw HTML, CSS and JS code can be entered. Most useful for people wanting to shoot themselves in the foot. """ from django.db import models from django.utils.html import mark_safe from django.utils.translation import ugettext_lazy as _ from content_editor.admin import ContentEditorInline __all__ = ("HTML", "HTMLInline", "render_html") class HTML(models.Model): """ Raw HTML plugin """ html = models.TextField( "HTML", help_text=_( "The content will be inserted directly into the page." " It is VERY important that the HTML snippet is well-formed!" ), ) class Meta: abstract = True verbose_name = "HTML" verbose_name_plural = "HTML" def __str__(self): return "" class HTMLInline(ContentEditorInline): """ Just available for consistency, absolutely no difference to a standard ``ContentEditorInline``. """ pass def render_html(plugin, **kwargs): """ Return the HTML code as safe string so that it is not escaped. Of course the contents are not guaranteed to be safe at all """ return mark_safe(plugin.html) Make the default HTML textarea smaller
""" Plugin providing a simple textarea where raw HTML, CSS and JS code can be entered. Most useful for people wanting to shoot themselves in the foot. """ from django import forms from django.db import models from django.utils.html import mark_safe from django.utils.translation import ugettext_lazy as _ from content_editor.admin import ContentEditorInline __all__ = ("HTML", "HTMLInline", "render_html") class HTML(models.Model): """ Raw HTML plugin """ html = models.TextField( "HTML", help_text=_( "The content will be inserted directly into the page." " It is VERY important that the HTML snippet is well-formed!" ), ) class Meta: abstract = True verbose_name = "HTML" verbose_name_plural = "HTML" def __str__(self): return "" class HTMLInline(ContentEditorInline): """ Just available for consistency, absolutely no difference to a standard ``ContentEditorInline``. """ formfield_overrides = { models.TextField: { "widget": forms.Textarea( attrs={"rows": 3, "cols": 40, "class": "vLargeTextField"} ) } } def render_html(plugin, **kwargs): """ Return the HTML code as safe string so that it is not escaped. Of course the contents are not guaranteed to be safe at all """ return mark_safe(plugin.html)
<commit_before>""" Plugin providing a simple textarea where raw HTML, CSS and JS code can be entered. Most useful for people wanting to shoot themselves in the foot. """ from django.db import models from django.utils.html import mark_safe from django.utils.translation import ugettext_lazy as _ from content_editor.admin import ContentEditorInline __all__ = ("HTML", "HTMLInline", "render_html") class HTML(models.Model): """ Raw HTML plugin """ html = models.TextField( "HTML", help_text=_( "The content will be inserted directly into the page." " It is VERY important that the HTML snippet is well-formed!" ), ) class Meta: abstract = True verbose_name = "HTML" verbose_name_plural = "HTML" def __str__(self): return "" class HTMLInline(ContentEditorInline): """ Just available for consistency, absolutely no difference to a standard ``ContentEditorInline``. """ pass def render_html(plugin, **kwargs): """ Return the HTML code as safe string so that it is not escaped. Of course the contents are not guaranteed to be safe at all """ return mark_safe(plugin.html) <commit_msg>Make the default HTML textarea smaller<commit_after>
""" Plugin providing a simple textarea where raw HTML, CSS and JS code can be entered. Most useful for people wanting to shoot themselves in the foot. """ from django import forms from django.db import models from django.utils.html import mark_safe from django.utils.translation import ugettext_lazy as _ from content_editor.admin import ContentEditorInline __all__ = ("HTML", "HTMLInline", "render_html") class HTML(models.Model): """ Raw HTML plugin """ html = models.TextField( "HTML", help_text=_( "The content will be inserted directly into the page." " It is VERY important that the HTML snippet is well-formed!" ), ) class Meta: abstract = True verbose_name = "HTML" verbose_name_plural = "HTML" def __str__(self): return "" class HTMLInline(ContentEditorInline): """ Just available for consistency, absolutely no difference to a standard ``ContentEditorInline``. """ formfield_overrides = { models.TextField: { "widget": forms.Textarea( attrs={"rows": 3, "cols": 40, "class": "vLargeTextField"} ) } } def render_html(plugin, **kwargs): """ Return the HTML code as safe string so that it is not escaped. Of course the contents are not guaranteed to be safe at all """ return mark_safe(plugin.html)
""" Plugin providing a simple textarea where raw HTML, CSS and JS code can be entered. Most useful for people wanting to shoot themselves in the foot. """ from django.db import models from django.utils.html import mark_safe from django.utils.translation import ugettext_lazy as _ from content_editor.admin import ContentEditorInline __all__ = ("HTML", "HTMLInline", "render_html") class HTML(models.Model): """ Raw HTML plugin """ html = models.TextField( "HTML", help_text=_( "The content will be inserted directly into the page." " It is VERY important that the HTML snippet is well-formed!" ), ) class Meta: abstract = True verbose_name = "HTML" verbose_name_plural = "HTML" def __str__(self): return "" class HTMLInline(ContentEditorInline): """ Just available for consistency, absolutely no difference to a standard ``ContentEditorInline``. """ pass def render_html(plugin, **kwargs): """ Return the HTML code as safe string so that it is not escaped. Of course the contents are not guaranteed to be safe at all """ return mark_safe(plugin.html) Make the default HTML textarea smaller""" Plugin providing a simple textarea where raw HTML, CSS and JS code can be entered. Most useful for people wanting to shoot themselves in the foot. """ from django import forms from django.db import models from django.utils.html import mark_safe from django.utils.translation import ugettext_lazy as _ from content_editor.admin import ContentEditorInline __all__ = ("HTML", "HTMLInline", "render_html") class HTML(models.Model): """ Raw HTML plugin """ html = models.TextField( "HTML", help_text=_( "The content will be inserted directly into the page." " It is VERY important that the HTML snippet is well-formed!" ), ) class Meta: abstract = True verbose_name = "HTML" verbose_name_plural = "HTML" def __str__(self): return "" class HTMLInline(ContentEditorInline): """ Just available for consistency, absolutely no difference to a standard ``ContentEditorInline``. """ formfield_overrides = { models.TextField: { "widget": forms.Textarea( attrs={"rows": 3, "cols": 40, "class": "vLargeTextField"} ) } } def render_html(plugin, **kwargs): """ Return the HTML code as safe string so that it is not escaped. Of course the contents are not guaranteed to be safe at all """ return mark_safe(plugin.html)
<commit_before>""" Plugin providing a simple textarea where raw HTML, CSS and JS code can be entered. Most useful for people wanting to shoot themselves in the foot. """ from django.db import models from django.utils.html import mark_safe from django.utils.translation import ugettext_lazy as _ from content_editor.admin import ContentEditorInline __all__ = ("HTML", "HTMLInline", "render_html") class HTML(models.Model): """ Raw HTML plugin """ html = models.TextField( "HTML", help_text=_( "The content will be inserted directly into the page." " It is VERY important that the HTML snippet is well-formed!" ), ) class Meta: abstract = True verbose_name = "HTML" verbose_name_plural = "HTML" def __str__(self): return "" class HTMLInline(ContentEditorInline): """ Just available for consistency, absolutely no difference to a standard ``ContentEditorInline``. """ pass def render_html(plugin, **kwargs): """ Return the HTML code as safe string so that it is not escaped. Of course the contents are not guaranteed to be safe at all """ return mark_safe(plugin.html) <commit_msg>Make the default HTML textarea smaller<commit_after>""" Plugin providing a simple textarea where raw HTML, CSS and JS code can be entered. Most useful for people wanting to shoot themselves in the foot. """ from django import forms from django.db import models from django.utils.html import mark_safe from django.utils.translation import ugettext_lazy as _ from content_editor.admin import ContentEditorInline __all__ = ("HTML", "HTMLInline", "render_html") class HTML(models.Model): """ Raw HTML plugin """ html = models.TextField( "HTML", help_text=_( "The content will be inserted directly into the page." " It is VERY important that the HTML snippet is well-formed!" ), ) class Meta: abstract = True verbose_name = "HTML" verbose_name_plural = "HTML" def __str__(self): return "" class HTMLInline(ContentEditorInline): """ Just available for consistency, absolutely no difference to a standard ``ContentEditorInline``. """ formfield_overrides = { models.TextField: { "widget": forms.Textarea( attrs={"rows": 3, "cols": 40, "class": "vLargeTextField"} ) } } def render_html(plugin, **kwargs): """ Return the HTML code as safe string so that it is not escaped. Of course the contents are not guaranteed to be safe at all """ return mark_safe(plugin.html)
720833e96e24ffe73822a3a1280e3dc901e52829
anchorhub/lib/filetolist.py
anchorhub/lib/filetolist.py
""" Class for FileToList """ class FileToList(object): """ FileToList is a helper class used to import text files and turn them into lists, with each index in the list representing a single line from the text file. """ @staticmethod def to_list(file_path): """ Static method. Takes in a file path, and outputs a list of stings. Each element in the list corresponds to a line in the file. :param file_path: string file path :return: A list of strings, with elements in the list corresponding to lines in the file pointed to in file_path """ l = [] f = open(file_path, 'rb') for line in f: l.append(line) f.close() return l
""" Class for FileToList """ class FileToList(object): """ FileToList is a helper class used to import text files and turn them into lists, with each index in the list representing a single line from the text file. """ @staticmethod def to_list(file_path): """ Static method. Takes in a file path, and outputs a list of stings. Each element in the list corresponds to a line in the file. :param file_path: string file path :return: A list of strings, with elements in the list corresponding to lines in the file pointed to in file_path """ l = [] f = open(file_path, 'r') for line in f: l.append(line) f.close() return l
Remove 'b' classifer on FileToList's read() usage
Remove 'b' classifer on FileToList's read() usage
Python
apache-2.0
samjabrahams/anchorhub
""" Class for FileToList """ class FileToList(object): """ FileToList is a helper class used to import text files and turn them into lists, with each index in the list representing a single line from the text file. """ @staticmethod def to_list(file_path): """ Static method. Takes in a file path, and outputs a list of stings. Each element in the list corresponds to a line in the file. :param file_path: string file path :return: A list of strings, with elements in the list corresponding to lines in the file pointed to in file_path """ l = [] f = open(file_path, 'rb') for line in f: l.append(line) f.close() return l Remove 'b' classifer on FileToList's read() usage
""" Class for FileToList """ class FileToList(object): """ FileToList is a helper class used to import text files and turn them into lists, with each index in the list representing a single line from the text file. """ @staticmethod def to_list(file_path): """ Static method. Takes in a file path, and outputs a list of stings. Each element in the list corresponds to a line in the file. :param file_path: string file path :return: A list of strings, with elements in the list corresponding to lines in the file pointed to in file_path """ l = [] f = open(file_path, 'r') for line in f: l.append(line) f.close() return l
<commit_before>""" Class for FileToList """ class FileToList(object): """ FileToList is a helper class used to import text files and turn them into lists, with each index in the list representing a single line from the text file. """ @staticmethod def to_list(file_path): """ Static method. Takes in a file path, and outputs a list of stings. Each element in the list corresponds to a line in the file. :param file_path: string file path :return: A list of strings, with elements in the list corresponding to lines in the file pointed to in file_path """ l = [] f = open(file_path, 'rb') for line in f: l.append(line) f.close() return l <commit_msg>Remove 'b' classifer on FileToList's read() usage<commit_after>
""" Class for FileToList """ class FileToList(object): """ FileToList is a helper class used to import text files and turn them into lists, with each index in the list representing a single line from the text file. """ @staticmethod def to_list(file_path): """ Static method. Takes in a file path, and outputs a list of stings. Each element in the list corresponds to a line in the file. :param file_path: string file path :return: A list of strings, with elements in the list corresponding to lines in the file pointed to in file_path """ l = [] f = open(file_path, 'r') for line in f: l.append(line) f.close() return l
""" Class for FileToList """ class FileToList(object): """ FileToList is a helper class used to import text files and turn them into lists, with each index in the list representing a single line from the text file. """ @staticmethod def to_list(file_path): """ Static method. Takes in a file path, and outputs a list of stings. Each element in the list corresponds to a line in the file. :param file_path: string file path :return: A list of strings, with elements in the list corresponding to lines in the file pointed to in file_path """ l = [] f = open(file_path, 'rb') for line in f: l.append(line) f.close() return l Remove 'b' classifer on FileToList's read() usage""" Class for FileToList """ class FileToList(object): """ FileToList is a helper class used to import text files and turn them into lists, with each index in the list representing a single line from the text file. """ @staticmethod def to_list(file_path): """ Static method. Takes in a file path, and outputs a list of stings. Each element in the list corresponds to a line in the file. :param file_path: string file path :return: A list of strings, with elements in the list corresponding to lines in the file pointed to in file_path """ l = [] f = open(file_path, 'r') for line in f: l.append(line) f.close() return l
<commit_before>""" Class for FileToList """ class FileToList(object): """ FileToList is a helper class used to import text files and turn them into lists, with each index in the list representing a single line from the text file. """ @staticmethod def to_list(file_path): """ Static method. Takes in a file path, and outputs a list of stings. Each element in the list corresponds to a line in the file. :param file_path: string file path :return: A list of strings, with elements in the list corresponding to lines in the file pointed to in file_path """ l = [] f = open(file_path, 'rb') for line in f: l.append(line) f.close() return l <commit_msg>Remove 'b' classifer on FileToList's read() usage<commit_after>""" Class for FileToList """ class FileToList(object): """ FileToList is a helper class used to import text files and turn them into lists, with each index in the list representing a single line from the text file. """ @staticmethod def to_list(file_path): """ Static method. Takes in a file path, and outputs a list of stings. Each element in the list corresponds to a line in the file. :param file_path: string file path :return: A list of strings, with elements in the list corresponding to lines in the file pointed to in file_path """ l = [] f = open(file_path, 'r') for line in f: l.append(line) f.close() return l
614f83d826c51a51ebb4feb01371a441473af423
featureflow/__init__.py
featureflow/__init__.py
__version__ = '1.16.14' from model import BaseModel from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \ PickleFeature from extractor import Node, Graph, Aggregator, NotEnoughData from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip from data import \ IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \ KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \ InMemoryDatabase from datawriter import DataWriter from database_iterator import DatabaseIterator from encoder import IdentityEncoder from decoder import Decoder, PickleDecoder from lmdbstore import LmdbDatabase from objectstore import ObjectStoreDatabase from persistence import PersistenceSettings from iteratornode import IteratorNode from eventlog import EventLog, RedisChannel try: from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \ BaseNumpyDecoder, NumpyMetaData, NumpyFeature except ImportError: pass
__version__ = '1.16.14' from model import BaseModel from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \ PickleFeature from extractor import Node, Graph, Aggregator, NotEnoughData from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip from data import \ IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \ KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \ InMemoryDatabase from datawriter import DataWriter from database_iterator import DatabaseIterator from encoder import IdentityEncoder, PickleEncoder from decoder import Decoder, PickleDecoder from lmdbstore import LmdbDatabase from objectstore import ObjectStoreDatabase from persistence import PersistenceSettings from iteratornode import IteratorNode from eventlog import EventLog, RedisChannel try: from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \ BaseNumpyDecoder, NumpyMetaData, NumpyFeature except ImportError: pass
Add PickleEncoder to the public API
Add PickleEncoder to the public API
Python
mit
JohnVinyard/featureflow,JohnVinyard/featureflow
__version__ = '1.16.14' from model import BaseModel from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \ PickleFeature from extractor import Node, Graph, Aggregator, NotEnoughData from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip from data import \ IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \ KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \ InMemoryDatabase from datawriter import DataWriter from database_iterator import DatabaseIterator from encoder import IdentityEncoder from decoder import Decoder, PickleDecoder from lmdbstore import LmdbDatabase from objectstore import ObjectStoreDatabase from persistence import PersistenceSettings from iteratornode import IteratorNode from eventlog import EventLog, RedisChannel try: from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \ BaseNumpyDecoder, NumpyMetaData, NumpyFeature except ImportError: pass Add PickleEncoder to the public API
__version__ = '1.16.14' from model import BaseModel from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \ PickleFeature from extractor import Node, Graph, Aggregator, NotEnoughData from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip from data import \ IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \ KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \ InMemoryDatabase from datawriter import DataWriter from database_iterator import DatabaseIterator from encoder import IdentityEncoder, PickleEncoder from decoder import Decoder, PickleDecoder from lmdbstore import LmdbDatabase from objectstore import ObjectStoreDatabase from persistence import PersistenceSettings from iteratornode import IteratorNode from eventlog import EventLog, RedisChannel try: from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \ BaseNumpyDecoder, NumpyMetaData, NumpyFeature except ImportError: pass
<commit_before>__version__ = '1.16.14' from model import BaseModel from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \ PickleFeature from extractor import Node, Graph, Aggregator, NotEnoughData from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip from data import \ IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \ KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \ InMemoryDatabase from datawriter import DataWriter from database_iterator import DatabaseIterator from encoder import IdentityEncoder from decoder import Decoder, PickleDecoder from lmdbstore import LmdbDatabase from objectstore import ObjectStoreDatabase from persistence import PersistenceSettings from iteratornode import IteratorNode from eventlog import EventLog, RedisChannel try: from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \ BaseNumpyDecoder, NumpyMetaData, NumpyFeature except ImportError: pass <commit_msg>Add PickleEncoder to the public API<commit_after>
__version__ = '1.16.14' from model import BaseModel from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \ PickleFeature from extractor import Node, Graph, Aggregator, NotEnoughData from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip from data import \ IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \ KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \ InMemoryDatabase from datawriter import DataWriter from database_iterator import DatabaseIterator from encoder import IdentityEncoder, PickleEncoder from decoder import Decoder, PickleDecoder from lmdbstore import LmdbDatabase from objectstore import ObjectStoreDatabase from persistence import PersistenceSettings from iteratornode import IteratorNode from eventlog import EventLog, RedisChannel try: from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \ BaseNumpyDecoder, NumpyMetaData, NumpyFeature except ImportError: pass
__version__ = '1.16.14' from model import BaseModel from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \ PickleFeature from extractor import Node, Graph, Aggregator, NotEnoughData from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip from data import \ IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \ KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \ InMemoryDatabase from datawriter import DataWriter from database_iterator import DatabaseIterator from encoder import IdentityEncoder from decoder import Decoder, PickleDecoder from lmdbstore import LmdbDatabase from objectstore import ObjectStoreDatabase from persistence import PersistenceSettings from iteratornode import IteratorNode from eventlog import EventLog, RedisChannel try: from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \ BaseNumpyDecoder, NumpyMetaData, NumpyFeature except ImportError: pass Add PickleEncoder to the public API__version__ = '1.16.14' from model import BaseModel from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \ PickleFeature from extractor import Node, Graph, Aggregator, NotEnoughData from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip from data import \ IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \ KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \ InMemoryDatabase from datawriter import DataWriter from database_iterator import DatabaseIterator from encoder import IdentityEncoder, PickleEncoder from decoder import Decoder, PickleDecoder from lmdbstore import LmdbDatabase from objectstore import ObjectStoreDatabase from persistence import PersistenceSettings from iteratornode import IteratorNode from eventlog import EventLog, RedisChannel try: from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \ BaseNumpyDecoder, NumpyMetaData, NumpyFeature except ImportError: pass
<commit_before>__version__ = '1.16.14' from model import BaseModel from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \ PickleFeature from extractor import Node, Graph, Aggregator, NotEnoughData from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip from data import \ IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \ KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \ InMemoryDatabase from datawriter import DataWriter from database_iterator import DatabaseIterator from encoder import IdentityEncoder from decoder import Decoder, PickleDecoder from lmdbstore import LmdbDatabase from objectstore import ObjectStoreDatabase from persistence import PersistenceSettings from iteratornode import IteratorNode from eventlog import EventLog, RedisChannel try: from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \ BaseNumpyDecoder, NumpyMetaData, NumpyFeature except ImportError: pass <commit_msg>Add PickleEncoder to the public API<commit_after>__version__ = '1.16.14' from model import BaseModel from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \ PickleFeature from extractor import Node, Graph, Aggregator, NotEnoughData from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip from data import \ IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \ KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \ InMemoryDatabase from datawriter import DataWriter from database_iterator import DatabaseIterator from encoder import IdentityEncoder, PickleEncoder from decoder import Decoder, PickleDecoder from lmdbstore import LmdbDatabase from objectstore import ObjectStoreDatabase from persistence import PersistenceSettings from iteratornode import IteratorNode from eventlog import EventLog, RedisChannel try: from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \ BaseNumpyDecoder, NumpyMetaData, NumpyFeature except ImportError: pass
d087e0cc47697e6b7f222de90a4143e3bb612a66
radar/models/forms.py
radar/models/forms.py
from sqlalchemy import Column, Integer, ForeignKey, String from sqlalchemy.orm import relationship from sqlalchemy.dialects import postgresql from radar.database import db from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship from radar.models.logs import log_changes class Form(db.Model): __tablename__ = 'forms' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) data = Column(postgresql.JSONB, nullable=False) @log_changes class Entry(db.Model, MetaModelMixin): __tablename__ = 'entries' id = uuid_pk_column() patient_id = patient_id_column() patient = patient_relationship('entries') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form') data = Column(postgresql.JSONB, nullable=False) class GroupForm(db.Model): __tablename__ = 'group_forms' id = Column(Integer, primary_key=True) group_id = Column(Integer, ForeignKey('groups.id'), nullable=False) group = relationship('Group') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form')
from sqlalchemy import Column, Integer, ForeignKey, String from sqlalchemy.orm import relationship from sqlalchemy.dialects import postgresql from radar.database import db from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship from radar.models.logs import log_changes class Form(db.Model): __tablename__ = 'forms' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) data = Column(postgresql.JSONB, nullable=False) @log_changes class Entry(db.Model, MetaModelMixin): __tablename__ = 'entries' id = uuid_pk_column() patient_id = patient_id_column() patient = patient_relationship('entries') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form') data = Column(postgresql.JSONB, nullable=False) Index('entries_patient_idx', Entry.patient_id) class GroupForm(db.Model): __tablename__ = 'group_forms' id = Column(Integer, primary_key=True) group_id = Column(Integer, ForeignKey('groups.id'), nullable=False) group = relationship('Group') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form')
Add index on patient id
Add index on patient id
Python
agpl-3.0
renalreg/radar,renalreg/radar,renalreg/radar,renalreg/radar
from sqlalchemy import Column, Integer, ForeignKey, String from sqlalchemy.orm import relationship from sqlalchemy.dialects import postgresql from radar.database import db from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship from radar.models.logs import log_changes class Form(db.Model): __tablename__ = 'forms' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) data = Column(postgresql.JSONB, nullable=False) @log_changes class Entry(db.Model, MetaModelMixin): __tablename__ = 'entries' id = uuid_pk_column() patient_id = patient_id_column() patient = patient_relationship('entries') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form') data = Column(postgresql.JSONB, nullable=False) class GroupForm(db.Model): __tablename__ = 'group_forms' id = Column(Integer, primary_key=True) group_id = Column(Integer, ForeignKey('groups.id'), nullable=False) group = relationship('Group') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form')Add index on patient id
from sqlalchemy import Column, Integer, ForeignKey, String from sqlalchemy.orm import relationship from sqlalchemy.dialects import postgresql from radar.database import db from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship from radar.models.logs import log_changes class Form(db.Model): __tablename__ = 'forms' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) data = Column(postgresql.JSONB, nullable=False) @log_changes class Entry(db.Model, MetaModelMixin): __tablename__ = 'entries' id = uuid_pk_column() patient_id = patient_id_column() patient = patient_relationship('entries') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form') data = Column(postgresql.JSONB, nullable=False) Index('entries_patient_idx', Entry.patient_id) class GroupForm(db.Model): __tablename__ = 'group_forms' id = Column(Integer, primary_key=True) group_id = Column(Integer, ForeignKey('groups.id'), nullable=False) group = relationship('Group') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form')
<commit_before>from sqlalchemy import Column, Integer, ForeignKey, String from sqlalchemy.orm import relationship from sqlalchemy.dialects import postgresql from radar.database import db from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship from radar.models.logs import log_changes class Form(db.Model): __tablename__ = 'forms' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) data = Column(postgresql.JSONB, nullable=False) @log_changes class Entry(db.Model, MetaModelMixin): __tablename__ = 'entries' id = uuid_pk_column() patient_id = patient_id_column() patient = patient_relationship('entries') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form') data = Column(postgresql.JSONB, nullable=False) class GroupForm(db.Model): __tablename__ = 'group_forms' id = Column(Integer, primary_key=True) group_id = Column(Integer, ForeignKey('groups.id'), nullable=False) group = relationship('Group') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form')<commit_msg>Add index on patient id<commit_after>
from sqlalchemy import Column, Integer, ForeignKey, String from sqlalchemy.orm import relationship from sqlalchemy.dialects import postgresql from radar.database import db from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship from radar.models.logs import log_changes class Form(db.Model): __tablename__ = 'forms' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) data = Column(postgresql.JSONB, nullable=False) @log_changes class Entry(db.Model, MetaModelMixin): __tablename__ = 'entries' id = uuid_pk_column() patient_id = patient_id_column() patient = patient_relationship('entries') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form') data = Column(postgresql.JSONB, nullable=False) Index('entries_patient_idx', Entry.patient_id) class GroupForm(db.Model): __tablename__ = 'group_forms' id = Column(Integer, primary_key=True) group_id = Column(Integer, ForeignKey('groups.id'), nullable=False) group = relationship('Group') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form')
from sqlalchemy import Column, Integer, ForeignKey, String from sqlalchemy.orm import relationship from sqlalchemy.dialects import postgresql from radar.database import db from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship from radar.models.logs import log_changes class Form(db.Model): __tablename__ = 'forms' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) data = Column(postgresql.JSONB, nullable=False) @log_changes class Entry(db.Model, MetaModelMixin): __tablename__ = 'entries' id = uuid_pk_column() patient_id = patient_id_column() patient = patient_relationship('entries') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form') data = Column(postgresql.JSONB, nullable=False) class GroupForm(db.Model): __tablename__ = 'group_forms' id = Column(Integer, primary_key=True) group_id = Column(Integer, ForeignKey('groups.id'), nullable=False) group = relationship('Group') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form')Add index on patient idfrom sqlalchemy import Column, Integer, ForeignKey, String from sqlalchemy.orm import relationship from sqlalchemy.dialects import postgresql from radar.database import db from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship from radar.models.logs import log_changes class Form(db.Model): __tablename__ = 'forms' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) data = Column(postgresql.JSONB, nullable=False) @log_changes class Entry(db.Model, MetaModelMixin): __tablename__ = 'entries' id = uuid_pk_column() patient_id = patient_id_column() patient = patient_relationship('entries') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form') data = Column(postgresql.JSONB, nullable=False) Index('entries_patient_idx', Entry.patient_id) class GroupForm(db.Model): __tablename__ = 'group_forms' id = Column(Integer, primary_key=True) group_id = Column(Integer, ForeignKey('groups.id'), nullable=False) group = relationship('Group') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form')
<commit_before>from sqlalchemy import Column, Integer, ForeignKey, String from sqlalchemy.orm import relationship from sqlalchemy.dialects import postgresql from radar.database import db from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship from radar.models.logs import log_changes class Form(db.Model): __tablename__ = 'forms' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) data = Column(postgresql.JSONB, nullable=False) @log_changes class Entry(db.Model, MetaModelMixin): __tablename__ = 'entries' id = uuid_pk_column() patient_id = patient_id_column() patient = patient_relationship('entries') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form') data = Column(postgresql.JSONB, nullable=False) class GroupForm(db.Model): __tablename__ = 'group_forms' id = Column(Integer, primary_key=True) group_id = Column(Integer, ForeignKey('groups.id'), nullable=False) group = relationship('Group') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form')<commit_msg>Add index on patient id<commit_after>from sqlalchemy import Column, Integer, ForeignKey, String from sqlalchemy.orm import relationship from sqlalchemy.dialects import postgresql from radar.database import db from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship from radar.models.logs import log_changes class Form(db.Model): __tablename__ = 'forms' id = Column(Integer, primary_key=True) name = Column(String, nullable=False) data = Column(postgresql.JSONB, nullable=False) @log_changes class Entry(db.Model, MetaModelMixin): __tablename__ = 'entries' id = uuid_pk_column() patient_id = patient_id_column() patient = patient_relationship('entries') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form') data = Column(postgresql.JSONB, nullable=False) Index('entries_patient_idx', Entry.patient_id) class GroupForm(db.Model): __tablename__ = 'group_forms' id = Column(Integer, primary_key=True) group_id = Column(Integer, ForeignKey('groups.id'), nullable=False) group = relationship('Group') form_id = Column(Integer, ForeignKey('forms.id'), nullable=False) form = relationship('Form')
e00fb0d87b60a982c2d932864a67a70e7d5b4312
src/apps/rDSN.monitor/rDSN.Monitor.py
src/apps/rDSN.monitor/rDSN.Monitor.py
import sys import os import threading import time sys.path.append(os.getcwd() + '/app_package') from MonitorApp import * def start_dsn(): service_app = ServiceApp() app_dict['monitor'] = MonitorService service_app.register_app('monitor') if len(sys.argv) < 2: #rDSN.Monitor run as an embedded service print "rDSN.Monitor runs in embedded mode" Native.dsn_app_loader_signal() time.sleep(1) elif sys.argv[1] == 'standalone': #rDSN.Monitor run as a caller calling the monitored program print "rDSN.Monitor runs in standalone mode" argv = (c_char_p*2)() argv[0] = b'rDSN.Monitor.exe' argv[1] = b'config.ini' Native.dsn_run(2, argv, c_bool(1)) if __name__ == '__main__': start_dsn()
import sys import os import threading sys.path.append(os.getcwd() + '/app_package') from MonitorApp import * def start_dsn(): service_app = ServiceApp() app_dict['monitor'] = MonitorService service_app.register_app('monitor') if len(sys.argv) < 2: #rDSN.Monitor run as an embedded service print "rDSN.Monitor runs in embedded mode" Native.dsn_app_loader_signal() #to be fix, hangs forever now to keep python interpreter alive dummy_event = threading.Event() dummy_event.wait() elif sys.argv[1] == 'standalone': #rDSN.Monitor run as a caller calling the monitored program print "rDSN.Monitor runs in standalone mode" argv = (c_char_p*2)() argv[0] = b'rDSN.Monitor.exe' argv[1] = b'config.ini' Native.dsn_run(2, argv, c_bool(1)) if __name__ == '__main__': start_dsn()
Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app starts
Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app starts
Python
mit
mcfatealan/rDSN.Python,rDSN-Projects/rDSN.Python,mcfatealan/rDSN.Python,mcfatealan/rDSN.Python,rDSN-Projects/rDSN.Python,mcfatealan/rDSN.Python,rDSN-Projects/rDSN.Python,rDSN-Projects/rDSN.Python,rDSN-Projects/rDSN.Python,mcfatealan/rDSN.Python
import sys import os import threading import time sys.path.append(os.getcwd() + '/app_package') from MonitorApp import * def start_dsn(): service_app = ServiceApp() app_dict['monitor'] = MonitorService service_app.register_app('monitor') if len(sys.argv) < 2: #rDSN.Monitor run as an embedded service print "rDSN.Monitor runs in embedded mode" Native.dsn_app_loader_signal() time.sleep(1) elif sys.argv[1] == 'standalone': #rDSN.Monitor run as a caller calling the monitored program print "rDSN.Monitor runs in standalone mode" argv = (c_char_p*2)() argv[0] = b'rDSN.Monitor.exe' argv[1] = b'config.ini' Native.dsn_run(2, argv, c_bool(1)) if __name__ == '__main__': start_dsn() Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app starts
import sys import os import threading sys.path.append(os.getcwd() + '/app_package') from MonitorApp import * def start_dsn(): service_app = ServiceApp() app_dict['monitor'] = MonitorService service_app.register_app('monitor') if len(sys.argv) < 2: #rDSN.Monitor run as an embedded service print "rDSN.Monitor runs in embedded mode" Native.dsn_app_loader_signal() #to be fix, hangs forever now to keep python interpreter alive dummy_event = threading.Event() dummy_event.wait() elif sys.argv[1] == 'standalone': #rDSN.Monitor run as a caller calling the monitored program print "rDSN.Monitor runs in standalone mode" argv = (c_char_p*2)() argv[0] = b'rDSN.Monitor.exe' argv[1] = b'config.ini' Native.dsn_run(2, argv, c_bool(1)) if __name__ == '__main__': start_dsn()
<commit_before>import sys import os import threading import time sys.path.append(os.getcwd() + '/app_package') from MonitorApp import * def start_dsn(): service_app = ServiceApp() app_dict['monitor'] = MonitorService service_app.register_app('monitor') if len(sys.argv) < 2: #rDSN.Monitor run as an embedded service print "rDSN.Monitor runs in embedded mode" Native.dsn_app_loader_signal() time.sleep(1) elif sys.argv[1] == 'standalone': #rDSN.Monitor run as a caller calling the monitored program print "rDSN.Monitor runs in standalone mode" argv = (c_char_p*2)() argv[0] = b'rDSN.Monitor.exe' argv[1] = b'config.ini' Native.dsn_run(2, argv, c_bool(1)) if __name__ == '__main__': start_dsn() <commit_msg>Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app starts<commit_after>
import sys import os import threading sys.path.append(os.getcwd() + '/app_package') from MonitorApp import * def start_dsn(): service_app = ServiceApp() app_dict['monitor'] = MonitorService service_app.register_app('monitor') if len(sys.argv) < 2: #rDSN.Monitor run as an embedded service print "rDSN.Monitor runs in embedded mode" Native.dsn_app_loader_signal() #to be fix, hangs forever now to keep python interpreter alive dummy_event = threading.Event() dummy_event.wait() elif sys.argv[1] == 'standalone': #rDSN.Monitor run as a caller calling the monitored program print "rDSN.Monitor runs in standalone mode" argv = (c_char_p*2)() argv[0] = b'rDSN.Monitor.exe' argv[1] = b'config.ini' Native.dsn_run(2, argv, c_bool(1)) if __name__ == '__main__': start_dsn()
import sys import os import threading import time sys.path.append(os.getcwd() + '/app_package') from MonitorApp import * def start_dsn(): service_app = ServiceApp() app_dict['monitor'] = MonitorService service_app.register_app('monitor') if len(sys.argv) < 2: #rDSN.Monitor run as an embedded service print "rDSN.Monitor runs in embedded mode" Native.dsn_app_loader_signal() time.sleep(1) elif sys.argv[1] == 'standalone': #rDSN.Monitor run as a caller calling the monitored program print "rDSN.Monitor runs in standalone mode" argv = (c_char_p*2)() argv[0] = b'rDSN.Monitor.exe' argv[1] = b'config.ini' Native.dsn_run(2, argv, c_bool(1)) if __name__ == '__main__': start_dsn() Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app startsimport sys import os import threading sys.path.append(os.getcwd() + '/app_package') from MonitorApp import * def start_dsn(): service_app = ServiceApp() app_dict['monitor'] = MonitorService service_app.register_app('monitor') if len(sys.argv) < 2: #rDSN.Monitor run as an embedded service print "rDSN.Monitor runs in embedded mode" Native.dsn_app_loader_signal() #to be fix, hangs forever now to keep python interpreter alive dummy_event = threading.Event() dummy_event.wait() elif sys.argv[1] == 'standalone': #rDSN.Monitor run as a caller calling the monitored program print "rDSN.Monitor runs in standalone mode" argv = (c_char_p*2)() argv[0] = b'rDSN.Monitor.exe' argv[1] = b'config.ini' Native.dsn_run(2, argv, c_bool(1)) if __name__ == '__main__': start_dsn()
<commit_before>import sys import os import threading import time sys.path.append(os.getcwd() + '/app_package') from MonitorApp import * def start_dsn(): service_app = ServiceApp() app_dict['monitor'] = MonitorService service_app.register_app('monitor') if len(sys.argv) < 2: #rDSN.Monitor run as an embedded service print "rDSN.Monitor runs in embedded mode" Native.dsn_app_loader_signal() time.sleep(1) elif sys.argv[1] == 'standalone': #rDSN.Monitor run as a caller calling the monitored program print "rDSN.Monitor runs in standalone mode" argv = (c_char_p*2)() argv[0] = b'rDSN.Monitor.exe' argv[1] = b'config.ini' Native.dsn_run(2, argv, c_bool(1)) if __name__ == '__main__': start_dsn() <commit_msg>Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app starts<commit_after>import sys import os import threading sys.path.append(os.getcwd() + '/app_package') from MonitorApp import * def start_dsn(): service_app = ServiceApp() app_dict['monitor'] = MonitorService service_app.register_app('monitor') if len(sys.argv) < 2: #rDSN.Monitor run as an embedded service print "rDSN.Monitor runs in embedded mode" Native.dsn_app_loader_signal() #to be fix, hangs forever now to keep python interpreter alive dummy_event = threading.Event() dummy_event.wait() elif sys.argv[1] == 'standalone': #rDSN.Monitor run as a caller calling the monitored program print "rDSN.Monitor runs in standalone mode" argv = (c_char_p*2)() argv[0] = b'rDSN.Monitor.exe' argv[1] = b'config.ini' Native.dsn_run(2, argv, c_bool(1)) if __name__ == '__main__': start_dsn()
65daee8f169e8bb6e721ce016c7bcf6cb9893016
froide/problem/utils.py
froide/problem/utils.py
from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n{}'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
Add problem admin URL to problem report email
Add problem admin URL to problem report email
Python
mit
stefanw/froide,stefanw/froide,stefanw/froide,fin/froide,fin/froide,stefanw/froide,fin/froide,stefanw/froide,fin/froide
from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True Add problem admin URL to problem report email
from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n{}'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
<commit_before>from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True <commit_msg>Add problem admin URL to problem report email<commit_after>
from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n{}'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True Add problem admin URL to problem report emailfrom django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n{}'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
<commit_before>from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True <commit_msg>Add problem admin URL to problem report email<commit_after>from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n{}'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
6a4e16f9afa373233c03cc8f1ede7076e9a44058
basics/utils.py
basics/utils.py
import numpy as np def arctan_transform(array, thresh): return np.arctan(array/thresh)
import numpy as np from functools import partial def arctan_transform(array, thresh): return np.arctan(array/thresh) def dist_uppertri(cond_arr, shape): dist_arr = np.zeros((shape, ) * 2, dtype=cond_arr.dtype) def unrav_ind(i, j, n): return n*j - j*(j+1)/2 + i - 1 - j arr_ind = partial(unrav_ind, n=shape) for i in xrange(shape): for j in xrange(i): dist_arr[i, j] = cond_arr[arr_ind(i, j)] return dist_arr
Convert a condensed distance matrix (pdist) into an upper triangular matrix
Convert a condensed distance matrix (pdist) into an upper triangular matrix
Python
mit
e-koch/BaSiCs
import numpy as np def arctan_transform(array, thresh): return np.arctan(array/thresh) Convert a condensed distance matrix (pdist) into an upper triangular matrix
import numpy as np from functools import partial def arctan_transform(array, thresh): return np.arctan(array/thresh) def dist_uppertri(cond_arr, shape): dist_arr = np.zeros((shape, ) * 2, dtype=cond_arr.dtype) def unrav_ind(i, j, n): return n*j - j*(j+1)/2 + i - 1 - j arr_ind = partial(unrav_ind, n=shape) for i in xrange(shape): for j in xrange(i): dist_arr[i, j] = cond_arr[arr_ind(i, j)] return dist_arr
<commit_before> import numpy as np def arctan_transform(array, thresh): return np.arctan(array/thresh) <commit_msg>Convert a condensed distance matrix (pdist) into an upper triangular matrix<commit_after>
import numpy as np from functools import partial def arctan_transform(array, thresh): return np.arctan(array/thresh) def dist_uppertri(cond_arr, shape): dist_arr = np.zeros((shape, ) * 2, dtype=cond_arr.dtype) def unrav_ind(i, j, n): return n*j - j*(j+1)/2 + i - 1 - j arr_ind = partial(unrav_ind, n=shape) for i in xrange(shape): for j in xrange(i): dist_arr[i, j] = cond_arr[arr_ind(i, j)] return dist_arr
import numpy as np def arctan_transform(array, thresh): return np.arctan(array/thresh) Convert a condensed distance matrix (pdist) into an upper triangular matrix import numpy as np from functools import partial def arctan_transform(array, thresh): return np.arctan(array/thresh) def dist_uppertri(cond_arr, shape): dist_arr = np.zeros((shape, ) * 2, dtype=cond_arr.dtype) def unrav_ind(i, j, n): return n*j - j*(j+1)/2 + i - 1 - j arr_ind = partial(unrav_ind, n=shape) for i in xrange(shape): for j in xrange(i): dist_arr[i, j] = cond_arr[arr_ind(i, j)] return dist_arr
<commit_before> import numpy as np def arctan_transform(array, thresh): return np.arctan(array/thresh) <commit_msg>Convert a condensed distance matrix (pdist) into an upper triangular matrix<commit_after> import numpy as np from functools import partial def arctan_transform(array, thresh): return np.arctan(array/thresh) def dist_uppertri(cond_arr, shape): dist_arr = np.zeros((shape, ) * 2, dtype=cond_arr.dtype) def unrav_ind(i, j, n): return n*j - j*(j+1)/2 + i - 1 - j arr_ind = partial(unrav_ind, n=shape) for i in xrange(shape): for j in xrange(i): dist_arr[i, j] = cond_arr[arr_ind(i, j)] return dist_arr
e97dee6ec7c49cf3d33803504c7269a41c4d0a0f
authentication_app/views.py
authentication_app/views.py
from django.shortcuts import render from django.http import HttpResponse from .models import Greeting # Create your views here. def index(request): return HttpResponse('Hello from Python!') def db(request): greeting = Greeting() greeting.save() greetings = Greeting.objects.all() return render(request, 'db.html', {'greetings': greetings})
from rest_framework import permissions, viewsets from authentication_app.models import Account from authentication_app.permissions import IsAccountOwner from authentication_app.serializers import AccountSerializer ''' @name : AccountViewSerializer @desc : Defines the serializer for the account view. ''' class AccountViewSerializer(viewsets.ModelViewSet): lookup_field = 'username' queryset = Account.objects.all() serializer_class = AccountSerializer def get_permissions(self): if self.reqiest.method in permissions.SAFE_METHODS: return (permissions.AllowAny(),) if self.request.method == 'POST': return (permissions.AllowAny(),) return (permissions.IsAuthenticated(), IsAccountOwner(),) def create(self, request): serializer = self.serializer_class(data=reqiest.data) if serializer.is_valid(): Account.objects.create_user(**serializer.validated_data) return Response(serializer.validated_data, status=status.HTTP_201_CREATED) return Response({ 'status' : 'Bad Request', 'message' : 'Account could not be created with the received data.' }, status=status.HTTP_400_BAD_REQUEST)
Add the view serializer for the account model.
Add the view serializer for the account model.
Python
mit
mvpgomes/shopit-app,mvpgomes/shopit-app,mvpgomes/shopit-app,mvpgomes/shopit-app
from django.shortcuts import render from django.http import HttpResponse from .models import Greeting # Create your views here. def index(request): return HttpResponse('Hello from Python!') def db(request): greeting = Greeting() greeting.save() greetings = Greeting.objects.all() return render(request, 'db.html', {'greetings': greetings}) Add the view serializer for the account model.
from rest_framework import permissions, viewsets from authentication_app.models import Account from authentication_app.permissions import IsAccountOwner from authentication_app.serializers import AccountSerializer ''' @name : AccountViewSerializer @desc : Defines the serializer for the account view. ''' class AccountViewSerializer(viewsets.ModelViewSet): lookup_field = 'username' queryset = Account.objects.all() serializer_class = AccountSerializer def get_permissions(self): if self.reqiest.method in permissions.SAFE_METHODS: return (permissions.AllowAny(),) if self.request.method == 'POST': return (permissions.AllowAny(),) return (permissions.IsAuthenticated(), IsAccountOwner(),) def create(self, request): serializer = self.serializer_class(data=reqiest.data) if serializer.is_valid(): Account.objects.create_user(**serializer.validated_data) return Response(serializer.validated_data, status=status.HTTP_201_CREATED) return Response({ 'status' : 'Bad Request', 'message' : 'Account could not be created with the received data.' }, status=status.HTTP_400_BAD_REQUEST)
<commit_before>from django.shortcuts import render from django.http import HttpResponse from .models import Greeting # Create your views here. def index(request): return HttpResponse('Hello from Python!') def db(request): greeting = Greeting() greeting.save() greetings = Greeting.objects.all() return render(request, 'db.html', {'greetings': greetings}) <commit_msg>Add the view serializer for the account model.<commit_after>
from rest_framework import permissions, viewsets from authentication_app.models import Account from authentication_app.permissions import IsAccountOwner from authentication_app.serializers import AccountSerializer ''' @name : AccountViewSerializer @desc : Defines the serializer for the account view. ''' class AccountViewSerializer(viewsets.ModelViewSet): lookup_field = 'username' queryset = Account.objects.all() serializer_class = AccountSerializer def get_permissions(self): if self.reqiest.method in permissions.SAFE_METHODS: return (permissions.AllowAny(),) if self.request.method == 'POST': return (permissions.AllowAny(),) return (permissions.IsAuthenticated(), IsAccountOwner(),) def create(self, request): serializer = self.serializer_class(data=reqiest.data) if serializer.is_valid(): Account.objects.create_user(**serializer.validated_data) return Response(serializer.validated_data, status=status.HTTP_201_CREATED) return Response({ 'status' : 'Bad Request', 'message' : 'Account could not be created with the received data.' }, status=status.HTTP_400_BAD_REQUEST)
from django.shortcuts import render from django.http import HttpResponse from .models import Greeting # Create your views here. def index(request): return HttpResponse('Hello from Python!') def db(request): greeting = Greeting() greeting.save() greetings = Greeting.objects.all() return render(request, 'db.html', {'greetings': greetings}) Add the view serializer for the account model.from rest_framework import permissions, viewsets from authentication_app.models import Account from authentication_app.permissions import IsAccountOwner from authentication_app.serializers import AccountSerializer ''' @name : AccountViewSerializer @desc : Defines the serializer for the account view. ''' class AccountViewSerializer(viewsets.ModelViewSet): lookup_field = 'username' queryset = Account.objects.all() serializer_class = AccountSerializer def get_permissions(self): if self.reqiest.method in permissions.SAFE_METHODS: return (permissions.AllowAny(),) if self.request.method == 'POST': return (permissions.AllowAny(),) return (permissions.IsAuthenticated(), IsAccountOwner(),) def create(self, request): serializer = self.serializer_class(data=reqiest.data) if serializer.is_valid(): Account.objects.create_user(**serializer.validated_data) return Response(serializer.validated_data, status=status.HTTP_201_CREATED) return Response({ 'status' : 'Bad Request', 'message' : 'Account could not be created with the received data.' }, status=status.HTTP_400_BAD_REQUEST)
<commit_before>from django.shortcuts import render from django.http import HttpResponse from .models import Greeting # Create your views here. def index(request): return HttpResponse('Hello from Python!') def db(request): greeting = Greeting() greeting.save() greetings = Greeting.objects.all() return render(request, 'db.html', {'greetings': greetings}) <commit_msg>Add the view serializer for the account model.<commit_after>from rest_framework import permissions, viewsets from authentication_app.models import Account from authentication_app.permissions import IsAccountOwner from authentication_app.serializers import AccountSerializer ''' @name : AccountViewSerializer @desc : Defines the serializer for the account view. ''' class AccountViewSerializer(viewsets.ModelViewSet): lookup_field = 'username' queryset = Account.objects.all() serializer_class = AccountSerializer def get_permissions(self): if self.reqiest.method in permissions.SAFE_METHODS: return (permissions.AllowAny(),) if self.request.method == 'POST': return (permissions.AllowAny(),) return (permissions.IsAuthenticated(), IsAccountOwner(),) def create(self, request): serializer = self.serializer_class(data=reqiest.data) if serializer.is_valid(): Account.objects.create_user(**serializer.validated_data) return Response(serializer.validated_data, status=status.HTTP_201_CREATED) return Response({ 'status' : 'Bad Request', 'message' : 'Account could not be created with the received data.' }, status=status.HTTP_400_BAD_REQUEST)
f8304bb26151fdb999a77da9afbea8ff653a37f8
artists/views.py
artists/views.py
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): # TODO re-update old cumulative similarity if artist name changed cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
Add note to update old similarities
Add note to update old similarities
Python
bsd-3-clause
FreeMusicNinja/api.freemusic.ninja
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity]) Add note to update old similarities
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): # TODO re-update old cumulative similarity if artist name changed cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
<commit_before>from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity]) <commit_msg>Add note to update old similarities<commit_after>
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): # TODO re-update old cumulative similarity if artist name changed cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity]) Add note to update old similaritiesfrom django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): # TODO re-update old cumulative similarity if artist name changed cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
<commit_before>from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity]) <commit_msg>Add note to update old similarities<commit_after>from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): # TODO re-update old cumulative similarity if artist name changed cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
261fb861015ee96771e4c387bcd2b2c7d5c369db
hellopython/__init__.py
hellopython/__init__.py
__version__ = '1.0.0' from story.story import BaseStory from . import print_method class Story(BaseStory): name = 'hellopython' adventures = [ print_method ]
__version__ = '1.0.0' from story.story import BaseStory from . import print_method class Story(BaseStory): name = 'hellopython' title = 'Introuction to python' adventures = [ print_method ]
Add a title to the story the story
Add a title to the story the story
Python
mit
pyschool/hipyschool
__version__ = '1.0.0' from story.story import BaseStory from . import print_method class Story(BaseStory): name = 'hellopython' adventures = [ print_method ] Add a title to the story the story
__version__ = '1.0.0' from story.story import BaseStory from . import print_method class Story(BaseStory): name = 'hellopython' title = 'Introuction to python' adventures = [ print_method ]
<commit_before>__version__ = '1.0.0' from story.story import BaseStory from . import print_method class Story(BaseStory): name = 'hellopython' adventures = [ print_method ] <commit_msg>Add a title to the story the story<commit_after>
__version__ = '1.0.0' from story.story import BaseStory from . import print_method class Story(BaseStory): name = 'hellopython' title = 'Introuction to python' adventures = [ print_method ]
__version__ = '1.0.0' from story.story import BaseStory from . import print_method class Story(BaseStory): name = 'hellopython' adventures = [ print_method ] Add a title to the story the story__version__ = '1.0.0' from story.story import BaseStory from . import print_method class Story(BaseStory): name = 'hellopython' title = 'Introuction to python' adventures = [ print_method ]
<commit_before>__version__ = '1.0.0' from story.story import BaseStory from . import print_method class Story(BaseStory): name = 'hellopython' adventures = [ print_method ] <commit_msg>Add a title to the story the story<commit_after>__version__ = '1.0.0' from story.story import BaseStory from . import print_method class Story(BaseStory): name = 'hellopython' title = 'Introuction to python' adventures = [ print_method ]
8d11c6854e9c2309abb74a2e4b960a5206a27a0c
funbox/iterators_ordered.py
funbox/iterators_ordered.py
#! /usr/bin/env python """Functions on iterators, optimised for case when iterators are sorted. Note sift_o is hidden as _sift_o at the moment because it is broken. Please don't use it. Once fixed, I'll remove the leading underscore again. """ import itertools import iterators def partition_o(left_function, items): """Return a pair of iterators: left and right Items for which left_function returns a true value go into left. Items for which left_function returns a false value go into right. Items must be sorted such that left_function may be true for an initial set of items, but once an item is found such that left_function(item) is false, it will remain false for the rest of the items. In other words the following must hold: for all N where 0 <= N < (len(items) - 1) : not(func(item[n])) => not(func(item[n+1])) For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but not for [1,3,4,2,5]. >>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5]) >>> list(left) [-1, 0, 2] >>> list(right) [3, 4, 5] """ left = itertools.takewhile(left_function, items) right = itertools.dropwhile(left_function, items) return left, right if __name__ == "__main__": import doctest doctest.testmod()
#! /usr/bin/env python """Functions on iterators, optimised for case when iterators are sorted. """ import itertools import iterators def partition_o(left_function, items): """Return a pair of iterators: left and right Items for which left_function returns a true value go into left. Items for which left_function returns a false value go into right. Items must be sorted such that left_function may be true for an initial set of items, but once an item is found such that left_function(item) is false, it will remain false for the rest of the items. In other words the following must hold: for all N where 0 <= N < (len(items) - 1) : not(func(item[n])) => not(func(item[n+1])) For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but not for [1,3,4,2,5]. >>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5]) >>> list(left) [-1, 0, 2] >>> list(right) [3, 4, 5] """ left = itertools.takewhile(left_function, items) right = itertools.dropwhile(left_function, items) return left, right if __name__ == "__main__": import doctest doctest.testmod()
Remove reference in docs to removed function.
Remove reference in docs to removed function.
Python
mit
nmbooker/python-funbox,nmbooker/python-funbox
#! /usr/bin/env python """Functions on iterators, optimised for case when iterators are sorted. Note sift_o is hidden as _sift_o at the moment because it is broken. Please don't use it. Once fixed, I'll remove the leading underscore again. """ import itertools import iterators def partition_o(left_function, items): """Return a pair of iterators: left and right Items for which left_function returns a true value go into left. Items for which left_function returns a false value go into right. Items must be sorted such that left_function may be true for an initial set of items, but once an item is found such that left_function(item) is false, it will remain false for the rest of the items. In other words the following must hold: for all N where 0 <= N < (len(items) - 1) : not(func(item[n])) => not(func(item[n+1])) For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but not for [1,3,4,2,5]. >>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5]) >>> list(left) [-1, 0, 2] >>> list(right) [3, 4, 5] """ left = itertools.takewhile(left_function, items) right = itertools.dropwhile(left_function, items) return left, right if __name__ == "__main__": import doctest doctest.testmod() Remove reference in docs to removed function.
#! /usr/bin/env python """Functions on iterators, optimised for case when iterators are sorted. """ import itertools import iterators def partition_o(left_function, items): """Return a pair of iterators: left and right Items for which left_function returns a true value go into left. Items for which left_function returns a false value go into right. Items must be sorted such that left_function may be true for an initial set of items, but once an item is found such that left_function(item) is false, it will remain false for the rest of the items. In other words the following must hold: for all N where 0 <= N < (len(items) - 1) : not(func(item[n])) => not(func(item[n+1])) For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but not for [1,3,4,2,5]. >>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5]) >>> list(left) [-1, 0, 2] >>> list(right) [3, 4, 5] """ left = itertools.takewhile(left_function, items) right = itertools.dropwhile(left_function, items) return left, right if __name__ == "__main__": import doctest doctest.testmod()
<commit_before>#! /usr/bin/env python """Functions on iterators, optimised for case when iterators are sorted. Note sift_o is hidden as _sift_o at the moment because it is broken. Please don't use it. Once fixed, I'll remove the leading underscore again. """ import itertools import iterators def partition_o(left_function, items): """Return a pair of iterators: left and right Items for which left_function returns a true value go into left. Items for which left_function returns a false value go into right. Items must be sorted such that left_function may be true for an initial set of items, but once an item is found such that left_function(item) is false, it will remain false for the rest of the items. In other words the following must hold: for all N where 0 <= N < (len(items) - 1) : not(func(item[n])) => not(func(item[n+1])) For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but not for [1,3,4,2,5]. >>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5]) >>> list(left) [-1, 0, 2] >>> list(right) [3, 4, 5] """ left = itertools.takewhile(left_function, items) right = itertools.dropwhile(left_function, items) return left, right if __name__ == "__main__": import doctest doctest.testmod() <commit_msg>Remove reference in docs to removed function.<commit_after>
#! /usr/bin/env python """Functions on iterators, optimised for case when iterators are sorted. """ import itertools import iterators def partition_o(left_function, items): """Return a pair of iterators: left and right Items for which left_function returns a true value go into left. Items for which left_function returns a false value go into right. Items must be sorted such that left_function may be true for an initial set of items, but once an item is found such that left_function(item) is false, it will remain false for the rest of the items. In other words the following must hold: for all N where 0 <= N < (len(items) - 1) : not(func(item[n])) => not(func(item[n+1])) For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but not for [1,3,4,2,5]. >>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5]) >>> list(left) [-1, 0, 2] >>> list(right) [3, 4, 5] """ left = itertools.takewhile(left_function, items) right = itertools.dropwhile(left_function, items) return left, right if __name__ == "__main__": import doctest doctest.testmod()
#! /usr/bin/env python """Functions on iterators, optimised for case when iterators are sorted. Note sift_o is hidden as _sift_o at the moment because it is broken. Please don't use it. Once fixed, I'll remove the leading underscore again. """ import itertools import iterators def partition_o(left_function, items): """Return a pair of iterators: left and right Items for which left_function returns a true value go into left. Items for which left_function returns a false value go into right. Items must be sorted such that left_function may be true for an initial set of items, but once an item is found such that left_function(item) is false, it will remain false for the rest of the items. In other words the following must hold: for all N where 0 <= N < (len(items) - 1) : not(func(item[n])) => not(func(item[n+1])) For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but not for [1,3,4,2,5]. >>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5]) >>> list(left) [-1, 0, 2] >>> list(right) [3, 4, 5] """ left = itertools.takewhile(left_function, items) right = itertools.dropwhile(left_function, items) return left, right if __name__ == "__main__": import doctest doctest.testmod() Remove reference in docs to removed function.#! /usr/bin/env python """Functions on iterators, optimised for case when iterators are sorted. """ import itertools import iterators def partition_o(left_function, items): """Return a pair of iterators: left and right Items for which left_function returns a true value go into left. Items for which left_function returns a false value go into right. Items must be sorted such that left_function may be true for an initial set of items, but once an item is found such that left_function(item) is false, it will remain false for the rest of the items. In other words the following must hold: for all N where 0 <= N < (len(items) - 1) : not(func(item[n])) => not(func(item[n+1])) For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but not for [1,3,4,2,5]. >>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5]) >>> list(left) [-1, 0, 2] >>> list(right) [3, 4, 5] """ left = itertools.takewhile(left_function, items) right = itertools.dropwhile(left_function, items) return left, right if __name__ == "__main__": import doctest doctest.testmod()
<commit_before>#! /usr/bin/env python """Functions on iterators, optimised for case when iterators are sorted. Note sift_o is hidden as _sift_o at the moment because it is broken. Please don't use it. Once fixed, I'll remove the leading underscore again. """ import itertools import iterators def partition_o(left_function, items): """Return a pair of iterators: left and right Items for which left_function returns a true value go into left. Items for which left_function returns a false value go into right. Items must be sorted such that left_function may be true for an initial set of items, but once an item is found such that left_function(item) is false, it will remain false for the rest of the items. In other words the following must hold: for all N where 0 <= N < (len(items) - 1) : not(func(item[n])) => not(func(item[n+1])) For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but not for [1,3,4,2,5]. >>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5]) >>> list(left) [-1, 0, 2] >>> list(right) [3, 4, 5] """ left = itertools.takewhile(left_function, items) right = itertools.dropwhile(left_function, items) return left, right if __name__ == "__main__": import doctest doctest.testmod() <commit_msg>Remove reference in docs to removed function.<commit_after>#! /usr/bin/env python """Functions on iterators, optimised for case when iterators are sorted. """ import itertools import iterators def partition_o(left_function, items): """Return a pair of iterators: left and right Items for which left_function returns a true value go into left. Items for which left_function returns a false value go into right. Items must be sorted such that left_function may be true for an initial set of items, but once an item is found such that left_function(item) is false, it will remain false for the rest of the items. In other words the following must hold: for all N where 0 <= N < (len(items) - 1) : not(func(item[n])) => not(func(item[n+1])) For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but not for [1,3,4,2,5]. >>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5]) >>> list(left) [-1, 0, 2] >>> list(right) [3, 4, 5] """ left = itertools.takewhile(left_function, items) right = itertools.dropwhile(left_function, items) return left, right if __name__ == "__main__": import doctest doctest.testmod()
276df9f8fbb5ad15fd768db6a13040a37037e7d6
service/urls.py
service/urls.py
from django.conf.urls import url, include from rest_framework import routers import service.authors.views import service.friendrequest.views import service.users.views import service.posts.views router = routers.DefaultRouter() router.register(r'users', service.users.views.UserViewSet) router.register(r'nodes', service.nodes.views.NodeViewSet) router.register(r'author', service.authors.views.AuthorViewSet, base_name="author") router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post") # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browseable API. urlpatterns = [ url(r'^', include(router.urls)), url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'), ]
from django.conf.urls import url, include from rest_framework import routers import service.authors.views import service.friendrequest.views import service.users.views import service.nodes.views import service.posts.views router = routers.DefaultRouter() router.register(r'users', service.users.views.UserViewSet) router.register(r'nodes', service.nodes.views.NodeViewSet) router.register(r'author', service.authors.views.AuthorViewSet, base_name="author") router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post") # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browseable API. urlpatterns = [ url(r'^', include(router.urls)), url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'), ]
Add missing Node view import
Add missing Node view import
Python
apache-2.0
TeamAADGT/CMPUT404-project-socialdistribution,TeamAADGT/CMPUT404-project-socialdistribution,TeamAADGT/CMPUT404-project-socialdistribution
from django.conf.urls import url, include from rest_framework import routers import service.authors.views import service.friendrequest.views import service.users.views import service.posts.views router = routers.DefaultRouter() router.register(r'users', service.users.views.UserViewSet) router.register(r'nodes', service.nodes.views.NodeViewSet) router.register(r'author', service.authors.views.AuthorViewSet, base_name="author") router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post") # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browseable API. urlpatterns = [ url(r'^', include(router.urls)), url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'), ] Add missing Node view import
from django.conf.urls import url, include from rest_framework import routers import service.authors.views import service.friendrequest.views import service.users.views import service.nodes.views import service.posts.views router = routers.DefaultRouter() router.register(r'users', service.users.views.UserViewSet) router.register(r'nodes', service.nodes.views.NodeViewSet) router.register(r'author', service.authors.views.AuthorViewSet, base_name="author") router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post") # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browseable API. urlpatterns = [ url(r'^', include(router.urls)), url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'), ]
<commit_before>from django.conf.urls import url, include from rest_framework import routers import service.authors.views import service.friendrequest.views import service.users.views import service.posts.views router = routers.DefaultRouter() router.register(r'users', service.users.views.UserViewSet) router.register(r'nodes', service.nodes.views.NodeViewSet) router.register(r'author', service.authors.views.AuthorViewSet, base_name="author") router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post") # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browseable API. urlpatterns = [ url(r'^', include(router.urls)), url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'), ] <commit_msg>Add missing Node view import<commit_after>
from django.conf.urls import url, include from rest_framework import routers import service.authors.views import service.friendrequest.views import service.users.views import service.nodes.views import service.posts.views router = routers.DefaultRouter() router.register(r'users', service.users.views.UserViewSet) router.register(r'nodes', service.nodes.views.NodeViewSet) router.register(r'author', service.authors.views.AuthorViewSet, base_name="author") router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post") # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browseable API. urlpatterns = [ url(r'^', include(router.urls)), url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'), ]
from django.conf.urls import url, include from rest_framework import routers import service.authors.views import service.friendrequest.views import service.users.views import service.posts.views router = routers.DefaultRouter() router.register(r'users', service.users.views.UserViewSet) router.register(r'nodes', service.nodes.views.NodeViewSet) router.register(r'author', service.authors.views.AuthorViewSet, base_name="author") router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post") # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browseable API. urlpatterns = [ url(r'^', include(router.urls)), url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'), ] Add missing Node view importfrom django.conf.urls import url, include from rest_framework import routers import service.authors.views import service.friendrequest.views import service.users.views import service.nodes.views import service.posts.views router = routers.DefaultRouter() router.register(r'users', service.users.views.UserViewSet) router.register(r'nodes', service.nodes.views.NodeViewSet) router.register(r'author', service.authors.views.AuthorViewSet, base_name="author") router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post") # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browseable API. urlpatterns = [ url(r'^', include(router.urls)), url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'), ]
<commit_before>from django.conf.urls import url, include from rest_framework import routers import service.authors.views import service.friendrequest.views import service.users.views import service.posts.views router = routers.DefaultRouter() router.register(r'users', service.users.views.UserViewSet) router.register(r'nodes', service.nodes.views.NodeViewSet) router.register(r'author', service.authors.views.AuthorViewSet, base_name="author") router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post") # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browseable API. urlpatterns = [ url(r'^', include(router.urls)), url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'), ] <commit_msg>Add missing Node view import<commit_after>from django.conf.urls import url, include from rest_framework import routers import service.authors.views import service.friendrequest.views import service.users.views import service.nodes.views import service.posts.views router = routers.DefaultRouter() router.register(r'users', service.users.views.UserViewSet) router.register(r'nodes', service.nodes.views.NodeViewSet) router.register(r'author', service.authors.views.AuthorViewSet, base_name="author") router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post") # Wire up our API using automatic URL routing. # Additionally, we include login URLs for the browseable API. urlpatterns = [ url(r'^', include(router.urls)), url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'), ]
73d22cc63a2a37bd3c99774bf098ca12c81d54ae
funnels.py
funnels.py
import pyglet from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels window = pyglet.window.Window()#fullscreen=True) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run()
import pyglet import argparse from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels def main(fullscreen): window = pyglet.window.Window(fullscreen=fullscreen) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() if __name__ == '__main__': parser = argparse.ArgumentParser(description="Arithemetic practice game.") parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True') parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False') parser.set_defaults(fullscreen=True) results = parser.parse_args() main(results.fullscreen)
Add argparse to turn on/off fullscreen behavior
Add argparse to turn on/off fullscreen behavior
Python
mit
simeonf/claire
import pyglet from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels window = pyglet.window.Window()#fullscreen=True) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() Add argparse to turn on/off fullscreen behavior
import pyglet import argparse from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels def main(fullscreen): window = pyglet.window.Window(fullscreen=fullscreen) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() if __name__ == '__main__': parser = argparse.ArgumentParser(description="Arithemetic practice game.") parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True') parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False') parser.set_defaults(fullscreen=True) results = parser.parse_args() main(results.fullscreen)
<commit_before>import pyglet from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels window = pyglet.window.Window()#fullscreen=True) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() <commit_msg>Add argparse to turn on/off fullscreen behavior<commit_after>
import pyglet import argparse from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels def main(fullscreen): window = pyglet.window.Window(fullscreen=fullscreen) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() if __name__ == '__main__': parser = argparse.ArgumentParser(description="Arithemetic practice game.") parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True') parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False') parser.set_defaults(fullscreen=True) results = parser.parse_args() main(results.fullscreen)
import pyglet from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels window = pyglet.window.Window()#fullscreen=True) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() Add argparse to turn on/off fullscreen behaviorimport pyglet import argparse from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels def main(fullscreen): window = pyglet.window.Window(fullscreen=fullscreen) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() if __name__ == '__main__': parser = argparse.ArgumentParser(description="Arithemetic practice game.") parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True') parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False') parser.set_defaults(fullscreen=True) results = parser.parse_args() main(results.fullscreen)
<commit_before>import pyglet from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels window = pyglet.window.Window()#fullscreen=True) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() <commit_msg>Add argparse to turn on/off fullscreen behavior<commit_after>import pyglet import argparse from levels import GameOver, IntroScreen, TheGame from levels.levels import Levels def main(fullscreen): window = pyglet.window.Window(fullscreen=fullscreen) levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)]) pyglet.clock.schedule(levels.clock) @window.event def on_key_press(symbol, modifiers): levels.key(symbol, modifiers) @window.event def on_draw(): levels.draw() pyglet.app.run() if __name__ == '__main__': parser = argparse.ArgumentParser(description="Arithemetic practice game.") parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True') parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False') parser.set_defaults(fullscreen=True) results = parser.parse_args() main(results.fullscreen)
86edd9a5d060d88b011d280b72e208716e001c3a
phy/__init__.py
phy/__init__.py
# -*- coding: utf-8 -*- # flake8: noqa """ phy is an open source electrophysiological data analysis package in Python for neuronal recordings made with high-density multielectrode arrays containing up to thousands of channels. """ #------------------------------------------------------------------------------ # Imports #------------------------------------------------------------------------------ import os.path as op from pkg_resources import get_distribution, DistributionNotFound from .utils.logging import _default_logger, set_level from .utils.datasets import download_test_data from .utils.dock import enable_qt, qt_app #------------------------------------------------------------------------------ # Global variables and functions #------------------------------------------------------------------------------ __author__ = 'Kwik team' __email__ = 'cyrille.rossant at gmail.com' __version__ = '0.1.0.dev0' __all__ = ['debug', 'set_level'] # Set up the default logger. _default_logger() def debug(enable=True): """Enable debug logging mode.""" if enable: set_level('debug') else: set_level('info')
# -*- coding: utf-8 -*- # flake8: noqa """ phy is an open source electrophysiological data analysis package in Python for neuronal recordings made with high-density multielectrode arrays containing up to thousands of channels. """ #------------------------------------------------------------------------------ # Imports #------------------------------------------------------------------------------ # Define a mock @profile decorator to avoid NameErrors when profiling. import sys if sys.version_info[0] == 3: import builtins else: import __builtin__ as builtins builtins.__dict__['profile'] = lambda func: func import os.path as op from pkg_resources import get_distribution, DistributionNotFound from .utils.logging import _default_logger, set_level from .utils.datasets import download_test_data from .utils.dock import enable_qt, qt_app #------------------------------------------------------------------------------ # Global variables and functions #------------------------------------------------------------------------------ __author__ = 'Kwik team' __email__ = 'cyrille.rossant at gmail.com' __version__ = '0.1.0.dev0' __all__ = ['debug', 'set_level'] # Set up the default logger. _default_logger() def debug(enable=True): """Enable debug logging mode.""" if enable: set_level('debug') else: set_level('info')
Define mock @profile decorator in builtins.
Define mock @profile decorator in builtins.
Python
bsd-3-clause
nippoo/phy,kwikteam/phy,nsteinme/phy,rossant/phy,kwikteam/phy,rossant/phy,nsteinme/phy,rossant/phy,nippoo/phy,kwikteam/phy
# -*- coding: utf-8 -*- # flake8: noqa """ phy is an open source electrophysiological data analysis package in Python for neuronal recordings made with high-density multielectrode arrays containing up to thousands of channels. """ #------------------------------------------------------------------------------ # Imports #------------------------------------------------------------------------------ import os.path as op from pkg_resources import get_distribution, DistributionNotFound from .utils.logging import _default_logger, set_level from .utils.datasets import download_test_data from .utils.dock import enable_qt, qt_app #------------------------------------------------------------------------------ # Global variables and functions #------------------------------------------------------------------------------ __author__ = 'Kwik team' __email__ = 'cyrille.rossant at gmail.com' __version__ = '0.1.0.dev0' __all__ = ['debug', 'set_level'] # Set up the default logger. _default_logger() def debug(enable=True): """Enable debug logging mode.""" if enable: set_level('debug') else: set_level('info') Define mock @profile decorator in builtins.
# -*- coding: utf-8 -*- # flake8: noqa """ phy is an open source electrophysiological data analysis package in Python for neuronal recordings made with high-density multielectrode arrays containing up to thousands of channels. """ #------------------------------------------------------------------------------ # Imports #------------------------------------------------------------------------------ # Define a mock @profile decorator to avoid NameErrors when profiling. import sys if sys.version_info[0] == 3: import builtins else: import __builtin__ as builtins builtins.__dict__['profile'] = lambda func: func import os.path as op from pkg_resources import get_distribution, DistributionNotFound from .utils.logging import _default_logger, set_level from .utils.datasets import download_test_data from .utils.dock import enable_qt, qt_app #------------------------------------------------------------------------------ # Global variables and functions #------------------------------------------------------------------------------ __author__ = 'Kwik team' __email__ = 'cyrille.rossant at gmail.com' __version__ = '0.1.0.dev0' __all__ = ['debug', 'set_level'] # Set up the default logger. _default_logger() def debug(enable=True): """Enable debug logging mode.""" if enable: set_level('debug') else: set_level('info')
<commit_before># -*- coding: utf-8 -*- # flake8: noqa """ phy is an open source electrophysiological data analysis package in Python for neuronal recordings made with high-density multielectrode arrays containing up to thousands of channels. """ #------------------------------------------------------------------------------ # Imports #------------------------------------------------------------------------------ import os.path as op from pkg_resources import get_distribution, DistributionNotFound from .utils.logging import _default_logger, set_level from .utils.datasets import download_test_data from .utils.dock import enable_qt, qt_app #------------------------------------------------------------------------------ # Global variables and functions #------------------------------------------------------------------------------ __author__ = 'Kwik team' __email__ = 'cyrille.rossant at gmail.com' __version__ = '0.1.0.dev0' __all__ = ['debug', 'set_level'] # Set up the default logger. _default_logger() def debug(enable=True): """Enable debug logging mode.""" if enable: set_level('debug') else: set_level('info') <commit_msg>Define mock @profile decorator in builtins.<commit_after>
# -*- coding: utf-8 -*- # flake8: noqa """ phy is an open source electrophysiological data analysis package in Python for neuronal recordings made with high-density multielectrode arrays containing up to thousands of channels. """ #------------------------------------------------------------------------------ # Imports #------------------------------------------------------------------------------ # Define a mock @profile decorator to avoid NameErrors when profiling. import sys if sys.version_info[0] == 3: import builtins else: import __builtin__ as builtins builtins.__dict__['profile'] = lambda func: func import os.path as op from pkg_resources import get_distribution, DistributionNotFound from .utils.logging import _default_logger, set_level from .utils.datasets import download_test_data from .utils.dock import enable_qt, qt_app #------------------------------------------------------------------------------ # Global variables and functions #------------------------------------------------------------------------------ __author__ = 'Kwik team' __email__ = 'cyrille.rossant at gmail.com' __version__ = '0.1.0.dev0' __all__ = ['debug', 'set_level'] # Set up the default logger. _default_logger() def debug(enable=True): """Enable debug logging mode.""" if enable: set_level('debug') else: set_level('info')
# -*- coding: utf-8 -*- # flake8: noqa """ phy is an open source electrophysiological data analysis package in Python for neuronal recordings made with high-density multielectrode arrays containing up to thousands of channels. """ #------------------------------------------------------------------------------ # Imports #------------------------------------------------------------------------------ import os.path as op from pkg_resources import get_distribution, DistributionNotFound from .utils.logging import _default_logger, set_level from .utils.datasets import download_test_data from .utils.dock import enable_qt, qt_app #------------------------------------------------------------------------------ # Global variables and functions #------------------------------------------------------------------------------ __author__ = 'Kwik team' __email__ = 'cyrille.rossant at gmail.com' __version__ = '0.1.0.dev0' __all__ = ['debug', 'set_level'] # Set up the default logger. _default_logger() def debug(enable=True): """Enable debug logging mode.""" if enable: set_level('debug') else: set_level('info') Define mock @profile decorator in builtins.# -*- coding: utf-8 -*- # flake8: noqa """ phy is an open source electrophysiological data analysis package in Python for neuronal recordings made with high-density multielectrode arrays containing up to thousands of channels. """ #------------------------------------------------------------------------------ # Imports #------------------------------------------------------------------------------ # Define a mock @profile decorator to avoid NameErrors when profiling. import sys if sys.version_info[0] == 3: import builtins else: import __builtin__ as builtins builtins.__dict__['profile'] = lambda func: func import os.path as op from pkg_resources import get_distribution, DistributionNotFound from .utils.logging import _default_logger, set_level from .utils.datasets import download_test_data from .utils.dock import enable_qt, qt_app #------------------------------------------------------------------------------ # Global variables and functions #------------------------------------------------------------------------------ __author__ = 'Kwik team' __email__ = 'cyrille.rossant at gmail.com' __version__ = '0.1.0.dev0' __all__ = ['debug', 'set_level'] # Set up the default logger. _default_logger() def debug(enable=True): """Enable debug logging mode.""" if enable: set_level('debug') else: set_level('info')
<commit_before># -*- coding: utf-8 -*- # flake8: noqa """ phy is an open source electrophysiological data analysis package in Python for neuronal recordings made with high-density multielectrode arrays containing up to thousands of channels. """ #------------------------------------------------------------------------------ # Imports #------------------------------------------------------------------------------ import os.path as op from pkg_resources import get_distribution, DistributionNotFound from .utils.logging import _default_logger, set_level from .utils.datasets import download_test_data from .utils.dock import enable_qt, qt_app #------------------------------------------------------------------------------ # Global variables and functions #------------------------------------------------------------------------------ __author__ = 'Kwik team' __email__ = 'cyrille.rossant at gmail.com' __version__ = '0.1.0.dev0' __all__ = ['debug', 'set_level'] # Set up the default logger. _default_logger() def debug(enable=True): """Enable debug logging mode.""" if enable: set_level('debug') else: set_level('info') <commit_msg>Define mock @profile decorator in builtins.<commit_after># -*- coding: utf-8 -*- # flake8: noqa """ phy is an open source electrophysiological data analysis package in Python for neuronal recordings made with high-density multielectrode arrays containing up to thousands of channels. """ #------------------------------------------------------------------------------ # Imports #------------------------------------------------------------------------------ # Define a mock @profile decorator to avoid NameErrors when profiling. import sys if sys.version_info[0] == 3: import builtins else: import __builtin__ as builtins builtins.__dict__['profile'] = lambda func: func import os.path as op from pkg_resources import get_distribution, DistributionNotFound from .utils.logging import _default_logger, set_level from .utils.datasets import download_test_data from .utils.dock import enable_qt, qt_app #------------------------------------------------------------------------------ # Global variables and functions #------------------------------------------------------------------------------ __author__ = 'Kwik team' __email__ = 'cyrille.rossant at gmail.com' __version__ = '0.1.0.dev0' __all__ = ['debug', 'set_level'] # Set up the default logger. _default_logger() def debug(enable=True): """Enable debug logging mode.""" if enable: set_level('debug') else: set_level('info')
1b23e939a40652f8ef870e3ee7146f62fd131933
getlost.py
getlost.py
from os import environ from urllib2 import urlopen from math import log from flask import Flask, json, jsonify app = Flask(__name__) from hip import get_ranking_array from utils import jsonp url = 'http://open.mapquestapi.com/directions/v2/route' params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian' rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}' @app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>") @jsonp def route(from_lat, from_lng, to_lat, to_lng): resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'], flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng)) route = json.loads(resp.read().decode("utf-8")) coords = [(man['startPoint']['lat'], man['startPoint']['lng']) for leg in route['route']['legs'] for man in leg['maneuvers']] hip_rank, total_rank = get_ranking_array(coords) return jsonify(route=route, hip_rank=list(hip_rank), total_rank=log(total_rank)) if __name__ == "__main__": app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
from os import environ from urllib2 import urlopen from math import log, sqrt from flask import Flask, json, jsonify app = Flask(__name__) from hip import get_ranking_array from utils import jsonp url = 'http://open.mapquestapi.com/directions/v2/route' params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian' rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}' @app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>") @jsonp def route(from_lat, from_lng, to_lat, to_lng): resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'], flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng)) route = json.loads(resp.read().decode("utf-8")) coords = [(man['startPoint']['lat'], man['startPoint']['lng']) for leg in route['route']['legs'] for man in leg['maneuvers']] hip_rank, total_rank = get_ranking_array(coords) total_rank /= sqrt((float(to_lng) - float(from_lng))**2 + (float(to_lat) - float(from_lat))**2) return jsonify(route=route, hip_rank=list(hip_rank), total_rank=log(total_rank)) if __name__ == "__main__": app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
Normalize total rank by distance between start and end
Normalize total rank by distance between start and end
Python
apache-2.0
kynan/GetLost
from os import environ from urllib2 import urlopen from math import log from flask import Flask, json, jsonify app = Flask(__name__) from hip import get_ranking_array from utils import jsonp url = 'http://open.mapquestapi.com/directions/v2/route' params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian' rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}' @app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>") @jsonp def route(from_lat, from_lng, to_lat, to_lng): resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'], flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng)) route = json.loads(resp.read().decode("utf-8")) coords = [(man['startPoint']['lat'], man['startPoint']['lng']) for leg in route['route']['legs'] for man in leg['maneuvers']] hip_rank, total_rank = get_ranking_array(coords) return jsonify(route=route, hip_rank=list(hip_rank), total_rank=log(total_rank)) if __name__ == "__main__": app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False)) Normalize total rank by distance between start and end
from os import environ from urllib2 import urlopen from math import log, sqrt from flask import Flask, json, jsonify app = Flask(__name__) from hip import get_ranking_array from utils import jsonp url = 'http://open.mapquestapi.com/directions/v2/route' params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian' rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}' @app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>") @jsonp def route(from_lat, from_lng, to_lat, to_lng): resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'], flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng)) route = json.loads(resp.read().decode("utf-8")) coords = [(man['startPoint']['lat'], man['startPoint']['lng']) for leg in route['route']['legs'] for man in leg['maneuvers']] hip_rank, total_rank = get_ranking_array(coords) total_rank /= sqrt((float(to_lng) - float(from_lng))**2 + (float(to_lat) - float(from_lat))**2) return jsonify(route=route, hip_rank=list(hip_rank), total_rank=log(total_rank)) if __name__ == "__main__": app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
<commit_before>from os import environ from urllib2 import urlopen from math import log from flask import Flask, json, jsonify app = Flask(__name__) from hip import get_ranking_array from utils import jsonp url = 'http://open.mapquestapi.com/directions/v2/route' params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian' rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}' @app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>") @jsonp def route(from_lat, from_lng, to_lat, to_lng): resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'], flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng)) route = json.loads(resp.read().decode("utf-8")) coords = [(man['startPoint']['lat'], man['startPoint']['lng']) for leg in route['route']['legs'] for man in leg['maneuvers']] hip_rank, total_rank = get_ranking_array(coords) return jsonify(route=route, hip_rank=list(hip_rank), total_rank=log(total_rank)) if __name__ == "__main__": app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False)) <commit_msg>Normalize total rank by distance between start and end<commit_after>
from os import environ from urllib2 import urlopen from math import log, sqrt from flask import Flask, json, jsonify app = Flask(__name__) from hip import get_ranking_array from utils import jsonp url = 'http://open.mapquestapi.com/directions/v2/route' params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian' rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}' @app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>") @jsonp def route(from_lat, from_lng, to_lat, to_lng): resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'], flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng)) route = json.loads(resp.read().decode("utf-8")) coords = [(man['startPoint']['lat'], man['startPoint']['lng']) for leg in route['route']['legs'] for man in leg['maneuvers']] hip_rank, total_rank = get_ranking_array(coords) total_rank /= sqrt((float(to_lng) - float(from_lng))**2 + (float(to_lat) - float(from_lat))**2) return jsonify(route=route, hip_rank=list(hip_rank), total_rank=log(total_rank)) if __name__ == "__main__": app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
from os import environ from urllib2 import urlopen from math import log from flask import Flask, json, jsonify app = Flask(__name__) from hip import get_ranking_array from utils import jsonp url = 'http://open.mapquestapi.com/directions/v2/route' params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian' rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}' @app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>") @jsonp def route(from_lat, from_lng, to_lat, to_lng): resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'], flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng)) route = json.loads(resp.read().decode("utf-8")) coords = [(man['startPoint']['lat'], man['startPoint']['lng']) for leg in route['route']['legs'] for man in leg['maneuvers']] hip_rank, total_rank = get_ranking_array(coords) return jsonify(route=route, hip_rank=list(hip_rank), total_rank=log(total_rank)) if __name__ == "__main__": app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False)) Normalize total rank by distance between start and endfrom os import environ from urllib2 import urlopen from math import log, sqrt from flask import Flask, json, jsonify app = Flask(__name__) from hip import get_ranking_array from utils import jsonp url = 'http://open.mapquestapi.com/directions/v2/route' params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian' rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}' @app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>") @jsonp def route(from_lat, from_lng, to_lat, to_lng): resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'], flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng)) route = json.loads(resp.read().decode("utf-8")) coords = [(man['startPoint']['lat'], man['startPoint']['lng']) for leg in route['route']['legs'] for man in leg['maneuvers']] hip_rank, total_rank = get_ranking_array(coords) total_rank /= sqrt((float(to_lng) - float(from_lng))**2 + (float(to_lat) - float(from_lat))**2) return jsonify(route=route, hip_rank=list(hip_rank), total_rank=log(total_rank)) if __name__ == "__main__": app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
<commit_before>from os import environ from urllib2 import urlopen from math import log from flask import Flask, json, jsonify app = Flask(__name__) from hip import get_ranking_array from utils import jsonp url = 'http://open.mapquestapi.com/directions/v2/route' params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian' rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}' @app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>") @jsonp def route(from_lat, from_lng, to_lat, to_lng): resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'], flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng)) route = json.loads(resp.read().decode("utf-8")) coords = [(man['startPoint']['lat'], man['startPoint']['lng']) for leg in route['route']['legs'] for man in leg['maneuvers']] hip_rank, total_rank = get_ranking_array(coords) return jsonify(route=route, hip_rank=list(hip_rank), total_rank=log(total_rank)) if __name__ == "__main__": app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False)) <commit_msg>Normalize total rank by distance between start and end<commit_after>from os import environ from urllib2 import urlopen from math import log, sqrt from flask import Flask, json, jsonify app = Flask(__name__) from hip import get_ranking_array from utils import jsonp url = 'http://open.mapquestapi.com/directions/v2/route' params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian' rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}' @app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>") @jsonp def route(from_lat, from_lng, to_lat, to_lng): resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'], flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng)) route = json.loads(resp.read().decode("utf-8")) coords = [(man['startPoint']['lat'], man['startPoint']['lng']) for leg in route['route']['legs'] for man in leg['maneuvers']] hip_rank, total_rank = get_ranking_array(coords) total_rank /= sqrt((float(to_lng) - float(from_lng))**2 + (float(to_lat) - float(from_lat))**2) return jsonify(route=route, hip_rank=list(hip_rank), total_rank=log(total_rank)) if __name__ == "__main__": app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
96bcf7f55a50895dead660add9fc949af197f550
networking_sfc/tests/functional/services/sfc/agent/extensions/test_ovs_agent_sfc_extension.py
networking_sfc/tests/functional/services/sfc/agent/extensions/test_ovs_agent_sfc_extension.py
# Copyright (c) 2016 Red Hat, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron.tests.functional.agent.l2 import base class TestOVSAgentSfcExtension(base.OVSAgentTestFramework): def setUp(self): super(TestOVSAgentSfcExtension, self).setUp() self.config.set_override('extensions', ['sfc'], 'agent') def test_run(self): agent = self.create_agent() self.start_agent(agent) agent_state = agent.state_rpc.report_state.call_args[0][1] self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
# Copyright (c) 2016 Red Hat, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron.tests.functional.agent.l2 import base class TestOVSAgentSfcExtension(base.OVSAgentTestFramework): def setUp(self): super(TestOVSAgentSfcExtension, self).setUp() self.config.set_override('extensions', ['sfc'], 'agent') self.agent = self.create_agent() def test_run(self): self.agent._report_state() agent_state = self.agent.state_rpc.report_state.call_args[0][1] self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
Fix extension loading functional test
Fix extension loading functional test Call the agent _report_state() before checking the report state itself Change-Id: Idbf552d5ca5968bc95b0a3c395499c3f2d215729 Closes-Bug: 1658089
Python
apache-2.0
openstack/networking-sfc,openstack/networking-sfc
# Copyright (c) 2016 Red Hat, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron.tests.functional.agent.l2 import base class TestOVSAgentSfcExtension(base.OVSAgentTestFramework): def setUp(self): super(TestOVSAgentSfcExtension, self).setUp() self.config.set_override('extensions', ['sfc'], 'agent') def test_run(self): agent = self.create_agent() self.start_agent(agent) agent_state = agent.state_rpc.report_state.call_args[0][1] self.assertEqual(['sfc'], agent_state['configurations']['extensions']) Fix extension loading functional test Call the agent _report_state() before checking the report state itself Change-Id: Idbf552d5ca5968bc95b0a3c395499c3f2d215729 Closes-Bug: 1658089
# Copyright (c) 2016 Red Hat, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron.tests.functional.agent.l2 import base class TestOVSAgentSfcExtension(base.OVSAgentTestFramework): def setUp(self): super(TestOVSAgentSfcExtension, self).setUp() self.config.set_override('extensions', ['sfc'], 'agent') self.agent = self.create_agent() def test_run(self): self.agent._report_state() agent_state = self.agent.state_rpc.report_state.call_args[0][1] self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
<commit_before># Copyright (c) 2016 Red Hat, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron.tests.functional.agent.l2 import base class TestOVSAgentSfcExtension(base.OVSAgentTestFramework): def setUp(self): super(TestOVSAgentSfcExtension, self).setUp() self.config.set_override('extensions', ['sfc'], 'agent') def test_run(self): agent = self.create_agent() self.start_agent(agent) agent_state = agent.state_rpc.report_state.call_args[0][1] self.assertEqual(['sfc'], agent_state['configurations']['extensions']) <commit_msg>Fix extension loading functional test Call the agent _report_state() before checking the report state itself Change-Id: Idbf552d5ca5968bc95b0a3c395499c3f2d215729 Closes-Bug: 1658089<commit_after>
# Copyright (c) 2016 Red Hat, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron.tests.functional.agent.l2 import base class TestOVSAgentSfcExtension(base.OVSAgentTestFramework): def setUp(self): super(TestOVSAgentSfcExtension, self).setUp() self.config.set_override('extensions', ['sfc'], 'agent') self.agent = self.create_agent() def test_run(self): self.agent._report_state() agent_state = self.agent.state_rpc.report_state.call_args[0][1] self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
# Copyright (c) 2016 Red Hat, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron.tests.functional.agent.l2 import base class TestOVSAgentSfcExtension(base.OVSAgentTestFramework): def setUp(self): super(TestOVSAgentSfcExtension, self).setUp() self.config.set_override('extensions', ['sfc'], 'agent') def test_run(self): agent = self.create_agent() self.start_agent(agent) agent_state = agent.state_rpc.report_state.call_args[0][1] self.assertEqual(['sfc'], agent_state['configurations']['extensions']) Fix extension loading functional test Call the agent _report_state() before checking the report state itself Change-Id: Idbf552d5ca5968bc95b0a3c395499c3f2d215729 Closes-Bug: 1658089# Copyright (c) 2016 Red Hat, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron.tests.functional.agent.l2 import base class TestOVSAgentSfcExtension(base.OVSAgentTestFramework): def setUp(self): super(TestOVSAgentSfcExtension, self).setUp() self.config.set_override('extensions', ['sfc'], 'agent') self.agent = self.create_agent() def test_run(self): self.agent._report_state() agent_state = self.agent.state_rpc.report_state.call_args[0][1] self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
<commit_before># Copyright (c) 2016 Red Hat, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron.tests.functional.agent.l2 import base class TestOVSAgentSfcExtension(base.OVSAgentTestFramework): def setUp(self): super(TestOVSAgentSfcExtension, self).setUp() self.config.set_override('extensions', ['sfc'], 'agent') def test_run(self): agent = self.create_agent() self.start_agent(agent) agent_state = agent.state_rpc.report_state.call_args[0][1] self.assertEqual(['sfc'], agent_state['configurations']['extensions']) <commit_msg>Fix extension loading functional test Call the agent _report_state() before checking the report state itself Change-Id: Idbf552d5ca5968bc95b0a3c395499c3f2d215729 Closes-Bug: 1658089<commit_after># Copyright (c) 2016 Red Hat, Inc. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron.tests.functional.agent.l2 import base class TestOVSAgentSfcExtension(base.OVSAgentTestFramework): def setUp(self): super(TestOVSAgentSfcExtension, self).setUp() self.config.set_override('extensions', ['sfc'], 'agent') self.agent = self.create_agent() def test_run(self): self.agent._report_state() agent_state = self.agent.state_rpc.report_state.call_args[0][1] self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
907298a325e966f6e03c766c90f22e1b03c25c1e
data/propaganda2mongo.py
data/propaganda2mongo.py
import bson.json_util from bson.objectid import ObjectId import json import sys def main(): node_table = {} while True: line = sys.stdin.readline() if not line: break record = json.loads(line) ident = str(record["twitter_id"]) aoid = node_table.get(ident) if aoid is None: node_table[ident] = aoid = ObjectId() print bson.json_util.dumps({"_id": aoid, "type": "node", "data": {"twitter_id": ident, "type": "audience", "propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"], "geos": record["geos"], "timestamps_of_propaganda": record["timestamps_of_propaganda"]}}) for p in record["propagandists_followed"]: oid = node_table.get(p) if oid is None: node_table[ident] = oid = ObjectId() print bson.json_util.dumps({"_id": oid, "type": "node", "data": {"twitter_id": ident, "type": "propagandist"}}) print bson.json_util.dumps({"_id": ObjectId(), "type": "link", "source": aoid, "target": oid, "data": {}}) if __name__ == "__main__": sys.exit(main())
import bson.json_util from bson.objectid import ObjectId import json import sys def main(): node_table = {} while True: line = sys.stdin.readline() if not line: break record = json.loads(line) ident = str(record["twitter_id"]) aoid = node_table.get(ident) if aoid is None: node_table[ident] = aoid = ObjectId() print bson.json_util.dumps({"_id": aoid, "type": "node", "data": {"twitter_id": ident, "type": "audience", "propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"], "geos": record["geos"], "timestamps_of_propaganda": record["timestamps_of_propaganda"]}}) for p in record["propagandists_followed"]: oid = node_table.get(p) if oid is None: node_table[ident] = oid = ObjectId() print bson.json_util.dumps({"_id": oid, "type": "node", "data": {"twitter_id": p, "type": "propagandist"}}) print bson.json_util.dumps({"_id": ObjectId(), "type": "link", "source": aoid, "target": oid, "data": {}}) if __name__ == "__main__": sys.exit(main())
Fix bug in data collection
Fix bug in data collection
Python
apache-2.0
XDATA-Year-3/clique-propaganda,XDATA-Year-3/clique-propaganda,XDATA-Year-3/clique-propaganda
import bson.json_util from bson.objectid import ObjectId import json import sys def main(): node_table = {} while True: line = sys.stdin.readline() if not line: break record = json.loads(line) ident = str(record["twitter_id"]) aoid = node_table.get(ident) if aoid is None: node_table[ident] = aoid = ObjectId() print bson.json_util.dumps({"_id": aoid, "type": "node", "data": {"twitter_id": ident, "type": "audience", "propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"], "geos": record["geos"], "timestamps_of_propaganda": record["timestamps_of_propaganda"]}}) for p in record["propagandists_followed"]: oid = node_table.get(p) if oid is None: node_table[ident] = oid = ObjectId() print bson.json_util.dumps({"_id": oid, "type": "node", "data": {"twitter_id": ident, "type": "propagandist"}}) print bson.json_util.dumps({"_id": ObjectId(), "type": "link", "source": aoid, "target": oid, "data": {}}) if __name__ == "__main__": sys.exit(main()) Fix bug in data collection
import bson.json_util from bson.objectid import ObjectId import json import sys def main(): node_table = {} while True: line = sys.stdin.readline() if not line: break record = json.loads(line) ident = str(record["twitter_id"]) aoid = node_table.get(ident) if aoid is None: node_table[ident] = aoid = ObjectId() print bson.json_util.dumps({"_id": aoid, "type": "node", "data": {"twitter_id": ident, "type": "audience", "propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"], "geos": record["geos"], "timestamps_of_propaganda": record["timestamps_of_propaganda"]}}) for p in record["propagandists_followed"]: oid = node_table.get(p) if oid is None: node_table[ident] = oid = ObjectId() print bson.json_util.dumps({"_id": oid, "type": "node", "data": {"twitter_id": p, "type": "propagandist"}}) print bson.json_util.dumps({"_id": ObjectId(), "type": "link", "source": aoid, "target": oid, "data": {}}) if __name__ == "__main__": sys.exit(main())
<commit_before>import bson.json_util from bson.objectid import ObjectId import json import sys def main(): node_table = {} while True: line = sys.stdin.readline() if not line: break record = json.loads(line) ident = str(record["twitter_id"]) aoid = node_table.get(ident) if aoid is None: node_table[ident] = aoid = ObjectId() print bson.json_util.dumps({"_id": aoid, "type": "node", "data": {"twitter_id": ident, "type": "audience", "propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"], "geos": record["geos"], "timestamps_of_propaganda": record["timestamps_of_propaganda"]}}) for p in record["propagandists_followed"]: oid = node_table.get(p) if oid is None: node_table[ident] = oid = ObjectId() print bson.json_util.dumps({"_id": oid, "type": "node", "data": {"twitter_id": ident, "type": "propagandist"}}) print bson.json_util.dumps({"_id": ObjectId(), "type": "link", "source": aoid, "target": oid, "data": {}}) if __name__ == "__main__": sys.exit(main()) <commit_msg>Fix bug in data collection<commit_after>
import bson.json_util from bson.objectid import ObjectId import json import sys def main(): node_table = {} while True: line = sys.stdin.readline() if not line: break record = json.loads(line) ident = str(record["twitter_id"]) aoid = node_table.get(ident) if aoid is None: node_table[ident] = aoid = ObjectId() print bson.json_util.dumps({"_id": aoid, "type": "node", "data": {"twitter_id": ident, "type": "audience", "propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"], "geos": record["geos"], "timestamps_of_propaganda": record["timestamps_of_propaganda"]}}) for p in record["propagandists_followed"]: oid = node_table.get(p) if oid is None: node_table[ident] = oid = ObjectId() print bson.json_util.dumps({"_id": oid, "type": "node", "data": {"twitter_id": p, "type": "propagandist"}}) print bson.json_util.dumps({"_id": ObjectId(), "type": "link", "source": aoid, "target": oid, "data": {}}) if __name__ == "__main__": sys.exit(main())
import bson.json_util from bson.objectid import ObjectId import json import sys def main(): node_table = {} while True: line = sys.stdin.readline() if not line: break record = json.loads(line) ident = str(record["twitter_id"]) aoid = node_table.get(ident) if aoid is None: node_table[ident] = aoid = ObjectId() print bson.json_util.dumps({"_id": aoid, "type": "node", "data": {"twitter_id": ident, "type": "audience", "propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"], "geos": record["geos"], "timestamps_of_propaganda": record["timestamps_of_propaganda"]}}) for p in record["propagandists_followed"]: oid = node_table.get(p) if oid is None: node_table[ident] = oid = ObjectId() print bson.json_util.dumps({"_id": oid, "type": "node", "data": {"twitter_id": ident, "type": "propagandist"}}) print bson.json_util.dumps({"_id": ObjectId(), "type": "link", "source": aoid, "target": oid, "data": {}}) if __name__ == "__main__": sys.exit(main()) Fix bug in data collectionimport bson.json_util from bson.objectid import ObjectId import json import sys def main(): node_table = {} while True: line = sys.stdin.readline() if not line: break record = json.loads(line) ident = str(record["twitter_id"]) aoid = node_table.get(ident) if aoid is None: node_table[ident] = aoid = ObjectId() print bson.json_util.dumps({"_id": aoid, "type": "node", "data": {"twitter_id": ident, "type": "audience", "propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"], "geos": record["geos"], "timestamps_of_propaganda": record["timestamps_of_propaganda"]}}) for p in record["propagandists_followed"]: oid = node_table.get(p) if oid is None: node_table[ident] = oid = ObjectId() print bson.json_util.dumps({"_id": oid, "type": "node", "data": {"twitter_id": p, "type": "propagandist"}}) print bson.json_util.dumps({"_id": ObjectId(), "type": "link", "source": aoid, "target": oid, "data": {}}) if __name__ == "__main__": sys.exit(main())
<commit_before>import bson.json_util from bson.objectid import ObjectId import json import sys def main(): node_table = {} while True: line = sys.stdin.readline() if not line: break record = json.loads(line) ident = str(record["twitter_id"]) aoid = node_table.get(ident) if aoid is None: node_table[ident] = aoid = ObjectId() print bson.json_util.dumps({"_id": aoid, "type": "node", "data": {"twitter_id": ident, "type": "audience", "propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"], "geos": record["geos"], "timestamps_of_propaganda": record["timestamps_of_propaganda"]}}) for p in record["propagandists_followed"]: oid = node_table.get(p) if oid is None: node_table[ident] = oid = ObjectId() print bson.json_util.dumps({"_id": oid, "type": "node", "data": {"twitter_id": ident, "type": "propagandist"}}) print bson.json_util.dumps({"_id": ObjectId(), "type": "link", "source": aoid, "target": oid, "data": {}}) if __name__ == "__main__": sys.exit(main()) <commit_msg>Fix bug in data collection<commit_after>import bson.json_util from bson.objectid import ObjectId import json import sys def main(): node_table = {} while True: line = sys.stdin.readline() if not line: break record = json.loads(line) ident = str(record["twitter_id"]) aoid = node_table.get(ident) if aoid is None: node_table[ident] = aoid = ObjectId() print bson.json_util.dumps({"_id": aoid, "type": "node", "data": {"twitter_id": ident, "type": "audience", "propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"], "geos": record["geos"], "timestamps_of_propaganda": record["timestamps_of_propaganda"]}}) for p in record["propagandists_followed"]: oid = node_table.get(p) if oid is None: node_table[ident] = oid = ObjectId() print bson.json_util.dumps({"_id": oid, "type": "node", "data": {"twitter_id": p, "type": "propagandist"}}) print bson.json_util.dumps({"_id": ObjectId(), "type": "link", "source": aoid, "target": oid, "data": {}}) if __name__ == "__main__": sys.exit(main())
2e5a8adb47491be58d3cdc48a4984812538f55a6
golang/main.py
golang/main.py
from evolution_master.runners import pkg, download # Install for Arch with pkg.pacman() as pkg_man: pkg_man.install('go') # Install for Debian & Ubuntu with pkg.apt() as pkg_man: pkg_man.install('golang') # TODO: make this a runner and require a switch to enable this pkg_man.install('golang-go-darwin-amd64', 'golang-go-freebsd-amd64', 'golang-go-netbsd-amd64', 'golang-go-windows-amd64') # Install for OSX with pkg.brew() as pkg_man: pkg_man.install('go') # Install for Windows with download.https() as downloader, pkg.msiexec() as installer: downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi') downloader.checksum('sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba') installer.install_flags('/qn' '/norestart') installer.await(downloader.finished())
from genes import apt, brew, pacman, http_downloader, checksum, msiexec import platform opsys = platform.system() dist = platform.linux_distribution() if platform == 'Linux' and dist == 'Arch': pacman.update() pacman.sync('go') if platform == 'Linux' and (dist == 'Debian' or dist == 'Ubuntu'): apt.update() apt.install('golang') # TODO: make this a runner and require a switch to enable this apt.install('golang-go-darwin-amd64', 'golang-go-freebsd-amd64', 'golang-go-netbsd-amd64', 'golang-go-windows-amd64') if platform == 'Darwin': brew.update() brew.install('go') if platform == 'Windows': installer = http_downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi') checksum.check(installer, 'sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba') install_flags = ('/qn' '/norestart') msiexec.run(installer, install_flags)
Format go to the new design
Format go to the new design
Python
mit
hatchery/Genepool2,hatchery/genepool
from evolution_master.runners import pkg, download # Install for Arch with pkg.pacman() as pkg_man: pkg_man.install('go') # Install for Debian & Ubuntu with pkg.apt() as pkg_man: pkg_man.install('golang') # TODO: make this a runner and require a switch to enable this pkg_man.install('golang-go-darwin-amd64', 'golang-go-freebsd-amd64', 'golang-go-netbsd-amd64', 'golang-go-windows-amd64') # Install for OSX with pkg.brew() as pkg_man: pkg_man.install('go') # Install for Windows with download.https() as downloader, pkg.msiexec() as installer: downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi') downloader.checksum('sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba') installer.install_flags('/qn' '/norestart') installer.await(downloader.finished()) Format go to the new design
from genes import apt, brew, pacman, http_downloader, checksum, msiexec import platform opsys = platform.system() dist = platform.linux_distribution() if platform == 'Linux' and dist == 'Arch': pacman.update() pacman.sync('go') if platform == 'Linux' and (dist == 'Debian' or dist == 'Ubuntu'): apt.update() apt.install('golang') # TODO: make this a runner and require a switch to enable this apt.install('golang-go-darwin-amd64', 'golang-go-freebsd-amd64', 'golang-go-netbsd-amd64', 'golang-go-windows-amd64') if platform == 'Darwin': brew.update() brew.install('go') if platform == 'Windows': installer = http_downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi') checksum.check(installer, 'sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba') install_flags = ('/qn' '/norestart') msiexec.run(installer, install_flags)
<commit_before>from evolution_master.runners import pkg, download # Install for Arch with pkg.pacman() as pkg_man: pkg_man.install('go') # Install for Debian & Ubuntu with pkg.apt() as pkg_man: pkg_man.install('golang') # TODO: make this a runner and require a switch to enable this pkg_man.install('golang-go-darwin-amd64', 'golang-go-freebsd-amd64', 'golang-go-netbsd-amd64', 'golang-go-windows-amd64') # Install for OSX with pkg.brew() as pkg_man: pkg_man.install('go') # Install for Windows with download.https() as downloader, pkg.msiexec() as installer: downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi') downloader.checksum('sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba') installer.install_flags('/qn' '/norestart') installer.await(downloader.finished()) <commit_msg>Format go to the new design<commit_after>
from genes import apt, brew, pacman, http_downloader, checksum, msiexec import platform opsys = platform.system() dist = platform.linux_distribution() if platform == 'Linux' and dist == 'Arch': pacman.update() pacman.sync('go') if platform == 'Linux' and (dist == 'Debian' or dist == 'Ubuntu'): apt.update() apt.install('golang') # TODO: make this a runner and require a switch to enable this apt.install('golang-go-darwin-amd64', 'golang-go-freebsd-amd64', 'golang-go-netbsd-amd64', 'golang-go-windows-amd64') if platform == 'Darwin': brew.update() brew.install('go') if platform == 'Windows': installer = http_downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi') checksum.check(installer, 'sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba') install_flags = ('/qn' '/norestart') msiexec.run(installer, install_flags)
from evolution_master.runners import pkg, download # Install for Arch with pkg.pacman() as pkg_man: pkg_man.install('go') # Install for Debian & Ubuntu with pkg.apt() as pkg_man: pkg_man.install('golang') # TODO: make this a runner and require a switch to enable this pkg_man.install('golang-go-darwin-amd64', 'golang-go-freebsd-amd64', 'golang-go-netbsd-amd64', 'golang-go-windows-amd64') # Install for OSX with pkg.brew() as pkg_man: pkg_man.install('go') # Install for Windows with download.https() as downloader, pkg.msiexec() as installer: downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi') downloader.checksum('sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba') installer.install_flags('/qn' '/norestart') installer.await(downloader.finished()) Format go to the new designfrom genes import apt, brew, pacman, http_downloader, checksum, msiexec import platform opsys = platform.system() dist = platform.linux_distribution() if platform == 'Linux' and dist == 'Arch': pacman.update() pacman.sync('go') if platform == 'Linux' and (dist == 'Debian' or dist == 'Ubuntu'): apt.update() apt.install('golang') # TODO: make this a runner and require a switch to enable this apt.install('golang-go-darwin-amd64', 'golang-go-freebsd-amd64', 'golang-go-netbsd-amd64', 'golang-go-windows-amd64') if platform == 'Darwin': brew.update() brew.install('go') if platform == 'Windows': installer = http_downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi') checksum.check(installer, 'sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba') install_flags = ('/qn' '/norestart') msiexec.run(installer, install_flags)
<commit_before>from evolution_master.runners import pkg, download # Install for Arch with pkg.pacman() as pkg_man: pkg_man.install('go') # Install for Debian & Ubuntu with pkg.apt() as pkg_man: pkg_man.install('golang') # TODO: make this a runner and require a switch to enable this pkg_man.install('golang-go-darwin-amd64', 'golang-go-freebsd-amd64', 'golang-go-netbsd-amd64', 'golang-go-windows-amd64') # Install for OSX with pkg.brew() as pkg_man: pkg_man.install('go') # Install for Windows with download.https() as downloader, pkg.msiexec() as installer: downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi') downloader.checksum('sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba') installer.install_flags('/qn' '/norestart') installer.await(downloader.finished()) <commit_msg>Format go to the new design<commit_after>from genes import apt, brew, pacman, http_downloader, checksum, msiexec import platform opsys = platform.system() dist = platform.linux_distribution() if platform == 'Linux' and dist == 'Arch': pacman.update() pacman.sync('go') if platform == 'Linux' and (dist == 'Debian' or dist == 'Ubuntu'): apt.update() apt.install('golang') # TODO: make this a runner and require a switch to enable this apt.install('golang-go-darwin-amd64', 'golang-go-freebsd-amd64', 'golang-go-netbsd-amd64', 'golang-go-windows-amd64') if platform == 'Darwin': brew.update() brew.install('go') if platform == 'Windows': installer = http_downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi') checksum.check(installer, 'sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba') install_flags = ('/qn' '/norestart') msiexec.run(installer, install_flags)