commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 0
2.94k
| new_contents
stringlengths 1
4.43k
| subject
stringlengths 15
444
| message
stringlengths 16
3.45k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 5
43.2k
| prompt
stringlengths 17
4.58k
| response
stringlengths 1
4.43k
| prompt_tagged
stringlengths 58
4.62k
| response_tagged
stringlengths 1
4.43k
| text
stringlengths 132
7.29k
| text_tagged
stringlengths 173
7.33k
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
364731a5986a629f934ae8f82743385c5e4b7226
|
main.py
|
main.py
|
import argparse
from microproxy import proxy
from microproxy.viewer import log as log_viewer
def proxy_handler(args):
proxy.start_proxy_server(args.host, args.port, args.proxy_mode)
def viewer_handler(args):
if args.viewer_type == "log":
log_viewer.start()
def main():
parser = argparse.ArgumentParser(description="")
subparser = parser.add_subparsers()
proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server")
proxy_parser.add_argument("--host", default="127.0.0.1")
proxy_parser.add_argument("--port", type=int, default=5580)
proxy_parser.add_argument("--proxy_mode",
choices=["socks", "transparent"],
default="socks")
proxy_parser.set_defaults(func=proxy_handler)
viewer_parser = subparser.add_parser("viewer", help="Open Viewer")
viewer_parser.add_argument("--viewer_type",
choices=["log"],
default="log")
viewer_parser.set_defaults(func=viewer_handler)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()
|
import argparse
from microproxy import proxy
from microproxy.viewer import log as log_viewer
def proxy_handler(args):
proxy.start_proxy_server(args.host, args.port, args.proxy_mode)
def viewer_handler(args):
if args.viewer_mode == "log":
log_viewer.start()
def main():
parser = argparse.ArgumentParser(description="")
subparser = parser.add_subparsers()
proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server")
proxy_parser.add_argument("--host", default="127.0.0.1")
proxy_parser.add_argument("--port", type=int, default=5580)
proxy_parser.add_argument("--proxy-mode",
choices=["socks", "transparent"],
default="socks")
proxy_parser.set_defaults(func=proxy_handler)
viewer_parser = subparser.add_parser("viewer", help="Open Viewer")
viewer_parser.add_argument("--viewer-mode",
choices=["log"],
default="log")
viewer_parser.set_defaults(func=viewer_handler)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()
|
Change command option format. - proxy_mode into proxy-mode - viewer_type into viewer-mode
|
Change command option format.
- proxy_mode into proxy-mode
- viewer_type into viewer-mode
|
Python
|
mit
|
mike820324/microProxy,mike820324/microProxy
|
import argparse
from microproxy import proxy
from microproxy.viewer import log as log_viewer
def proxy_handler(args):
proxy.start_proxy_server(args.host, args.port, args.proxy_mode)
def viewer_handler(args):
if args.viewer_type == "log":
log_viewer.start()
def main():
parser = argparse.ArgumentParser(description="")
subparser = parser.add_subparsers()
proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server")
proxy_parser.add_argument("--host", default="127.0.0.1")
proxy_parser.add_argument("--port", type=int, default=5580)
proxy_parser.add_argument("--proxy_mode",
choices=["socks", "transparent"],
default="socks")
proxy_parser.set_defaults(func=proxy_handler)
viewer_parser = subparser.add_parser("viewer", help="Open Viewer")
viewer_parser.add_argument("--viewer_type",
choices=["log"],
default="log")
viewer_parser.set_defaults(func=viewer_handler)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()
Change command option format.
- proxy_mode into proxy-mode
- viewer_type into viewer-mode
|
import argparse
from microproxy import proxy
from microproxy.viewer import log as log_viewer
def proxy_handler(args):
proxy.start_proxy_server(args.host, args.port, args.proxy_mode)
def viewer_handler(args):
if args.viewer_mode == "log":
log_viewer.start()
def main():
parser = argparse.ArgumentParser(description="")
subparser = parser.add_subparsers()
proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server")
proxy_parser.add_argument("--host", default="127.0.0.1")
proxy_parser.add_argument("--port", type=int, default=5580)
proxy_parser.add_argument("--proxy-mode",
choices=["socks", "transparent"],
default="socks")
proxy_parser.set_defaults(func=proxy_handler)
viewer_parser = subparser.add_parser("viewer", help="Open Viewer")
viewer_parser.add_argument("--viewer-mode",
choices=["log"],
default="log")
viewer_parser.set_defaults(func=viewer_handler)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()
|
<commit_before>import argparse
from microproxy import proxy
from microproxy.viewer import log as log_viewer
def proxy_handler(args):
proxy.start_proxy_server(args.host, args.port, args.proxy_mode)
def viewer_handler(args):
if args.viewer_type == "log":
log_viewer.start()
def main():
parser = argparse.ArgumentParser(description="")
subparser = parser.add_subparsers()
proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server")
proxy_parser.add_argument("--host", default="127.0.0.1")
proxy_parser.add_argument("--port", type=int, default=5580)
proxy_parser.add_argument("--proxy_mode",
choices=["socks", "transparent"],
default="socks")
proxy_parser.set_defaults(func=proxy_handler)
viewer_parser = subparser.add_parser("viewer", help="Open Viewer")
viewer_parser.add_argument("--viewer_type",
choices=["log"],
default="log")
viewer_parser.set_defaults(func=viewer_handler)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()
<commit_msg>Change command option format.
- proxy_mode into proxy-mode
- viewer_type into viewer-mode<commit_after>
|
import argparse
from microproxy import proxy
from microproxy.viewer import log as log_viewer
def proxy_handler(args):
proxy.start_proxy_server(args.host, args.port, args.proxy_mode)
def viewer_handler(args):
if args.viewer_mode == "log":
log_viewer.start()
def main():
parser = argparse.ArgumentParser(description="")
subparser = parser.add_subparsers()
proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server")
proxy_parser.add_argument("--host", default="127.0.0.1")
proxy_parser.add_argument("--port", type=int, default=5580)
proxy_parser.add_argument("--proxy-mode",
choices=["socks", "transparent"],
default="socks")
proxy_parser.set_defaults(func=proxy_handler)
viewer_parser = subparser.add_parser("viewer", help="Open Viewer")
viewer_parser.add_argument("--viewer-mode",
choices=["log"],
default="log")
viewer_parser.set_defaults(func=viewer_handler)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()
|
import argparse
from microproxy import proxy
from microproxy.viewer import log as log_viewer
def proxy_handler(args):
proxy.start_proxy_server(args.host, args.port, args.proxy_mode)
def viewer_handler(args):
if args.viewer_type == "log":
log_viewer.start()
def main():
parser = argparse.ArgumentParser(description="")
subparser = parser.add_subparsers()
proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server")
proxy_parser.add_argument("--host", default="127.0.0.1")
proxy_parser.add_argument("--port", type=int, default=5580)
proxy_parser.add_argument("--proxy_mode",
choices=["socks", "transparent"],
default="socks")
proxy_parser.set_defaults(func=proxy_handler)
viewer_parser = subparser.add_parser("viewer", help="Open Viewer")
viewer_parser.add_argument("--viewer_type",
choices=["log"],
default="log")
viewer_parser.set_defaults(func=viewer_handler)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()
Change command option format.
- proxy_mode into proxy-mode
- viewer_type into viewer-modeimport argparse
from microproxy import proxy
from microproxy.viewer import log as log_viewer
def proxy_handler(args):
proxy.start_proxy_server(args.host, args.port, args.proxy_mode)
def viewer_handler(args):
if args.viewer_mode == "log":
log_viewer.start()
def main():
parser = argparse.ArgumentParser(description="")
subparser = parser.add_subparsers()
proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server")
proxy_parser.add_argument("--host", default="127.0.0.1")
proxy_parser.add_argument("--port", type=int, default=5580)
proxy_parser.add_argument("--proxy-mode",
choices=["socks", "transparent"],
default="socks")
proxy_parser.set_defaults(func=proxy_handler)
viewer_parser = subparser.add_parser("viewer", help="Open Viewer")
viewer_parser.add_argument("--viewer-mode",
choices=["log"],
default="log")
viewer_parser.set_defaults(func=viewer_handler)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()
|
<commit_before>import argparse
from microproxy import proxy
from microproxy.viewer import log as log_viewer
def proxy_handler(args):
proxy.start_proxy_server(args.host, args.port, args.proxy_mode)
def viewer_handler(args):
if args.viewer_type == "log":
log_viewer.start()
def main():
parser = argparse.ArgumentParser(description="")
subparser = parser.add_subparsers()
proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server")
proxy_parser.add_argument("--host", default="127.0.0.1")
proxy_parser.add_argument("--port", type=int, default=5580)
proxy_parser.add_argument("--proxy_mode",
choices=["socks", "transparent"],
default="socks")
proxy_parser.set_defaults(func=proxy_handler)
viewer_parser = subparser.add_parser("viewer", help="Open Viewer")
viewer_parser.add_argument("--viewer_type",
choices=["log"],
default="log")
viewer_parser.set_defaults(func=viewer_handler)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()
<commit_msg>Change command option format.
- proxy_mode into proxy-mode
- viewer_type into viewer-mode<commit_after>import argparse
from microproxy import proxy
from microproxy.viewer import log as log_viewer
def proxy_handler(args):
proxy.start_proxy_server(args.host, args.port, args.proxy_mode)
def viewer_handler(args):
if args.viewer_mode == "log":
log_viewer.start()
def main():
parser = argparse.ArgumentParser(description="")
subparser = parser.add_subparsers()
proxy_parser = subparser.add_parser('proxy', help="Enable Proxy Server")
proxy_parser.add_argument("--host", default="127.0.0.1")
proxy_parser.add_argument("--port", type=int, default=5580)
proxy_parser.add_argument("--proxy-mode",
choices=["socks", "transparent"],
default="socks")
proxy_parser.set_defaults(func=proxy_handler)
viewer_parser = subparser.add_parser("viewer", help="Open Viewer")
viewer_parser.add_argument("--viewer-mode",
choices=["log"],
default="log")
viewer_parser.set_defaults(func=viewer_handler)
args = parser.parse_args()
args.func(args)
if __name__ == "__main__":
main()
|
aa34f571e93d298884f08014865a86a4c92dfcbd
|
main.py
|
main.py
|
__author__ = 'Chad Peterson'
__email__ = 'chapeter@cisco.com'
from CHROnIC_Portal import app
app.secret_key = '1234'
app.run(host='0.0.0.0', port=5001, debug=True)
|
__author__ = 'Chad Peterson'
__email__ = 'chapeter@cisco.com'
from CHROnIC_Portal import app
app.secret_key = '1234'
app.run(host='0.0.0.0', port=5000, debug=True)
|
Put port back to 5000. Missed change from integration testing...
|
Put port back to 5000. Missed change from integration testing...
|
Python
|
mit
|
chapeter/CHROnIC_Portal,chapeter/CHROnIC_Portal
|
__author__ = 'Chad Peterson'
__email__ = 'chapeter@cisco.com'
from CHROnIC_Portal import app
app.secret_key = '1234'
app.run(host='0.0.0.0', port=5001, debug=True)
Put port back to 5000. Missed change from integration testing...
|
__author__ = 'Chad Peterson'
__email__ = 'chapeter@cisco.com'
from CHROnIC_Portal import app
app.secret_key = '1234'
app.run(host='0.0.0.0', port=5000, debug=True)
|
<commit_before>__author__ = 'Chad Peterson'
__email__ = 'chapeter@cisco.com'
from CHROnIC_Portal import app
app.secret_key = '1234'
app.run(host='0.0.0.0', port=5001, debug=True)
<commit_msg>Put port back to 5000. Missed change from integration testing...<commit_after>
|
__author__ = 'Chad Peterson'
__email__ = 'chapeter@cisco.com'
from CHROnIC_Portal import app
app.secret_key = '1234'
app.run(host='0.0.0.0', port=5000, debug=True)
|
__author__ = 'Chad Peterson'
__email__ = 'chapeter@cisco.com'
from CHROnIC_Portal import app
app.secret_key = '1234'
app.run(host='0.0.0.0', port=5001, debug=True)
Put port back to 5000. Missed change from integration testing...__author__ = 'Chad Peterson'
__email__ = 'chapeter@cisco.com'
from CHROnIC_Portal import app
app.secret_key = '1234'
app.run(host='0.0.0.0', port=5000, debug=True)
|
<commit_before>__author__ = 'Chad Peterson'
__email__ = 'chapeter@cisco.com'
from CHROnIC_Portal import app
app.secret_key = '1234'
app.run(host='0.0.0.0', port=5001, debug=True)
<commit_msg>Put port back to 5000. Missed change from integration testing...<commit_after>__author__ = 'Chad Peterson'
__email__ = 'chapeter@cisco.com'
from CHROnIC_Portal import app
app.secret_key = '1234'
app.run(host='0.0.0.0', port=5000, debug=True)
|
a4f010ed53615dcbe48c08a445e7d64045001133
|
base_comment_template/tests/test_base_comment_template.py
|
base_comment_template/tests/test_base_comment_template.py
|
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl).
from odoo.tests.common import TransactionCase
class TestResPartner(TransactionCase):
def setUp(self):
self.template_id = self.env['base.comment.template'].create({
'name': 'Comment before lines',
'position': 'before_lines',
'text': 'Text before lines',
})
def test_commercial_partner_fields(self):
# Azure Interior
partner_id = self.env.ref('base.res_partner_12')
partner_id.property_comment_template_id = self.template_id.id
# Test childs propagation of commercial partner field
for child_id in partner_id.child_ids:
self.assertEqual(
child_id.property_comment_template_id == self.template_id)
|
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl).
from odoo.tests.common import TransactionCase
class TestResPartner(TransactionCase):
def setUp(self):
super(TestResPartner, self).setUp()
self.template_id = self.env['base.comment.template'].create({
'name': 'Comment before lines',
'position': 'before_lines',
'text': 'Text before lines',
})
def test_commercial_partner_fields(self):
# Azure Interior
partner_id = self.env.ref('base.res_partner_12')
partner_id.property_comment_template_id = self.template_id.id
# Test childs propagation of commercial partner field
for child_id in partner_id.child_ids:
self.assertEqual(
child_id.property_comment_template_id, self.template_id)
|
Move comment_template_id field to the Invoicing tab
|
[IMP] account_invoice_comment_template: Move comment_template_id field to the Invoicing tab
[IMP] account_invoice_comment_template: rename partner field name from comment_template_id to invoice_comment_template_id
[IMP] account_invoice_comment_template: Make partner field company_dependant and move domain definition of invoice fields from the view to the model
[MOV] account_invoice_comment_template: comment_template_id to base_comment_template
[IMP] account_invoice_comment_template: Translate templates when partner changes
|
Python
|
agpl-3.0
|
OCA/reporting-engine,OCA/reporting-engine,OCA/reporting-engine,OCA/reporting-engine
|
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl).
from odoo.tests.common import TransactionCase
class TestResPartner(TransactionCase):
def setUp(self):
self.template_id = self.env['base.comment.template'].create({
'name': 'Comment before lines',
'position': 'before_lines',
'text': 'Text before lines',
})
def test_commercial_partner_fields(self):
# Azure Interior
partner_id = self.env.ref('base.res_partner_12')
partner_id.property_comment_template_id = self.template_id.id
# Test childs propagation of commercial partner field
for child_id in partner_id.child_ids:
self.assertEqual(
child_id.property_comment_template_id == self.template_id)
[IMP] account_invoice_comment_template: Move comment_template_id field to the Invoicing tab
[IMP] account_invoice_comment_template: rename partner field name from comment_template_id to invoice_comment_template_id
[IMP] account_invoice_comment_template: Make partner field company_dependant and move domain definition of invoice fields from the view to the model
[MOV] account_invoice_comment_template: comment_template_id to base_comment_template
[IMP] account_invoice_comment_template: Translate templates when partner changes
|
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl).
from odoo.tests.common import TransactionCase
class TestResPartner(TransactionCase):
def setUp(self):
super(TestResPartner, self).setUp()
self.template_id = self.env['base.comment.template'].create({
'name': 'Comment before lines',
'position': 'before_lines',
'text': 'Text before lines',
})
def test_commercial_partner_fields(self):
# Azure Interior
partner_id = self.env.ref('base.res_partner_12')
partner_id.property_comment_template_id = self.template_id.id
# Test childs propagation of commercial partner field
for child_id in partner_id.child_ids:
self.assertEqual(
child_id.property_comment_template_id, self.template_id)
|
<commit_before># License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl).
from odoo.tests.common import TransactionCase
class TestResPartner(TransactionCase):
def setUp(self):
self.template_id = self.env['base.comment.template'].create({
'name': 'Comment before lines',
'position': 'before_lines',
'text': 'Text before lines',
})
def test_commercial_partner_fields(self):
# Azure Interior
partner_id = self.env.ref('base.res_partner_12')
partner_id.property_comment_template_id = self.template_id.id
# Test childs propagation of commercial partner field
for child_id in partner_id.child_ids:
self.assertEqual(
child_id.property_comment_template_id == self.template_id)
<commit_msg>[IMP] account_invoice_comment_template: Move comment_template_id field to the Invoicing tab
[IMP] account_invoice_comment_template: rename partner field name from comment_template_id to invoice_comment_template_id
[IMP] account_invoice_comment_template: Make partner field company_dependant and move domain definition of invoice fields from the view to the model
[MOV] account_invoice_comment_template: comment_template_id to base_comment_template
[IMP] account_invoice_comment_template: Translate templates when partner changes<commit_after>
|
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl).
from odoo.tests.common import TransactionCase
class TestResPartner(TransactionCase):
def setUp(self):
super(TestResPartner, self).setUp()
self.template_id = self.env['base.comment.template'].create({
'name': 'Comment before lines',
'position': 'before_lines',
'text': 'Text before lines',
})
def test_commercial_partner_fields(self):
# Azure Interior
partner_id = self.env.ref('base.res_partner_12')
partner_id.property_comment_template_id = self.template_id.id
# Test childs propagation of commercial partner field
for child_id in partner_id.child_ids:
self.assertEqual(
child_id.property_comment_template_id, self.template_id)
|
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl).
from odoo.tests.common import TransactionCase
class TestResPartner(TransactionCase):
def setUp(self):
self.template_id = self.env['base.comment.template'].create({
'name': 'Comment before lines',
'position': 'before_lines',
'text': 'Text before lines',
})
def test_commercial_partner_fields(self):
# Azure Interior
partner_id = self.env.ref('base.res_partner_12')
partner_id.property_comment_template_id = self.template_id.id
# Test childs propagation of commercial partner field
for child_id in partner_id.child_ids:
self.assertEqual(
child_id.property_comment_template_id == self.template_id)
[IMP] account_invoice_comment_template: Move comment_template_id field to the Invoicing tab
[IMP] account_invoice_comment_template: rename partner field name from comment_template_id to invoice_comment_template_id
[IMP] account_invoice_comment_template: Make partner field company_dependant and move domain definition of invoice fields from the view to the model
[MOV] account_invoice_comment_template: comment_template_id to base_comment_template
[IMP] account_invoice_comment_template: Translate templates when partner changes# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl).
from odoo.tests.common import TransactionCase
class TestResPartner(TransactionCase):
def setUp(self):
super(TestResPartner, self).setUp()
self.template_id = self.env['base.comment.template'].create({
'name': 'Comment before lines',
'position': 'before_lines',
'text': 'Text before lines',
})
def test_commercial_partner_fields(self):
# Azure Interior
partner_id = self.env.ref('base.res_partner_12')
partner_id.property_comment_template_id = self.template_id.id
# Test childs propagation of commercial partner field
for child_id in partner_id.child_ids:
self.assertEqual(
child_id.property_comment_template_id, self.template_id)
|
<commit_before># License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl).
from odoo.tests.common import TransactionCase
class TestResPartner(TransactionCase):
def setUp(self):
self.template_id = self.env['base.comment.template'].create({
'name': 'Comment before lines',
'position': 'before_lines',
'text': 'Text before lines',
})
def test_commercial_partner_fields(self):
# Azure Interior
partner_id = self.env.ref('base.res_partner_12')
partner_id.property_comment_template_id = self.template_id.id
# Test childs propagation of commercial partner field
for child_id in partner_id.child_ids:
self.assertEqual(
child_id.property_comment_template_id == self.template_id)
<commit_msg>[IMP] account_invoice_comment_template: Move comment_template_id field to the Invoicing tab
[IMP] account_invoice_comment_template: rename partner field name from comment_template_id to invoice_comment_template_id
[IMP] account_invoice_comment_template: Make partner field company_dependant and move domain definition of invoice fields from the view to the model
[MOV] account_invoice_comment_template: comment_template_id to base_comment_template
[IMP] account_invoice_comment_template: Translate templates when partner changes<commit_after># License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl).
from odoo.tests.common import TransactionCase
class TestResPartner(TransactionCase):
def setUp(self):
super(TestResPartner, self).setUp()
self.template_id = self.env['base.comment.template'].create({
'name': 'Comment before lines',
'position': 'before_lines',
'text': 'Text before lines',
})
def test_commercial_partner_fields(self):
# Azure Interior
partner_id = self.env.ref('base.res_partner_12')
partner_id.property_comment_template_id = self.template_id.id
# Test childs propagation of commercial partner field
for child_id in partner_id.child_ids:
self.assertEqual(
child_id.property_comment_template_id, self.template_id)
|
0fd4ba14a16c6bfc100856dec0af6b17eb6917f2
|
codewars/valid_braces.py
|
codewars/valid_braces.py
|
# Valid Braces
# http://www.codewars.com/kata/5277c8a221e209d3f6000b56/train/python
import unittest
def valid_braces(string: str) -> bool:
stack = []
braces = {')': '(', '}': '{', ']': '['}
for l in string:
if l in braces.values():
stack.append(l)
elif stack:
if stack.pop() != braces[l]:
# We are closing a brace which we didn't open last.
return False
else:
# The stack was empty and we tried to remove a brace which is not valid.
return False
if stack:
# List is not empty.
return False
else:
return True
class ValidBracesTest(unittest.TestCase):
def test(self):
self.assertEqual(True, valid_braces("()"))
self.assertEqual(valid_braces("[(])"), False)
if __name__ == '__main__':
unittest.main()
|
Add solution for `Valid Braces`
|
Add solution for `Valid Braces`
|
Python
|
mit
|
davidlukac/codekata-python
|
Add solution for `Valid Braces`
|
# Valid Braces
# http://www.codewars.com/kata/5277c8a221e209d3f6000b56/train/python
import unittest
def valid_braces(string: str) -> bool:
stack = []
braces = {')': '(', '}': '{', ']': '['}
for l in string:
if l in braces.values():
stack.append(l)
elif stack:
if stack.pop() != braces[l]:
# We are closing a brace which we didn't open last.
return False
else:
# The stack was empty and we tried to remove a brace which is not valid.
return False
if stack:
# List is not empty.
return False
else:
return True
class ValidBracesTest(unittest.TestCase):
def test(self):
self.assertEqual(True, valid_braces("()"))
self.assertEqual(valid_braces("[(])"), False)
if __name__ == '__main__':
unittest.main()
|
<commit_before><commit_msg>Add solution for `Valid Braces`<commit_after>
|
# Valid Braces
# http://www.codewars.com/kata/5277c8a221e209d3f6000b56/train/python
import unittest
def valid_braces(string: str) -> bool:
stack = []
braces = {')': '(', '}': '{', ']': '['}
for l in string:
if l in braces.values():
stack.append(l)
elif stack:
if stack.pop() != braces[l]:
# We are closing a brace which we didn't open last.
return False
else:
# The stack was empty and we tried to remove a brace which is not valid.
return False
if stack:
# List is not empty.
return False
else:
return True
class ValidBracesTest(unittest.TestCase):
def test(self):
self.assertEqual(True, valid_braces("()"))
self.assertEqual(valid_braces("[(])"), False)
if __name__ == '__main__':
unittest.main()
|
Add solution for `Valid Braces`# Valid Braces
# http://www.codewars.com/kata/5277c8a221e209d3f6000b56/train/python
import unittest
def valid_braces(string: str) -> bool:
stack = []
braces = {')': '(', '}': '{', ']': '['}
for l in string:
if l in braces.values():
stack.append(l)
elif stack:
if stack.pop() != braces[l]:
# We are closing a brace which we didn't open last.
return False
else:
# The stack was empty and we tried to remove a brace which is not valid.
return False
if stack:
# List is not empty.
return False
else:
return True
class ValidBracesTest(unittest.TestCase):
def test(self):
self.assertEqual(True, valid_braces("()"))
self.assertEqual(valid_braces("[(])"), False)
if __name__ == '__main__':
unittest.main()
|
<commit_before><commit_msg>Add solution for `Valid Braces`<commit_after># Valid Braces
# http://www.codewars.com/kata/5277c8a221e209d3f6000b56/train/python
import unittest
def valid_braces(string: str) -> bool:
stack = []
braces = {')': '(', '}': '{', ']': '['}
for l in string:
if l in braces.values():
stack.append(l)
elif stack:
if stack.pop() != braces[l]:
# We are closing a brace which we didn't open last.
return False
else:
# The stack was empty and we tried to remove a brace which is not valid.
return False
if stack:
# List is not empty.
return False
else:
return True
class ValidBracesTest(unittest.TestCase):
def test(self):
self.assertEqual(True, valid_braces("()"))
self.assertEqual(valid_braces("[(])"), False)
if __name__ == '__main__':
unittest.main()
|
|
886d6e56e53742ec6cd2c59440459b17b093f4e0
|
blockbuster/__init__.py
|
blockbuster/__init__.py
|
__author__ = 'Matt Stibbs'
__version__ = '1.26.04'
target_schema_version = '1.25.00'
from flask import Flask
app = Flask(__name__)
def startup():
import blockbuster.bb_dbconnector_factory
import blockbuster.bb_logging as log
import blockbuster.bb_auditlogger as audit
blockbuster.app.debug = blockbuster.config.debug_mode
blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}",
blockbuster.__version__, target_schema_version))
blockbuster.bb_logging.logger.info(
'Application Setting - Time Restriction Disabled') \
if not blockbuster.config.timerestriction else blockbuster.bb_logging.logger.info(
'Application Setting - Time Restriction Enabled')
if blockbuster.config.debug_mode:
blockbuster.bb_logging.logger.info("========= APPLICATION IS RUNNING IN DEBUG MODE ==========")
try:
if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check():
import blockbuster.bb_routes
print("Running...")
else:
raise RuntimeError("Incorrect database schema version. Wanted ")
except RuntimeError, e:
log.logger.exception(e)
audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e))
startup()
|
__author__ = 'Matt Stibbs'
__version__ = '1.26.04'
target_schema_version = '1.25.00'
from flask import Flask
app = Flask(__name__)
def startup():
import blockbuster.bb_dbconnector_factory
import blockbuster.bb_logging as log
import blockbuster.bb_auditlogger as audit
blockbuster.app.debug = blockbuster.config.debug_mode
blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}",
blockbuster.__version__, target_schema_version))
time_setting = "Application Setting - Time Restriction Disabled" if not blockbuster.config.timerestriction else "Application Setting - Time Restriction Enabled"
print(time_setting)
if blockbuster.config.debug_mode:
print("========= APPLICATION IS RUNNING IN DEBUG MODE ==========")
try:
if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check():
import blockbuster.bb_routes
print("Running...")
else:
raise RuntimeError("Incorrect database schema version. Wanted ")
except RuntimeError, e:
log.logger.exception(e)
audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e))
startup()
|
Tweak console output on startup
|
Tweak console output on startup
|
Python
|
mit
|
mattstibbs/blockbuster-server,mattstibbs/blockbuster-server
|
__author__ = 'Matt Stibbs'
__version__ = '1.26.04'
target_schema_version = '1.25.00'
from flask import Flask
app = Flask(__name__)
def startup():
import blockbuster.bb_dbconnector_factory
import blockbuster.bb_logging as log
import blockbuster.bb_auditlogger as audit
blockbuster.app.debug = blockbuster.config.debug_mode
blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}",
blockbuster.__version__, target_schema_version))
blockbuster.bb_logging.logger.info(
'Application Setting - Time Restriction Disabled') \
if not blockbuster.config.timerestriction else blockbuster.bb_logging.logger.info(
'Application Setting - Time Restriction Enabled')
if blockbuster.config.debug_mode:
blockbuster.bb_logging.logger.info("========= APPLICATION IS RUNNING IN DEBUG MODE ==========")
try:
if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check():
import blockbuster.bb_routes
print("Running...")
else:
raise RuntimeError("Incorrect database schema version. Wanted ")
except RuntimeError, e:
log.logger.exception(e)
audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e))
startup()
Tweak console output on startup
|
__author__ = 'Matt Stibbs'
__version__ = '1.26.04'
target_schema_version = '1.25.00'
from flask import Flask
app = Flask(__name__)
def startup():
import blockbuster.bb_dbconnector_factory
import blockbuster.bb_logging as log
import blockbuster.bb_auditlogger as audit
blockbuster.app.debug = blockbuster.config.debug_mode
blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}",
blockbuster.__version__, target_schema_version))
time_setting = "Application Setting - Time Restriction Disabled" if not blockbuster.config.timerestriction else "Application Setting - Time Restriction Enabled"
print(time_setting)
if blockbuster.config.debug_mode:
print("========= APPLICATION IS RUNNING IN DEBUG MODE ==========")
try:
if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check():
import blockbuster.bb_routes
print("Running...")
else:
raise RuntimeError("Incorrect database schema version. Wanted ")
except RuntimeError, e:
log.logger.exception(e)
audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e))
startup()
|
<commit_before>__author__ = 'Matt Stibbs'
__version__ = '1.26.04'
target_schema_version = '1.25.00'
from flask import Flask
app = Flask(__name__)
def startup():
import blockbuster.bb_dbconnector_factory
import blockbuster.bb_logging as log
import blockbuster.bb_auditlogger as audit
blockbuster.app.debug = blockbuster.config.debug_mode
blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}",
blockbuster.__version__, target_schema_version))
blockbuster.bb_logging.logger.info(
'Application Setting - Time Restriction Disabled') \
if not blockbuster.config.timerestriction else blockbuster.bb_logging.logger.info(
'Application Setting - Time Restriction Enabled')
if blockbuster.config.debug_mode:
blockbuster.bb_logging.logger.info("========= APPLICATION IS RUNNING IN DEBUG MODE ==========")
try:
if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check():
import blockbuster.bb_routes
print("Running...")
else:
raise RuntimeError("Incorrect database schema version. Wanted ")
except RuntimeError, e:
log.logger.exception(e)
audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e))
startup()
<commit_msg>Tweak console output on startup<commit_after>
|
__author__ = 'Matt Stibbs'
__version__ = '1.26.04'
target_schema_version = '1.25.00'
from flask import Flask
app = Flask(__name__)
def startup():
import blockbuster.bb_dbconnector_factory
import blockbuster.bb_logging as log
import blockbuster.bb_auditlogger as audit
blockbuster.app.debug = blockbuster.config.debug_mode
blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}",
blockbuster.__version__, target_schema_version))
time_setting = "Application Setting - Time Restriction Disabled" if not blockbuster.config.timerestriction else "Application Setting - Time Restriction Enabled"
print(time_setting)
if blockbuster.config.debug_mode:
print("========= APPLICATION IS RUNNING IN DEBUG MODE ==========")
try:
if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check():
import blockbuster.bb_routes
print("Running...")
else:
raise RuntimeError("Incorrect database schema version. Wanted ")
except RuntimeError, e:
log.logger.exception(e)
audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e))
startup()
|
__author__ = 'Matt Stibbs'
__version__ = '1.26.04'
target_schema_version = '1.25.00'
from flask import Flask
app = Flask(__name__)
def startup():
import blockbuster.bb_dbconnector_factory
import blockbuster.bb_logging as log
import blockbuster.bb_auditlogger as audit
blockbuster.app.debug = blockbuster.config.debug_mode
blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}",
blockbuster.__version__, target_schema_version))
blockbuster.bb_logging.logger.info(
'Application Setting - Time Restriction Disabled') \
if not blockbuster.config.timerestriction else blockbuster.bb_logging.logger.info(
'Application Setting - Time Restriction Enabled')
if blockbuster.config.debug_mode:
blockbuster.bb_logging.logger.info("========= APPLICATION IS RUNNING IN DEBUG MODE ==========")
try:
if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check():
import blockbuster.bb_routes
print("Running...")
else:
raise RuntimeError("Incorrect database schema version. Wanted ")
except RuntimeError, e:
log.logger.exception(e)
audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e))
startup()
Tweak console output on startup__author__ = 'Matt Stibbs'
__version__ = '1.26.04'
target_schema_version = '1.25.00'
from flask import Flask
app = Flask(__name__)
def startup():
import blockbuster.bb_dbconnector_factory
import blockbuster.bb_logging as log
import blockbuster.bb_auditlogger as audit
blockbuster.app.debug = blockbuster.config.debug_mode
blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}",
blockbuster.__version__, target_schema_version))
time_setting = "Application Setting - Time Restriction Disabled" if not blockbuster.config.timerestriction else "Application Setting - Time Restriction Enabled"
print(time_setting)
if blockbuster.config.debug_mode:
print("========= APPLICATION IS RUNNING IN DEBUG MODE ==========")
try:
if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check():
import blockbuster.bb_routes
print("Running...")
else:
raise RuntimeError("Incorrect database schema version. Wanted ")
except RuntimeError, e:
log.logger.exception(e)
audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e))
startup()
|
<commit_before>__author__ = 'Matt Stibbs'
__version__ = '1.26.04'
target_schema_version = '1.25.00'
from flask import Flask
app = Flask(__name__)
def startup():
import blockbuster.bb_dbconnector_factory
import blockbuster.bb_logging as log
import blockbuster.bb_auditlogger as audit
blockbuster.app.debug = blockbuster.config.debug_mode
blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}",
blockbuster.__version__, target_schema_version))
blockbuster.bb_logging.logger.info(
'Application Setting - Time Restriction Disabled') \
if not blockbuster.config.timerestriction else blockbuster.bb_logging.logger.info(
'Application Setting - Time Restriction Enabled')
if blockbuster.config.debug_mode:
blockbuster.bb_logging.logger.info("========= APPLICATION IS RUNNING IN DEBUG MODE ==========")
try:
if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check():
import blockbuster.bb_routes
print("Running...")
else:
raise RuntimeError("Incorrect database schema version. Wanted ")
except RuntimeError, e:
log.logger.exception(e)
audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e))
startup()
<commit_msg>Tweak console output on startup<commit_after>__author__ = 'Matt Stibbs'
__version__ = '1.26.04'
target_schema_version = '1.25.00'
from flask import Flask
app = Flask(__name__)
def startup():
import blockbuster.bb_dbconnector_factory
import blockbuster.bb_logging as log
import blockbuster.bb_auditlogger as audit
blockbuster.app.debug = blockbuster.config.debug_mode
blockbuster.bb_logging.logger.info(str.format("Application Startup - BlockBuster v{0} Schema v{1}",
blockbuster.__version__, target_schema_version))
time_setting = "Application Setting - Time Restriction Disabled" if not blockbuster.config.timerestriction else "Application Setting - Time Restriction Enabled"
print(time_setting)
if blockbuster.config.debug_mode:
print("========= APPLICATION IS RUNNING IN DEBUG MODE ==========")
try:
if blockbuster.bb_dbconnector_factory.DBConnectorInterfaceFactory().create().db_version_check():
import blockbuster.bb_routes
print("Running...")
else:
raise RuntimeError("Incorrect database schema version. Wanted ")
except RuntimeError, e:
log.logger.exception(e)
audit.BBAuditLoggerFactory().create().logException('app', 'STARTUP', str(e))
startup()
|
40095b001ab95fda4cc80bcc807508e9580ebf2d
|
fireplace/cards/gvg/neutral_legendary.py
|
fireplace/cards/gvg/neutral_legendary.py
|
from ..utils import *
##
# Minions
# Dr. Boom
class GVG_110:
action = [Summon(CONTROLLER, "GVG_110t") * 2]
# Boom Bot
class GVG_110t:
def deathrattle(self):
return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))]
# Sneed's Old Shredder
class GVG_114:
def deathrattle(self):
legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY)
return [Summon(CONTROLLER, legendary)]
|
from ..utils import *
##
# Minions
# Dr. Boom
class GVG_110:
action = [Summon(CONTROLLER, "GVG_110t") * 2]
# Boom Bot
class GVG_110t:
def deathrattle(self):
return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))]
# Sneed's Old Shredder
class GVG_114:
def deathrattle(self):
legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY)
return [Summon(CONTROLLER, legendary)]
# Toshley
class GVG_115:
action = [GiveSparePart(CONTROLLER)]
deathrattle = [GiveSparePart(CONTROLLER)]
# Mekgineer Thermaplugg
class GVG_116:
def MINION_DESTROY(self, minion):
if minion.controller is not self.controller:
return [Summon(CONTROLLER, "EX1_029")]
# Gazlowe
class GVG_117:
def OWN_CARD_PLAYED(self, card):
if card.type == CardType.SPELL and card.cost == 1:
return [Give(CONTROLLER, randomCollectible(race=Race.MECHANICAL))]
|
Implement Toshley, Mekgineer Thermaplugg and Gazlowe
|
Implement Toshley, Mekgineer Thermaplugg and Gazlowe
|
Python
|
agpl-3.0
|
amw2104/fireplace,oftc-ftw/fireplace,smallnamespace/fireplace,smallnamespace/fireplace,liujimj/fireplace,oftc-ftw/fireplace,amw2104/fireplace,butozerca/fireplace,Ragowit/fireplace,liujimj/fireplace,jleclanche/fireplace,butozerca/fireplace,Ragowit/fireplace,NightKev/fireplace,beheh/fireplace,Meerkov/fireplace,Meerkov/fireplace
|
from ..utils import *
##
# Minions
# Dr. Boom
class GVG_110:
action = [Summon(CONTROLLER, "GVG_110t") * 2]
# Boom Bot
class GVG_110t:
def deathrattle(self):
return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))]
# Sneed's Old Shredder
class GVG_114:
def deathrattle(self):
legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY)
return [Summon(CONTROLLER, legendary)]
Implement Toshley, Mekgineer Thermaplugg and Gazlowe
|
from ..utils import *
##
# Minions
# Dr. Boom
class GVG_110:
action = [Summon(CONTROLLER, "GVG_110t") * 2]
# Boom Bot
class GVG_110t:
def deathrattle(self):
return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))]
# Sneed's Old Shredder
class GVG_114:
def deathrattle(self):
legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY)
return [Summon(CONTROLLER, legendary)]
# Toshley
class GVG_115:
action = [GiveSparePart(CONTROLLER)]
deathrattle = [GiveSparePart(CONTROLLER)]
# Mekgineer Thermaplugg
class GVG_116:
def MINION_DESTROY(self, minion):
if minion.controller is not self.controller:
return [Summon(CONTROLLER, "EX1_029")]
# Gazlowe
class GVG_117:
def OWN_CARD_PLAYED(self, card):
if card.type == CardType.SPELL and card.cost == 1:
return [Give(CONTROLLER, randomCollectible(race=Race.MECHANICAL))]
|
<commit_before>from ..utils import *
##
# Minions
# Dr. Boom
class GVG_110:
action = [Summon(CONTROLLER, "GVG_110t") * 2]
# Boom Bot
class GVG_110t:
def deathrattle(self):
return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))]
# Sneed's Old Shredder
class GVG_114:
def deathrattle(self):
legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY)
return [Summon(CONTROLLER, legendary)]
<commit_msg>Implement Toshley, Mekgineer Thermaplugg and Gazlowe<commit_after>
|
from ..utils import *
##
# Minions
# Dr. Boom
class GVG_110:
action = [Summon(CONTROLLER, "GVG_110t") * 2]
# Boom Bot
class GVG_110t:
def deathrattle(self):
return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))]
# Sneed's Old Shredder
class GVG_114:
def deathrattle(self):
legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY)
return [Summon(CONTROLLER, legendary)]
# Toshley
class GVG_115:
action = [GiveSparePart(CONTROLLER)]
deathrattle = [GiveSparePart(CONTROLLER)]
# Mekgineer Thermaplugg
class GVG_116:
def MINION_DESTROY(self, minion):
if minion.controller is not self.controller:
return [Summon(CONTROLLER, "EX1_029")]
# Gazlowe
class GVG_117:
def OWN_CARD_PLAYED(self, card):
if card.type == CardType.SPELL and card.cost == 1:
return [Give(CONTROLLER, randomCollectible(race=Race.MECHANICAL))]
|
from ..utils import *
##
# Minions
# Dr. Boom
class GVG_110:
action = [Summon(CONTROLLER, "GVG_110t") * 2]
# Boom Bot
class GVG_110t:
def deathrattle(self):
return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))]
# Sneed's Old Shredder
class GVG_114:
def deathrattle(self):
legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY)
return [Summon(CONTROLLER, legendary)]
Implement Toshley, Mekgineer Thermaplugg and Gazlowefrom ..utils import *
##
# Minions
# Dr. Boom
class GVG_110:
action = [Summon(CONTROLLER, "GVG_110t") * 2]
# Boom Bot
class GVG_110t:
def deathrattle(self):
return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))]
# Sneed's Old Shredder
class GVG_114:
def deathrattle(self):
legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY)
return [Summon(CONTROLLER, legendary)]
# Toshley
class GVG_115:
action = [GiveSparePart(CONTROLLER)]
deathrattle = [GiveSparePart(CONTROLLER)]
# Mekgineer Thermaplugg
class GVG_116:
def MINION_DESTROY(self, minion):
if minion.controller is not self.controller:
return [Summon(CONTROLLER, "EX1_029")]
# Gazlowe
class GVG_117:
def OWN_CARD_PLAYED(self, card):
if card.type == CardType.SPELL and card.cost == 1:
return [Give(CONTROLLER, randomCollectible(race=Race.MECHANICAL))]
|
<commit_before>from ..utils import *
##
# Minions
# Dr. Boom
class GVG_110:
action = [Summon(CONTROLLER, "GVG_110t") * 2]
# Boom Bot
class GVG_110t:
def deathrattle(self):
return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))]
# Sneed's Old Shredder
class GVG_114:
def deathrattle(self):
legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY)
return [Summon(CONTROLLER, legendary)]
<commit_msg>Implement Toshley, Mekgineer Thermaplugg and Gazlowe<commit_after>from ..utils import *
##
# Minions
# Dr. Boom
class GVG_110:
action = [Summon(CONTROLLER, "GVG_110t") * 2]
# Boom Bot
class GVG_110t:
def deathrattle(self):
return [Hit(RANDOM_ENEMY_CHARACTER, random.randint(1, 4))]
# Sneed's Old Shredder
class GVG_114:
def deathrattle(self):
legendary = randomCollectible(type=CardType.MINION, rarity=Rarity.LEGENDARY)
return [Summon(CONTROLLER, legendary)]
# Toshley
class GVG_115:
action = [GiveSparePart(CONTROLLER)]
deathrattle = [GiveSparePart(CONTROLLER)]
# Mekgineer Thermaplugg
class GVG_116:
def MINION_DESTROY(self, minion):
if minion.controller is not self.controller:
return [Summon(CONTROLLER, "EX1_029")]
# Gazlowe
class GVG_117:
def OWN_CARD_PLAYED(self, card):
if card.type == CardType.SPELL and card.cost == 1:
return [Give(CONTROLLER, randomCollectible(race=Race.MECHANICAL))]
|
dd445cbf33268ece3a6b006d3d31d6169fec03b8
|
acoustid/scripts/backfill_meta_created.py
|
acoustid/scripts/backfill_meta_created.py
|
#!/usr/bin/env python
# Copyright (C) 2019 Lukas Lalinsky
# Distributed under the MIT license, see the LICENSE file for details.
import logging
logger = logging.getLogger(__name__)
def run_backfill_meta_created(script, opts, args):
if script.config.cluster.role != 'master':
logger.info('Not running backfill_meta_created in slave mode')
return
query = """
WITH meta_created AS (
SELECT meta_id, min(created) AS created
FROM track_meta
WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000)
GROUP BY meta_id
)
UPDATE meta
SET created = meta_created.created
FROM meta_created
WHERE meta.id = meta_created.meta_id AND meta.created IS NULL
"""
for i in range(10):
with script.context() as ctx:
fingerprint_db = ctx.db.get_fingerprint_db()
fingerprint_db.execute(query)
ctx.db.session.commit()
|
#!/usr/bin/env python
# Copyright (C) 2019 Lukas Lalinsky
# Distributed under the MIT license, see the LICENSE file for details.
import logging
logger = logging.getLogger(__name__)
def run_backfill_meta_created(script, opts, args):
if script.config.cluster.role != 'master':
logger.info('Not running backfill_meta_created in slave mode')
return
query = """
WITH meta_created AS (
SELECT meta_id, min(created) AS created
FROM track_meta
WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000)
GROUP BY meta_id
)
UPDATE meta
SET created = meta_created.created
FROM meta_created
WHERE meta.id = meta_created.meta_id AND meta.created IS NULL
"""
for i in range(100):
with script.context() as ctx:
fingerprint_db = ctx.db.get_fingerprint_db()
fingerprint_db.execute(query)
ctx.db.session.commit()
|
Increase the number of backill_meta_created iterations
|
Increase the number of backill_meta_created iterations
|
Python
|
mit
|
lalinsky/acoustid-server,lalinsky/acoustid-server,lalinsky/acoustid-server,lalinsky/acoustid-server
|
#!/usr/bin/env python
# Copyright (C) 2019 Lukas Lalinsky
# Distributed under the MIT license, see the LICENSE file for details.
import logging
logger = logging.getLogger(__name__)
def run_backfill_meta_created(script, opts, args):
if script.config.cluster.role != 'master':
logger.info('Not running backfill_meta_created in slave mode')
return
query = """
WITH meta_created AS (
SELECT meta_id, min(created) AS created
FROM track_meta
WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000)
GROUP BY meta_id
)
UPDATE meta
SET created = meta_created.created
FROM meta_created
WHERE meta.id = meta_created.meta_id AND meta.created IS NULL
"""
for i in range(10):
with script.context() as ctx:
fingerprint_db = ctx.db.get_fingerprint_db()
fingerprint_db.execute(query)
ctx.db.session.commit()
Increase the number of backill_meta_created iterations
|
#!/usr/bin/env python
# Copyright (C) 2019 Lukas Lalinsky
# Distributed under the MIT license, see the LICENSE file for details.
import logging
logger = logging.getLogger(__name__)
def run_backfill_meta_created(script, opts, args):
if script.config.cluster.role != 'master':
logger.info('Not running backfill_meta_created in slave mode')
return
query = """
WITH meta_created AS (
SELECT meta_id, min(created) AS created
FROM track_meta
WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000)
GROUP BY meta_id
)
UPDATE meta
SET created = meta_created.created
FROM meta_created
WHERE meta.id = meta_created.meta_id AND meta.created IS NULL
"""
for i in range(100):
with script.context() as ctx:
fingerprint_db = ctx.db.get_fingerprint_db()
fingerprint_db.execute(query)
ctx.db.session.commit()
|
<commit_before>#!/usr/bin/env python
# Copyright (C) 2019 Lukas Lalinsky
# Distributed under the MIT license, see the LICENSE file for details.
import logging
logger = logging.getLogger(__name__)
def run_backfill_meta_created(script, opts, args):
if script.config.cluster.role != 'master':
logger.info('Not running backfill_meta_created in slave mode')
return
query = """
WITH meta_created AS (
SELECT meta_id, min(created) AS created
FROM track_meta
WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000)
GROUP BY meta_id
)
UPDATE meta
SET created = meta_created.created
FROM meta_created
WHERE meta.id = meta_created.meta_id AND meta.created IS NULL
"""
for i in range(10):
with script.context() as ctx:
fingerprint_db = ctx.db.get_fingerprint_db()
fingerprint_db.execute(query)
ctx.db.session.commit()
<commit_msg>Increase the number of backill_meta_created iterations<commit_after>
|
#!/usr/bin/env python
# Copyright (C) 2019 Lukas Lalinsky
# Distributed under the MIT license, see the LICENSE file for details.
import logging
logger = logging.getLogger(__name__)
def run_backfill_meta_created(script, opts, args):
if script.config.cluster.role != 'master':
logger.info('Not running backfill_meta_created in slave mode')
return
query = """
WITH meta_created AS (
SELECT meta_id, min(created) AS created
FROM track_meta
WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000)
GROUP BY meta_id
)
UPDATE meta
SET created = meta_created.created
FROM meta_created
WHERE meta.id = meta_created.meta_id AND meta.created IS NULL
"""
for i in range(100):
with script.context() as ctx:
fingerprint_db = ctx.db.get_fingerprint_db()
fingerprint_db.execute(query)
ctx.db.session.commit()
|
#!/usr/bin/env python
# Copyright (C) 2019 Lukas Lalinsky
# Distributed under the MIT license, see the LICENSE file for details.
import logging
logger = logging.getLogger(__name__)
def run_backfill_meta_created(script, opts, args):
if script.config.cluster.role != 'master':
logger.info('Not running backfill_meta_created in slave mode')
return
query = """
WITH meta_created AS (
SELECT meta_id, min(created) AS created
FROM track_meta
WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000)
GROUP BY meta_id
)
UPDATE meta
SET created = meta_created.created
FROM meta_created
WHERE meta.id = meta_created.meta_id AND meta.created IS NULL
"""
for i in range(10):
with script.context() as ctx:
fingerprint_db = ctx.db.get_fingerprint_db()
fingerprint_db.execute(query)
ctx.db.session.commit()
Increase the number of backill_meta_created iterations#!/usr/bin/env python
# Copyright (C) 2019 Lukas Lalinsky
# Distributed under the MIT license, see the LICENSE file for details.
import logging
logger = logging.getLogger(__name__)
def run_backfill_meta_created(script, opts, args):
if script.config.cluster.role != 'master':
logger.info('Not running backfill_meta_created in slave mode')
return
query = """
WITH meta_created AS (
SELECT meta_id, min(created) AS created
FROM track_meta
WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000)
GROUP BY meta_id
)
UPDATE meta
SET created = meta_created.created
FROM meta_created
WHERE meta.id = meta_created.meta_id AND meta.created IS NULL
"""
for i in range(100):
with script.context() as ctx:
fingerprint_db = ctx.db.get_fingerprint_db()
fingerprint_db.execute(query)
ctx.db.session.commit()
|
<commit_before>#!/usr/bin/env python
# Copyright (C) 2019 Lukas Lalinsky
# Distributed under the MIT license, see the LICENSE file for details.
import logging
logger = logging.getLogger(__name__)
def run_backfill_meta_created(script, opts, args):
if script.config.cluster.role != 'master':
logger.info('Not running backfill_meta_created in slave mode')
return
query = """
WITH meta_created AS (
SELECT meta_id, min(created) AS created
FROM track_meta
WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000)
GROUP BY meta_id
)
UPDATE meta
SET created = meta_created.created
FROM meta_created
WHERE meta.id = meta_created.meta_id AND meta.created IS NULL
"""
for i in range(10):
with script.context() as ctx:
fingerprint_db = ctx.db.get_fingerprint_db()
fingerprint_db.execute(query)
ctx.db.session.commit()
<commit_msg>Increase the number of backill_meta_created iterations<commit_after>#!/usr/bin/env python
# Copyright (C) 2019 Lukas Lalinsky
# Distributed under the MIT license, see the LICENSE file for details.
import logging
logger = logging.getLogger(__name__)
def run_backfill_meta_created(script, opts, args):
if script.config.cluster.role != 'master':
logger.info('Not running backfill_meta_created in slave mode')
return
query = """
WITH meta_created AS (
SELECT meta_id, min(created) AS created
FROM track_meta
WHERE meta_id IN (SELECT id FROM meta WHERE created IS NULL LIMIT 10000)
GROUP BY meta_id
)
UPDATE meta
SET created = meta_created.created
FROM meta_created
WHERE meta.id = meta_created.meta_id AND meta.created IS NULL
"""
for i in range(100):
with script.context() as ctx:
fingerprint_db = ctx.db.get_fingerprint_db()
fingerprint_db.execute(query)
ctx.db.session.commit()
|
1ad03769569d86d1eda45f7c6582234ed455ea88
|
src/main.py
|
src/main.py
|
"""Where player runs the game"""
import random
import time
import board
import conversion
import games
if __name__ == '__main__':
NUMBER_OF_TRIALS = 1
for i in range(NUMBER_OF_TRIALS):
X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square()
LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS)
ORIENTATIONS = (True, False)
GAMES = games.GamePositions()
BOARD_DESCRIPTION = GAMES.random_game()
BOARD = board.Board(BOARD_DESCRIPTION,
orientation=random.choice(ORIENTATIONS))
BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS)
BOARD.update_board_string()
print(BOARD)
start = time.time()
ATTEMPT = input('Enter highlighted square: ')
end = time.time()
if ATTEMPT == LOCATION:
print("Correct!")
else:
print("Wrong, the answer was {0}".format(LOCATION))
print("Attempt took {0} seconds".format(end - start))
|
"""Where player runs the game"""
import random
import time
import board
import conversion
import games
from settings import RECORD_FILE
def write_record_to_file(a_string, file_name):
with open(file_name, 'w') as f:
f.write(a_string)
def get_record_from_file(file_name):
with open(file_name, 'r') as f:
result = f.readline()
return result
if __name__ == '__main__':
NUMBER_OF_TRIALS = 1
for i in range(NUMBER_OF_TRIALS):
X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square()
LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS)
ORIENTATIONS = (True, False)
GAMES = games.GamePositions()
BOARD_DESCRIPTION = GAMES.random_game()
BOARD = board.Board(BOARD_DESCRIPTION,
orientation=random.choice(ORIENTATIONS))
BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS)
BOARD.update_board_string()
print(BOARD)
start = time.time()
ATTEMPT = input('Enter highlighted square: ')
end = time.time()
if ATTEMPT == LOCATION:
print("Correct!")
else:
print("Wrong, the answer was {0}".format(LOCATION))
print("Attempt took {0} seconds".format(end - start))
|
Add functions to read/write to a record file
|
Add functions to read/write to a record file
|
Python
|
mit
|
blairck/chess_notation
|
"""Where player runs the game"""
import random
import time
import board
import conversion
import games
if __name__ == '__main__':
NUMBER_OF_TRIALS = 1
for i in range(NUMBER_OF_TRIALS):
X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square()
LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS)
ORIENTATIONS = (True, False)
GAMES = games.GamePositions()
BOARD_DESCRIPTION = GAMES.random_game()
BOARD = board.Board(BOARD_DESCRIPTION,
orientation=random.choice(ORIENTATIONS))
BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS)
BOARD.update_board_string()
print(BOARD)
start = time.time()
ATTEMPT = input('Enter highlighted square: ')
end = time.time()
if ATTEMPT == LOCATION:
print("Correct!")
else:
print("Wrong, the answer was {0}".format(LOCATION))
print("Attempt took {0} seconds".format(end - start))
Add functions to read/write to a record file
|
"""Where player runs the game"""
import random
import time
import board
import conversion
import games
from settings import RECORD_FILE
def write_record_to_file(a_string, file_name):
with open(file_name, 'w') as f:
f.write(a_string)
def get_record_from_file(file_name):
with open(file_name, 'r') as f:
result = f.readline()
return result
if __name__ == '__main__':
NUMBER_OF_TRIALS = 1
for i in range(NUMBER_OF_TRIALS):
X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square()
LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS)
ORIENTATIONS = (True, False)
GAMES = games.GamePositions()
BOARD_DESCRIPTION = GAMES.random_game()
BOARD = board.Board(BOARD_DESCRIPTION,
orientation=random.choice(ORIENTATIONS))
BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS)
BOARD.update_board_string()
print(BOARD)
start = time.time()
ATTEMPT = input('Enter highlighted square: ')
end = time.time()
if ATTEMPT == LOCATION:
print("Correct!")
else:
print("Wrong, the answer was {0}".format(LOCATION))
print("Attempt took {0} seconds".format(end - start))
|
<commit_before>"""Where player runs the game"""
import random
import time
import board
import conversion
import games
if __name__ == '__main__':
NUMBER_OF_TRIALS = 1
for i in range(NUMBER_OF_TRIALS):
X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square()
LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS)
ORIENTATIONS = (True, False)
GAMES = games.GamePositions()
BOARD_DESCRIPTION = GAMES.random_game()
BOARD = board.Board(BOARD_DESCRIPTION,
orientation=random.choice(ORIENTATIONS))
BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS)
BOARD.update_board_string()
print(BOARD)
start = time.time()
ATTEMPT = input('Enter highlighted square: ')
end = time.time()
if ATTEMPT == LOCATION:
print("Correct!")
else:
print("Wrong, the answer was {0}".format(LOCATION))
print("Attempt took {0} seconds".format(end - start))
<commit_msg>Add functions to read/write to a record file<commit_after>
|
"""Where player runs the game"""
import random
import time
import board
import conversion
import games
from settings import RECORD_FILE
def write_record_to_file(a_string, file_name):
with open(file_name, 'w') as f:
f.write(a_string)
def get_record_from_file(file_name):
with open(file_name, 'r') as f:
result = f.readline()
return result
if __name__ == '__main__':
NUMBER_OF_TRIALS = 1
for i in range(NUMBER_OF_TRIALS):
X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square()
LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS)
ORIENTATIONS = (True, False)
GAMES = games.GamePositions()
BOARD_DESCRIPTION = GAMES.random_game()
BOARD = board.Board(BOARD_DESCRIPTION,
orientation=random.choice(ORIENTATIONS))
BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS)
BOARD.update_board_string()
print(BOARD)
start = time.time()
ATTEMPT = input('Enter highlighted square: ')
end = time.time()
if ATTEMPT == LOCATION:
print("Correct!")
else:
print("Wrong, the answer was {0}".format(LOCATION))
print("Attempt took {0} seconds".format(end - start))
|
"""Where player runs the game"""
import random
import time
import board
import conversion
import games
if __name__ == '__main__':
NUMBER_OF_TRIALS = 1
for i in range(NUMBER_OF_TRIALS):
X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square()
LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS)
ORIENTATIONS = (True, False)
GAMES = games.GamePositions()
BOARD_DESCRIPTION = GAMES.random_game()
BOARD = board.Board(BOARD_DESCRIPTION,
orientation=random.choice(ORIENTATIONS))
BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS)
BOARD.update_board_string()
print(BOARD)
start = time.time()
ATTEMPT = input('Enter highlighted square: ')
end = time.time()
if ATTEMPT == LOCATION:
print("Correct!")
else:
print("Wrong, the answer was {0}".format(LOCATION))
print("Attempt took {0} seconds".format(end - start))
Add functions to read/write to a record file"""Where player runs the game"""
import random
import time
import board
import conversion
import games
from settings import RECORD_FILE
def write_record_to_file(a_string, file_name):
with open(file_name, 'w') as f:
f.write(a_string)
def get_record_from_file(file_name):
with open(file_name, 'r') as f:
result = f.readline()
return result
if __name__ == '__main__':
NUMBER_OF_TRIALS = 1
for i in range(NUMBER_OF_TRIALS):
X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square()
LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS)
ORIENTATIONS = (True, False)
GAMES = games.GamePositions()
BOARD_DESCRIPTION = GAMES.random_game()
BOARD = board.Board(BOARD_DESCRIPTION,
orientation=random.choice(ORIENTATIONS))
BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS)
BOARD.update_board_string()
print(BOARD)
start = time.time()
ATTEMPT = input('Enter highlighted square: ')
end = time.time()
if ATTEMPT == LOCATION:
print("Correct!")
else:
print("Wrong, the answer was {0}".format(LOCATION))
print("Attempt took {0} seconds".format(end - start))
|
<commit_before>"""Where player runs the game"""
import random
import time
import board
import conversion
import games
if __name__ == '__main__':
NUMBER_OF_TRIALS = 1
for i in range(NUMBER_OF_TRIALS):
X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square()
LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS)
ORIENTATIONS = (True, False)
GAMES = games.GamePositions()
BOARD_DESCRIPTION = GAMES.random_game()
BOARD = board.Board(BOARD_DESCRIPTION,
orientation=random.choice(ORIENTATIONS))
BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS)
BOARD.update_board_string()
print(BOARD)
start = time.time()
ATTEMPT = input('Enter highlighted square: ')
end = time.time()
if ATTEMPT == LOCATION:
print("Correct!")
else:
print("Wrong, the answer was {0}".format(LOCATION))
print("Attempt took {0} seconds".format(end - start))
<commit_msg>Add functions to read/write to a record file<commit_after>"""Where player runs the game"""
import random
import time
import board
import conversion
import games
from settings import RECORD_FILE
def write_record_to_file(a_string, file_name):
with open(file_name, 'w') as f:
f.write(a_string)
def get_record_from_file(file_name):
with open(file_name, 'r') as f:
result = f.readline()
return result
if __name__ == '__main__':
NUMBER_OF_TRIALS = 1
for i in range(NUMBER_OF_TRIALS):
X_LOC_CHESS, Y_LOC_CHESS = board.identify_random_square()
LOCATION = conversion.coordinate_to_alg(X_LOC_CHESS, Y_LOC_CHESS)
ORIENTATIONS = (True, False)
GAMES = games.GamePositions()
BOARD_DESCRIPTION = GAMES.random_game()
BOARD = board.Board(BOARD_DESCRIPTION,
orientation=random.choice(ORIENTATIONS))
BOARD.highlight_square(X_LOC_CHESS, Y_LOC_CHESS)
BOARD.update_board_string()
print(BOARD)
start = time.time()
ATTEMPT = input('Enter highlighted square: ')
end = time.time()
if ATTEMPT == LOCATION:
print("Correct!")
else:
print("Wrong, the answer was {0}".format(LOCATION))
print("Attempt took {0} seconds".format(end - start))
|
96d12496e425806a635ba345a534c0ca2790754d
|
satchmo/apps/payment/modules/giftcertificate/processor.py
|
satchmo/apps/payment/modules/giftcertificate/processor.py
|
"""
GiftCertificate processor
"""
from django.utils.translation import ugettext as _
from l10n.utils import moneyfmt
from models import GiftCertificate
from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET
class PaymentProcessor(BasePaymentProcessor):
def __init__(self, settings):
super(PaymentProcessor, self).__init__('giftcertificate', settings)
def capture_payment(self, testing=False, order=None, amount=NOTSET):
"""
Process the transaction and return a ProcessorResponse
"""
if not order:
order = self.order
if amount==NOTSET:
amount = order.balance
payment = None
if self.order.paid_in_full:
success = True
reason_code = "0"
response_text = _("No balance to pay")
else:
try:
gc = GiftCertificate.objects.from_order(self.order)
except GiftCertificate.DoesNotExist:
success = False
reason_code="1"
response_text = _("No such Gift Certificate")
if not gc.valid:
success = False
reason_code="2"
response_text = _("Bad Gift Certificate")
else:
gc.apply_to_order(self.order)
payment = gc.orderpayment
reason_code = "0"
response_text = _("Success")
success = True
if not self.order.paid_in_full:
response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance)
return ProcessorResult(self.key, success, response_text, payment=payment)
|
"""
GiftCertificate processor
"""
from django.utils.translation import ugettext as _
from l10n.utils import moneyfmt
from models import GiftCertificate
from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET
class PaymentProcessor(BasePaymentProcessor):
def __init__(self, settings):
super(PaymentProcessor, self).__init__('giftcertificate', settings)
def capture_payment(self, testing=False, order=None, amount=NOTSET):
"""
Process the transaction and return a ProcessorResponse
"""
if not order:
order = self.order
if amount==NOTSET:
amount = order.balance
payment = None
valid_gc = False
if self.order.paid_in_full:
success = True
reason_code = "0"
response_text = _("No balance to pay")
else:
try:
gc = GiftCertificate.objects.from_order(self.order)
valid_gc = gc.valid
except GiftCertificate.DoesNotExist:
success = False
reason_code="1"
response_text = _("No such Gift Certificate")
if not valid_gc:
success = False
reason_code="2"
response_text = _("Bad Gift Certificate")
else:
gc.apply_to_order(self.order)
payment = gc.orderpayment
reason_code = "0"
response_text = _("Success")
success = True
if not self.order.paid_in_full:
response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance)
return ProcessorResult(self.key, success, response_text, payment=payment)
|
Fix the gift certificate module so that an invalid code won't throw an exception.
|
Fix the gift certificate module so that an invalid code won't throw an exception.
|
Python
|
bsd-3-clause
|
twidi/satchmo,ringemup/satchmo,ringemup/satchmo,dokterbob/satchmo,twidi/satchmo,dokterbob/satchmo,Ryati/satchmo,Ryati/satchmo
|
"""
GiftCertificate processor
"""
from django.utils.translation import ugettext as _
from l10n.utils import moneyfmt
from models import GiftCertificate
from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET
class PaymentProcessor(BasePaymentProcessor):
def __init__(self, settings):
super(PaymentProcessor, self).__init__('giftcertificate', settings)
def capture_payment(self, testing=False, order=None, amount=NOTSET):
"""
Process the transaction and return a ProcessorResponse
"""
if not order:
order = self.order
if amount==NOTSET:
amount = order.balance
payment = None
if self.order.paid_in_full:
success = True
reason_code = "0"
response_text = _("No balance to pay")
else:
try:
gc = GiftCertificate.objects.from_order(self.order)
except GiftCertificate.DoesNotExist:
success = False
reason_code="1"
response_text = _("No such Gift Certificate")
if not gc.valid:
success = False
reason_code="2"
response_text = _("Bad Gift Certificate")
else:
gc.apply_to_order(self.order)
payment = gc.orderpayment
reason_code = "0"
response_text = _("Success")
success = True
if not self.order.paid_in_full:
response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance)
return ProcessorResult(self.key, success, response_text, payment=payment)
Fix the gift certificate module so that an invalid code won't throw an exception.
|
"""
GiftCertificate processor
"""
from django.utils.translation import ugettext as _
from l10n.utils import moneyfmt
from models import GiftCertificate
from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET
class PaymentProcessor(BasePaymentProcessor):
def __init__(self, settings):
super(PaymentProcessor, self).__init__('giftcertificate', settings)
def capture_payment(self, testing=False, order=None, amount=NOTSET):
"""
Process the transaction and return a ProcessorResponse
"""
if not order:
order = self.order
if amount==NOTSET:
amount = order.balance
payment = None
valid_gc = False
if self.order.paid_in_full:
success = True
reason_code = "0"
response_text = _("No balance to pay")
else:
try:
gc = GiftCertificate.objects.from_order(self.order)
valid_gc = gc.valid
except GiftCertificate.DoesNotExist:
success = False
reason_code="1"
response_text = _("No such Gift Certificate")
if not valid_gc:
success = False
reason_code="2"
response_text = _("Bad Gift Certificate")
else:
gc.apply_to_order(self.order)
payment = gc.orderpayment
reason_code = "0"
response_text = _("Success")
success = True
if not self.order.paid_in_full:
response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance)
return ProcessorResult(self.key, success, response_text, payment=payment)
|
<commit_before>"""
GiftCertificate processor
"""
from django.utils.translation import ugettext as _
from l10n.utils import moneyfmt
from models import GiftCertificate
from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET
class PaymentProcessor(BasePaymentProcessor):
def __init__(self, settings):
super(PaymentProcessor, self).__init__('giftcertificate', settings)
def capture_payment(self, testing=False, order=None, amount=NOTSET):
"""
Process the transaction and return a ProcessorResponse
"""
if not order:
order = self.order
if amount==NOTSET:
amount = order.balance
payment = None
if self.order.paid_in_full:
success = True
reason_code = "0"
response_text = _("No balance to pay")
else:
try:
gc = GiftCertificate.objects.from_order(self.order)
except GiftCertificate.DoesNotExist:
success = False
reason_code="1"
response_text = _("No such Gift Certificate")
if not gc.valid:
success = False
reason_code="2"
response_text = _("Bad Gift Certificate")
else:
gc.apply_to_order(self.order)
payment = gc.orderpayment
reason_code = "0"
response_text = _("Success")
success = True
if not self.order.paid_in_full:
response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance)
return ProcessorResult(self.key, success, response_text, payment=payment)
<commit_msg>Fix the gift certificate module so that an invalid code won't throw an exception.<commit_after>
|
"""
GiftCertificate processor
"""
from django.utils.translation import ugettext as _
from l10n.utils import moneyfmt
from models import GiftCertificate
from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET
class PaymentProcessor(BasePaymentProcessor):
def __init__(self, settings):
super(PaymentProcessor, self).__init__('giftcertificate', settings)
def capture_payment(self, testing=False, order=None, amount=NOTSET):
"""
Process the transaction and return a ProcessorResponse
"""
if not order:
order = self.order
if amount==NOTSET:
amount = order.balance
payment = None
valid_gc = False
if self.order.paid_in_full:
success = True
reason_code = "0"
response_text = _("No balance to pay")
else:
try:
gc = GiftCertificate.objects.from_order(self.order)
valid_gc = gc.valid
except GiftCertificate.DoesNotExist:
success = False
reason_code="1"
response_text = _("No such Gift Certificate")
if not valid_gc:
success = False
reason_code="2"
response_text = _("Bad Gift Certificate")
else:
gc.apply_to_order(self.order)
payment = gc.orderpayment
reason_code = "0"
response_text = _("Success")
success = True
if not self.order.paid_in_full:
response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance)
return ProcessorResult(self.key, success, response_text, payment=payment)
|
"""
GiftCertificate processor
"""
from django.utils.translation import ugettext as _
from l10n.utils import moneyfmt
from models import GiftCertificate
from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET
class PaymentProcessor(BasePaymentProcessor):
def __init__(self, settings):
super(PaymentProcessor, self).__init__('giftcertificate', settings)
def capture_payment(self, testing=False, order=None, amount=NOTSET):
"""
Process the transaction and return a ProcessorResponse
"""
if not order:
order = self.order
if amount==NOTSET:
amount = order.balance
payment = None
if self.order.paid_in_full:
success = True
reason_code = "0"
response_text = _("No balance to pay")
else:
try:
gc = GiftCertificate.objects.from_order(self.order)
except GiftCertificate.DoesNotExist:
success = False
reason_code="1"
response_text = _("No such Gift Certificate")
if not gc.valid:
success = False
reason_code="2"
response_text = _("Bad Gift Certificate")
else:
gc.apply_to_order(self.order)
payment = gc.orderpayment
reason_code = "0"
response_text = _("Success")
success = True
if not self.order.paid_in_full:
response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance)
return ProcessorResult(self.key, success, response_text, payment=payment)
Fix the gift certificate module so that an invalid code won't throw an exception."""
GiftCertificate processor
"""
from django.utils.translation import ugettext as _
from l10n.utils import moneyfmt
from models import GiftCertificate
from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET
class PaymentProcessor(BasePaymentProcessor):
def __init__(self, settings):
super(PaymentProcessor, self).__init__('giftcertificate', settings)
def capture_payment(self, testing=False, order=None, amount=NOTSET):
"""
Process the transaction and return a ProcessorResponse
"""
if not order:
order = self.order
if amount==NOTSET:
amount = order.balance
payment = None
valid_gc = False
if self.order.paid_in_full:
success = True
reason_code = "0"
response_text = _("No balance to pay")
else:
try:
gc = GiftCertificate.objects.from_order(self.order)
valid_gc = gc.valid
except GiftCertificate.DoesNotExist:
success = False
reason_code="1"
response_text = _("No such Gift Certificate")
if not valid_gc:
success = False
reason_code="2"
response_text = _("Bad Gift Certificate")
else:
gc.apply_to_order(self.order)
payment = gc.orderpayment
reason_code = "0"
response_text = _("Success")
success = True
if not self.order.paid_in_full:
response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance)
return ProcessorResult(self.key, success, response_text, payment=payment)
|
<commit_before>"""
GiftCertificate processor
"""
from django.utils.translation import ugettext as _
from l10n.utils import moneyfmt
from models import GiftCertificate
from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET
class PaymentProcessor(BasePaymentProcessor):
def __init__(self, settings):
super(PaymentProcessor, self).__init__('giftcertificate', settings)
def capture_payment(self, testing=False, order=None, amount=NOTSET):
"""
Process the transaction and return a ProcessorResponse
"""
if not order:
order = self.order
if amount==NOTSET:
amount = order.balance
payment = None
if self.order.paid_in_full:
success = True
reason_code = "0"
response_text = _("No balance to pay")
else:
try:
gc = GiftCertificate.objects.from_order(self.order)
except GiftCertificate.DoesNotExist:
success = False
reason_code="1"
response_text = _("No such Gift Certificate")
if not gc.valid:
success = False
reason_code="2"
response_text = _("Bad Gift Certificate")
else:
gc.apply_to_order(self.order)
payment = gc.orderpayment
reason_code = "0"
response_text = _("Success")
success = True
if not self.order.paid_in_full:
response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance)
return ProcessorResult(self.key, success, response_text, payment=payment)
<commit_msg>Fix the gift certificate module so that an invalid code won't throw an exception.<commit_after>"""
GiftCertificate processor
"""
from django.utils.translation import ugettext as _
from l10n.utils import moneyfmt
from models import GiftCertificate
from payment.modules.base import BasePaymentProcessor, ProcessorResult, NOTSET
class PaymentProcessor(BasePaymentProcessor):
def __init__(self, settings):
super(PaymentProcessor, self).__init__('giftcertificate', settings)
def capture_payment(self, testing=False, order=None, amount=NOTSET):
"""
Process the transaction and return a ProcessorResponse
"""
if not order:
order = self.order
if amount==NOTSET:
amount = order.balance
payment = None
valid_gc = False
if self.order.paid_in_full:
success = True
reason_code = "0"
response_text = _("No balance to pay")
else:
try:
gc = GiftCertificate.objects.from_order(self.order)
valid_gc = gc.valid
except GiftCertificate.DoesNotExist:
success = False
reason_code="1"
response_text = _("No such Gift Certificate")
if not valid_gc:
success = False
reason_code="2"
response_text = _("Bad Gift Certificate")
else:
gc.apply_to_order(self.order)
payment = gc.orderpayment
reason_code = "0"
response_text = _("Success")
success = True
if not self.order.paid_in_full:
response_text = _("%s balance remains after gift certificate was applied") % moneyfmt(self.order.balance)
return ProcessorResult(self.key, success, response_text, payment=payment)
|
056d82002c133736a800b08bd071b71c9f5615f8
|
ci/generate_pipeline_yml.py
|
ci/generate_pipeline_yml.py
|
#!/usr/bin/env python
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
|
#!/usr/bin/env python
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
|
Update TAS versions we test against
|
Update TAS versions we test against
|
Python
|
apache-2.0
|
cf-platform-eng/tile-generator,cf-platform-eng/tile-generator,cf-platform-eng/tile-generator,cf-platform-eng/tile-generator
|
#!/usr/bin/env python
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
Update TAS versions we test against
|
#!/usr/bin/env python
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
|
<commit_before>#!/usr/bin/env python
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
<commit_msg>Update TAS versions we test against<commit_after>
|
#!/usr/bin/env python
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
|
#!/usr/bin/env python
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
Update TAS versions we test against#!/usr/bin/env python
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
|
<commit_before>#!/usr/bin/env python
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_9', '2_10', '2_11_lts2']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
<commit_msg>Update TAS versions we test against<commit_after>#!/usr/bin/env python
import os
from jinja2 import Template
clusters = ['2_7_lts', '2_11_lts2', '2_12', '2_13']
# Commenting out this as we only have one example and it breaks
tiles = [] # [d for d in os.listdir('../examples') if os.path.isdir(os.path.join('../examples', d))]
with open('pipeline.yml.jinja2', 'r') as f:
t = Template(f.read());
with open('pipeline.yml', 'w') as f:
f.write(t.render(clusters=clusters, tiles=tiles))
print("Successfully generated pipeline.yml")
|
bc8d7a7572fcde45ae95176301522979fa54aa87
|
carnifex/test/unit/mocks.py
|
carnifex/test/unit/mocks.py
|
from twisted.internet._baseprocess import BaseProcess
from carnifex.inductor import ProcessInductor
from twisted.internet.error import ProcessTerminated, ProcessDone
class MockProcess(BaseProcess):
def run(self, fauxProcessData):
for childFd, data in fauxProcessData:
self.proto.childDataReceived(childFd, data)
def terminate(self, signal):
"""Simulate that the process was terminated with a signal
"""
self.processEnded((None, signal))
def exit(self, exitCode):
"""Simulate that the process exited
"""
self.processEnded((exitCode, None))
def _getReason(self, status):
exitCode, signal = status
if exitCode or signal:
return ProcessTerminated(exitCode, signal, status)
return ProcessDone(status)
class MockProcessInductor(ProcessInductor):
def __init__(self, reactor, fauxProcessData):
self.reactor = reactor
self.fauxProcessData = fauxProcessData
def execute(self, processProtocol, executable, args=(), env={},
path=None, uid=None, gid=None, usePTY=0, childFDs=None):
process = MockProcess(processProtocol)
process.run(self.fauxProcessData)
processProtocol.makeConnection(process)
process.exit(0)
return process
|
from twisted.internet._baseprocess import BaseProcess
from carnifex.inductor import ProcessInductor
from twisted.internet.error import ProcessTerminated, ProcessDone
class MockProcess(BaseProcess):
def run(self, fauxProcessData):
for childFd, data in fauxProcessData:
self.proto.childDataReceived(childFd, data)
def terminate(self, signal):
"""Simulate that the process was terminated with a signal
"""
self.processEnded((None, signal))
def exit(self, exitCode):
"""Simulate that the process exited
"""
self.processEnded((exitCode, None))
def _getReason(self, status):
exitCode, signal = status
if exitCode or signal:
return ProcessTerminated(exitCode, signal, status)
return ProcessDone(status)
class MockProcessInductor(ProcessInductor):
def __init__(self, reactor, fauxProcessData, exitCode=0):
self.reactor = reactor
self.fauxProcessData = fauxProcessData
self.exitCode = exitCode
def execute(self, processProtocol, executable, args=(), env={},
path=None, uid=None, gid=None, usePTY=0, childFDs=None):
process = MockProcess(processProtocol)
process.run(self.fauxProcessData)
processProtocol.makeConnection(process)
process.exit(self.exitCode)
return process
|
Allow specifying what exit code to use when emulating process exit
|
Allow specifying what exit code to use when emulating process exit
|
Python
|
mit
|
sporsh/carnifex
|
from twisted.internet._baseprocess import BaseProcess
from carnifex.inductor import ProcessInductor
from twisted.internet.error import ProcessTerminated, ProcessDone
class MockProcess(BaseProcess):
def run(self, fauxProcessData):
for childFd, data in fauxProcessData:
self.proto.childDataReceived(childFd, data)
def terminate(self, signal):
"""Simulate that the process was terminated with a signal
"""
self.processEnded((None, signal))
def exit(self, exitCode):
"""Simulate that the process exited
"""
self.processEnded((exitCode, None))
def _getReason(self, status):
exitCode, signal = status
if exitCode or signal:
return ProcessTerminated(exitCode, signal, status)
return ProcessDone(status)
class MockProcessInductor(ProcessInductor):
def __init__(self, reactor, fauxProcessData):
self.reactor = reactor
self.fauxProcessData = fauxProcessData
def execute(self, processProtocol, executable, args=(), env={},
path=None, uid=None, gid=None, usePTY=0, childFDs=None):
process = MockProcess(processProtocol)
process.run(self.fauxProcessData)
processProtocol.makeConnection(process)
process.exit(0)
return process
Allow specifying what exit code to use when emulating process exit
|
from twisted.internet._baseprocess import BaseProcess
from carnifex.inductor import ProcessInductor
from twisted.internet.error import ProcessTerminated, ProcessDone
class MockProcess(BaseProcess):
def run(self, fauxProcessData):
for childFd, data in fauxProcessData:
self.proto.childDataReceived(childFd, data)
def terminate(self, signal):
"""Simulate that the process was terminated with a signal
"""
self.processEnded((None, signal))
def exit(self, exitCode):
"""Simulate that the process exited
"""
self.processEnded((exitCode, None))
def _getReason(self, status):
exitCode, signal = status
if exitCode or signal:
return ProcessTerminated(exitCode, signal, status)
return ProcessDone(status)
class MockProcessInductor(ProcessInductor):
def __init__(self, reactor, fauxProcessData, exitCode=0):
self.reactor = reactor
self.fauxProcessData = fauxProcessData
self.exitCode = exitCode
def execute(self, processProtocol, executable, args=(), env={},
path=None, uid=None, gid=None, usePTY=0, childFDs=None):
process = MockProcess(processProtocol)
process.run(self.fauxProcessData)
processProtocol.makeConnection(process)
process.exit(self.exitCode)
return process
|
<commit_before>from twisted.internet._baseprocess import BaseProcess
from carnifex.inductor import ProcessInductor
from twisted.internet.error import ProcessTerminated, ProcessDone
class MockProcess(BaseProcess):
def run(self, fauxProcessData):
for childFd, data in fauxProcessData:
self.proto.childDataReceived(childFd, data)
def terminate(self, signal):
"""Simulate that the process was terminated with a signal
"""
self.processEnded((None, signal))
def exit(self, exitCode):
"""Simulate that the process exited
"""
self.processEnded((exitCode, None))
def _getReason(self, status):
exitCode, signal = status
if exitCode or signal:
return ProcessTerminated(exitCode, signal, status)
return ProcessDone(status)
class MockProcessInductor(ProcessInductor):
def __init__(self, reactor, fauxProcessData):
self.reactor = reactor
self.fauxProcessData = fauxProcessData
def execute(self, processProtocol, executable, args=(), env={},
path=None, uid=None, gid=None, usePTY=0, childFDs=None):
process = MockProcess(processProtocol)
process.run(self.fauxProcessData)
processProtocol.makeConnection(process)
process.exit(0)
return process
<commit_msg>Allow specifying what exit code to use when emulating process exit<commit_after>
|
from twisted.internet._baseprocess import BaseProcess
from carnifex.inductor import ProcessInductor
from twisted.internet.error import ProcessTerminated, ProcessDone
class MockProcess(BaseProcess):
def run(self, fauxProcessData):
for childFd, data in fauxProcessData:
self.proto.childDataReceived(childFd, data)
def terminate(self, signal):
"""Simulate that the process was terminated with a signal
"""
self.processEnded((None, signal))
def exit(self, exitCode):
"""Simulate that the process exited
"""
self.processEnded((exitCode, None))
def _getReason(self, status):
exitCode, signal = status
if exitCode or signal:
return ProcessTerminated(exitCode, signal, status)
return ProcessDone(status)
class MockProcessInductor(ProcessInductor):
def __init__(self, reactor, fauxProcessData, exitCode=0):
self.reactor = reactor
self.fauxProcessData = fauxProcessData
self.exitCode = exitCode
def execute(self, processProtocol, executable, args=(), env={},
path=None, uid=None, gid=None, usePTY=0, childFDs=None):
process = MockProcess(processProtocol)
process.run(self.fauxProcessData)
processProtocol.makeConnection(process)
process.exit(self.exitCode)
return process
|
from twisted.internet._baseprocess import BaseProcess
from carnifex.inductor import ProcessInductor
from twisted.internet.error import ProcessTerminated, ProcessDone
class MockProcess(BaseProcess):
def run(self, fauxProcessData):
for childFd, data in fauxProcessData:
self.proto.childDataReceived(childFd, data)
def terminate(self, signal):
"""Simulate that the process was terminated with a signal
"""
self.processEnded((None, signal))
def exit(self, exitCode):
"""Simulate that the process exited
"""
self.processEnded((exitCode, None))
def _getReason(self, status):
exitCode, signal = status
if exitCode or signal:
return ProcessTerminated(exitCode, signal, status)
return ProcessDone(status)
class MockProcessInductor(ProcessInductor):
def __init__(self, reactor, fauxProcessData):
self.reactor = reactor
self.fauxProcessData = fauxProcessData
def execute(self, processProtocol, executable, args=(), env={},
path=None, uid=None, gid=None, usePTY=0, childFDs=None):
process = MockProcess(processProtocol)
process.run(self.fauxProcessData)
processProtocol.makeConnection(process)
process.exit(0)
return process
Allow specifying what exit code to use when emulating process exitfrom twisted.internet._baseprocess import BaseProcess
from carnifex.inductor import ProcessInductor
from twisted.internet.error import ProcessTerminated, ProcessDone
class MockProcess(BaseProcess):
def run(self, fauxProcessData):
for childFd, data in fauxProcessData:
self.proto.childDataReceived(childFd, data)
def terminate(self, signal):
"""Simulate that the process was terminated with a signal
"""
self.processEnded((None, signal))
def exit(self, exitCode):
"""Simulate that the process exited
"""
self.processEnded((exitCode, None))
def _getReason(self, status):
exitCode, signal = status
if exitCode or signal:
return ProcessTerminated(exitCode, signal, status)
return ProcessDone(status)
class MockProcessInductor(ProcessInductor):
def __init__(self, reactor, fauxProcessData, exitCode=0):
self.reactor = reactor
self.fauxProcessData = fauxProcessData
self.exitCode = exitCode
def execute(self, processProtocol, executable, args=(), env={},
path=None, uid=None, gid=None, usePTY=0, childFDs=None):
process = MockProcess(processProtocol)
process.run(self.fauxProcessData)
processProtocol.makeConnection(process)
process.exit(self.exitCode)
return process
|
<commit_before>from twisted.internet._baseprocess import BaseProcess
from carnifex.inductor import ProcessInductor
from twisted.internet.error import ProcessTerminated, ProcessDone
class MockProcess(BaseProcess):
def run(self, fauxProcessData):
for childFd, data in fauxProcessData:
self.proto.childDataReceived(childFd, data)
def terminate(self, signal):
"""Simulate that the process was terminated with a signal
"""
self.processEnded((None, signal))
def exit(self, exitCode):
"""Simulate that the process exited
"""
self.processEnded((exitCode, None))
def _getReason(self, status):
exitCode, signal = status
if exitCode or signal:
return ProcessTerminated(exitCode, signal, status)
return ProcessDone(status)
class MockProcessInductor(ProcessInductor):
def __init__(self, reactor, fauxProcessData):
self.reactor = reactor
self.fauxProcessData = fauxProcessData
def execute(self, processProtocol, executable, args=(), env={},
path=None, uid=None, gid=None, usePTY=0, childFDs=None):
process = MockProcess(processProtocol)
process.run(self.fauxProcessData)
processProtocol.makeConnection(process)
process.exit(0)
return process
<commit_msg>Allow specifying what exit code to use when emulating process exit<commit_after>from twisted.internet._baseprocess import BaseProcess
from carnifex.inductor import ProcessInductor
from twisted.internet.error import ProcessTerminated, ProcessDone
class MockProcess(BaseProcess):
def run(self, fauxProcessData):
for childFd, data in fauxProcessData:
self.proto.childDataReceived(childFd, data)
def terminate(self, signal):
"""Simulate that the process was terminated with a signal
"""
self.processEnded((None, signal))
def exit(self, exitCode):
"""Simulate that the process exited
"""
self.processEnded((exitCode, None))
def _getReason(self, status):
exitCode, signal = status
if exitCode or signal:
return ProcessTerminated(exitCode, signal, status)
return ProcessDone(status)
class MockProcessInductor(ProcessInductor):
def __init__(self, reactor, fauxProcessData, exitCode=0):
self.reactor = reactor
self.fauxProcessData = fauxProcessData
self.exitCode = exitCode
def execute(self, processProtocol, executable, args=(), env={},
path=None, uid=None, gid=None, usePTY=0, childFDs=None):
process = MockProcess(processProtocol)
process.run(self.fauxProcessData)
processProtocol.makeConnection(process)
process.exit(self.exitCode)
return process
|
fdaabeaa3694103153c81a18971e6b55597cd66e
|
Sketches/JT/Jam/library/trunk/Kamaelia/Apps/Jam/Audio/Synth.py
|
Sketches/JT/Jam/library/trunk/Kamaelia/Apps/Jam/Audio/Synth.py
|
import Axon
from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser
from Kamaelia.Apps.Jam.Audio.Mixer import MonoMixer
class Synth(Axon.Component.component):
polyphony = 8
polyphoniser = Polyphoniser
def __init__(self, voiceGenerator, **argd):
super(Synth, self).__init__(**argd)
polyphoniser = self.polyphoniser(**argd).activate()
mixer = MonoMixer(channels=self.polyphony, **argd).activate()
self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1)
self.link((mixer, "outbox"), (self, "outbox"), passthrough=2)
for index, voice in enumerate(voiceGenerator()):
voice = voice.activate()
self.link((polyphoniser, "voice%i" % index), (voice, "inbox"))
self.link((voice, "outbox"), (mixer, "in%i" % index))
def main(self):
while 1:
if not self.anyReady():
self.pause()
yield 1
|
import Axon
from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser
class Synth(Axon.Component.component):
polyphony = 8
polyphoniser = Polyphoniser
def __init__(self, voiceGenerator, **argd):
super(Synth, self).__init__(**argd)
polyphoniser = self.polyphoniser(**argd).activate()
self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1)
for index, voice in enumerate(voiceGenerator()):
voice = voice.activate()
self.link((polyphoniser, "voice%i" % index), (voice, "inbox"))
def main(self):
while 1:
if not self.anyReady():
self.pause()
yield 1
|
Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.
|
Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.
|
Python
|
apache-2.0
|
sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia
|
import Axon
from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser
from Kamaelia.Apps.Jam.Audio.Mixer import MonoMixer
class Synth(Axon.Component.component):
polyphony = 8
polyphoniser = Polyphoniser
def __init__(self, voiceGenerator, **argd):
super(Synth, self).__init__(**argd)
polyphoniser = self.polyphoniser(**argd).activate()
mixer = MonoMixer(channels=self.polyphony, **argd).activate()
self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1)
self.link((mixer, "outbox"), (self, "outbox"), passthrough=2)
for index, voice in enumerate(voiceGenerator()):
voice = voice.activate()
self.link((polyphoniser, "voice%i" % index), (voice, "inbox"))
self.link((voice, "outbox"), (mixer, "in%i" % index))
def main(self):
while 1:
if not self.anyReady():
self.pause()
yield 1
Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.
|
import Axon
from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser
class Synth(Axon.Component.component):
polyphony = 8
polyphoniser = Polyphoniser
def __init__(self, voiceGenerator, **argd):
super(Synth, self).__init__(**argd)
polyphoniser = self.polyphoniser(**argd).activate()
self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1)
for index, voice in enumerate(voiceGenerator()):
voice = voice.activate()
self.link((polyphoniser, "voice%i" % index), (voice, "inbox"))
def main(self):
while 1:
if not self.anyReady():
self.pause()
yield 1
|
<commit_before>import Axon
from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser
from Kamaelia.Apps.Jam.Audio.Mixer import MonoMixer
class Synth(Axon.Component.component):
polyphony = 8
polyphoniser = Polyphoniser
def __init__(self, voiceGenerator, **argd):
super(Synth, self).__init__(**argd)
polyphoniser = self.polyphoniser(**argd).activate()
mixer = MonoMixer(channels=self.polyphony, **argd).activate()
self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1)
self.link((mixer, "outbox"), (self, "outbox"), passthrough=2)
for index, voice in enumerate(voiceGenerator()):
voice = voice.activate()
self.link((polyphoniser, "voice%i" % index), (voice, "inbox"))
self.link((voice, "outbox"), (mixer, "in%i" % index))
def main(self):
while 1:
if not self.anyReady():
self.pause()
yield 1
<commit_msg>Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.<commit_after>
|
import Axon
from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser
class Synth(Axon.Component.component):
polyphony = 8
polyphoniser = Polyphoniser
def __init__(self, voiceGenerator, **argd):
super(Synth, self).__init__(**argd)
polyphoniser = self.polyphoniser(**argd).activate()
self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1)
for index, voice in enumerate(voiceGenerator()):
voice = voice.activate()
self.link((polyphoniser, "voice%i" % index), (voice, "inbox"))
def main(self):
while 1:
if not self.anyReady():
self.pause()
yield 1
|
import Axon
from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser
from Kamaelia.Apps.Jam.Audio.Mixer import MonoMixer
class Synth(Axon.Component.component):
polyphony = 8
polyphoniser = Polyphoniser
def __init__(self, voiceGenerator, **argd):
super(Synth, self).__init__(**argd)
polyphoniser = self.polyphoniser(**argd).activate()
mixer = MonoMixer(channels=self.polyphony, **argd).activate()
self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1)
self.link((mixer, "outbox"), (self, "outbox"), passthrough=2)
for index, voice in enumerate(voiceGenerator()):
voice = voice.activate()
self.link((polyphoniser, "voice%i" % index), (voice, "inbox"))
self.link((voice, "outbox"), (mixer, "in%i" % index))
def main(self):
while 1:
if not self.anyReady():
self.pause()
yield 1
Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.import Axon
from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser
class Synth(Axon.Component.component):
polyphony = 8
polyphoniser = Polyphoniser
def __init__(self, voiceGenerator, **argd):
super(Synth, self).__init__(**argd)
polyphoniser = self.polyphoniser(**argd).activate()
self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1)
for index, voice in enumerate(voiceGenerator()):
voice = voice.activate()
self.link((polyphoniser, "voice%i" % index), (voice, "inbox"))
def main(self):
while 1:
if not self.anyReady():
self.pause()
yield 1
|
<commit_before>import Axon
from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser
from Kamaelia.Apps.Jam.Audio.Mixer import MonoMixer
class Synth(Axon.Component.component):
polyphony = 8
polyphoniser = Polyphoniser
def __init__(self, voiceGenerator, **argd):
super(Synth, self).__init__(**argd)
polyphoniser = self.polyphoniser(**argd).activate()
mixer = MonoMixer(channels=self.polyphony, **argd).activate()
self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1)
self.link((mixer, "outbox"), (self, "outbox"), passthrough=2)
for index, voice in enumerate(voiceGenerator()):
voice = voice.activate()
self.link((polyphoniser, "voice%i" % index), (voice, "inbox"))
self.link((voice, "outbox"), (mixer, "in%i" % index))
def main(self):
while 1:
if not self.anyReady():
self.pause()
yield 1
<commit_msg>Remove mixer section from synth code to reflect the components directly calling pygame mixer methods.<commit_after>import Axon
from Kamaelia.Apps.Jam.Audio.Polyphony import Polyphoniser
class Synth(Axon.Component.component):
polyphony = 8
polyphoniser = Polyphoniser
def __init__(self, voiceGenerator, **argd):
super(Synth, self).__init__(**argd)
polyphoniser = self.polyphoniser(**argd).activate()
self.link((self, "inbox"), (polyphoniser, "inbox"), passthrough=1)
for index, voice in enumerate(voiceGenerator()):
voice = voice.activate()
self.link((polyphoniser, "voice%i" % index), (voice, "inbox"))
def main(self):
while 1:
if not self.anyReady():
self.pause()
yield 1
|
75e14847fe2c0f0c40897e449bab093f4be1b17c
|
cineapp/jinja_filters.py
|
cineapp/jinja_filters.py
|
# -*- coding: utf-8 -*-
from cineapp import app
@app.template_filter()
def minutes_to_human_duration(minutes_duration):
"""
Convert a duration in minutes into a duration in a cool format human readable
"""
try:
hours,minutes = divmod(minutes_duration,60)
return "%sh %smin" %(hours,minutes)
except TypeError:
return None
@app.template_filter()
def date_format(date,format_date):
"""
Convert a date object into a custom format
"""
try:
return date.strftime(format_date)
except:
return None
|
# -*- coding: utf-8 -*-
from cineapp import app
import datetime
@app.template_filter()
def minutes_to_human_duration(minutes_duration):
"""
Convert a duration in minutes into a duration in a cool format human readable
"""
try:
hours,minutes = divmod(minutes_duration,60)
return "%sh %smin" %(hours,minutes)
except TypeError:
return None
@app.template_filter()
def date_format(date,format_date):
"""
Convert a date object into a custom format
"""
try:
if isinstance(date, basestring):
date_array=date.split('-')
date_to_convert=datetime.datetime(int(date_array[0]),int(date_array[1]),int(date_array[2]))
return date_to_convert.strftime(format_date)
else:
return date.strftime(format_date)
except Exception,e:
print e
return None
|
Improve jinja filter date converter
|
Improve jinja filter date converter
The filter now can convert date which are strings and not datetime objects.
|
Python
|
mit
|
ptitoliv/cineapp,ptitoliv/cineapp,ptitoliv/cineapp
|
# -*- coding: utf-8 -*-
from cineapp import app
@app.template_filter()
def minutes_to_human_duration(minutes_duration):
"""
Convert a duration in minutes into a duration in a cool format human readable
"""
try:
hours,minutes = divmod(minutes_duration,60)
return "%sh %smin" %(hours,minutes)
except TypeError:
return None
@app.template_filter()
def date_format(date,format_date):
"""
Convert a date object into a custom format
"""
try:
return date.strftime(format_date)
except:
return None
Improve jinja filter date converter
The filter now can convert date which are strings and not datetime objects.
|
# -*- coding: utf-8 -*-
from cineapp import app
import datetime
@app.template_filter()
def minutes_to_human_duration(minutes_duration):
"""
Convert a duration in minutes into a duration in a cool format human readable
"""
try:
hours,minutes = divmod(minutes_duration,60)
return "%sh %smin" %(hours,minutes)
except TypeError:
return None
@app.template_filter()
def date_format(date,format_date):
"""
Convert a date object into a custom format
"""
try:
if isinstance(date, basestring):
date_array=date.split('-')
date_to_convert=datetime.datetime(int(date_array[0]),int(date_array[1]),int(date_array[2]))
return date_to_convert.strftime(format_date)
else:
return date.strftime(format_date)
except Exception,e:
print e
return None
|
<commit_before># -*- coding: utf-8 -*-
from cineapp import app
@app.template_filter()
def minutes_to_human_duration(minutes_duration):
"""
Convert a duration in minutes into a duration in a cool format human readable
"""
try:
hours,minutes = divmod(minutes_duration,60)
return "%sh %smin" %(hours,minutes)
except TypeError:
return None
@app.template_filter()
def date_format(date,format_date):
"""
Convert a date object into a custom format
"""
try:
return date.strftime(format_date)
except:
return None
<commit_msg>Improve jinja filter date converter
The filter now can convert date which are strings and not datetime objects.<commit_after>
|
# -*- coding: utf-8 -*-
from cineapp import app
import datetime
@app.template_filter()
def minutes_to_human_duration(minutes_duration):
"""
Convert a duration in minutes into a duration in a cool format human readable
"""
try:
hours,minutes = divmod(minutes_duration,60)
return "%sh %smin" %(hours,minutes)
except TypeError:
return None
@app.template_filter()
def date_format(date,format_date):
"""
Convert a date object into a custom format
"""
try:
if isinstance(date, basestring):
date_array=date.split('-')
date_to_convert=datetime.datetime(int(date_array[0]),int(date_array[1]),int(date_array[2]))
return date_to_convert.strftime(format_date)
else:
return date.strftime(format_date)
except Exception,e:
print e
return None
|
# -*- coding: utf-8 -*-
from cineapp import app
@app.template_filter()
def minutes_to_human_duration(minutes_duration):
"""
Convert a duration in minutes into a duration in a cool format human readable
"""
try:
hours,minutes = divmod(minutes_duration,60)
return "%sh %smin" %(hours,minutes)
except TypeError:
return None
@app.template_filter()
def date_format(date,format_date):
"""
Convert a date object into a custom format
"""
try:
return date.strftime(format_date)
except:
return None
Improve jinja filter date converter
The filter now can convert date which are strings and not datetime objects.# -*- coding: utf-8 -*-
from cineapp import app
import datetime
@app.template_filter()
def minutes_to_human_duration(minutes_duration):
"""
Convert a duration in minutes into a duration in a cool format human readable
"""
try:
hours,minutes = divmod(minutes_duration,60)
return "%sh %smin" %(hours,minutes)
except TypeError:
return None
@app.template_filter()
def date_format(date,format_date):
"""
Convert a date object into a custom format
"""
try:
if isinstance(date, basestring):
date_array=date.split('-')
date_to_convert=datetime.datetime(int(date_array[0]),int(date_array[1]),int(date_array[2]))
return date_to_convert.strftime(format_date)
else:
return date.strftime(format_date)
except Exception,e:
print e
return None
|
<commit_before># -*- coding: utf-8 -*-
from cineapp import app
@app.template_filter()
def minutes_to_human_duration(minutes_duration):
"""
Convert a duration in minutes into a duration in a cool format human readable
"""
try:
hours,minutes = divmod(minutes_duration,60)
return "%sh %smin" %(hours,minutes)
except TypeError:
return None
@app.template_filter()
def date_format(date,format_date):
"""
Convert a date object into a custom format
"""
try:
return date.strftime(format_date)
except:
return None
<commit_msg>Improve jinja filter date converter
The filter now can convert date which are strings and not datetime objects.<commit_after># -*- coding: utf-8 -*-
from cineapp import app
import datetime
@app.template_filter()
def minutes_to_human_duration(minutes_duration):
"""
Convert a duration in minutes into a duration in a cool format human readable
"""
try:
hours,minutes = divmod(minutes_duration,60)
return "%sh %smin" %(hours,minutes)
except TypeError:
return None
@app.template_filter()
def date_format(date,format_date):
"""
Convert a date object into a custom format
"""
try:
if isinstance(date, basestring):
date_array=date.split('-')
date_to_convert=datetime.datetime(int(date_array[0]),int(date_array[1]),int(date_array[2]))
return date_to_convert.strftime(format_date)
else:
return date.strftime(format_date)
except Exception,e:
print e
return None
|
4974f83d9ed1e085ef2daaeba4db56a4001055cf
|
comics/comics/ctrlaltdel.py
|
comics/comics/ctrlaltdel.py
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Ctrl+Alt+Del"
language = "en"
url = "http://www.cad-comic.com/cad/"
start_date = "2002-10-23"
rights = "Tim Buckley"
class Crawler(CrawlerBase):
history_capable_date = "2002-10-23"
schedule = "Mo,We,Fr"
time_zone = "US/Eastern"
# Without User-Agent set, the server returns empty responses
headers = {"User-Agent": "Mozilla/4.0"}
def crawl(self, pub_date):
page = self.parse_page(
"http://www.cad-comic.com/cad/%s" % pub_date.strftime("%Y%m%d")
)
url = page.src('img[src*="/comics/"]')
title = page.alt('img[src*="/comics/"]')
return CrawlerImage(url, title)
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Ctrl+Alt+Del"
language = "en"
url = "https://cad-comic.com/category/ctrl-alt-del/"
start_date = "2002-10-23"
rights = "Tim Buckley"
class Crawler(CrawlerBase):
# history_capable_date = "2002-10-23"
history_capable_days = 20
schedule = "Mo,We,Fr"
time_zone = "US/Eastern"
# Without User-Agent set, the server returns empty responses
headers = {"User-Agent": "Mozilla/4.0"}
def crawl(self, pub_date):
feed = self.parse_feed("https://cad-comic.com/feed/")
for entry in feed.for_date(pub_date):
url = entry.summary.src("img")
title = entry.title
return CrawlerImage(url, title)
|
Update "Ctrl+Alt+Del" after site change
|
Update "Ctrl+Alt+Del" after site change
|
Python
|
agpl-3.0
|
datagutten/comics,jodal/comics,datagutten/comics,datagutten/comics,datagutten/comics,jodal/comics,jodal/comics,jodal/comics
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Ctrl+Alt+Del"
language = "en"
url = "http://www.cad-comic.com/cad/"
start_date = "2002-10-23"
rights = "Tim Buckley"
class Crawler(CrawlerBase):
history_capable_date = "2002-10-23"
schedule = "Mo,We,Fr"
time_zone = "US/Eastern"
# Without User-Agent set, the server returns empty responses
headers = {"User-Agent": "Mozilla/4.0"}
def crawl(self, pub_date):
page = self.parse_page(
"http://www.cad-comic.com/cad/%s" % pub_date.strftime("%Y%m%d")
)
url = page.src('img[src*="/comics/"]')
title = page.alt('img[src*="/comics/"]')
return CrawlerImage(url, title)
Update "Ctrl+Alt+Del" after site change
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Ctrl+Alt+Del"
language = "en"
url = "https://cad-comic.com/category/ctrl-alt-del/"
start_date = "2002-10-23"
rights = "Tim Buckley"
class Crawler(CrawlerBase):
# history_capable_date = "2002-10-23"
history_capable_days = 20
schedule = "Mo,We,Fr"
time_zone = "US/Eastern"
# Without User-Agent set, the server returns empty responses
headers = {"User-Agent": "Mozilla/4.0"}
def crawl(self, pub_date):
feed = self.parse_feed("https://cad-comic.com/feed/")
for entry in feed.for_date(pub_date):
url = entry.summary.src("img")
title = entry.title
return CrawlerImage(url, title)
|
<commit_before>from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Ctrl+Alt+Del"
language = "en"
url = "http://www.cad-comic.com/cad/"
start_date = "2002-10-23"
rights = "Tim Buckley"
class Crawler(CrawlerBase):
history_capable_date = "2002-10-23"
schedule = "Mo,We,Fr"
time_zone = "US/Eastern"
# Without User-Agent set, the server returns empty responses
headers = {"User-Agent": "Mozilla/4.0"}
def crawl(self, pub_date):
page = self.parse_page(
"http://www.cad-comic.com/cad/%s" % pub_date.strftime("%Y%m%d")
)
url = page.src('img[src*="/comics/"]')
title = page.alt('img[src*="/comics/"]')
return CrawlerImage(url, title)
<commit_msg>Update "Ctrl+Alt+Del" after site change<commit_after>
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Ctrl+Alt+Del"
language = "en"
url = "https://cad-comic.com/category/ctrl-alt-del/"
start_date = "2002-10-23"
rights = "Tim Buckley"
class Crawler(CrawlerBase):
# history_capable_date = "2002-10-23"
history_capable_days = 20
schedule = "Mo,We,Fr"
time_zone = "US/Eastern"
# Without User-Agent set, the server returns empty responses
headers = {"User-Agent": "Mozilla/4.0"}
def crawl(self, pub_date):
feed = self.parse_feed("https://cad-comic.com/feed/")
for entry in feed.for_date(pub_date):
url = entry.summary.src("img")
title = entry.title
return CrawlerImage(url, title)
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Ctrl+Alt+Del"
language = "en"
url = "http://www.cad-comic.com/cad/"
start_date = "2002-10-23"
rights = "Tim Buckley"
class Crawler(CrawlerBase):
history_capable_date = "2002-10-23"
schedule = "Mo,We,Fr"
time_zone = "US/Eastern"
# Without User-Agent set, the server returns empty responses
headers = {"User-Agent": "Mozilla/4.0"}
def crawl(self, pub_date):
page = self.parse_page(
"http://www.cad-comic.com/cad/%s" % pub_date.strftime("%Y%m%d")
)
url = page.src('img[src*="/comics/"]')
title = page.alt('img[src*="/comics/"]')
return CrawlerImage(url, title)
Update "Ctrl+Alt+Del" after site changefrom comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Ctrl+Alt+Del"
language = "en"
url = "https://cad-comic.com/category/ctrl-alt-del/"
start_date = "2002-10-23"
rights = "Tim Buckley"
class Crawler(CrawlerBase):
# history_capable_date = "2002-10-23"
history_capable_days = 20
schedule = "Mo,We,Fr"
time_zone = "US/Eastern"
# Without User-Agent set, the server returns empty responses
headers = {"User-Agent": "Mozilla/4.0"}
def crawl(self, pub_date):
feed = self.parse_feed("https://cad-comic.com/feed/")
for entry in feed.for_date(pub_date):
url = entry.summary.src("img")
title = entry.title
return CrawlerImage(url, title)
|
<commit_before>from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Ctrl+Alt+Del"
language = "en"
url = "http://www.cad-comic.com/cad/"
start_date = "2002-10-23"
rights = "Tim Buckley"
class Crawler(CrawlerBase):
history_capable_date = "2002-10-23"
schedule = "Mo,We,Fr"
time_zone = "US/Eastern"
# Without User-Agent set, the server returns empty responses
headers = {"User-Agent": "Mozilla/4.0"}
def crawl(self, pub_date):
page = self.parse_page(
"http://www.cad-comic.com/cad/%s" % pub_date.strftime("%Y%m%d")
)
url = page.src('img[src*="/comics/"]')
title = page.alt('img[src*="/comics/"]')
return CrawlerImage(url, title)
<commit_msg>Update "Ctrl+Alt+Del" after site change<commit_after>from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Ctrl+Alt+Del"
language = "en"
url = "https://cad-comic.com/category/ctrl-alt-del/"
start_date = "2002-10-23"
rights = "Tim Buckley"
class Crawler(CrawlerBase):
# history_capable_date = "2002-10-23"
history_capable_days = 20
schedule = "Mo,We,Fr"
time_zone = "US/Eastern"
# Without User-Agent set, the server returns empty responses
headers = {"User-Agent": "Mozilla/4.0"}
def crawl(self, pub_date):
feed = self.parse_feed("https://cad-comic.com/feed/")
for entry in feed.for_date(pub_date):
url = entry.summary.src("img")
title = entry.title
return CrawlerImage(url, title)
|
4007ecdc66e361bcb81bb5b661e682eeef0a6ea5
|
remo/profiles/migrations/0011_groups_new_onboarding_group.py
|
remo/profiles/migrations/0011_groups_new_onboarding_group.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
def forwards(apps, schema_editor):
"""Create Onboarding group."""
Group = apps.get_model('auth', 'Group')
Group.objects.create(name='Onboarding')
def backwards(apps, schema_editor):
"""Delete Onboarding group."""
Group = apps.get_model('auth', 'Group')
Group.objects.filter(name='Onboarding').delete()
class Migration(migrations.Migration):
dependencies = [
('profiles', '0010_auto_20171221_0112'),
]
operations = [
migrations.RunPython(forwards, backwards)
]
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
def forwards(apps, schema_editor):
"""Create Onboarding group."""
Group = apps.get_model('auth', 'Group')
if not Group.objects.filter(name='Onboarding').exists():
Group.objects.create(name='Onboarding')
def backwards(apps, schema_editor):
"""Delete Onboarding group."""
Group = apps.get_model('auth', 'Group')
if Group.objects.filter(name='Onboarding').exists():
Group.objects.filter(name='Onboarding').delete()
class Migration(migrations.Migration):
dependencies = [
('profiles', '0010_auto_20171221_0112'),
]
operations = [
migrations.RunPython(forwards, backwards)
]
|
Check if Onboarding exists before creating.
|
Check if Onboarding exists before creating.
|
Python
|
bsd-3-clause
|
mozilla/remo,akatsoulas/remo,Mte90/remo,mozilla/remo,Mte90/remo,akatsoulas/remo,mozilla/remo,Mte90/remo,mozilla/remo,akatsoulas/remo,akatsoulas/remo,Mte90/remo
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
def forwards(apps, schema_editor):
"""Create Onboarding group."""
Group = apps.get_model('auth', 'Group')
Group.objects.create(name='Onboarding')
def backwards(apps, schema_editor):
"""Delete Onboarding group."""
Group = apps.get_model('auth', 'Group')
Group.objects.filter(name='Onboarding').delete()
class Migration(migrations.Migration):
dependencies = [
('profiles', '0010_auto_20171221_0112'),
]
operations = [
migrations.RunPython(forwards, backwards)
]
Check if Onboarding exists before creating.
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
def forwards(apps, schema_editor):
"""Create Onboarding group."""
Group = apps.get_model('auth', 'Group')
if not Group.objects.filter(name='Onboarding').exists():
Group.objects.create(name='Onboarding')
def backwards(apps, schema_editor):
"""Delete Onboarding group."""
Group = apps.get_model('auth', 'Group')
if Group.objects.filter(name='Onboarding').exists():
Group.objects.filter(name='Onboarding').delete()
class Migration(migrations.Migration):
dependencies = [
('profiles', '0010_auto_20171221_0112'),
]
operations = [
migrations.RunPython(forwards, backwards)
]
|
<commit_before># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
def forwards(apps, schema_editor):
"""Create Onboarding group."""
Group = apps.get_model('auth', 'Group')
Group.objects.create(name='Onboarding')
def backwards(apps, schema_editor):
"""Delete Onboarding group."""
Group = apps.get_model('auth', 'Group')
Group.objects.filter(name='Onboarding').delete()
class Migration(migrations.Migration):
dependencies = [
('profiles', '0010_auto_20171221_0112'),
]
operations = [
migrations.RunPython(forwards, backwards)
]
<commit_msg>Check if Onboarding exists before creating.<commit_after>
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
def forwards(apps, schema_editor):
"""Create Onboarding group."""
Group = apps.get_model('auth', 'Group')
if not Group.objects.filter(name='Onboarding').exists():
Group.objects.create(name='Onboarding')
def backwards(apps, schema_editor):
"""Delete Onboarding group."""
Group = apps.get_model('auth', 'Group')
if Group.objects.filter(name='Onboarding').exists():
Group.objects.filter(name='Onboarding').delete()
class Migration(migrations.Migration):
dependencies = [
('profiles', '0010_auto_20171221_0112'),
]
operations = [
migrations.RunPython(forwards, backwards)
]
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
def forwards(apps, schema_editor):
"""Create Onboarding group."""
Group = apps.get_model('auth', 'Group')
Group.objects.create(name='Onboarding')
def backwards(apps, schema_editor):
"""Delete Onboarding group."""
Group = apps.get_model('auth', 'Group')
Group.objects.filter(name='Onboarding').delete()
class Migration(migrations.Migration):
dependencies = [
('profiles', '0010_auto_20171221_0112'),
]
operations = [
migrations.RunPython(forwards, backwards)
]
Check if Onboarding exists before creating.# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
def forwards(apps, schema_editor):
"""Create Onboarding group."""
Group = apps.get_model('auth', 'Group')
if not Group.objects.filter(name='Onboarding').exists():
Group.objects.create(name='Onboarding')
def backwards(apps, schema_editor):
"""Delete Onboarding group."""
Group = apps.get_model('auth', 'Group')
if Group.objects.filter(name='Onboarding').exists():
Group.objects.filter(name='Onboarding').delete()
class Migration(migrations.Migration):
dependencies = [
('profiles', '0010_auto_20171221_0112'),
]
operations = [
migrations.RunPython(forwards, backwards)
]
|
<commit_before># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
def forwards(apps, schema_editor):
"""Create Onboarding group."""
Group = apps.get_model('auth', 'Group')
Group.objects.create(name='Onboarding')
def backwards(apps, schema_editor):
"""Delete Onboarding group."""
Group = apps.get_model('auth', 'Group')
Group.objects.filter(name='Onboarding').delete()
class Migration(migrations.Migration):
dependencies = [
('profiles', '0010_auto_20171221_0112'),
]
operations = [
migrations.RunPython(forwards, backwards)
]
<commit_msg>Check if Onboarding exists before creating.<commit_after># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
def forwards(apps, schema_editor):
"""Create Onboarding group."""
Group = apps.get_model('auth', 'Group')
if not Group.objects.filter(name='Onboarding').exists():
Group.objects.create(name='Onboarding')
def backwards(apps, schema_editor):
"""Delete Onboarding group."""
Group = apps.get_model('auth', 'Group')
if Group.objects.filter(name='Onboarding').exists():
Group.objects.filter(name='Onboarding').delete()
class Migration(migrations.Migration):
dependencies = [
('profiles', '0010_auto_20171221_0112'),
]
operations = [
migrations.RunPython(forwards, backwards)
]
|
5e3be1d123063495f21d0c0068c7132d43fd9724
|
account/models.py
|
account/models.py
|
from django.db import models
from django.db.models import signals
from django.contrib.auth.models import User
from course.models import Term
class Profile(models.Model):
user = models.OneToOneField(User)
student_id = models.CharField(max_length=10, null=True)
default_term = models.ForeignKey(Term, null=True)
facebook_id = models.CharField(max_length=50, null=True)
def create_profile(sender, instance, created, **kwargs):
profile = Profile.objects.get(user=instance)
if not profile:
Profile(user=instance).save()
signals.post_save.connect(create_profile, sender=User)
|
from django.db import models
from django.db.models import signals
from django.contrib.auth.models import User
from course.models import Term
class Profile(models.Model):
user = models.OneToOneField(User)
student_id = models.CharField(max_length=10, null=True)
default_term = models.ForeignKey(Term, null=True)
facebook_id = models.CharField(max_length=50, null=True)
def create_profile(sender, instance, created, **kwargs):
try:
profile = Profile.objects.get(user=instance)
except Profile.DoesNotExist:
Profile(user=instance).save()
signals.post_save.connect(create_profile, sender=User)
|
Fix login error for new accounts where a profile doesn't exist
|
Fix login error for new accounts where a profile doesn't exist
|
Python
|
apache-2.0
|
OpenCourseProject/OpenCourse,gravitylow/OpenCourse,gravitylow/OpenCourse,gravitylow/OpenCourse,OpenCourseProject/OpenCourse,OpenCourseProject/OpenCourse
|
from django.db import models
from django.db.models import signals
from django.contrib.auth.models import User
from course.models import Term
class Profile(models.Model):
user = models.OneToOneField(User)
student_id = models.CharField(max_length=10, null=True)
default_term = models.ForeignKey(Term, null=True)
facebook_id = models.CharField(max_length=50, null=True)
def create_profile(sender, instance, created, **kwargs):
profile = Profile.objects.get(user=instance)
if not profile:
Profile(user=instance).save()
signals.post_save.connect(create_profile, sender=User)
Fix login error for new accounts where a profile doesn't exist
|
from django.db import models
from django.db.models import signals
from django.contrib.auth.models import User
from course.models import Term
class Profile(models.Model):
user = models.OneToOneField(User)
student_id = models.CharField(max_length=10, null=True)
default_term = models.ForeignKey(Term, null=True)
facebook_id = models.CharField(max_length=50, null=True)
def create_profile(sender, instance, created, **kwargs):
try:
profile = Profile.objects.get(user=instance)
except Profile.DoesNotExist:
Profile(user=instance).save()
signals.post_save.connect(create_profile, sender=User)
|
<commit_before>from django.db import models
from django.db.models import signals
from django.contrib.auth.models import User
from course.models import Term
class Profile(models.Model):
user = models.OneToOneField(User)
student_id = models.CharField(max_length=10, null=True)
default_term = models.ForeignKey(Term, null=True)
facebook_id = models.CharField(max_length=50, null=True)
def create_profile(sender, instance, created, **kwargs):
profile = Profile.objects.get(user=instance)
if not profile:
Profile(user=instance).save()
signals.post_save.connect(create_profile, sender=User)
<commit_msg>Fix login error for new accounts where a profile doesn't exist<commit_after>
|
from django.db import models
from django.db.models import signals
from django.contrib.auth.models import User
from course.models import Term
class Profile(models.Model):
user = models.OneToOneField(User)
student_id = models.CharField(max_length=10, null=True)
default_term = models.ForeignKey(Term, null=True)
facebook_id = models.CharField(max_length=50, null=True)
def create_profile(sender, instance, created, **kwargs):
try:
profile = Profile.objects.get(user=instance)
except Profile.DoesNotExist:
Profile(user=instance).save()
signals.post_save.connect(create_profile, sender=User)
|
from django.db import models
from django.db.models import signals
from django.contrib.auth.models import User
from course.models import Term
class Profile(models.Model):
user = models.OneToOneField(User)
student_id = models.CharField(max_length=10, null=True)
default_term = models.ForeignKey(Term, null=True)
facebook_id = models.CharField(max_length=50, null=True)
def create_profile(sender, instance, created, **kwargs):
profile = Profile.objects.get(user=instance)
if not profile:
Profile(user=instance).save()
signals.post_save.connect(create_profile, sender=User)
Fix login error for new accounts where a profile doesn't existfrom django.db import models
from django.db.models import signals
from django.contrib.auth.models import User
from course.models import Term
class Profile(models.Model):
user = models.OneToOneField(User)
student_id = models.CharField(max_length=10, null=True)
default_term = models.ForeignKey(Term, null=True)
facebook_id = models.CharField(max_length=50, null=True)
def create_profile(sender, instance, created, **kwargs):
try:
profile = Profile.objects.get(user=instance)
except Profile.DoesNotExist:
Profile(user=instance).save()
signals.post_save.connect(create_profile, sender=User)
|
<commit_before>from django.db import models
from django.db.models import signals
from django.contrib.auth.models import User
from course.models import Term
class Profile(models.Model):
user = models.OneToOneField(User)
student_id = models.CharField(max_length=10, null=True)
default_term = models.ForeignKey(Term, null=True)
facebook_id = models.CharField(max_length=50, null=True)
def create_profile(sender, instance, created, **kwargs):
profile = Profile.objects.get(user=instance)
if not profile:
Profile(user=instance).save()
signals.post_save.connect(create_profile, sender=User)
<commit_msg>Fix login error for new accounts where a profile doesn't exist<commit_after>from django.db import models
from django.db.models import signals
from django.contrib.auth.models import User
from course.models import Term
class Profile(models.Model):
user = models.OneToOneField(User)
student_id = models.CharField(max_length=10, null=True)
default_term = models.ForeignKey(Term, null=True)
facebook_id = models.CharField(max_length=50, null=True)
def create_profile(sender, instance, created, **kwargs):
try:
profile = Profile.objects.get(user=instance)
except Profile.DoesNotExist:
Profile(user=instance).save()
signals.post_save.connect(create_profile, sender=User)
|
7a07a89d59250127fce21b5f1b68492046b3eb60
|
pyshelf/search/metadata.py
|
pyshelf/search/metadata.py
|
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer
# Required for case sensitivity
metadata_analyzer = analyzer("metadata_analyzer", tokenizer=tokenizer("keyword"))
class Metadata(DocType):
property_list = Nested(
properties={
"name": String(),
"value": String(analyzer=metadata_analyzer),
"immutable": Boolean()
}
)
def update_all(self, metadata):
"""
Updates all metadata related to an artifact.
Args
metadata(dict): collection of metadata for document.
"""
self.property_list = metadata.values()
|
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer
# Required for case sensitivity
# To add an analyzer to an existing mapping requires mapping to be "closed"
case_sensitive_analyzer = analyzer("case_sensitive_analyzer", tokenizer=tokenizer("keyword"))
class Metadata(DocType):
property_list = Nested(
properties={
"name": String(analyzer=case_sensitive_analyzer),
"value": String(analyzer=case_sensitive_analyzer),
"immutable": Boolean()
}
)
def update_all(self, metadata):
"""
Updates all metadata related to an artifact.
Args
metadata(dict): collection of metadata for document.
"""
self.property_list = metadata.values()
|
Add case sensitivity to field and clarify analyzer.
|
Add case sensitivity to field and clarify analyzer.
|
Python
|
mit
|
not-nexus/shelf,kyle-long/pyshelf,kyle-long/pyshelf,not-nexus/shelf
|
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer
# Required for case sensitivity
metadata_analyzer = analyzer("metadata_analyzer", tokenizer=tokenizer("keyword"))
class Metadata(DocType):
property_list = Nested(
properties={
"name": String(),
"value": String(analyzer=metadata_analyzer),
"immutable": Boolean()
}
)
def update_all(self, metadata):
"""
Updates all metadata related to an artifact.
Args
metadata(dict): collection of metadata for document.
"""
self.property_list = metadata.values()
Add case sensitivity to field and clarify analyzer.
|
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer
# Required for case sensitivity
# To add an analyzer to an existing mapping requires mapping to be "closed"
case_sensitive_analyzer = analyzer("case_sensitive_analyzer", tokenizer=tokenizer("keyword"))
class Metadata(DocType):
property_list = Nested(
properties={
"name": String(analyzer=case_sensitive_analyzer),
"value": String(analyzer=case_sensitive_analyzer),
"immutable": Boolean()
}
)
def update_all(self, metadata):
"""
Updates all metadata related to an artifact.
Args
metadata(dict): collection of metadata for document.
"""
self.property_list = metadata.values()
|
<commit_before>from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer
# Required for case sensitivity
metadata_analyzer = analyzer("metadata_analyzer", tokenizer=tokenizer("keyword"))
class Metadata(DocType):
property_list = Nested(
properties={
"name": String(),
"value": String(analyzer=metadata_analyzer),
"immutable": Boolean()
}
)
def update_all(self, metadata):
"""
Updates all metadata related to an artifact.
Args
metadata(dict): collection of metadata for document.
"""
self.property_list = metadata.values()
<commit_msg>Add case sensitivity to field and clarify analyzer.<commit_after>
|
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer
# Required for case sensitivity
# To add an analyzer to an existing mapping requires mapping to be "closed"
case_sensitive_analyzer = analyzer("case_sensitive_analyzer", tokenizer=tokenizer("keyword"))
class Metadata(DocType):
property_list = Nested(
properties={
"name": String(analyzer=case_sensitive_analyzer),
"value": String(analyzer=case_sensitive_analyzer),
"immutable": Boolean()
}
)
def update_all(self, metadata):
"""
Updates all metadata related to an artifact.
Args
metadata(dict): collection of metadata for document.
"""
self.property_list = metadata.values()
|
from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer
# Required for case sensitivity
metadata_analyzer = analyzer("metadata_analyzer", tokenizer=tokenizer("keyword"))
class Metadata(DocType):
property_list = Nested(
properties={
"name": String(),
"value": String(analyzer=metadata_analyzer),
"immutable": Boolean()
}
)
def update_all(self, metadata):
"""
Updates all metadata related to an artifact.
Args
metadata(dict): collection of metadata for document.
"""
self.property_list = metadata.values()
Add case sensitivity to field and clarify analyzer.from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer
# Required for case sensitivity
# To add an analyzer to an existing mapping requires mapping to be "closed"
case_sensitive_analyzer = analyzer("case_sensitive_analyzer", tokenizer=tokenizer("keyword"))
class Metadata(DocType):
property_list = Nested(
properties={
"name": String(analyzer=case_sensitive_analyzer),
"value": String(analyzer=case_sensitive_analyzer),
"immutable": Boolean()
}
)
def update_all(self, metadata):
"""
Updates all metadata related to an artifact.
Args
metadata(dict): collection of metadata for document.
"""
self.property_list = metadata.values()
|
<commit_before>from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer
# Required for case sensitivity
metadata_analyzer = analyzer("metadata_analyzer", tokenizer=tokenizer("keyword"))
class Metadata(DocType):
property_list = Nested(
properties={
"name": String(),
"value": String(analyzer=metadata_analyzer),
"immutable": Boolean()
}
)
def update_all(self, metadata):
"""
Updates all metadata related to an artifact.
Args
metadata(dict): collection of metadata for document.
"""
self.property_list = metadata.values()
<commit_msg>Add case sensitivity to field and clarify analyzer.<commit_after>from elasticsearch_dsl import String, Nested, Boolean, DocType, tokenizer, analyzer
# Required for case sensitivity
# To add an analyzer to an existing mapping requires mapping to be "closed"
case_sensitive_analyzer = analyzer("case_sensitive_analyzer", tokenizer=tokenizer("keyword"))
class Metadata(DocType):
property_list = Nested(
properties={
"name": String(analyzer=case_sensitive_analyzer),
"value": String(analyzer=case_sensitive_analyzer),
"immutable": Boolean()
}
)
def update_all(self, metadata):
"""
Updates all metadata related to an artifact.
Args
metadata(dict): collection of metadata for document.
"""
self.property_list = metadata.values()
|
7bc777a5e9fb15720dd6b41aa5e1fbcfd7d3141b
|
tests/test_postgres_processor.py
|
tests/test_postgres_processor.py
|
import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.postgres
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
|
import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.postgres
@pytest.mark.django_db
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
|
Add django pytest mark to process raw test
|
Add django pytest mark to process raw test
|
Python
|
apache-2.0
|
CenterForOpenScience/scrapi,felliott/scrapi,mehanig/scrapi,CenterForOpenScience/scrapi,fabianvf/scrapi,erinspace/scrapi,erinspace/scrapi,felliott/scrapi,mehanig/scrapi,fabianvf/scrapi
|
import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.postgres
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
Add django pytest mark to process raw test
|
import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.postgres
@pytest.mark.django_db
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
|
<commit_before>import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.postgres
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
<commit_msg>Add django pytest mark to process raw test<commit_after>
|
import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.postgres
@pytest.mark.django_db
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
|
import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.postgres
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
Add django pytest mark to process raw testimport pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.postgres
@pytest.mark.django_db
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
|
<commit_before>import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.postgres
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
<commit_msg>Add django pytest mark to process raw test<commit_after>import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.postgres
@pytest.mark.django_db
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
|
53646da453a4aa6d0e559ee3069626458f2fef78
|
common/urls.py
|
common/urls.py
|
import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
return re.sub("\\(.+\\)", "{id}", pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
|
import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
_pattern = pattern.replace('^', '').replace('$', '')
return re.sub("\\(.+\\)", "{id}", _pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
|
Fix backend home page url generator
|
Fix backend home page url generator
|
Python
|
mit
|
DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange,DemocracyLab/CivicTechExchange
|
import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
return re.sub("\\(.+\\)", "{id}", pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)Fix backend home page url generator
|
import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
_pattern = pattern.replace('^', '').replace('$', '')
return re.sub("\\(.+\\)", "{id}", _pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
|
<commit_before>import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
return re.sub("\\(.+\\)", "{id}", pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)<commit_msg>Fix backend home page url generator<commit_after>
|
import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
_pattern = pattern.replace('^', '').replace('$', '')
return re.sub("\\(.+\\)", "{id}", _pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
|
import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
return re.sub("\\(.+\\)", "{id}", pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)Fix backend home page url generatorimport json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
_pattern = pattern.replace('^', '').replace('$', '')
return re.sub("\\(.+\\)", "{id}", _pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
|
<commit_before>import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
return re.sub("\\(.+\\)", "{id}", pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)<commit_msg>Fix backend home page url generator<commit_after>import json
import os
import re
from django.urls import re_path
from civictechprojects import views
def url_generator_from_pattern(pattern):
_pattern = pattern.replace('^', '').replace('$', '')
return re.sub("\\(.+\\)", "{id}", _pattern)
def generate_url_patterns(spec_path, set_url_generators=False):
# Read json file
base_dir = os.path.dirname(__file__)
filename = os.path.join(base_dir, spec_path)
url_patterns = []
with open(filename, 'r', encoding='utf-8') as f:
urls_json = json.load(f)
for url_spec_json in urls_json:
url_patterns.append(re_path(url_spec_json['pattern'], views.index))
if set_url_generators:
url_generators[url_spec_json['name']] = {
'section': url_spec_json['name'],
'regex': re.compile(url_spec_json['pattern']),
'generator': url_generator_from_pattern(url_spec_json['pattern'])
}
return url_patterns
url_generators = {}
v1_urls = generate_url_patterns('./components/urls/urls_v1.json')
v2_urls = generate_url_patterns('./components/urls/urls_v2.json', set_url_generators=True)
|
0f8e2313d6f0ec06806ea05e861d1fc47d3c3016
|
utils/internal/zz_parse.py
|
utils/internal/zz_parse.py
|
import sys
sys.path.insert(0, '../..')
from pycparser import c_parser, c_ast, parse_file
if __name__ == "__main__":
#ast = parse_file('zc_pp.c', use_cpp=True, cpp_path="../cpp.exe")
parser = c_parser.CParser()
#code = r'''int ar[30];'''
code = r'''
char ***arr3d[40];
'''
#code = r'''
#int foo(int a, int arr[*]);
#'''
print(code)
ast = parser.parse(code)
ast.show(attrnames=True, nodenames=True)
print(ast.ext[0].__slots__)
print(dir(ast.ext[0]))
|
from __future__ import print_function
import sys
from pycparser import c_parser, c_generator, c_ast, parse_file
if __name__ == "__main__":
parser = c_parser.CParser()
code = r'''
void* ptr = (int[ ]){0};
'''
print(code)
ast = parser.parse(code)
ast.show(attrnames=True, nodenames=True)
print(ast.ext[0].__slots__)
print(dir(ast.ext[0]))
print("==== From C generator:")
generator = c_generator.CGenerator()
print(generator.visit(ast))
|
Clean up internal hacking util
|
Clean up internal hacking util
|
Python
|
bsd-3-clause
|
CtheSky/pycparser,CtheSky/pycparser,CtheSky/pycparser
|
import sys
sys.path.insert(0, '../..')
from pycparser import c_parser, c_ast, parse_file
if __name__ == "__main__":
#ast = parse_file('zc_pp.c', use_cpp=True, cpp_path="../cpp.exe")
parser = c_parser.CParser()
#code = r'''int ar[30];'''
code = r'''
char ***arr3d[40];
'''
#code = r'''
#int foo(int a, int arr[*]);
#'''
print(code)
ast = parser.parse(code)
ast.show(attrnames=True, nodenames=True)
print(ast.ext[0].__slots__)
print(dir(ast.ext[0]))
Clean up internal hacking util
|
from __future__ import print_function
import sys
from pycparser import c_parser, c_generator, c_ast, parse_file
if __name__ == "__main__":
parser = c_parser.CParser()
code = r'''
void* ptr = (int[ ]){0};
'''
print(code)
ast = parser.parse(code)
ast.show(attrnames=True, nodenames=True)
print(ast.ext[0].__slots__)
print(dir(ast.ext[0]))
print("==== From C generator:")
generator = c_generator.CGenerator()
print(generator.visit(ast))
|
<commit_before>import sys
sys.path.insert(0, '../..')
from pycparser import c_parser, c_ast, parse_file
if __name__ == "__main__":
#ast = parse_file('zc_pp.c', use_cpp=True, cpp_path="../cpp.exe")
parser = c_parser.CParser()
#code = r'''int ar[30];'''
code = r'''
char ***arr3d[40];
'''
#code = r'''
#int foo(int a, int arr[*]);
#'''
print(code)
ast = parser.parse(code)
ast.show(attrnames=True, nodenames=True)
print(ast.ext[0].__slots__)
print(dir(ast.ext[0]))
<commit_msg>Clean up internal hacking util<commit_after>
|
from __future__ import print_function
import sys
from pycparser import c_parser, c_generator, c_ast, parse_file
if __name__ == "__main__":
parser = c_parser.CParser()
code = r'''
void* ptr = (int[ ]){0};
'''
print(code)
ast = parser.parse(code)
ast.show(attrnames=True, nodenames=True)
print(ast.ext[0].__slots__)
print(dir(ast.ext[0]))
print("==== From C generator:")
generator = c_generator.CGenerator()
print(generator.visit(ast))
|
import sys
sys.path.insert(0, '../..')
from pycparser import c_parser, c_ast, parse_file
if __name__ == "__main__":
#ast = parse_file('zc_pp.c', use_cpp=True, cpp_path="../cpp.exe")
parser = c_parser.CParser()
#code = r'''int ar[30];'''
code = r'''
char ***arr3d[40];
'''
#code = r'''
#int foo(int a, int arr[*]);
#'''
print(code)
ast = parser.parse(code)
ast.show(attrnames=True, nodenames=True)
print(ast.ext[0].__slots__)
print(dir(ast.ext[0]))
Clean up internal hacking utilfrom __future__ import print_function
import sys
from pycparser import c_parser, c_generator, c_ast, parse_file
if __name__ == "__main__":
parser = c_parser.CParser()
code = r'''
void* ptr = (int[ ]){0};
'''
print(code)
ast = parser.parse(code)
ast.show(attrnames=True, nodenames=True)
print(ast.ext[0].__slots__)
print(dir(ast.ext[0]))
print("==== From C generator:")
generator = c_generator.CGenerator()
print(generator.visit(ast))
|
<commit_before>import sys
sys.path.insert(0, '../..')
from pycparser import c_parser, c_ast, parse_file
if __name__ == "__main__":
#ast = parse_file('zc_pp.c', use_cpp=True, cpp_path="../cpp.exe")
parser = c_parser.CParser()
#code = r'''int ar[30];'''
code = r'''
char ***arr3d[40];
'''
#code = r'''
#int foo(int a, int arr[*]);
#'''
print(code)
ast = parser.parse(code)
ast.show(attrnames=True, nodenames=True)
print(ast.ext[0].__slots__)
print(dir(ast.ext[0]))
<commit_msg>Clean up internal hacking util<commit_after>from __future__ import print_function
import sys
from pycparser import c_parser, c_generator, c_ast, parse_file
if __name__ == "__main__":
parser = c_parser.CParser()
code = r'''
void* ptr = (int[ ]){0};
'''
print(code)
ast = parser.parse(code)
ast.show(attrnames=True, nodenames=True)
print(ast.ext[0].__slots__)
print(dir(ast.ext[0]))
print("==== From C generator:")
generator = c_generator.CGenerator()
print(generator.visit(ast))
|
044a051c637f256613ff307caf3ae0126d09b049
|
backend/unichat/views.py
|
backend/unichat/views.py
|
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest
from django.views.decorators.csrf import csrf_exempt
import json
from helpers import get_school_list, check_signup_email
def get_schools(request):
resp = JsonResponse({'schools': get_school_list()})
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
@csrf_exempt
def signup(request):
if request.method == "POST":
signup_parameters = json.loads(request.body.decode('utf-8'))
if check_signup_email(signup_parameters['email']):
resp = HttpResponse('Signup OK')
else:
resp = HttpResponseBadRequest('Invalid univesity email')
elif request.method == "OPTIONS":
resp = HttpResponse('')
resp['Access-Control-Allow-Headers'] = 'Content-Type'
else:
resp = HttpResponseBadRequest('')
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
|
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest
from django.views.decorators.csrf import csrf_exempt
import json
from helpers import get_school_list, check_signup_email
def get_schools(request):
resp = JsonResponse({'schools': get_school_list()})
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
@csrf_exempt
def signup(request):
if request.method == "POST":
signup_parameters = json.loads(request.body.decode('utf-8'))
if check_signup_email(signup_parameters['email']):
resp = HttpResponse('Signup OK')
else:
resp = HttpResponseBadRequest('Invalid univesity email')
elif request.method == "OPTIONS":
resp = HttpResponse('')
resp['Access-Control-Allow-Headers'] = 'Content-Type'
else:
resp = HttpResponseBadRequest('Invalid request method')
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
|
Add error message to BadRequest signup response for invalid method
|
Add error message to BadRequest signup response for invalid method
|
Python
|
mit
|
dimkarakostas/unimeet,dimkarakostas/unimeet,dimkarakostas/unimeet,dimkarakostas/unimeet
|
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest
from django.views.decorators.csrf import csrf_exempt
import json
from helpers import get_school_list, check_signup_email
def get_schools(request):
resp = JsonResponse({'schools': get_school_list()})
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
@csrf_exempt
def signup(request):
if request.method == "POST":
signup_parameters = json.loads(request.body.decode('utf-8'))
if check_signup_email(signup_parameters['email']):
resp = HttpResponse('Signup OK')
else:
resp = HttpResponseBadRequest('Invalid univesity email')
elif request.method == "OPTIONS":
resp = HttpResponse('')
resp['Access-Control-Allow-Headers'] = 'Content-Type'
else:
resp = HttpResponseBadRequest('')
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
Add error message to BadRequest signup response for invalid method
|
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest
from django.views.decorators.csrf import csrf_exempt
import json
from helpers import get_school_list, check_signup_email
def get_schools(request):
resp = JsonResponse({'schools': get_school_list()})
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
@csrf_exempt
def signup(request):
if request.method == "POST":
signup_parameters = json.loads(request.body.decode('utf-8'))
if check_signup_email(signup_parameters['email']):
resp = HttpResponse('Signup OK')
else:
resp = HttpResponseBadRequest('Invalid univesity email')
elif request.method == "OPTIONS":
resp = HttpResponse('')
resp['Access-Control-Allow-Headers'] = 'Content-Type'
else:
resp = HttpResponseBadRequest('Invalid request method')
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
|
<commit_before>from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest
from django.views.decorators.csrf import csrf_exempt
import json
from helpers import get_school_list, check_signup_email
def get_schools(request):
resp = JsonResponse({'schools': get_school_list()})
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
@csrf_exempt
def signup(request):
if request.method == "POST":
signup_parameters = json.loads(request.body.decode('utf-8'))
if check_signup_email(signup_parameters['email']):
resp = HttpResponse('Signup OK')
else:
resp = HttpResponseBadRequest('Invalid univesity email')
elif request.method == "OPTIONS":
resp = HttpResponse('')
resp['Access-Control-Allow-Headers'] = 'Content-Type'
else:
resp = HttpResponseBadRequest('')
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
<commit_msg>Add error message to BadRequest signup response for invalid method<commit_after>
|
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest
from django.views.decorators.csrf import csrf_exempt
import json
from helpers import get_school_list, check_signup_email
def get_schools(request):
resp = JsonResponse({'schools': get_school_list()})
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
@csrf_exempt
def signup(request):
if request.method == "POST":
signup_parameters = json.loads(request.body.decode('utf-8'))
if check_signup_email(signup_parameters['email']):
resp = HttpResponse('Signup OK')
else:
resp = HttpResponseBadRequest('Invalid univesity email')
elif request.method == "OPTIONS":
resp = HttpResponse('')
resp['Access-Control-Allow-Headers'] = 'Content-Type'
else:
resp = HttpResponseBadRequest('Invalid request method')
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
|
from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest
from django.views.decorators.csrf import csrf_exempt
import json
from helpers import get_school_list, check_signup_email
def get_schools(request):
resp = JsonResponse({'schools': get_school_list()})
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
@csrf_exempt
def signup(request):
if request.method == "POST":
signup_parameters = json.loads(request.body.decode('utf-8'))
if check_signup_email(signup_parameters['email']):
resp = HttpResponse('Signup OK')
else:
resp = HttpResponseBadRequest('Invalid univesity email')
elif request.method == "OPTIONS":
resp = HttpResponse('')
resp['Access-Control-Allow-Headers'] = 'Content-Type'
else:
resp = HttpResponseBadRequest('')
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
Add error message to BadRequest signup response for invalid methodfrom django.http import JsonResponse, HttpResponse, HttpResponseBadRequest
from django.views.decorators.csrf import csrf_exempt
import json
from helpers import get_school_list, check_signup_email
def get_schools(request):
resp = JsonResponse({'schools': get_school_list()})
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
@csrf_exempt
def signup(request):
if request.method == "POST":
signup_parameters = json.loads(request.body.decode('utf-8'))
if check_signup_email(signup_parameters['email']):
resp = HttpResponse('Signup OK')
else:
resp = HttpResponseBadRequest('Invalid univesity email')
elif request.method == "OPTIONS":
resp = HttpResponse('')
resp['Access-Control-Allow-Headers'] = 'Content-Type'
else:
resp = HttpResponseBadRequest('Invalid request method')
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
|
<commit_before>from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest
from django.views.decorators.csrf import csrf_exempt
import json
from helpers import get_school_list, check_signup_email
def get_schools(request):
resp = JsonResponse({'schools': get_school_list()})
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
@csrf_exempt
def signup(request):
if request.method == "POST":
signup_parameters = json.loads(request.body.decode('utf-8'))
if check_signup_email(signup_parameters['email']):
resp = HttpResponse('Signup OK')
else:
resp = HttpResponseBadRequest('Invalid univesity email')
elif request.method == "OPTIONS":
resp = HttpResponse('')
resp['Access-Control-Allow-Headers'] = 'Content-Type'
else:
resp = HttpResponseBadRequest('')
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
<commit_msg>Add error message to BadRequest signup response for invalid method<commit_after>from django.http import JsonResponse, HttpResponse, HttpResponseBadRequest
from django.views.decorators.csrf import csrf_exempt
import json
from helpers import get_school_list, check_signup_email
def get_schools(request):
resp = JsonResponse({'schools': get_school_list()})
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
@csrf_exempt
def signup(request):
if request.method == "POST":
signup_parameters = json.loads(request.body.decode('utf-8'))
if check_signup_email(signup_parameters['email']):
resp = HttpResponse('Signup OK')
else:
resp = HttpResponseBadRequest('Invalid univesity email')
elif request.method == "OPTIONS":
resp = HttpResponse('')
resp['Access-Control-Allow-Headers'] = 'Content-Type'
else:
resp = HttpResponseBadRequest('Invalid request method')
resp['Access-Control-Allow-Origin'] = 'http://localhost:3000'
return resp
|
4c1c902010096d6d87d93b865d9c68794da51414
|
trex/parsers.py
|
trex/parsers.py
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from io import TextIOWrapper
from rest_framework.parsers import BaseParser
class PlainTextParser(BaseParser):
media_type = "text/plain"
def parse(self, stream, media_type=None, parser_context=None):
print "Running PlainTextParser"
charset = self.get_charset(media_type)
if charset:
stream = TextIOWrapper(stream, encoding=charset)
return stream
def get_charset(self, media_type):
if not media_type:
return None
charset = None
msplit = media_type.split(" ");
for m in msplit:
m = m.strip()
if "charset" in m:
csplit = m.split("=")
if len(csplit) > 1:
charset = csplit[1]
return charset.strip().lower()
return None
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from io import TextIOWrapper, BytesIO
from django.core.handlers.wsgi import WSGIRequest
from rest_framework.parsers import BaseParser
class PlainTextParser(BaseParser):
media_type = "text/plain"
def parse(self, stream, media_type=None, parser_context=None):
print "Running PlainTextParser"
charset = self.get_charset(media_type)
if charset:
if isinstance(stream, WSGIRequest):
stream = BytesIO(stream.read())
stream = TextIOWrapper(stream, encoding=charset)
return stream
def get_charset(self, media_type):
if not media_type:
return None
charset = None
msplit = media_type.split(" ");
for m in msplit:
m = m.strip()
if "charset" in m:
csplit = m.split("=")
if len(csplit) > 1:
charset = csplit[1]
return charset.strip().lower()
return None
|
Fix parsing data from request
|
Fix parsing data from request
The object passed to the parser method is not a real IOBase stream. It may only
be a Request object which has read, etc. methods. Therefore the real data must
be encapsulated in a BytesIO stream before changing the content type.
|
Python
|
mit
|
bjoernricks/trex,bjoernricks/trex
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from io import TextIOWrapper
from rest_framework.parsers import BaseParser
class PlainTextParser(BaseParser):
media_type = "text/plain"
def parse(self, stream, media_type=None, parser_context=None):
print "Running PlainTextParser"
charset = self.get_charset(media_type)
if charset:
stream = TextIOWrapper(stream, encoding=charset)
return stream
def get_charset(self, media_type):
if not media_type:
return None
charset = None
msplit = media_type.split(" ");
for m in msplit:
m = m.strip()
if "charset" in m:
csplit = m.split("=")
if len(csplit) > 1:
charset = csplit[1]
return charset.strip().lower()
return None
Fix parsing data from request
The object passed to the parser method is not a real IOBase stream. It may only
be a Request object which has read, etc. methods. Therefore the real data must
be encapsulated in a BytesIO stream before changing the content type.
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from io import TextIOWrapper, BytesIO
from django.core.handlers.wsgi import WSGIRequest
from rest_framework.parsers import BaseParser
class PlainTextParser(BaseParser):
media_type = "text/plain"
def parse(self, stream, media_type=None, parser_context=None):
print "Running PlainTextParser"
charset = self.get_charset(media_type)
if charset:
if isinstance(stream, WSGIRequest):
stream = BytesIO(stream.read())
stream = TextIOWrapper(stream, encoding=charset)
return stream
def get_charset(self, media_type):
if not media_type:
return None
charset = None
msplit = media_type.split(" ");
for m in msplit:
m = m.strip()
if "charset" in m:
csplit = m.split("=")
if len(csplit) > 1:
charset = csplit[1]
return charset.strip().lower()
return None
|
<commit_before># -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from io import TextIOWrapper
from rest_framework.parsers import BaseParser
class PlainTextParser(BaseParser):
media_type = "text/plain"
def parse(self, stream, media_type=None, parser_context=None):
print "Running PlainTextParser"
charset = self.get_charset(media_type)
if charset:
stream = TextIOWrapper(stream, encoding=charset)
return stream
def get_charset(self, media_type):
if not media_type:
return None
charset = None
msplit = media_type.split(" ");
for m in msplit:
m = m.strip()
if "charset" in m:
csplit = m.split("=")
if len(csplit) > 1:
charset = csplit[1]
return charset.strip().lower()
return None
<commit_msg>Fix parsing data from request
The object passed to the parser method is not a real IOBase stream. It may only
be a Request object which has read, etc. methods. Therefore the real data must
be encapsulated in a BytesIO stream before changing the content type.<commit_after>
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from io import TextIOWrapper, BytesIO
from django.core.handlers.wsgi import WSGIRequest
from rest_framework.parsers import BaseParser
class PlainTextParser(BaseParser):
media_type = "text/plain"
def parse(self, stream, media_type=None, parser_context=None):
print "Running PlainTextParser"
charset = self.get_charset(media_type)
if charset:
if isinstance(stream, WSGIRequest):
stream = BytesIO(stream.read())
stream = TextIOWrapper(stream, encoding=charset)
return stream
def get_charset(self, media_type):
if not media_type:
return None
charset = None
msplit = media_type.split(" ");
for m in msplit:
m = m.strip()
if "charset" in m:
csplit = m.split("=")
if len(csplit) > 1:
charset = csplit[1]
return charset.strip().lower()
return None
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from io import TextIOWrapper
from rest_framework.parsers import BaseParser
class PlainTextParser(BaseParser):
media_type = "text/plain"
def parse(self, stream, media_type=None, parser_context=None):
print "Running PlainTextParser"
charset = self.get_charset(media_type)
if charset:
stream = TextIOWrapper(stream, encoding=charset)
return stream
def get_charset(self, media_type):
if not media_type:
return None
charset = None
msplit = media_type.split(" ");
for m in msplit:
m = m.strip()
if "charset" in m:
csplit = m.split("=")
if len(csplit) > 1:
charset = csplit[1]
return charset.strip().lower()
return None
Fix parsing data from request
The object passed to the parser method is not a real IOBase stream. It may only
be a Request object which has read, etc. methods. Therefore the real data must
be encapsulated in a BytesIO stream before changing the content type.# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from io import TextIOWrapper, BytesIO
from django.core.handlers.wsgi import WSGIRequest
from rest_framework.parsers import BaseParser
class PlainTextParser(BaseParser):
media_type = "text/plain"
def parse(self, stream, media_type=None, parser_context=None):
print "Running PlainTextParser"
charset = self.get_charset(media_type)
if charset:
if isinstance(stream, WSGIRequest):
stream = BytesIO(stream.read())
stream = TextIOWrapper(stream, encoding=charset)
return stream
def get_charset(self, media_type):
if not media_type:
return None
charset = None
msplit = media_type.split(" ");
for m in msplit:
m = m.strip()
if "charset" in m:
csplit = m.split("=")
if len(csplit) > 1:
charset = csplit[1]
return charset.strip().lower()
return None
|
<commit_before># -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from io import TextIOWrapper
from rest_framework.parsers import BaseParser
class PlainTextParser(BaseParser):
media_type = "text/plain"
def parse(self, stream, media_type=None, parser_context=None):
print "Running PlainTextParser"
charset = self.get_charset(media_type)
if charset:
stream = TextIOWrapper(stream, encoding=charset)
return stream
def get_charset(self, media_type):
if not media_type:
return None
charset = None
msplit = media_type.split(" ");
for m in msplit:
m = m.strip()
if "charset" in m:
csplit = m.split("=")
if len(csplit) > 1:
charset = csplit[1]
return charset.strip().lower()
return None
<commit_msg>Fix parsing data from request
The object passed to the parser method is not a real IOBase stream. It may only
be a Request object which has read, etc. methods. Therefore the real data must
be encapsulated in a BytesIO stream before changing the content type.<commit_after># -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from io import TextIOWrapper, BytesIO
from django.core.handlers.wsgi import WSGIRequest
from rest_framework.parsers import BaseParser
class PlainTextParser(BaseParser):
media_type = "text/plain"
def parse(self, stream, media_type=None, parser_context=None):
print "Running PlainTextParser"
charset = self.get_charset(media_type)
if charset:
if isinstance(stream, WSGIRequest):
stream = BytesIO(stream.read())
stream = TextIOWrapper(stream, encoding=charset)
return stream
def get_charset(self, media_type):
if not media_type:
return None
charset = None
msplit = media_type.split(" ");
for m in msplit:
m = m.strip()
if "charset" in m:
csplit = m.split("=")
if len(csplit) > 1:
charset = csplit[1]
return charset.strip().lower()
return None
|
b86c53c388c39baee1ddfe3a615cdad20d272055
|
antcolony/util.py
|
antcolony/util.py
|
import json
def avg(iterable):
return sum(iterable) / len(iterable)
def nice_json_dump(data, filepath):
with open(filepath, 'w') as f:
json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
|
import json
def avg(iterable):
sum_ = 0
element_count = 0
for element in iterable:
sum_ += element
element_count += 1
return sum_ / element_count
def nice_json_dump(data, filepath):
with open(filepath, 'w') as f:
json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
|
Make avg() work with iterators
|
Make avg() work with iterators
|
Python
|
bsd-3-clause
|
ppolewicz/ant-colony,ppolewicz/ant-colony
|
import json
def avg(iterable):
return sum(iterable) / len(iterable)
def nice_json_dump(data, filepath):
with open(filepath, 'w') as f:
json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
Make avg() work with iterators
|
import json
def avg(iterable):
sum_ = 0
element_count = 0
for element in iterable:
sum_ += element
element_count += 1
return sum_ / element_count
def nice_json_dump(data, filepath):
with open(filepath, 'w') as f:
json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
|
<commit_before>import json
def avg(iterable):
return sum(iterable) / len(iterable)
def nice_json_dump(data, filepath):
with open(filepath, 'w') as f:
json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
<commit_msg>Make avg() work with iterators<commit_after>
|
import json
def avg(iterable):
sum_ = 0
element_count = 0
for element in iterable:
sum_ += element
element_count += 1
return sum_ / element_count
def nice_json_dump(data, filepath):
with open(filepath, 'w') as f:
json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
|
import json
def avg(iterable):
return sum(iterable) / len(iterable)
def nice_json_dump(data, filepath):
with open(filepath, 'w') as f:
json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
Make avg() work with iteratorsimport json
def avg(iterable):
sum_ = 0
element_count = 0
for element in iterable:
sum_ += element
element_count += 1
return sum_ / element_count
def nice_json_dump(data, filepath):
with open(filepath, 'w') as f:
json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
|
<commit_before>import json
def avg(iterable):
return sum(iterable) / len(iterable)
def nice_json_dump(data, filepath):
with open(filepath, 'w') as f:
json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
<commit_msg>Make avg() work with iterators<commit_after>import json
def avg(iterable):
sum_ = 0
element_count = 0
for element in iterable:
sum_ += element
element_count += 1
return sum_ / element_count
def nice_json_dump(data, filepath):
with open(filepath, 'w') as f:
json.dump(data, f, sort_keys=True, indent=4, separators=(',', ': '))
|
50a30ded8705b343478f85ea1c6c60e827982d37
|
auwsssp/urls.py
|
auwsssp/urls.py
|
from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
# Examples:
# url(r'^$', 'auwsssp.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
url(r'^admin/', include(admin.site.urls)),
)
|
from django.conf.urls import patterns, include, url
from django.conf import settings
from django.conf.urls.static import static
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
# Examples:
url(r'^$', 'signups.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
#url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
url(r'^admin/', include(admin.site.urls)),
)
if settings.DEBUG:
urlpatterns += static(settings.STATIC_URL,
document_root=settings.STATIC_ROOT)
urlpatterns += static(settings.MEDIA_URL,
document_root=settings.MEDIA_ROOT)
|
Modify index url and static folders
|
Modify index url and static folders
|
Python
|
mit
|
eyassug/au-water-sanitation-template,eyassug/au-water-sanitation-template,eyassug/au-water-sanitation-template
|
from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
# Examples:
# url(r'^$', 'auwsssp.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
url(r'^admin/', include(admin.site.urls)),
)
Modify index url and static folders
|
from django.conf.urls import patterns, include, url
from django.conf import settings
from django.conf.urls.static import static
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
# Examples:
url(r'^$', 'signups.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
#url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
url(r'^admin/', include(admin.site.urls)),
)
if settings.DEBUG:
urlpatterns += static(settings.STATIC_URL,
document_root=settings.STATIC_ROOT)
urlpatterns += static(settings.MEDIA_URL,
document_root=settings.MEDIA_ROOT)
|
<commit_before>from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
# Examples:
# url(r'^$', 'auwsssp.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<commit_msg>Modify index url and static folders<commit_after>
|
from django.conf.urls import patterns, include, url
from django.conf import settings
from django.conf.urls.static import static
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
# Examples:
url(r'^$', 'signups.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
#url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
url(r'^admin/', include(admin.site.urls)),
)
if settings.DEBUG:
urlpatterns += static(settings.STATIC_URL,
document_root=settings.STATIC_ROOT)
urlpatterns += static(settings.MEDIA_URL,
document_root=settings.MEDIA_ROOT)
|
from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
# Examples:
# url(r'^$', 'auwsssp.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
url(r'^admin/', include(admin.site.urls)),
)
Modify index url and static foldersfrom django.conf.urls import patterns, include, url
from django.conf import settings
from django.conf.urls.static import static
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
# Examples:
url(r'^$', 'signups.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
#url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
url(r'^admin/', include(admin.site.urls)),
)
if settings.DEBUG:
urlpatterns += static(settings.STATIC_URL,
document_root=settings.STATIC_ROOT)
urlpatterns += static(settings.MEDIA_URL,
document_root=settings.MEDIA_ROOT)
|
<commit_before>from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
# Examples:
# url(r'^$', 'auwsssp.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<commit_msg>Modify index url and static folders<commit_after>from django.conf.urls import patterns, include, url
from django.conf import settings
from django.conf.urls.static import static
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
# Examples:
url(r'^$', 'signups.views.home', name='home'),
# url(r'^blog/', include('blog.urls')),
#url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
url(r'^admin/', include(admin.site.urls)),
)
if settings.DEBUG:
urlpatterns += static(settings.STATIC_URL,
document_root=settings.STATIC_ROOT)
urlpatterns += static(settings.MEDIA_URL,
document_root=settings.MEDIA_ROOT)
|
b43c6604163e18ae03c6ef206c4892e0beb873f7
|
django_cradmin/demo/uimock_demo/urls.py
|
django_cradmin/demo/uimock_demo/urls.py
|
from django.urls import path
from django_cradmin import viewhelpers
from .views import overview
urlpatterns = [
path('simple/<str:mockname>',
viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'),
name='cradmin_uimock_demo_simple'),
path('',
overview.Overview.as_view(),
name='cradmin_uimock_demo'),
]
|
from django.urls import path, re_path
from django_cradmin import viewhelpers
from .views import overview
urlpatterns = [
re_path(r'^simple/(?P<mockname>.+)?$',
viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'),
name='cradmin_uimock_demo_simple'),
path('',
overview.Overview.as_view(),
name='cradmin_uimock_demo'),
]
|
Fix url that was wrongly converted to django3.
|
Fix url that was wrongly converted to django3.
|
Python
|
bsd-3-clause
|
appressoas/django_cradmin,appressoas/django_cradmin,appressoas/django_cradmin
|
from django.urls import path
from django_cradmin import viewhelpers
from .views import overview
urlpatterns = [
path('simple/<str:mockname>',
viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'),
name='cradmin_uimock_demo_simple'),
path('',
overview.Overview.as_view(),
name='cradmin_uimock_demo'),
]
Fix url that was wrongly converted to django3.
|
from django.urls import path, re_path
from django_cradmin import viewhelpers
from .views import overview
urlpatterns = [
re_path(r'^simple/(?P<mockname>.+)?$',
viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'),
name='cradmin_uimock_demo_simple'),
path('',
overview.Overview.as_view(),
name='cradmin_uimock_demo'),
]
|
<commit_before>from django.urls import path
from django_cradmin import viewhelpers
from .views import overview
urlpatterns = [
path('simple/<str:mockname>',
viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'),
name='cradmin_uimock_demo_simple'),
path('',
overview.Overview.as_view(),
name='cradmin_uimock_demo'),
]
<commit_msg>Fix url that was wrongly converted to django3.<commit_after>
|
from django.urls import path, re_path
from django_cradmin import viewhelpers
from .views import overview
urlpatterns = [
re_path(r'^simple/(?P<mockname>.+)?$',
viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'),
name='cradmin_uimock_demo_simple'),
path('',
overview.Overview.as_view(),
name='cradmin_uimock_demo'),
]
|
from django.urls import path
from django_cradmin import viewhelpers
from .views import overview
urlpatterns = [
path('simple/<str:mockname>',
viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'),
name='cradmin_uimock_demo_simple'),
path('',
overview.Overview.as_view(),
name='cradmin_uimock_demo'),
]
Fix url that was wrongly converted to django3.from django.urls import path, re_path
from django_cradmin import viewhelpers
from .views import overview
urlpatterns = [
re_path(r'^simple/(?P<mockname>.+)?$',
viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'),
name='cradmin_uimock_demo_simple'),
path('',
overview.Overview.as_view(),
name='cradmin_uimock_demo'),
]
|
<commit_before>from django.urls import path
from django_cradmin import viewhelpers
from .views import overview
urlpatterns = [
path('simple/<str:mockname>',
viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'),
name='cradmin_uimock_demo_simple'),
path('',
overview.Overview.as_view(),
name='cradmin_uimock_demo'),
]
<commit_msg>Fix url that was wrongly converted to django3.<commit_after>from django.urls import path, re_path
from django_cradmin import viewhelpers
from .views import overview
urlpatterns = [
re_path(r'^simple/(?P<mockname>.+)?$',
viewhelpers.uimock.UiMock.as_view(template_directory='uimock_demo/simple/'),
name='cradmin_uimock_demo_simple'),
path('',
overview.Overview.as_view(),
name='cradmin_uimock_demo'),
]
|
ac3c8155abae010fb79866addd1e9cd50f5cae78
|
tests/test_impersonation.py
|
tests/test_impersonation.py
|
from django.core.urlresolvers import reverse
import pytest
from saleor.userprofile.impersonate import can_impersonate
from saleor.userprofile.models import User
def test_staff_with_permission_can_impersonate(
staff_client, customer_user, staff_user, permission_impersonate_user,
staff_group):
staff_group.permissions.add(permission_impersonate_user)
staff_user.groups.add(staff_group)
staff_user = User.objects.get(pk=staff_user.pk)
response = staff_client.get('/impersonate/{}'.format(customer_user.pk),
follow=True)
assert response.context['user'] == customer_user
assert response.context['user'].is_impersonate
assert response.context['request'].impersonator == staff_user
|
from django.core.urlresolvers import reverse
import pytest
from saleor.userprofile.impersonate import can_impersonate
from saleor.userprofile.models import User
def test_staff_with_permission_can_impersonate(
staff_client, customer_user, staff_user, permission_impersonate_user,
staff_group):
staff_group.permissions.add(permission_impersonate_user)
staff_user.groups.add(staff_group)
staff_user = User.objects.get(pk=staff_user.pk)
response = staff_client.get(reverse('impersonate-start',
args=[customer_user.pk]), follow=True)
assert response.context['user'] == customer_user
assert response.context['user'].is_impersonate
assert response.context['request'].impersonator == staff_user
|
Use reverse function in tests
|
Use reverse function in tests
|
Python
|
bsd-3-clause
|
UITools/saleor,maferelo/saleor,mociepka/saleor,UITools/saleor,mociepka/saleor,maferelo/saleor,UITools/saleor,UITools/saleor,maferelo/saleor,UITools/saleor,mociepka/saleor
|
from django.core.urlresolvers import reverse
import pytest
from saleor.userprofile.impersonate import can_impersonate
from saleor.userprofile.models import User
def test_staff_with_permission_can_impersonate(
staff_client, customer_user, staff_user, permission_impersonate_user,
staff_group):
staff_group.permissions.add(permission_impersonate_user)
staff_user.groups.add(staff_group)
staff_user = User.objects.get(pk=staff_user.pk)
response = staff_client.get('/impersonate/{}'.format(customer_user.pk),
follow=True)
assert response.context['user'] == customer_user
assert response.context['user'].is_impersonate
assert response.context['request'].impersonator == staff_user
Use reverse function in tests
|
from django.core.urlresolvers import reverse
import pytest
from saleor.userprofile.impersonate import can_impersonate
from saleor.userprofile.models import User
def test_staff_with_permission_can_impersonate(
staff_client, customer_user, staff_user, permission_impersonate_user,
staff_group):
staff_group.permissions.add(permission_impersonate_user)
staff_user.groups.add(staff_group)
staff_user = User.objects.get(pk=staff_user.pk)
response = staff_client.get(reverse('impersonate-start',
args=[customer_user.pk]), follow=True)
assert response.context['user'] == customer_user
assert response.context['user'].is_impersonate
assert response.context['request'].impersonator == staff_user
|
<commit_before>from django.core.urlresolvers import reverse
import pytest
from saleor.userprofile.impersonate import can_impersonate
from saleor.userprofile.models import User
def test_staff_with_permission_can_impersonate(
staff_client, customer_user, staff_user, permission_impersonate_user,
staff_group):
staff_group.permissions.add(permission_impersonate_user)
staff_user.groups.add(staff_group)
staff_user = User.objects.get(pk=staff_user.pk)
response = staff_client.get('/impersonate/{}'.format(customer_user.pk),
follow=True)
assert response.context['user'] == customer_user
assert response.context['user'].is_impersonate
assert response.context['request'].impersonator == staff_user
<commit_msg>Use reverse function in tests<commit_after>
|
from django.core.urlresolvers import reverse
import pytest
from saleor.userprofile.impersonate import can_impersonate
from saleor.userprofile.models import User
def test_staff_with_permission_can_impersonate(
staff_client, customer_user, staff_user, permission_impersonate_user,
staff_group):
staff_group.permissions.add(permission_impersonate_user)
staff_user.groups.add(staff_group)
staff_user = User.objects.get(pk=staff_user.pk)
response = staff_client.get(reverse('impersonate-start',
args=[customer_user.pk]), follow=True)
assert response.context['user'] == customer_user
assert response.context['user'].is_impersonate
assert response.context['request'].impersonator == staff_user
|
from django.core.urlresolvers import reverse
import pytest
from saleor.userprofile.impersonate import can_impersonate
from saleor.userprofile.models import User
def test_staff_with_permission_can_impersonate(
staff_client, customer_user, staff_user, permission_impersonate_user,
staff_group):
staff_group.permissions.add(permission_impersonate_user)
staff_user.groups.add(staff_group)
staff_user = User.objects.get(pk=staff_user.pk)
response = staff_client.get('/impersonate/{}'.format(customer_user.pk),
follow=True)
assert response.context['user'] == customer_user
assert response.context['user'].is_impersonate
assert response.context['request'].impersonator == staff_user
Use reverse function in testsfrom django.core.urlresolvers import reverse
import pytest
from saleor.userprofile.impersonate import can_impersonate
from saleor.userprofile.models import User
def test_staff_with_permission_can_impersonate(
staff_client, customer_user, staff_user, permission_impersonate_user,
staff_group):
staff_group.permissions.add(permission_impersonate_user)
staff_user.groups.add(staff_group)
staff_user = User.objects.get(pk=staff_user.pk)
response = staff_client.get(reverse('impersonate-start',
args=[customer_user.pk]), follow=True)
assert response.context['user'] == customer_user
assert response.context['user'].is_impersonate
assert response.context['request'].impersonator == staff_user
|
<commit_before>from django.core.urlresolvers import reverse
import pytest
from saleor.userprofile.impersonate import can_impersonate
from saleor.userprofile.models import User
def test_staff_with_permission_can_impersonate(
staff_client, customer_user, staff_user, permission_impersonate_user,
staff_group):
staff_group.permissions.add(permission_impersonate_user)
staff_user.groups.add(staff_group)
staff_user = User.objects.get(pk=staff_user.pk)
response = staff_client.get('/impersonate/{}'.format(customer_user.pk),
follow=True)
assert response.context['user'] == customer_user
assert response.context['user'].is_impersonate
assert response.context['request'].impersonator == staff_user
<commit_msg>Use reverse function in tests<commit_after>from django.core.urlresolvers import reverse
import pytest
from saleor.userprofile.impersonate import can_impersonate
from saleor.userprofile.models import User
def test_staff_with_permission_can_impersonate(
staff_client, customer_user, staff_user, permission_impersonate_user,
staff_group):
staff_group.permissions.add(permission_impersonate_user)
staff_user.groups.add(staff_group)
staff_user = User.objects.get(pk=staff_user.pk)
response = staff_client.get(reverse('impersonate-start',
args=[customer_user.pk]), follow=True)
assert response.context['user'] == customer_user
assert response.context['user'].is_impersonate
assert response.context['request'].impersonator == staff_user
|
8b545ee63ec695a77ba08fa5ff45b7d6dd3d94f8
|
cuteshop/downloaders/git.py
|
cuteshop/downloaders/git.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess
from ..utils import DEVNULL, change_working_directory
from .base import DOWNLOAD_CONTAINER
def download(source_info):
url = source_info['git']
subprocess.call(
('git', 'clone', url, DOWNLOAD_CONTAINER),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
if 'tag' in source_info:
with change_working_directory(DOWNLOAD_CONTAINER):
subprocess.call(
('git', 'checkout', source_info['tag']),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess
from ..utils import DEVNULL, change_working_directory
from .base import DOWNLOAD_CONTAINER
def _checkout(name):
with change_working_directory(DOWNLOAD_CONTAINER):
subprocess.call(
('git', 'checkout', name),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
def download(source_info):
url = source_info['git']
subprocess.call(
('git', 'clone', url, DOWNLOAD_CONTAINER),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
if 'tag' in source_info:
_checkout(source_info['tag'])
elif 'branch' in source_info:
_checkout(source_info['branch'])
|
Add auto branch checkout functionality
|
Add auto branch checkout functionality
|
Python
|
mit
|
uranusjr/cuteshop
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess
from ..utils import DEVNULL, change_working_directory
from .base import DOWNLOAD_CONTAINER
def download(source_info):
url = source_info['git']
subprocess.call(
('git', 'clone', url, DOWNLOAD_CONTAINER),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
if 'tag' in source_info:
with change_working_directory(DOWNLOAD_CONTAINER):
subprocess.call(
('git', 'checkout', source_info['tag']),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
Add auto branch checkout functionality
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess
from ..utils import DEVNULL, change_working_directory
from .base import DOWNLOAD_CONTAINER
def _checkout(name):
with change_working_directory(DOWNLOAD_CONTAINER):
subprocess.call(
('git', 'checkout', name),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
def download(source_info):
url = source_info['git']
subprocess.call(
('git', 'clone', url, DOWNLOAD_CONTAINER),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
if 'tag' in source_info:
_checkout(source_info['tag'])
elif 'branch' in source_info:
_checkout(source_info['branch'])
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess
from ..utils import DEVNULL, change_working_directory
from .base import DOWNLOAD_CONTAINER
def download(source_info):
url = source_info['git']
subprocess.call(
('git', 'clone', url, DOWNLOAD_CONTAINER),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
if 'tag' in source_info:
with change_working_directory(DOWNLOAD_CONTAINER):
subprocess.call(
('git', 'checkout', source_info['tag']),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
<commit_msg>Add auto branch checkout functionality<commit_after>
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess
from ..utils import DEVNULL, change_working_directory
from .base import DOWNLOAD_CONTAINER
def _checkout(name):
with change_working_directory(DOWNLOAD_CONTAINER):
subprocess.call(
('git', 'checkout', name),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
def download(source_info):
url = source_info['git']
subprocess.call(
('git', 'clone', url, DOWNLOAD_CONTAINER),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
if 'tag' in source_info:
_checkout(source_info['tag'])
elif 'branch' in source_info:
_checkout(source_info['branch'])
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess
from ..utils import DEVNULL, change_working_directory
from .base import DOWNLOAD_CONTAINER
def download(source_info):
url = source_info['git']
subprocess.call(
('git', 'clone', url, DOWNLOAD_CONTAINER),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
if 'tag' in source_info:
with change_working_directory(DOWNLOAD_CONTAINER):
subprocess.call(
('git', 'checkout', source_info['tag']),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
Add auto branch checkout functionality#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess
from ..utils import DEVNULL, change_working_directory
from .base import DOWNLOAD_CONTAINER
def _checkout(name):
with change_working_directory(DOWNLOAD_CONTAINER):
subprocess.call(
('git', 'checkout', name),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
def download(source_info):
url = source_info['git']
subprocess.call(
('git', 'clone', url, DOWNLOAD_CONTAINER),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
if 'tag' in source_info:
_checkout(source_info['tag'])
elif 'branch' in source_info:
_checkout(source_info['branch'])
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess
from ..utils import DEVNULL, change_working_directory
from .base import DOWNLOAD_CONTAINER
def download(source_info):
url = source_info['git']
subprocess.call(
('git', 'clone', url, DOWNLOAD_CONTAINER),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
if 'tag' in source_info:
with change_working_directory(DOWNLOAD_CONTAINER):
subprocess.call(
('git', 'checkout', source_info['tag']),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
<commit_msg>Add auto branch checkout functionality<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess
from ..utils import DEVNULL, change_working_directory
from .base import DOWNLOAD_CONTAINER
def _checkout(name):
with change_working_directory(DOWNLOAD_CONTAINER):
subprocess.call(
('git', 'checkout', name),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
def download(source_info):
url = source_info['git']
subprocess.call(
('git', 'clone', url, DOWNLOAD_CONTAINER),
stdout=DEVNULL, stderr=subprocess.STDOUT,
)
if 'tag' in source_info:
_checkout(source_info['tag'])
elif 'branch' in source_info:
_checkout(source_info['branch'])
|
687f48ca94b67321a1576a1dbb1d7ae89fe6f0b7
|
tests/test_pubannotation.py
|
tests/test_pubannotation.py
|
import kindred
def test_pubannotation_groST():
corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 50
assert relationCount == 1454
assert entityCount == 2657
def test_pubannotation_wikiPain():
corpus = kindred.pubannotation.load('WikiPainGoldStandard')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 49
assert relationCount == 715
assert entityCount == 878
if __name__ == '__main__':
test_pubannotation()
|
import kindred
def test_pubannotation():
corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 50
assert relationCount == 1454
assert entityCount == 2657
if __name__ == '__main__':
test_pubannotation()
|
Remove one of the pubannotation tests as their data seems to change
|
Remove one of the pubannotation tests as their data seems to change
|
Python
|
mit
|
jakelever/kindred,jakelever/kindred
|
import kindred
def test_pubannotation_groST():
corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 50
assert relationCount == 1454
assert entityCount == 2657
def test_pubannotation_wikiPain():
corpus = kindred.pubannotation.load('WikiPainGoldStandard')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 49
assert relationCount == 715
assert entityCount == 878
if __name__ == '__main__':
test_pubannotation()
Remove one of the pubannotation tests as their data seems to change
|
import kindred
def test_pubannotation():
corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 50
assert relationCount == 1454
assert entityCount == 2657
if __name__ == '__main__':
test_pubannotation()
|
<commit_before>
import kindred
def test_pubannotation_groST():
corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 50
assert relationCount == 1454
assert entityCount == 2657
def test_pubannotation_wikiPain():
corpus = kindred.pubannotation.load('WikiPainGoldStandard')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 49
assert relationCount == 715
assert entityCount == 878
if __name__ == '__main__':
test_pubannotation()
<commit_msg>Remove one of the pubannotation tests as their data seems to change<commit_after>
|
import kindred
def test_pubannotation():
corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 50
assert relationCount == 1454
assert entityCount == 2657
if __name__ == '__main__':
test_pubannotation()
|
import kindred
def test_pubannotation_groST():
corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 50
assert relationCount == 1454
assert entityCount == 2657
def test_pubannotation_wikiPain():
corpus = kindred.pubannotation.load('WikiPainGoldStandard')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 49
assert relationCount == 715
assert entityCount == 878
if __name__ == '__main__':
test_pubannotation()
Remove one of the pubannotation tests as their data seems to change
import kindred
def test_pubannotation():
corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 50
assert relationCount == 1454
assert entityCount == 2657
if __name__ == '__main__':
test_pubannotation()
|
<commit_before>
import kindred
def test_pubannotation_groST():
corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 50
assert relationCount == 1454
assert entityCount == 2657
def test_pubannotation_wikiPain():
corpus = kindred.pubannotation.load('WikiPainGoldStandard')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 49
assert relationCount == 715
assert entityCount == 878
if __name__ == '__main__':
test_pubannotation()
<commit_msg>Remove one of the pubannotation tests as their data seems to change<commit_after>
import kindred
def test_pubannotation():
corpus = kindred.pubannotation.load('bionlp-st-gro-2013-development')
assert isinstance(corpus,kindred.Corpus)
fileCount = len(corpus.documents)
entityCount = sum([ len(d.entities) for d in corpus.documents ])
relationCount = sum([ len(d.relations) for d in corpus.documents ])
assert fileCount == 50
assert relationCount == 1454
assert entityCount == 2657
if __name__ == '__main__':
test_pubannotation()
|
462700e3b1158fef187732007125a0930841dafd
|
bugsy/errors.py
|
bugsy/errors.py
|
class BugsyException(Exception):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
def __init__(self, msg, error_code=None):
self.msg = msg
self.code = error_code
def __str__(self):
return "Message: {message} Code: {code}".format(message=self.msg,
code=self.code)
class LoginException(BugsyException):
"""
If a username and password are passed in but we don't receive a token
then this error will be raised.
"""
pass
class BugException(BugsyException):
"""
If we try do something that is not allowed to a bug then
this error is raised
"""
pass
class SearchException(BugsyException):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
pass
|
class BugsyException(Exception):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
def __init__(self, msg, error_code=None):
self.msg = msg
self.code = error_code
def __str__(self):
return "Message: {message} Code: {code}".format(message=self.msg,
code=self.code)
class LoginException(BugsyException):
"""
If a username and password are passed in but we don't receive a token
then this error will be raised.
"""
pass
class AttachmentException(BugsyException):
"""
If we try do something that is not allowed to an attachment then
this error is raised
"""
pass
class BugException(BugsyException):
"""
If we try do something that is not allowed to a bug then
this error is raised
"""
pass
class SearchException(BugsyException):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
pass
|
Add exception handler for attachment related operations
|
Add exception handler for attachment related operations
|
Python
|
apache-2.0
|
AutomatedTester/Bugsy
|
class BugsyException(Exception):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
def __init__(self, msg, error_code=None):
self.msg = msg
self.code = error_code
def __str__(self):
return "Message: {message} Code: {code}".format(message=self.msg,
code=self.code)
class LoginException(BugsyException):
"""
If a username and password are passed in but we don't receive a token
then this error will be raised.
"""
pass
class BugException(BugsyException):
"""
If we try do something that is not allowed to a bug then
this error is raised
"""
pass
class SearchException(BugsyException):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
pass
Add exception handler for attachment related operations
|
class BugsyException(Exception):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
def __init__(self, msg, error_code=None):
self.msg = msg
self.code = error_code
def __str__(self):
return "Message: {message} Code: {code}".format(message=self.msg,
code=self.code)
class LoginException(BugsyException):
"""
If a username and password are passed in but we don't receive a token
then this error will be raised.
"""
pass
class AttachmentException(BugsyException):
"""
If we try do something that is not allowed to an attachment then
this error is raised
"""
pass
class BugException(BugsyException):
"""
If we try do something that is not allowed to a bug then
this error is raised
"""
pass
class SearchException(BugsyException):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
pass
|
<commit_before>class BugsyException(Exception):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
def __init__(self, msg, error_code=None):
self.msg = msg
self.code = error_code
def __str__(self):
return "Message: {message} Code: {code}".format(message=self.msg,
code=self.code)
class LoginException(BugsyException):
"""
If a username and password are passed in but we don't receive a token
then this error will be raised.
"""
pass
class BugException(BugsyException):
"""
If we try do something that is not allowed to a bug then
this error is raised
"""
pass
class SearchException(BugsyException):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
pass
<commit_msg>Add exception handler for attachment related operations<commit_after>
|
class BugsyException(Exception):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
def __init__(self, msg, error_code=None):
self.msg = msg
self.code = error_code
def __str__(self):
return "Message: {message} Code: {code}".format(message=self.msg,
code=self.code)
class LoginException(BugsyException):
"""
If a username and password are passed in but we don't receive a token
then this error will be raised.
"""
pass
class AttachmentException(BugsyException):
"""
If we try do something that is not allowed to an attachment then
this error is raised
"""
pass
class BugException(BugsyException):
"""
If we try do something that is not allowed to a bug then
this error is raised
"""
pass
class SearchException(BugsyException):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
pass
|
class BugsyException(Exception):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
def __init__(self, msg, error_code=None):
self.msg = msg
self.code = error_code
def __str__(self):
return "Message: {message} Code: {code}".format(message=self.msg,
code=self.code)
class LoginException(BugsyException):
"""
If a username and password are passed in but we don't receive a token
then this error will be raised.
"""
pass
class BugException(BugsyException):
"""
If we try do something that is not allowed to a bug then
this error is raised
"""
pass
class SearchException(BugsyException):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
pass
Add exception handler for attachment related operationsclass BugsyException(Exception):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
def __init__(self, msg, error_code=None):
self.msg = msg
self.code = error_code
def __str__(self):
return "Message: {message} Code: {code}".format(message=self.msg,
code=self.code)
class LoginException(BugsyException):
"""
If a username and password are passed in but we don't receive a token
then this error will be raised.
"""
pass
class AttachmentException(BugsyException):
"""
If we try do something that is not allowed to an attachment then
this error is raised
"""
pass
class BugException(BugsyException):
"""
If we try do something that is not allowed to a bug then
this error is raised
"""
pass
class SearchException(BugsyException):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
pass
|
<commit_before>class BugsyException(Exception):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
def __init__(self, msg, error_code=None):
self.msg = msg
self.code = error_code
def __str__(self):
return "Message: {message} Code: {code}".format(message=self.msg,
code=self.code)
class LoginException(BugsyException):
"""
If a username and password are passed in but we don't receive a token
then this error will be raised.
"""
pass
class BugException(BugsyException):
"""
If we try do something that is not allowed to a bug then
this error is raised
"""
pass
class SearchException(BugsyException):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
pass
<commit_msg>Add exception handler for attachment related operations<commit_after>class BugsyException(Exception):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
def __init__(self, msg, error_code=None):
self.msg = msg
self.code = error_code
def __str__(self):
return "Message: {message} Code: {code}".format(message=self.msg,
code=self.code)
class LoginException(BugsyException):
"""
If a username and password are passed in but we don't receive a token
then this error will be raised.
"""
pass
class AttachmentException(BugsyException):
"""
If we try do something that is not allowed to an attachment then
this error is raised
"""
pass
class BugException(BugsyException):
"""
If we try do something that is not allowed to a bug then
this error is raised
"""
pass
class SearchException(BugsyException):
"""
If while interacting with Bugzilla and we try do something that is not
supported this error will be raised.
"""
pass
|
31c921f0f88df5bc532db0f326ba9ef53318feb9
|
codejail/django_integration.py
|
codejail/django_integration.py
|
"""Django integration for codejail"""
from django.core.exceptions import MiddlewareNotUsed
from django.conf import settings
import codejail.jail_code
class ConfigureCodeJailMiddleware(object):
"""Middleware to configure codejail on startup."""
def __init__(self):
python_bin = settings.CODE_JAIL.get('python_bin')
if python_bin:
user = settings.CODE_JAIL['user']
codejail.jail_code.configure("python", python_bin, user=user)
raise MiddlewareNotUsed
|
"""Django integration for codejail.
Code to glue codejail into a Django environment.
"""
from django.core.exceptions import MiddlewareNotUsed
from django.conf import settings
import codejail.jail_code
class ConfigureCodeJailMiddleware(object):
"""
Middleware to configure codejail on startup.
This is a Django idiom to have code run once on server startup: put the
code in the `__init__` of some middleware, and have it do the work, then
raise `MiddlewareNotUsed` to disable the middleware.
"""
def __init__(self):
python_bin = settings.CODE_JAIL.get('python_bin')
if python_bin:
user = settings.CODE_JAIL['user']
codejail.jail_code.configure("python", python_bin, user=user)
raise MiddlewareNotUsed
|
Add more detail in docstring
|
Add more detail in docstring
|
Python
|
agpl-3.0
|
StepicOrg/codejail,edx/codejail
|
"""Django integration for codejail"""
from django.core.exceptions import MiddlewareNotUsed
from django.conf import settings
import codejail.jail_code
class ConfigureCodeJailMiddleware(object):
"""Middleware to configure codejail on startup."""
def __init__(self):
python_bin = settings.CODE_JAIL.get('python_bin')
if python_bin:
user = settings.CODE_JAIL['user']
codejail.jail_code.configure("python", python_bin, user=user)
raise MiddlewareNotUsed
Add more detail in docstring
|
"""Django integration for codejail.
Code to glue codejail into a Django environment.
"""
from django.core.exceptions import MiddlewareNotUsed
from django.conf import settings
import codejail.jail_code
class ConfigureCodeJailMiddleware(object):
"""
Middleware to configure codejail on startup.
This is a Django idiom to have code run once on server startup: put the
code in the `__init__` of some middleware, and have it do the work, then
raise `MiddlewareNotUsed` to disable the middleware.
"""
def __init__(self):
python_bin = settings.CODE_JAIL.get('python_bin')
if python_bin:
user = settings.CODE_JAIL['user']
codejail.jail_code.configure("python", python_bin, user=user)
raise MiddlewareNotUsed
|
<commit_before>"""Django integration for codejail"""
from django.core.exceptions import MiddlewareNotUsed
from django.conf import settings
import codejail.jail_code
class ConfigureCodeJailMiddleware(object):
"""Middleware to configure codejail on startup."""
def __init__(self):
python_bin = settings.CODE_JAIL.get('python_bin')
if python_bin:
user = settings.CODE_JAIL['user']
codejail.jail_code.configure("python", python_bin, user=user)
raise MiddlewareNotUsed
<commit_msg>Add more detail in docstring<commit_after>
|
"""Django integration for codejail.
Code to glue codejail into a Django environment.
"""
from django.core.exceptions import MiddlewareNotUsed
from django.conf import settings
import codejail.jail_code
class ConfigureCodeJailMiddleware(object):
"""
Middleware to configure codejail on startup.
This is a Django idiom to have code run once on server startup: put the
code in the `__init__` of some middleware, and have it do the work, then
raise `MiddlewareNotUsed` to disable the middleware.
"""
def __init__(self):
python_bin = settings.CODE_JAIL.get('python_bin')
if python_bin:
user = settings.CODE_JAIL['user']
codejail.jail_code.configure("python", python_bin, user=user)
raise MiddlewareNotUsed
|
"""Django integration for codejail"""
from django.core.exceptions import MiddlewareNotUsed
from django.conf import settings
import codejail.jail_code
class ConfigureCodeJailMiddleware(object):
"""Middleware to configure codejail on startup."""
def __init__(self):
python_bin = settings.CODE_JAIL.get('python_bin')
if python_bin:
user = settings.CODE_JAIL['user']
codejail.jail_code.configure("python", python_bin, user=user)
raise MiddlewareNotUsed
Add more detail in docstring"""Django integration for codejail.
Code to glue codejail into a Django environment.
"""
from django.core.exceptions import MiddlewareNotUsed
from django.conf import settings
import codejail.jail_code
class ConfigureCodeJailMiddleware(object):
"""
Middleware to configure codejail on startup.
This is a Django idiom to have code run once on server startup: put the
code in the `__init__` of some middleware, and have it do the work, then
raise `MiddlewareNotUsed` to disable the middleware.
"""
def __init__(self):
python_bin = settings.CODE_JAIL.get('python_bin')
if python_bin:
user = settings.CODE_JAIL['user']
codejail.jail_code.configure("python", python_bin, user=user)
raise MiddlewareNotUsed
|
<commit_before>"""Django integration for codejail"""
from django.core.exceptions import MiddlewareNotUsed
from django.conf import settings
import codejail.jail_code
class ConfigureCodeJailMiddleware(object):
"""Middleware to configure codejail on startup."""
def __init__(self):
python_bin = settings.CODE_JAIL.get('python_bin')
if python_bin:
user = settings.CODE_JAIL['user']
codejail.jail_code.configure("python", python_bin, user=user)
raise MiddlewareNotUsed
<commit_msg>Add more detail in docstring<commit_after>"""Django integration for codejail.
Code to glue codejail into a Django environment.
"""
from django.core.exceptions import MiddlewareNotUsed
from django.conf import settings
import codejail.jail_code
class ConfigureCodeJailMiddleware(object):
"""
Middleware to configure codejail on startup.
This is a Django idiom to have code run once on server startup: put the
code in the `__init__` of some middleware, and have it do the work, then
raise `MiddlewareNotUsed` to disable the middleware.
"""
def __init__(self):
python_bin = settings.CODE_JAIL.get('python_bin')
if python_bin:
user = settings.CODE_JAIL['user']
codejail.jail_code.configure("python", python_bin, user=user)
raise MiddlewareNotUsed
|
55ba2c2310a0f3a4a413801ce8edf52e001c9ffd
|
tornado_srv.py
|
tornado_srv.py
|
import tornado.web
import tornado.wsgi
import tornado.httpserver
import tornado.ioloop
from mojibake.main import app
from mojibake.settings import PORT
container = tornado.wsgi.WSGIContainer(app)
http_server = tornado.httpserver.HTTPServer(container)
http_server.listen(PORT)
tornado.ioloop.IOLoop.instance().start()
|
import tornado.web
import tornado.wsgi
import tornado.httpserver
import tornado.ioloop
import os
from mojibake.main import app
from mojibake.settings import PORT
if os.name == 'posix':
import setproctitle
setproctitle.setproctitle('mojibake') # Set the process title to mojibake
print('Starting Mojibake...')
container = tornado.wsgi.WSGIContainer(app)
http_server = tornado.httpserver.HTTPServer(container)
http_server.listen(PORT)
tornado.ioloop.IOLoop.instance().start()
|
Set the process title on posix systems
|
Set the process title on posix systems
|
Python
|
mit
|
ardinor/mojibake,ardinor/mojibake,ardinor/mojibake
|
import tornado.web
import tornado.wsgi
import tornado.httpserver
import tornado.ioloop
from mojibake.main import app
from mojibake.settings import PORT
container = tornado.wsgi.WSGIContainer(app)
http_server = tornado.httpserver.HTTPServer(container)
http_server.listen(PORT)
tornado.ioloop.IOLoop.instance().start()
Set the process title on posix systems
|
import tornado.web
import tornado.wsgi
import tornado.httpserver
import tornado.ioloop
import os
from mojibake.main import app
from mojibake.settings import PORT
if os.name == 'posix':
import setproctitle
setproctitle.setproctitle('mojibake') # Set the process title to mojibake
print('Starting Mojibake...')
container = tornado.wsgi.WSGIContainer(app)
http_server = tornado.httpserver.HTTPServer(container)
http_server.listen(PORT)
tornado.ioloop.IOLoop.instance().start()
|
<commit_before>import tornado.web
import tornado.wsgi
import tornado.httpserver
import tornado.ioloop
from mojibake.main import app
from mojibake.settings import PORT
container = tornado.wsgi.WSGIContainer(app)
http_server = tornado.httpserver.HTTPServer(container)
http_server.listen(PORT)
tornado.ioloop.IOLoop.instance().start()
<commit_msg>Set the process title on posix systems<commit_after>
|
import tornado.web
import tornado.wsgi
import tornado.httpserver
import tornado.ioloop
import os
from mojibake.main import app
from mojibake.settings import PORT
if os.name == 'posix':
import setproctitle
setproctitle.setproctitle('mojibake') # Set the process title to mojibake
print('Starting Mojibake...')
container = tornado.wsgi.WSGIContainer(app)
http_server = tornado.httpserver.HTTPServer(container)
http_server.listen(PORT)
tornado.ioloop.IOLoop.instance().start()
|
import tornado.web
import tornado.wsgi
import tornado.httpserver
import tornado.ioloop
from mojibake.main import app
from mojibake.settings import PORT
container = tornado.wsgi.WSGIContainer(app)
http_server = tornado.httpserver.HTTPServer(container)
http_server.listen(PORT)
tornado.ioloop.IOLoop.instance().start()
Set the process title on posix systemsimport tornado.web
import tornado.wsgi
import tornado.httpserver
import tornado.ioloop
import os
from mojibake.main import app
from mojibake.settings import PORT
if os.name == 'posix':
import setproctitle
setproctitle.setproctitle('mojibake') # Set the process title to mojibake
print('Starting Mojibake...')
container = tornado.wsgi.WSGIContainer(app)
http_server = tornado.httpserver.HTTPServer(container)
http_server.listen(PORT)
tornado.ioloop.IOLoop.instance().start()
|
<commit_before>import tornado.web
import tornado.wsgi
import tornado.httpserver
import tornado.ioloop
from mojibake.main import app
from mojibake.settings import PORT
container = tornado.wsgi.WSGIContainer(app)
http_server = tornado.httpserver.HTTPServer(container)
http_server.listen(PORT)
tornado.ioloop.IOLoop.instance().start()
<commit_msg>Set the process title on posix systems<commit_after>import tornado.web
import tornado.wsgi
import tornado.httpserver
import tornado.ioloop
import os
from mojibake.main import app
from mojibake.settings import PORT
if os.name == 'posix':
import setproctitle
setproctitle.setproctitle('mojibake') # Set the process title to mojibake
print('Starting Mojibake...')
container = tornado.wsgi.WSGIContainer(app)
http_server = tornado.httpserver.HTTPServer(container)
http_server.listen(PORT)
tornado.ioloop.IOLoop.instance().start()
|
2e9e14980d87239f861377d1dac45bb04d3f9712
|
tests/basics/array_intbig.py
|
tests/basics/array_intbig.py
|
# test array('q') and array('Q')
try:
from array import array
except ImportError:
import sys
print("SKIP")
sys.exit()
print(array('q'))
print(array('Q'))
print(array('q', [0]))
print(array('Q', [0]))
print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1]))
print(array('Q', [0, 1, 2, 2**64-1]))
print(bytes(array('q', [-1])))
print(bytes(array('Q', [2**64-1])))
|
# test array types QqLl that require big-ints
try:
from array import array
except ImportError:
import sys
print("SKIP")
sys.exit()
print(array('L', [0, 2**32-1]))
print(array('l', [-2**31, 0, 2**31-1]))
print(array('q'))
print(array('Q'))
print(array('q', [0]))
print(array('Q', [0]))
print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1]))
print(array('Q', [0, 1, 2, 2**64-1]))
print(bytes(array('q', [-1])))
print(bytes(array('Q', [2**64-1])))
|
Update array test for big-int with lL typecodes.
|
tests/basics: Update array test for big-int with lL typecodes.
|
Python
|
mit
|
TDAbboud/micropython,tralamazza/micropython,hiway/micropython,AriZuu/micropython,puuu/micropython,lowRISC/micropython,torwag/micropython,ryannathans/micropython,bvernoux/micropython,pozetroninc/micropython,pramasoul/micropython,deshipu/micropython,tralamazza/micropython,trezor/micropython,pramasoul/micropython,swegener/micropython,adafruit/micropython,adafruit/circuitpython,SHA2017-badge/micropython-esp32,pfalcon/micropython,alex-robbins/micropython,AriZuu/micropython,chrisdearman/micropython,adafruit/circuitpython,trezor/micropython,pozetroninc/micropython,HenrikSolver/micropython,bvernoux/micropython,pozetroninc/micropython,ryannathans/micropython,oopy/micropython,ryannathans/micropython,kerneltask/micropython,lowRISC/micropython,dmazzella/micropython,SHA2017-badge/micropython-esp32,infinnovation/micropython,oopy/micropython,lowRISC/micropython,kerneltask/micropython,blazewicz/micropython,Timmenem/micropython,pramasoul/micropython,trezor/micropython,oopy/micropython,tobbad/micropython,Timmenem/micropython,adafruit/circuitpython,toolmacher/micropython,deshipu/micropython,TDAbboud/micropython,trezor/micropython,chrisdearman/micropython,MrSurly/micropython,dmazzella/micropython,blazewicz/micropython,adafruit/micropython,henriknelson/micropython,micropython/micropython-esp32,pfalcon/micropython,puuu/micropython,PappaPeppar/micropython,pfalcon/micropython,adafruit/circuitpython,MrSurly/micropython-esp32,selste/micropython,pozetroninc/micropython,micropython/micropython-esp32,Timmenem/micropython,torwag/micropython,HenrikSolver/micropython,dmazzella/micropython,alex-robbins/micropython,swegener/micropython,alex-robbins/micropython,dmazzella/micropython,MrSurly/micropython,PappaPeppar/micropython,oopy/micropython,MrSurly/micropython,micropython/micropython-esp32,bvernoux/micropython,toolmacher/micropython,hiway/micropython,pfalcon/micropython,pramasoul/micropython,HenrikSolver/micropython,cwyark/micropython,pfalcon/micropython,tobbad/micropython,alex-robbins/micropython,lowRISC/micropython,cwyark/micropython,AriZuu/micropython,selste/micropython,PappaPeppar/micropython,infinnovation/micropython,deshipu/micropython,torwag/micropython,HenrikSolver/micropython,MrSurly/micropython,lowRISC/micropython,selste/micropython,henriknelson/micropython,swegener/micropython,kerneltask/micropython,cwyark/micropython,TDAbboud/micropython,deshipu/micropython,torwag/micropython,tobbad/micropython,adafruit/micropython,tobbad/micropython,MrSurly/micropython-esp32,cwyark/micropython,puuu/micropython,MrSurly/micropython-esp32,pramasoul/micropython,HenrikSolver/micropython,MrSurly/micropython-esp32,bvernoux/micropython,kerneltask/micropython,SHA2017-badge/micropython-esp32,tralamazza/micropython,deshipu/micropython,chrisdearman/micropython,puuu/micropython,TDAbboud/micropython,bvernoux/micropython,toolmacher/micropython,ryannathans/micropython,toolmacher/micropython,alex-robbins/micropython,adafruit/circuitpython,PappaPeppar/micropython,micropython/micropython-esp32,hiway/micropython,swegener/micropython,blazewicz/micropython,selste/micropython,swegener/micropython,blazewicz/micropython,henriknelson/micropython,pozetroninc/micropython,PappaPeppar/micropython,TDAbboud/micropython,selste/micropython,toolmacher/micropython,puuu/micropython,SHA2017-badge/micropython-esp32,adafruit/circuitpython,blazewicz/micropython,SHA2017-badge/micropython-esp32,chrisdearman/micropython,infinnovation/micropython,trezor/micropython,adafruit/micropython,Timmenem/micropython,henriknelson/micropython,AriZuu/micropython,hiway/micropython,infinnovation/micropython,adafruit/micropython,tralamazza/micropython,infinnovation/micropython,cwyark/micropython,AriZuu/micropython,kerneltask/micropython,henriknelson/micropython,chrisdearman/micropython,Timmenem/micropython,tobbad/micropython,hiway/micropython,MrSurly/micropython,micropython/micropython-esp32,MrSurly/micropython-esp32,ryannathans/micropython,oopy/micropython,torwag/micropython
|
# test array('q') and array('Q')
try:
from array import array
except ImportError:
import sys
print("SKIP")
sys.exit()
print(array('q'))
print(array('Q'))
print(array('q', [0]))
print(array('Q', [0]))
print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1]))
print(array('Q', [0, 1, 2, 2**64-1]))
print(bytes(array('q', [-1])))
print(bytes(array('Q', [2**64-1])))
tests/basics: Update array test for big-int with lL typecodes.
|
# test array types QqLl that require big-ints
try:
from array import array
except ImportError:
import sys
print("SKIP")
sys.exit()
print(array('L', [0, 2**32-1]))
print(array('l', [-2**31, 0, 2**31-1]))
print(array('q'))
print(array('Q'))
print(array('q', [0]))
print(array('Q', [0]))
print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1]))
print(array('Q', [0, 1, 2, 2**64-1]))
print(bytes(array('q', [-1])))
print(bytes(array('Q', [2**64-1])))
|
<commit_before># test array('q') and array('Q')
try:
from array import array
except ImportError:
import sys
print("SKIP")
sys.exit()
print(array('q'))
print(array('Q'))
print(array('q', [0]))
print(array('Q', [0]))
print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1]))
print(array('Q', [0, 1, 2, 2**64-1]))
print(bytes(array('q', [-1])))
print(bytes(array('Q', [2**64-1])))
<commit_msg>tests/basics: Update array test for big-int with lL typecodes.<commit_after>
|
# test array types QqLl that require big-ints
try:
from array import array
except ImportError:
import sys
print("SKIP")
sys.exit()
print(array('L', [0, 2**32-1]))
print(array('l', [-2**31, 0, 2**31-1]))
print(array('q'))
print(array('Q'))
print(array('q', [0]))
print(array('Q', [0]))
print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1]))
print(array('Q', [0, 1, 2, 2**64-1]))
print(bytes(array('q', [-1])))
print(bytes(array('Q', [2**64-1])))
|
# test array('q') and array('Q')
try:
from array import array
except ImportError:
import sys
print("SKIP")
sys.exit()
print(array('q'))
print(array('Q'))
print(array('q', [0]))
print(array('Q', [0]))
print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1]))
print(array('Q', [0, 1, 2, 2**64-1]))
print(bytes(array('q', [-1])))
print(bytes(array('Q', [2**64-1])))
tests/basics: Update array test for big-int with lL typecodes.# test array types QqLl that require big-ints
try:
from array import array
except ImportError:
import sys
print("SKIP")
sys.exit()
print(array('L', [0, 2**32-1]))
print(array('l', [-2**31, 0, 2**31-1]))
print(array('q'))
print(array('Q'))
print(array('q', [0]))
print(array('Q', [0]))
print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1]))
print(array('Q', [0, 1, 2, 2**64-1]))
print(bytes(array('q', [-1])))
print(bytes(array('Q', [2**64-1])))
|
<commit_before># test array('q') and array('Q')
try:
from array import array
except ImportError:
import sys
print("SKIP")
sys.exit()
print(array('q'))
print(array('Q'))
print(array('q', [0]))
print(array('Q', [0]))
print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1]))
print(array('Q', [0, 1, 2, 2**64-1]))
print(bytes(array('q', [-1])))
print(bytes(array('Q', [2**64-1])))
<commit_msg>tests/basics: Update array test for big-int with lL typecodes.<commit_after># test array types QqLl that require big-ints
try:
from array import array
except ImportError:
import sys
print("SKIP")
sys.exit()
print(array('L', [0, 2**32-1]))
print(array('l', [-2**31, 0, 2**31-1]))
print(array('q'))
print(array('Q'))
print(array('q', [0]))
print(array('Q', [0]))
print(array('q', [-2**63, -1, 0, 1, 2, 2**63-1]))
print(array('Q', [0, 1, 2, 2**64-1]))
print(bytes(array('q', [-1])))
print(bytes(array('Q', [2**64-1])))
|
b9ef72138c5312fe8eb7cfa48abe48a8c477afdc
|
test/test_type_checker_creator.py
|
test/test_type_checker_creator.py
|
# encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import pytest
from dataproperty._type_checker_creator import IntegerTypeCheckerCreator
from dataproperty._type_checker_creator import FloatTypeCheckerCreator
from dataproperty._type_checker_creator import DateTimeTypeCheckerCreator
from dataproperty._type_checker import IntegerTypeChecker
from dataproperty._type_checker import FloatTypeChecker
from dataproperty._type_checker import DateTimeTypeChecker
class Test_TypeCheckerCreator(object):
@pytest.mark.parametrize(["value", "is_convert", "expected"], [
[IntegerTypeCheckerCreator, True, IntegerTypeChecker],
[IntegerTypeCheckerCreator, False, IntegerTypeChecker],
[FloatTypeCheckerCreator, True, FloatTypeChecker],
[FloatTypeCheckerCreator, False, FloatTypeChecker],
[DateTimeTypeCheckerCreator, True, DateTimeTypeChecker],
[DateTimeTypeCheckerCreator, False, DateTimeTypeChecker],
])
def test_normal(self, value, is_convert, expected):
creator = value()
type_checker = creator.create(None, is_convert)
assert isinstance(type_checker, expected)
|
# encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import pytest
import dataproperty._type_checker_creator as tcc
import dataproperty._type_checker as tc
class Test_TypeCheckerCreator(object):
@pytest.mark.parametrize(["value", "is_convert", "expected"], [
[tcc.NoneTypeCheckerCreator, True, tc.NoneTypeChecker],
[tcc.NoneTypeCheckerCreator, False, tc.NoneTypeChecker],
[tcc.IntegerTypeCheckerCreator, True, tc.IntegerTypeChecker],
[tcc.IntegerTypeCheckerCreator, False, tc.IntegerTypeChecker],
[tcc.FloatTypeCheckerCreator, True, tc.FloatTypeChecker],
[tcc.FloatTypeCheckerCreator, False, tc.FloatTypeChecker],
[tcc.DateTimeTypeCheckerCreator, True, tc.DateTimeTypeChecker],
[tcc.DateTimeTypeCheckerCreator, False, tc.DateTimeTypeChecker],
])
def test_normal(self, value, is_convert, expected):
creator = value()
type_checker = creator.create(None, is_convert)
assert isinstance(type_checker, expected)
|
Add tests for NoneTypeCheckerCreator class
|
Add tests for NoneTypeCheckerCreator class
|
Python
|
mit
|
thombashi/DataProperty
|
# encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import pytest
from dataproperty._type_checker_creator import IntegerTypeCheckerCreator
from dataproperty._type_checker_creator import FloatTypeCheckerCreator
from dataproperty._type_checker_creator import DateTimeTypeCheckerCreator
from dataproperty._type_checker import IntegerTypeChecker
from dataproperty._type_checker import FloatTypeChecker
from dataproperty._type_checker import DateTimeTypeChecker
class Test_TypeCheckerCreator(object):
@pytest.mark.parametrize(["value", "is_convert", "expected"], [
[IntegerTypeCheckerCreator, True, IntegerTypeChecker],
[IntegerTypeCheckerCreator, False, IntegerTypeChecker],
[FloatTypeCheckerCreator, True, FloatTypeChecker],
[FloatTypeCheckerCreator, False, FloatTypeChecker],
[DateTimeTypeCheckerCreator, True, DateTimeTypeChecker],
[DateTimeTypeCheckerCreator, False, DateTimeTypeChecker],
])
def test_normal(self, value, is_convert, expected):
creator = value()
type_checker = creator.create(None, is_convert)
assert isinstance(type_checker, expected)
Add tests for NoneTypeCheckerCreator class
|
# encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import pytest
import dataproperty._type_checker_creator as tcc
import dataproperty._type_checker as tc
class Test_TypeCheckerCreator(object):
@pytest.mark.parametrize(["value", "is_convert", "expected"], [
[tcc.NoneTypeCheckerCreator, True, tc.NoneTypeChecker],
[tcc.NoneTypeCheckerCreator, False, tc.NoneTypeChecker],
[tcc.IntegerTypeCheckerCreator, True, tc.IntegerTypeChecker],
[tcc.IntegerTypeCheckerCreator, False, tc.IntegerTypeChecker],
[tcc.FloatTypeCheckerCreator, True, tc.FloatTypeChecker],
[tcc.FloatTypeCheckerCreator, False, tc.FloatTypeChecker],
[tcc.DateTimeTypeCheckerCreator, True, tc.DateTimeTypeChecker],
[tcc.DateTimeTypeCheckerCreator, False, tc.DateTimeTypeChecker],
])
def test_normal(self, value, is_convert, expected):
creator = value()
type_checker = creator.create(None, is_convert)
assert isinstance(type_checker, expected)
|
<commit_before># encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import pytest
from dataproperty._type_checker_creator import IntegerTypeCheckerCreator
from dataproperty._type_checker_creator import FloatTypeCheckerCreator
from dataproperty._type_checker_creator import DateTimeTypeCheckerCreator
from dataproperty._type_checker import IntegerTypeChecker
from dataproperty._type_checker import FloatTypeChecker
from dataproperty._type_checker import DateTimeTypeChecker
class Test_TypeCheckerCreator(object):
@pytest.mark.parametrize(["value", "is_convert", "expected"], [
[IntegerTypeCheckerCreator, True, IntegerTypeChecker],
[IntegerTypeCheckerCreator, False, IntegerTypeChecker],
[FloatTypeCheckerCreator, True, FloatTypeChecker],
[FloatTypeCheckerCreator, False, FloatTypeChecker],
[DateTimeTypeCheckerCreator, True, DateTimeTypeChecker],
[DateTimeTypeCheckerCreator, False, DateTimeTypeChecker],
])
def test_normal(self, value, is_convert, expected):
creator = value()
type_checker = creator.create(None, is_convert)
assert isinstance(type_checker, expected)
<commit_msg>Add tests for NoneTypeCheckerCreator class<commit_after>
|
# encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import pytest
import dataproperty._type_checker_creator as tcc
import dataproperty._type_checker as tc
class Test_TypeCheckerCreator(object):
@pytest.mark.parametrize(["value", "is_convert", "expected"], [
[tcc.NoneTypeCheckerCreator, True, tc.NoneTypeChecker],
[tcc.NoneTypeCheckerCreator, False, tc.NoneTypeChecker],
[tcc.IntegerTypeCheckerCreator, True, tc.IntegerTypeChecker],
[tcc.IntegerTypeCheckerCreator, False, tc.IntegerTypeChecker],
[tcc.FloatTypeCheckerCreator, True, tc.FloatTypeChecker],
[tcc.FloatTypeCheckerCreator, False, tc.FloatTypeChecker],
[tcc.DateTimeTypeCheckerCreator, True, tc.DateTimeTypeChecker],
[tcc.DateTimeTypeCheckerCreator, False, tc.DateTimeTypeChecker],
])
def test_normal(self, value, is_convert, expected):
creator = value()
type_checker = creator.create(None, is_convert)
assert isinstance(type_checker, expected)
|
# encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import pytest
from dataproperty._type_checker_creator import IntegerTypeCheckerCreator
from dataproperty._type_checker_creator import FloatTypeCheckerCreator
from dataproperty._type_checker_creator import DateTimeTypeCheckerCreator
from dataproperty._type_checker import IntegerTypeChecker
from dataproperty._type_checker import FloatTypeChecker
from dataproperty._type_checker import DateTimeTypeChecker
class Test_TypeCheckerCreator(object):
@pytest.mark.parametrize(["value", "is_convert", "expected"], [
[IntegerTypeCheckerCreator, True, IntegerTypeChecker],
[IntegerTypeCheckerCreator, False, IntegerTypeChecker],
[FloatTypeCheckerCreator, True, FloatTypeChecker],
[FloatTypeCheckerCreator, False, FloatTypeChecker],
[DateTimeTypeCheckerCreator, True, DateTimeTypeChecker],
[DateTimeTypeCheckerCreator, False, DateTimeTypeChecker],
])
def test_normal(self, value, is_convert, expected):
creator = value()
type_checker = creator.create(None, is_convert)
assert isinstance(type_checker, expected)
Add tests for NoneTypeCheckerCreator class# encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import pytest
import dataproperty._type_checker_creator as tcc
import dataproperty._type_checker as tc
class Test_TypeCheckerCreator(object):
@pytest.mark.parametrize(["value", "is_convert", "expected"], [
[tcc.NoneTypeCheckerCreator, True, tc.NoneTypeChecker],
[tcc.NoneTypeCheckerCreator, False, tc.NoneTypeChecker],
[tcc.IntegerTypeCheckerCreator, True, tc.IntegerTypeChecker],
[tcc.IntegerTypeCheckerCreator, False, tc.IntegerTypeChecker],
[tcc.FloatTypeCheckerCreator, True, tc.FloatTypeChecker],
[tcc.FloatTypeCheckerCreator, False, tc.FloatTypeChecker],
[tcc.DateTimeTypeCheckerCreator, True, tc.DateTimeTypeChecker],
[tcc.DateTimeTypeCheckerCreator, False, tc.DateTimeTypeChecker],
])
def test_normal(self, value, is_convert, expected):
creator = value()
type_checker = creator.create(None, is_convert)
assert isinstance(type_checker, expected)
|
<commit_before># encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import pytest
from dataproperty._type_checker_creator import IntegerTypeCheckerCreator
from dataproperty._type_checker_creator import FloatTypeCheckerCreator
from dataproperty._type_checker_creator import DateTimeTypeCheckerCreator
from dataproperty._type_checker import IntegerTypeChecker
from dataproperty._type_checker import FloatTypeChecker
from dataproperty._type_checker import DateTimeTypeChecker
class Test_TypeCheckerCreator(object):
@pytest.mark.parametrize(["value", "is_convert", "expected"], [
[IntegerTypeCheckerCreator, True, IntegerTypeChecker],
[IntegerTypeCheckerCreator, False, IntegerTypeChecker],
[FloatTypeCheckerCreator, True, FloatTypeChecker],
[FloatTypeCheckerCreator, False, FloatTypeChecker],
[DateTimeTypeCheckerCreator, True, DateTimeTypeChecker],
[DateTimeTypeCheckerCreator, False, DateTimeTypeChecker],
])
def test_normal(self, value, is_convert, expected):
creator = value()
type_checker = creator.create(None, is_convert)
assert isinstance(type_checker, expected)
<commit_msg>Add tests for NoneTypeCheckerCreator class<commit_after># encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import pytest
import dataproperty._type_checker_creator as tcc
import dataproperty._type_checker as tc
class Test_TypeCheckerCreator(object):
@pytest.mark.parametrize(["value", "is_convert", "expected"], [
[tcc.NoneTypeCheckerCreator, True, tc.NoneTypeChecker],
[tcc.NoneTypeCheckerCreator, False, tc.NoneTypeChecker],
[tcc.IntegerTypeCheckerCreator, True, tc.IntegerTypeChecker],
[tcc.IntegerTypeCheckerCreator, False, tc.IntegerTypeChecker],
[tcc.FloatTypeCheckerCreator, True, tc.FloatTypeChecker],
[tcc.FloatTypeCheckerCreator, False, tc.FloatTypeChecker],
[tcc.DateTimeTypeCheckerCreator, True, tc.DateTimeTypeChecker],
[tcc.DateTimeTypeCheckerCreator, False, tc.DateTimeTypeChecker],
])
def test_normal(self, value, is_convert, expected):
creator = value()
type_checker = creator.create(None, is_convert)
assert isinstance(type_checker, expected)
|
927915f11ce536074920c515fab6e6ec3134d390
|
tests/test_huckle_install.py
|
tests/test_huckle_install.py
|
from __future__ import absolute_import, division, print_function
from subprocess import check_output
import os
def test_function():
setup = """
#!/bin/bash
huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf
echo '{"hello":"world"}' | jsonf go
"""
out = check_output(['bash', '-c', setup])
hello = """
#!/bin/bash
export PATH=$PATH:~/.huckle/bin
echo '{"hello":"world"}' | jsonf go
"""
out = check_output(['bash', '-c', hello])
assert('{\n "hello" : "world"\n}\n' in out)
|
from __future__ import absolute_import, division, print_function
import subprocess
import os
def test_function():
setup = """
#!/bin/bash
huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf
echo '{"hello":"world"}' | jsonf go
"""
p1 = subprocess.Popen(['bash', '-c', setup], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
out, err = p1.communicate()
hello = """
#!/bin/bash
export PATH=$PATH:~/.huckle/bin
echo '{"hello":"world"}' | jsonf go
"""
p2 = subprocess.Popen(['bash', '-c', hello], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
out, err = p2.communicate()
result = out.decode('utf-8')
assert('{\n "hello" : "world"\n}\n' in result)
|
Revert "fix test by switching to check_output"
|
Revert "fix test by switching to check_output"
This reverts commit 6cfd9d01d68c2f7ff4a8bba3351ee618e770d315.
|
Python
|
mit
|
cometaj2/huckle
|
from __future__ import absolute_import, division, print_function
from subprocess import check_output
import os
def test_function():
setup = """
#!/bin/bash
huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf
echo '{"hello":"world"}' | jsonf go
"""
out = check_output(['bash', '-c', setup])
hello = """
#!/bin/bash
export PATH=$PATH:~/.huckle/bin
echo '{"hello":"world"}' | jsonf go
"""
out = check_output(['bash', '-c', hello])
assert('{\n "hello" : "world"\n}\n' in out)
Revert "fix test by switching to check_output"
This reverts commit 6cfd9d01d68c2f7ff4a8bba3351ee618e770d315.
|
from __future__ import absolute_import, division, print_function
import subprocess
import os
def test_function():
setup = """
#!/bin/bash
huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf
echo '{"hello":"world"}' | jsonf go
"""
p1 = subprocess.Popen(['bash', '-c', setup], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
out, err = p1.communicate()
hello = """
#!/bin/bash
export PATH=$PATH:~/.huckle/bin
echo '{"hello":"world"}' | jsonf go
"""
p2 = subprocess.Popen(['bash', '-c', hello], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
out, err = p2.communicate()
result = out.decode('utf-8')
assert('{\n "hello" : "world"\n}\n' in result)
|
<commit_before>from __future__ import absolute_import, division, print_function
from subprocess import check_output
import os
def test_function():
setup = """
#!/bin/bash
huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf
echo '{"hello":"world"}' | jsonf go
"""
out = check_output(['bash', '-c', setup])
hello = """
#!/bin/bash
export PATH=$PATH:~/.huckle/bin
echo '{"hello":"world"}' | jsonf go
"""
out = check_output(['bash', '-c', hello])
assert('{\n "hello" : "world"\n}\n' in out)
<commit_msg>Revert "fix test by switching to check_output"
This reverts commit 6cfd9d01d68c2f7ff4a8bba3351ee618e770d315.<commit_after>
|
from __future__ import absolute_import, division, print_function
import subprocess
import os
def test_function():
setup = """
#!/bin/bash
huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf
echo '{"hello":"world"}' | jsonf go
"""
p1 = subprocess.Popen(['bash', '-c', setup], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
out, err = p1.communicate()
hello = """
#!/bin/bash
export PATH=$PATH:~/.huckle/bin
echo '{"hello":"world"}' | jsonf go
"""
p2 = subprocess.Popen(['bash', '-c', hello], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
out, err = p2.communicate()
result = out.decode('utf-8')
assert('{\n "hello" : "world"\n}\n' in result)
|
from __future__ import absolute_import, division, print_function
from subprocess import check_output
import os
def test_function():
setup = """
#!/bin/bash
huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf
echo '{"hello":"world"}' | jsonf go
"""
out = check_output(['bash', '-c', setup])
hello = """
#!/bin/bash
export PATH=$PATH:~/.huckle/bin
echo '{"hello":"world"}' | jsonf go
"""
out = check_output(['bash', '-c', hello])
assert('{\n "hello" : "world"\n}\n' in out)
Revert "fix test by switching to check_output"
This reverts commit 6cfd9d01d68c2f7ff4a8bba3351ee618e770d315.from __future__ import absolute_import, division, print_function
import subprocess
import os
def test_function():
setup = """
#!/bin/bash
huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf
echo '{"hello":"world"}' | jsonf go
"""
p1 = subprocess.Popen(['bash', '-c', setup], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
out, err = p1.communicate()
hello = """
#!/bin/bash
export PATH=$PATH:~/.huckle/bin
echo '{"hello":"world"}' | jsonf go
"""
p2 = subprocess.Popen(['bash', '-c', hello], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
out, err = p2.communicate()
result = out.decode('utf-8')
assert('{\n "hello" : "world"\n}\n' in result)
|
<commit_before>from __future__ import absolute_import, division, print_function
from subprocess import check_output
import os
def test_function():
setup = """
#!/bin/bash
huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf
echo '{"hello":"world"}' | jsonf go
"""
out = check_output(['bash', '-c', setup])
hello = """
#!/bin/bash
export PATH=$PATH:~/.huckle/bin
echo '{"hello":"world"}' | jsonf go
"""
out = check_output(['bash', '-c', hello])
assert('{\n "hello" : "world"\n}\n' in out)
<commit_msg>Revert "fix test by switching to check_output"
This reverts commit 6cfd9d01d68c2f7ff4a8bba3351ee618e770d315.<commit_after>from __future__ import absolute_import, division, print_function
import subprocess
import os
def test_function():
setup = """
#!/bin/bash
huckle install https://hcli.io/hcli/cli/jsonf?command=jsonf
echo '{"hello":"world"}' | jsonf go
"""
p1 = subprocess.Popen(['bash', '-c', setup], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
out, err = p1.communicate()
hello = """
#!/bin/bash
export PATH=$PATH:~/.huckle/bin
echo '{"hello":"world"}' | jsonf go
"""
p2 = subprocess.Popen(['bash', '-c', hello], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
out, err = p2.communicate()
result = out.decode('utf-8')
assert('{\n "hello" : "world"\n}\n' in result)
|
1c0f0decd5bdcea3174cee650ba08fb427b67016
|
tests/test_rover_instance.py
|
tests/test_rover_instance.py
|
from unittest import TestCase
from rover import Rover
class TestRover(TestCase):
def setUp(self):
self.rover = Rover()
def test_rover_compass(self):
assert self.rover.compass == ['N', 'E', 'S', 'W']
def test_rover_position(self):
assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction)
def test_rover_set_position(self):
self.rover.set_position(4, 9, 'W')
assert self.rover.position == (4, 9, 'W')
|
from unittest import TestCase
from rover import Rover
class TestRover(TestCase):
def setUp(self):
self.rover = Rover()
def test_rover_compass(self):
assert self.rover.compass == ['N', 'E', 'S', 'W']
def test_rover_position(self):
assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction)
def test_rover_set_position(self):
self.rover.set_position(4, 9, 'W')
assert self.rover.position == (4, 9, 'W')
def test_rover_move_forward_north(self):
self.rover.set_position(0, 0, 'N')
self.rover.move('F')
assert self.rover.position == (0, 1, 'N')
def test_rover_move_forward_south(self):
self.rover.set_position(0, 1, 'S')
self.rover.move('F')
assert self.rover.position == (0, 0, 'S')
def test_rover_move_forward_east(self):
self.rover.set_position(0, 0, 'E')
self.rover.move('F')
assert self.rover.position == (1, 0, 'E')
def test_rover_move_forward_west(self):
self.rover.set_position(1, 0, 'W')
self.rover.move('F')
assert self.rover.position == (0, 0, 'W')
|
Add failing tests for rover forward movement
|
Add failing tests for rover forward movement
|
Python
|
mit
|
authentik8/rover
|
from unittest import TestCase
from rover import Rover
class TestRover(TestCase):
def setUp(self):
self.rover = Rover()
def test_rover_compass(self):
assert self.rover.compass == ['N', 'E', 'S', 'W']
def test_rover_position(self):
assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction)
def test_rover_set_position(self):
self.rover.set_position(4, 9, 'W')
assert self.rover.position == (4, 9, 'W')
Add failing tests for rover forward movement
|
from unittest import TestCase
from rover import Rover
class TestRover(TestCase):
def setUp(self):
self.rover = Rover()
def test_rover_compass(self):
assert self.rover.compass == ['N', 'E', 'S', 'W']
def test_rover_position(self):
assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction)
def test_rover_set_position(self):
self.rover.set_position(4, 9, 'W')
assert self.rover.position == (4, 9, 'W')
def test_rover_move_forward_north(self):
self.rover.set_position(0, 0, 'N')
self.rover.move('F')
assert self.rover.position == (0, 1, 'N')
def test_rover_move_forward_south(self):
self.rover.set_position(0, 1, 'S')
self.rover.move('F')
assert self.rover.position == (0, 0, 'S')
def test_rover_move_forward_east(self):
self.rover.set_position(0, 0, 'E')
self.rover.move('F')
assert self.rover.position == (1, 0, 'E')
def test_rover_move_forward_west(self):
self.rover.set_position(1, 0, 'W')
self.rover.move('F')
assert self.rover.position == (0, 0, 'W')
|
<commit_before>
from unittest import TestCase
from rover import Rover
class TestRover(TestCase):
def setUp(self):
self.rover = Rover()
def test_rover_compass(self):
assert self.rover.compass == ['N', 'E', 'S', 'W']
def test_rover_position(self):
assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction)
def test_rover_set_position(self):
self.rover.set_position(4, 9, 'W')
assert self.rover.position == (4, 9, 'W')
<commit_msg>Add failing tests for rover forward movement<commit_after>
|
from unittest import TestCase
from rover import Rover
class TestRover(TestCase):
def setUp(self):
self.rover = Rover()
def test_rover_compass(self):
assert self.rover.compass == ['N', 'E', 'S', 'W']
def test_rover_position(self):
assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction)
def test_rover_set_position(self):
self.rover.set_position(4, 9, 'W')
assert self.rover.position == (4, 9, 'W')
def test_rover_move_forward_north(self):
self.rover.set_position(0, 0, 'N')
self.rover.move('F')
assert self.rover.position == (0, 1, 'N')
def test_rover_move_forward_south(self):
self.rover.set_position(0, 1, 'S')
self.rover.move('F')
assert self.rover.position == (0, 0, 'S')
def test_rover_move_forward_east(self):
self.rover.set_position(0, 0, 'E')
self.rover.move('F')
assert self.rover.position == (1, 0, 'E')
def test_rover_move_forward_west(self):
self.rover.set_position(1, 0, 'W')
self.rover.move('F')
assert self.rover.position == (0, 0, 'W')
|
from unittest import TestCase
from rover import Rover
class TestRover(TestCase):
def setUp(self):
self.rover = Rover()
def test_rover_compass(self):
assert self.rover.compass == ['N', 'E', 'S', 'W']
def test_rover_position(self):
assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction)
def test_rover_set_position(self):
self.rover.set_position(4, 9, 'W')
assert self.rover.position == (4, 9, 'W')
Add failing tests for rover forward movement
from unittest import TestCase
from rover import Rover
class TestRover(TestCase):
def setUp(self):
self.rover = Rover()
def test_rover_compass(self):
assert self.rover.compass == ['N', 'E', 'S', 'W']
def test_rover_position(self):
assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction)
def test_rover_set_position(self):
self.rover.set_position(4, 9, 'W')
assert self.rover.position == (4, 9, 'W')
def test_rover_move_forward_north(self):
self.rover.set_position(0, 0, 'N')
self.rover.move('F')
assert self.rover.position == (0, 1, 'N')
def test_rover_move_forward_south(self):
self.rover.set_position(0, 1, 'S')
self.rover.move('F')
assert self.rover.position == (0, 0, 'S')
def test_rover_move_forward_east(self):
self.rover.set_position(0, 0, 'E')
self.rover.move('F')
assert self.rover.position == (1, 0, 'E')
def test_rover_move_forward_west(self):
self.rover.set_position(1, 0, 'W')
self.rover.move('F')
assert self.rover.position == (0, 0, 'W')
|
<commit_before>
from unittest import TestCase
from rover import Rover
class TestRover(TestCase):
def setUp(self):
self.rover = Rover()
def test_rover_compass(self):
assert self.rover.compass == ['N', 'E', 'S', 'W']
def test_rover_position(self):
assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction)
def test_rover_set_position(self):
self.rover.set_position(4, 9, 'W')
assert self.rover.position == (4, 9, 'W')
<commit_msg>Add failing tests for rover forward movement<commit_after>
from unittest import TestCase
from rover import Rover
class TestRover(TestCase):
def setUp(self):
self.rover = Rover()
def test_rover_compass(self):
assert self.rover.compass == ['N', 'E', 'S', 'W']
def test_rover_position(self):
assert self.rover.position == (self.rover.x, self.rover.y, self.rover.direction)
def test_rover_set_position(self):
self.rover.set_position(4, 9, 'W')
assert self.rover.position == (4, 9, 'W')
def test_rover_move_forward_north(self):
self.rover.set_position(0, 0, 'N')
self.rover.move('F')
assert self.rover.position == (0, 1, 'N')
def test_rover_move_forward_south(self):
self.rover.set_position(0, 1, 'S')
self.rover.move('F')
assert self.rover.position == (0, 0, 'S')
def test_rover_move_forward_east(self):
self.rover.set_position(0, 0, 'E')
self.rover.move('F')
assert self.rover.position == (1, 0, 'E')
def test_rover_move_forward_west(self):
self.rover.set_position(1, 0, 'W')
self.rover.move('F')
assert self.rover.position == (0, 0, 'W')
|
05e61f1be4005edf2ff439ca2613bce8af217ff7
|
pubsubpull/models.py
|
pubsubpull/models.py
|
"""
Models.
"""
from django.contrib.auth.models import User
from django.core.exceptions import ValidationError
from django.db import models
from pubsubpull.fields import JSONB
class Request(models.Model):
"""A web request.
"""
user = models.ForeignKey(User, null=True, blank=True, related_name='requests')
method = models.CharField(max_length=20)
path = models.TextField()
OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE")
class UpdateLog(models.Model):
"""Store a change to a single row in a table.
"""
table = models.CharField(max_length=200)
type = models.CharField(max_length=1, choices=OPERATION_TYPE.items())
when = models.DateTimeField(auto_now_add=True)
request = models.ForeignKey(Request, null=True, blank=True,
related_name='changes')
old = JSONB(null=True, blank=True)
new = JSONB(null=True, blank=True)
def save(self, **kw):
raise ValidationError("Instances of this class cannot be using Django")
def __unicode__(self):
return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
|
"""
Models.
"""
from django.contrib.auth.models import User
from django.core.exceptions import ValidationError
from django.db import models
from pubsubpull.fields import JSONB
class Request(models.Model):
"""A web request.
"""
user = models.ForeignKey(User, null=True, blank=True, related_name='requests')
method = models.CharField(max_length=20)
path = models.TextField()
def __unicode__(self):
return "%s %s" % (self.method, self.path)
OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE")
class UpdateLog(models.Model):
"""Store a change to a single row in a table.
"""
table = models.CharField(max_length=200)
type = models.CharField(max_length=1, choices=OPERATION_TYPE.items())
when = models.DateTimeField(auto_now_add=True)
request = models.ForeignKey(Request, null=True, blank=True,
related_name='changes')
old = JSONB(null=True, blank=True)
new = JSONB(null=True, blank=True)
def save(self, **kw):
raise ValidationError("Instances of this class cannot be using Django")
def __unicode__(self):
return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
|
Add more useful display of the request data.
|
Add more useful display of the request data.
|
Python
|
mit
|
KayEss/django-pubsubpull,KayEss/django-pubsubpull,KayEss/django-pubsubpull
|
"""
Models.
"""
from django.contrib.auth.models import User
from django.core.exceptions import ValidationError
from django.db import models
from pubsubpull.fields import JSONB
class Request(models.Model):
"""A web request.
"""
user = models.ForeignKey(User, null=True, blank=True, related_name='requests')
method = models.CharField(max_length=20)
path = models.TextField()
OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE")
class UpdateLog(models.Model):
"""Store a change to a single row in a table.
"""
table = models.CharField(max_length=200)
type = models.CharField(max_length=1, choices=OPERATION_TYPE.items())
when = models.DateTimeField(auto_now_add=True)
request = models.ForeignKey(Request, null=True, blank=True,
related_name='changes')
old = JSONB(null=True, blank=True)
new = JSONB(null=True, blank=True)
def save(self, **kw):
raise ValidationError("Instances of this class cannot be using Django")
def __unicode__(self):
return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
Add more useful display of the request data.
|
"""
Models.
"""
from django.contrib.auth.models import User
from django.core.exceptions import ValidationError
from django.db import models
from pubsubpull.fields import JSONB
class Request(models.Model):
"""A web request.
"""
user = models.ForeignKey(User, null=True, blank=True, related_name='requests')
method = models.CharField(max_length=20)
path = models.TextField()
def __unicode__(self):
return "%s %s" % (self.method, self.path)
OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE")
class UpdateLog(models.Model):
"""Store a change to a single row in a table.
"""
table = models.CharField(max_length=200)
type = models.CharField(max_length=1, choices=OPERATION_TYPE.items())
when = models.DateTimeField(auto_now_add=True)
request = models.ForeignKey(Request, null=True, blank=True,
related_name='changes')
old = JSONB(null=True, blank=True)
new = JSONB(null=True, blank=True)
def save(self, **kw):
raise ValidationError("Instances of this class cannot be using Django")
def __unicode__(self):
return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
|
<commit_before>"""
Models.
"""
from django.contrib.auth.models import User
from django.core.exceptions import ValidationError
from django.db import models
from pubsubpull.fields import JSONB
class Request(models.Model):
"""A web request.
"""
user = models.ForeignKey(User, null=True, blank=True, related_name='requests')
method = models.CharField(max_length=20)
path = models.TextField()
OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE")
class UpdateLog(models.Model):
"""Store a change to a single row in a table.
"""
table = models.CharField(max_length=200)
type = models.CharField(max_length=1, choices=OPERATION_TYPE.items())
when = models.DateTimeField(auto_now_add=True)
request = models.ForeignKey(Request, null=True, blank=True,
related_name='changes')
old = JSONB(null=True, blank=True)
new = JSONB(null=True, blank=True)
def save(self, **kw):
raise ValidationError("Instances of this class cannot be using Django")
def __unicode__(self):
return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
<commit_msg>Add more useful display of the request data.<commit_after>
|
"""
Models.
"""
from django.contrib.auth.models import User
from django.core.exceptions import ValidationError
from django.db import models
from pubsubpull.fields import JSONB
class Request(models.Model):
"""A web request.
"""
user = models.ForeignKey(User, null=True, blank=True, related_name='requests')
method = models.CharField(max_length=20)
path = models.TextField()
def __unicode__(self):
return "%s %s" % (self.method, self.path)
OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE")
class UpdateLog(models.Model):
"""Store a change to a single row in a table.
"""
table = models.CharField(max_length=200)
type = models.CharField(max_length=1, choices=OPERATION_TYPE.items())
when = models.DateTimeField(auto_now_add=True)
request = models.ForeignKey(Request, null=True, blank=True,
related_name='changes')
old = JSONB(null=True, blank=True)
new = JSONB(null=True, blank=True)
def save(self, **kw):
raise ValidationError("Instances of this class cannot be using Django")
def __unicode__(self):
return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
|
"""
Models.
"""
from django.contrib.auth.models import User
from django.core.exceptions import ValidationError
from django.db import models
from pubsubpull.fields import JSONB
class Request(models.Model):
"""A web request.
"""
user = models.ForeignKey(User, null=True, blank=True, related_name='requests')
method = models.CharField(max_length=20)
path = models.TextField()
OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE")
class UpdateLog(models.Model):
"""Store a change to a single row in a table.
"""
table = models.CharField(max_length=200)
type = models.CharField(max_length=1, choices=OPERATION_TYPE.items())
when = models.DateTimeField(auto_now_add=True)
request = models.ForeignKey(Request, null=True, blank=True,
related_name='changes')
old = JSONB(null=True, blank=True)
new = JSONB(null=True, blank=True)
def save(self, **kw):
raise ValidationError("Instances of this class cannot be using Django")
def __unicode__(self):
return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
Add more useful display of the request data."""
Models.
"""
from django.contrib.auth.models import User
from django.core.exceptions import ValidationError
from django.db import models
from pubsubpull.fields import JSONB
class Request(models.Model):
"""A web request.
"""
user = models.ForeignKey(User, null=True, blank=True, related_name='requests')
method = models.CharField(max_length=20)
path = models.TextField()
def __unicode__(self):
return "%s %s" % (self.method, self.path)
OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE")
class UpdateLog(models.Model):
"""Store a change to a single row in a table.
"""
table = models.CharField(max_length=200)
type = models.CharField(max_length=1, choices=OPERATION_TYPE.items())
when = models.DateTimeField(auto_now_add=True)
request = models.ForeignKey(Request, null=True, blank=True,
related_name='changes')
old = JSONB(null=True, blank=True)
new = JSONB(null=True, blank=True)
def save(self, **kw):
raise ValidationError("Instances of this class cannot be using Django")
def __unicode__(self):
return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
|
<commit_before>"""
Models.
"""
from django.contrib.auth.models import User
from django.core.exceptions import ValidationError
from django.db import models
from pubsubpull.fields import JSONB
class Request(models.Model):
"""A web request.
"""
user = models.ForeignKey(User, null=True, blank=True, related_name='requests')
method = models.CharField(max_length=20)
path = models.TextField()
OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE")
class UpdateLog(models.Model):
"""Store a change to a single row in a table.
"""
table = models.CharField(max_length=200)
type = models.CharField(max_length=1, choices=OPERATION_TYPE.items())
when = models.DateTimeField(auto_now_add=True)
request = models.ForeignKey(Request, null=True, blank=True,
related_name='changes')
old = JSONB(null=True, blank=True)
new = JSONB(null=True, blank=True)
def save(self, **kw):
raise ValidationError("Instances of this class cannot be using Django")
def __unicode__(self):
return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
<commit_msg>Add more useful display of the request data.<commit_after>"""
Models.
"""
from django.contrib.auth.models import User
from django.core.exceptions import ValidationError
from django.db import models
from pubsubpull.fields import JSONB
class Request(models.Model):
"""A web request.
"""
user = models.ForeignKey(User, null=True, blank=True, related_name='requests')
method = models.CharField(max_length=20)
path = models.TextField()
def __unicode__(self):
return "%s %s" % (self.method, self.path)
OPERATION_TYPE = dict(I="INSERT", U="UPDATE", D="DELETE", T="TRUNCATE")
class UpdateLog(models.Model):
"""Store a change to a single row in a table.
"""
table = models.CharField(max_length=200)
type = models.CharField(max_length=1, choices=OPERATION_TYPE.items())
when = models.DateTimeField(auto_now_add=True)
request = models.ForeignKey(Request, null=True, blank=True,
related_name='changes')
old = JSONB(null=True, blank=True)
new = JSONB(null=True, blank=True)
def save(self, **kw):
raise ValidationError("Instances of this class cannot be using Django")
def __unicode__(self):
return u"%s %s @ %s" % (OPERATION_TYPE[self.type], self.table, self.when)
|
5b8241ad808bd11971d0d684bafd6f9019e58397
|
tests/contrib/flask/tests.py
|
tests/contrib/flask/tests.py
|
import logging
from flask import Flask
from raven.base import Client
from raven.contrib.flask import Sentry
from unittest2 import TestCase
class TempStoreClient(Client):
def __init__(self, *args, **kwargs):
self.events = []
super(TempStoreClient, self).__init__(*args, **kwargs)
def send(self, **kwargs):
self.events.append(kwargs)
def create_app():
app = Flask(__name__)
@app.route('/an-error/', methods=['GET', 'POST'])
def an_error():
raise ValueError('hello world')
return app
class FlaskTest(TestCase):
def setUp(self):
self.app = create_app()
self.client = self.app.test_client()
def test_error_handler(self):
client = TempStoreClient()
sentry = Sentry(self.app, client=client)
response = self.client.get('/an-error/?foo=bar')
self.assertEquals(response.status_code, 500)
self.assertEquals(len(client.events), 1)
event = client.events.pop(0)
self.assertEquals(event['class_name'], 'ValueError')
self.assertEquals(event['level'], logging.ERROR)
self.assertEquals(event['message'], 'hello world')
self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
|
import logging
from flask import Flask
from raven.base import Client
from raven.contrib.flask import Sentry
from unittest2 import TestCase
class TempStoreClient(Client):
def __init__(self, *args, **kwargs):
self.events = []
super(TempStoreClient, self).__init__(*args, **kwargs)
def send(self, **kwargs):
self.events.append(kwargs)
def create_app():
app = Flask(__name__)
@app.route('/an-error/', methods=['GET', 'POST'])
def an_error():
raise ValueError('hello world')
return app
class FlaskTest(TestCase):
def setUp(self):
self.app = create_app()
self.client = self.app.test_client()
def test_error_handler(self):
client = TempStoreClient()
sentry = Sentry(self.app, client=client)
response = self.client.get('/an-error/?foo=bar')
self.assertEquals(response.status_code, 500)
self.assertEquals(len(client.events), 1)
event = client.events.pop(0)
self.assertEquals(event['class_name'], 'ValueError')
self.assertEquals(event['level'], logging.ERROR)
self.assertEquals(event['message'], 'hello world')
self.assertEquals(event['url'], 'http://localhost/an-error/?foo=bar')
self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
|
Add url test for Flask
|
Add url test for Flask
|
Python
|
bsd-3-clause
|
nikolas/raven-python,Photonomie/raven-python,jmagnusson/raven-python,inspirehep/raven-python,danriti/raven-python,lopter/raven-python-old,nikolas/raven-python,johansteffner/raven-python,johansteffner/raven-python,daikeren/opbeat_python,daikeren/opbeat_python,someonehan/raven-python,inspirehep/raven-python,jmagnusson/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,jbarbuto/raven-python,johansteffner/raven-python,danriti/raven-python,icereval/raven-python,dirtycoder/opbeat_python,jmagnusson/raven-python,someonehan/raven-python,patrys/opbeat_python,nikolas/raven-python,mitsuhiko/raven,openlabs/raven,beniwohli/apm-agent-python,tarkatronic/opbeat_python,Photonomie/raven-python,inspirehep/raven-python,recht/raven-python,lepture/raven-python,getsentry/raven-python,nikolas/raven-python,beniwohli/apm-agent-python,jbarbuto/raven-python,arthurlogilab/raven-python,getsentry/raven-python,percipient/raven-python,ewdurbin/raven-python,icereval/raven-python,jmp0xf/raven-python,tarkatronic/opbeat_python,someonehan/raven-python,ronaldevers/raven-python,ewdurbin/raven-python,dirtycoder/opbeat_python,jmp0xf/raven-python,akheron/raven-python,akheron/raven-python,akalipetis/raven-python,tarkatronic/opbeat_python,Goldmund-Wyldebeast-Wunderliebe/raven-python,Photonomie/raven-python,beniwohli/apm-agent-python,beniwohli/apm-agent-python,lepture/raven-python,recht/raven-python,jbarbuto/raven-python,inspirehep/raven-python,daikeren/opbeat_python,smarkets/raven-python,smarkets/raven-python,akheron/raven-python,smarkets/raven-python,mitsuhiko/raven,jmp0xf/raven-python,ticosax/opbeat_python,patrys/opbeat_python,dirtycoder/opbeat_python,percipient/raven-python,ronaldevers/raven-python,patrys/opbeat_python,recht/raven-python,dbravender/raven-python,hzy/raven-python,ticosax/opbeat_python,patrys/opbeat_python,dbravender/raven-python,arthurlogilab/raven-python,smarkets/raven-python,akalipetis/raven-python,icereval/raven-python,arthurlogilab/raven-python,ticosax/opbeat_python,lepture/raven-python,percipient/raven-python,getsentry/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,collective/mr.poe,hzy/raven-python,jbarbuto/raven-python,akalipetis/raven-python,ronaldevers/raven-python,hzy/raven-python,danriti/raven-python,ewdurbin/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,alex/raven,arthurlogilab/raven-python,dbravender/raven-python,icereval/raven-python
|
import logging
from flask import Flask
from raven.base import Client
from raven.contrib.flask import Sentry
from unittest2 import TestCase
class TempStoreClient(Client):
def __init__(self, *args, **kwargs):
self.events = []
super(TempStoreClient, self).__init__(*args, **kwargs)
def send(self, **kwargs):
self.events.append(kwargs)
def create_app():
app = Flask(__name__)
@app.route('/an-error/', methods=['GET', 'POST'])
def an_error():
raise ValueError('hello world')
return app
class FlaskTest(TestCase):
def setUp(self):
self.app = create_app()
self.client = self.app.test_client()
def test_error_handler(self):
client = TempStoreClient()
sentry = Sentry(self.app, client=client)
response = self.client.get('/an-error/?foo=bar')
self.assertEquals(response.status_code, 500)
self.assertEquals(len(client.events), 1)
event = client.events.pop(0)
self.assertEquals(event['class_name'], 'ValueError')
self.assertEquals(event['level'], logging.ERROR)
self.assertEquals(event['message'], 'hello world')
self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
Add url test for Flask
|
import logging
from flask import Flask
from raven.base import Client
from raven.contrib.flask import Sentry
from unittest2 import TestCase
class TempStoreClient(Client):
def __init__(self, *args, **kwargs):
self.events = []
super(TempStoreClient, self).__init__(*args, **kwargs)
def send(self, **kwargs):
self.events.append(kwargs)
def create_app():
app = Flask(__name__)
@app.route('/an-error/', methods=['GET', 'POST'])
def an_error():
raise ValueError('hello world')
return app
class FlaskTest(TestCase):
def setUp(self):
self.app = create_app()
self.client = self.app.test_client()
def test_error_handler(self):
client = TempStoreClient()
sentry = Sentry(self.app, client=client)
response = self.client.get('/an-error/?foo=bar')
self.assertEquals(response.status_code, 500)
self.assertEquals(len(client.events), 1)
event = client.events.pop(0)
self.assertEquals(event['class_name'], 'ValueError')
self.assertEquals(event['level'], logging.ERROR)
self.assertEquals(event['message'], 'hello world')
self.assertEquals(event['url'], 'http://localhost/an-error/?foo=bar')
self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
|
<commit_before>import logging
from flask import Flask
from raven.base import Client
from raven.contrib.flask import Sentry
from unittest2 import TestCase
class TempStoreClient(Client):
def __init__(self, *args, **kwargs):
self.events = []
super(TempStoreClient, self).__init__(*args, **kwargs)
def send(self, **kwargs):
self.events.append(kwargs)
def create_app():
app = Flask(__name__)
@app.route('/an-error/', methods=['GET', 'POST'])
def an_error():
raise ValueError('hello world')
return app
class FlaskTest(TestCase):
def setUp(self):
self.app = create_app()
self.client = self.app.test_client()
def test_error_handler(self):
client = TempStoreClient()
sentry = Sentry(self.app, client=client)
response = self.client.get('/an-error/?foo=bar')
self.assertEquals(response.status_code, 500)
self.assertEquals(len(client.events), 1)
event = client.events.pop(0)
self.assertEquals(event['class_name'], 'ValueError')
self.assertEquals(event['level'], logging.ERROR)
self.assertEquals(event['message'], 'hello world')
self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
<commit_msg>Add url test for Flask<commit_after>
|
import logging
from flask import Flask
from raven.base import Client
from raven.contrib.flask import Sentry
from unittest2 import TestCase
class TempStoreClient(Client):
def __init__(self, *args, **kwargs):
self.events = []
super(TempStoreClient, self).__init__(*args, **kwargs)
def send(self, **kwargs):
self.events.append(kwargs)
def create_app():
app = Flask(__name__)
@app.route('/an-error/', methods=['GET', 'POST'])
def an_error():
raise ValueError('hello world')
return app
class FlaskTest(TestCase):
def setUp(self):
self.app = create_app()
self.client = self.app.test_client()
def test_error_handler(self):
client = TempStoreClient()
sentry = Sentry(self.app, client=client)
response = self.client.get('/an-error/?foo=bar')
self.assertEquals(response.status_code, 500)
self.assertEquals(len(client.events), 1)
event = client.events.pop(0)
self.assertEquals(event['class_name'], 'ValueError')
self.assertEquals(event['level'], logging.ERROR)
self.assertEquals(event['message'], 'hello world')
self.assertEquals(event['url'], 'http://localhost/an-error/?foo=bar')
self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
|
import logging
from flask import Flask
from raven.base import Client
from raven.contrib.flask import Sentry
from unittest2 import TestCase
class TempStoreClient(Client):
def __init__(self, *args, **kwargs):
self.events = []
super(TempStoreClient, self).__init__(*args, **kwargs)
def send(self, **kwargs):
self.events.append(kwargs)
def create_app():
app = Flask(__name__)
@app.route('/an-error/', methods=['GET', 'POST'])
def an_error():
raise ValueError('hello world')
return app
class FlaskTest(TestCase):
def setUp(self):
self.app = create_app()
self.client = self.app.test_client()
def test_error_handler(self):
client = TempStoreClient()
sentry = Sentry(self.app, client=client)
response = self.client.get('/an-error/?foo=bar')
self.assertEquals(response.status_code, 500)
self.assertEquals(len(client.events), 1)
event = client.events.pop(0)
self.assertEquals(event['class_name'], 'ValueError')
self.assertEquals(event['level'], logging.ERROR)
self.assertEquals(event['message'], 'hello world')
self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
Add url test for Flaskimport logging
from flask import Flask
from raven.base import Client
from raven.contrib.flask import Sentry
from unittest2 import TestCase
class TempStoreClient(Client):
def __init__(self, *args, **kwargs):
self.events = []
super(TempStoreClient, self).__init__(*args, **kwargs)
def send(self, **kwargs):
self.events.append(kwargs)
def create_app():
app = Flask(__name__)
@app.route('/an-error/', methods=['GET', 'POST'])
def an_error():
raise ValueError('hello world')
return app
class FlaskTest(TestCase):
def setUp(self):
self.app = create_app()
self.client = self.app.test_client()
def test_error_handler(self):
client = TempStoreClient()
sentry = Sentry(self.app, client=client)
response = self.client.get('/an-error/?foo=bar')
self.assertEquals(response.status_code, 500)
self.assertEquals(len(client.events), 1)
event = client.events.pop(0)
self.assertEquals(event['class_name'], 'ValueError')
self.assertEquals(event['level'], logging.ERROR)
self.assertEquals(event['message'], 'hello world')
self.assertEquals(event['url'], 'http://localhost/an-error/?foo=bar')
self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
|
<commit_before>import logging
from flask import Flask
from raven.base import Client
from raven.contrib.flask import Sentry
from unittest2 import TestCase
class TempStoreClient(Client):
def __init__(self, *args, **kwargs):
self.events = []
super(TempStoreClient, self).__init__(*args, **kwargs)
def send(self, **kwargs):
self.events.append(kwargs)
def create_app():
app = Flask(__name__)
@app.route('/an-error/', methods=['GET', 'POST'])
def an_error():
raise ValueError('hello world')
return app
class FlaskTest(TestCase):
def setUp(self):
self.app = create_app()
self.client = self.app.test_client()
def test_error_handler(self):
client = TempStoreClient()
sentry = Sentry(self.app, client=client)
response = self.client.get('/an-error/?foo=bar')
self.assertEquals(response.status_code, 500)
self.assertEquals(len(client.events), 1)
event = client.events.pop(0)
self.assertEquals(event['class_name'], 'ValueError')
self.assertEquals(event['level'], logging.ERROR)
self.assertEquals(event['message'], 'hello world')
self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
<commit_msg>Add url test for Flask<commit_after>import logging
from flask import Flask
from raven.base import Client
from raven.contrib.flask import Sentry
from unittest2 import TestCase
class TempStoreClient(Client):
def __init__(self, *args, **kwargs):
self.events = []
super(TempStoreClient, self).__init__(*args, **kwargs)
def send(self, **kwargs):
self.events.append(kwargs)
def create_app():
app = Flask(__name__)
@app.route('/an-error/', methods=['GET', 'POST'])
def an_error():
raise ValueError('hello world')
return app
class FlaskTest(TestCase):
def setUp(self):
self.app = create_app()
self.client = self.app.test_client()
def test_error_handler(self):
client = TempStoreClient()
sentry = Sentry(self.app, client=client)
response = self.client.get('/an-error/?foo=bar')
self.assertEquals(response.status_code, 500)
self.assertEquals(len(client.events), 1)
event = client.events.pop(0)
self.assertEquals(event['class_name'], 'ValueError')
self.assertEquals(event['level'], logging.ERROR)
self.assertEquals(event['message'], 'hello world')
self.assertEquals(event['url'], 'http://localhost/an-error/?foo=bar')
self.assertEquals(event['view'], 'tests.contrib.flask.tests.an_error')
|
4ccc5ea6cf25adb029f5e08cc0675e2b8415abdf
|
LayerView.py
|
LayerView.py
|
from UM.View.View import View
from UM.View.Renderer import Renderer
from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator
from UM.Resources import Resources
class LayerView(View):
def __init__(self):
super().__init__()
self._material = None
def beginRendering(self):
scene = self.getController().getScene()
renderer = self.getRenderer()
if not self._material:
self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'color.frag'))
self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0])
for node in DepthFirstIterator(scene.getRoot()):
if not node.render(renderer):
if node.getMeshData() and node.isVisible():
try:
layerData = node.getMeshData().layerData
except AttributeError:
continue
renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLineLoop)
def endRendering(self):
pass
|
from UM.View.View import View
from UM.View.Renderer import Renderer
from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator
from UM.Resources import Resources
class LayerView(View):
def __init__(self):
super().__init__()
self._material = None
def beginRendering(self):
scene = self.getController().getScene()
renderer = self.getRenderer()
if not self._material:
self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'vertexcolor.frag'))
self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0])
for node in DepthFirstIterator(scene.getRoot()):
if not node.render(renderer):
if node.getMeshData() and node.isVisible():
try:
layerData = node.getMeshData().layerData
except AttributeError:
continue
renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLines)
def endRendering(self):
pass
|
Support colours for rendering the layer view
|
Support colours for rendering the layer view
|
Python
|
agpl-3.0
|
markwal/Cura,DeskboxBrazil/Cura,ad1217/Cura,Curahelper/Cura,senttech/Cura,derekhe/Cura,fxtentacle/Cura,ynotstartups/Wanhao,ad1217/Cura,markwal/Cura,fxtentacle/Cura,Curahelper/Cura,quillford/Cura,hmflash/Cura,ynotstartups/Wanhao,fieldOfView/Cura,hmflash/Cura,totalretribution/Cura,quillford/Cura,lo0ol/Ultimaker-Cura,lo0ol/Ultimaker-Cura,senttech/Cura,fieldOfView/Cura,bq/Ultimaker-Cura,derekhe/Cura,DeskboxBrazil/Cura,bq/Ultimaker-Cura,totalretribution/Cura
|
from UM.View.View import View
from UM.View.Renderer import Renderer
from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator
from UM.Resources import Resources
class LayerView(View):
def __init__(self):
super().__init__()
self._material = None
def beginRendering(self):
scene = self.getController().getScene()
renderer = self.getRenderer()
if not self._material:
self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'color.frag'))
self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0])
for node in DepthFirstIterator(scene.getRoot()):
if not node.render(renderer):
if node.getMeshData() and node.isVisible():
try:
layerData = node.getMeshData().layerData
except AttributeError:
continue
renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLineLoop)
def endRendering(self):
pass
Support colours for rendering the layer view
|
from UM.View.View import View
from UM.View.Renderer import Renderer
from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator
from UM.Resources import Resources
class LayerView(View):
def __init__(self):
super().__init__()
self._material = None
def beginRendering(self):
scene = self.getController().getScene()
renderer = self.getRenderer()
if not self._material:
self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'vertexcolor.frag'))
self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0])
for node in DepthFirstIterator(scene.getRoot()):
if not node.render(renderer):
if node.getMeshData() and node.isVisible():
try:
layerData = node.getMeshData().layerData
except AttributeError:
continue
renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLines)
def endRendering(self):
pass
|
<commit_before>from UM.View.View import View
from UM.View.Renderer import Renderer
from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator
from UM.Resources import Resources
class LayerView(View):
def __init__(self):
super().__init__()
self._material = None
def beginRendering(self):
scene = self.getController().getScene()
renderer = self.getRenderer()
if not self._material:
self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'color.frag'))
self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0])
for node in DepthFirstIterator(scene.getRoot()):
if not node.render(renderer):
if node.getMeshData() and node.isVisible():
try:
layerData = node.getMeshData().layerData
except AttributeError:
continue
renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLineLoop)
def endRendering(self):
pass
<commit_msg>Support colours for rendering the layer view<commit_after>
|
from UM.View.View import View
from UM.View.Renderer import Renderer
from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator
from UM.Resources import Resources
class LayerView(View):
def __init__(self):
super().__init__()
self._material = None
def beginRendering(self):
scene = self.getController().getScene()
renderer = self.getRenderer()
if not self._material:
self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'vertexcolor.frag'))
self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0])
for node in DepthFirstIterator(scene.getRoot()):
if not node.render(renderer):
if node.getMeshData() and node.isVisible():
try:
layerData = node.getMeshData().layerData
except AttributeError:
continue
renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLines)
def endRendering(self):
pass
|
from UM.View.View import View
from UM.View.Renderer import Renderer
from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator
from UM.Resources import Resources
class LayerView(View):
def __init__(self):
super().__init__()
self._material = None
def beginRendering(self):
scene = self.getController().getScene()
renderer = self.getRenderer()
if not self._material:
self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'color.frag'))
self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0])
for node in DepthFirstIterator(scene.getRoot()):
if not node.render(renderer):
if node.getMeshData() and node.isVisible():
try:
layerData = node.getMeshData().layerData
except AttributeError:
continue
renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLineLoop)
def endRendering(self):
pass
Support colours for rendering the layer viewfrom UM.View.View import View
from UM.View.Renderer import Renderer
from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator
from UM.Resources import Resources
class LayerView(View):
def __init__(self):
super().__init__()
self._material = None
def beginRendering(self):
scene = self.getController().getScene()
renderer = self.getRenderer()
if not self._material:
self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'vertexcolor.frag'))
self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0])
for node in DepthFirstIterator(scene.getRoot()):
if not node.render(renderer):
if node.getMeshData() and node.isVisible():
try:
layerData = node.getMeshData().layerData
except AttributeError:
continue
renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLines)
def endRendering(self):
pass
|
<commit_before>from UM.View.View import View
from UM.View.Renderer import Renderer
from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator
from UM.Resources import Resources
class LayerView(View):
def __init__(self):
super().__init__()
self._material = None
def beginRendering(self):
scene = self.getController().getScene()
renderer = self.getRenderer()
if not self._material:
self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'color.frag'))
self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0])
for node in DepthFirstIterator(scene.getRoot()):
if not node.render(renderer):
if node.getMeshData() and node.isVisible():
try:
layerData = node.getMeshData().layerData
except AttributeError:
continue
renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLineLoop)
def endRendering(self):
pass
<commit_msg>Support colours for rendering the layer view<commit_after>from UM.View.View import View
from UM.View.Renderer import Renderer
from UM.Scene.Iterator.DepthFirstIterator import DepthFirstIterator
from UM.Resources import Resources
class LayerView(View):
def __init__(self):
super().__init__()
self._material = None
def beginRendering(self):
scene = self.getController().getScene()
renderer = self.getRenderer()
if not self._material:
self._material = renderer.createMaterial(Resources.getPath(Resources.ShadersLocation, 'basic.vert'), Resources.getPath(Resources.ShadersLocation, 'vertexcolor.frag'))
self._material.setUniformValue("u_color", [1.0, 0.0, 0.0, 1.0])
for node in DepthFirstIterator(scene.getRoot()):
if not node.render(renderer):
if node.getMeshData() and node.isVisible():
try:
layerData = node.getMeshData().layerData
except AttributeError:
continue
renderer.queueNode(node, mesh = layerData, material = self._material, mode = Renderer.RenderLines)
def endRendering(self):
pass
|
a3a5d2d6b76a4e903fea232b746b2df8b208ec9e
|
km3pipe/tests/test_plot.py
|
km3pipe/tests/test_plot.py
|
# Filename: test_plot.py
# pylint: disable=locally-disabled,C0111,R0904,C0103
import numpy as np
from km3pipe.testing import TestCase
from km3pipe.plot import bincenters
__author__ = "Moritz Lotze"
__copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration."
__credits__ = []
__license__ = "MIT"
__maintainer__ = "Moritz Lotze"
__email__ = "mlotze@km3net.de"
__status__ = "Development"
class TestBins(TestCase):
def test_binlims(self):
bins = np.linspace(0, 20, 21)
assert bincenters(bins).shape[0] == bins.shape[0] - 1
|
# Filename: test_plot.py
# pylint: disable=locally-disabled,C0111,R0904,C0103
import numpy as np
from km3pipe.testing import TestCase, patch
from km3pipe.plot import bincenters, meshgrid, automeshgrid, diag
__author__ = "Moritz Lotze"
__copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration."
__credits__ = []
__license__ = "MIT"
__maintainer__ = "Moritz Lotze"
__email__ = "mlotze@km3net.de"
__status__ = "Development"
class TestBins(TestCase):
def test_binlims(self):
bins = np.linspace(0, 20, 21)
assert bincenters(bins).shape[0] == bins.shape[0] - 1
class TestMeshStuff(TestCase):
def test_meshgrid(self):
xx, yy = meshgrid(-1, 1, 0.8)
assert np.allclose([[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6]], xx)
assert np.allclose([[-1.0, -1.0, -1.0],
[-0.2, -0.2, -0.2],
[0.6, 0.6, 0.6]], yy)
def test_meshgrid_with_y_specs(self):
xx, yy = meshgrid(-1, 1, 0.8, -10, 10, 8)
assert np.allclose([[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6]], xx)
assert np.allclose([[-10, -10, -10],
[-2, -2, -2],
[6, 6, 6]], yy)
class TestDiag(TestCase):
def test_call(self):
diag()
|
Add tests for plot functions
|
Add tests for plot functions
|
Python
|
mit
|
tamasgal/km3pipe,tamasgal/km3pipe
|
# Filename: test_plot.py
# pylint: disable=locally-disabled,C0111,R0904,C0103
import numpy as np
from km3pipe.testing import TestCase
from km3pipe.plot import bincenters
__author__ = "Moritz Lotze"
__copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration."
__credits__ = []
__license__ = "MIT"
__maintainer__ = "Moritz Lotze"
__email__ = "mlotze@km3net.de"
__status__ = "Development"
class TestBins(TestCase):
def test_binlims(self):
bins = np.linspace(0, 20, 21)
assert bincenters(bins).shape[0] == bins.shape[0] - 1
Add tests for plot functions
|
# Filename: test_plot.py
# pylint: disable=locally-disabled,C0111,R0904,C0103
import numpy as np
from km3pipe.testing import TestCase, patch
from km3pipe.plot import bincenters, meshgrid, automeshgrid, diag
__author__ = "Moritz Lotze"
__copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration."
__credits__ = []
__license__ = "MIT"
__maintainer__ = "Moritz Lotze"
__email__ = "mlotze@km3net.de"
__status__ = "Development"
class TestBins(TestCase):
def test_binlims(self):
bins = np.linspace(0, 20, 21)
assert bincenters(bins).shape[0] == bins.shape[0] - 1
class TestMeshStuff(TestCase):
def test_meshgrid(self):
xx, yy = meshgrid(-1, 1, 0.8)
assert np.allclose([[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6]], xx)
assert np.allclose([[-1.0, -1.0, -1.0],
[-0.2, -0.2, -0.2],
[0.6, 0.6, 0.6]], yy)
def test_meshgrid_with_y_specs(self):
xx, yy = meshgrid(-1, 1, 0.8, -10, 10, 8)
assert np.allclose([[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6]], xx)
assert np.allclose([[-10, -10, -10],
[-2, -2, -2],
[6, 6, 6]], yy)
class TestDiag(TestCase):
def test_call(self):
diag()
|
<commit_before># Filename: test_plot.py
# pylint: disable=locally-disabled,C0111,R0904,C0103
import numpy as np
from km3pipe.testing import TestCase
from km3pipe.plot import bincenters
__author__ = "Moritz Lotze"
__copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration."
__credits__ = []
__license__ = "MIT"
__maintainer__ = "Moritz Lotze"
__email__ = "mlotze@km3net.de"
__status__ = "Development"
class TestBins(TestCase):
def test_binlims(self):
bins = np.linspace(0, 20, 21)
assert bincenters(bins).shape[0] == bins.shape[0] - 1
<commit_msg>Add tests for plot functions<commit_after>
|
# Filename: test_plot.py
# pylint: disable=locally-disabled,C0111,R0904,C0103
import numpy as np
from km3pipe.testing import TestCase, patch
from km3pipe.plot import bincenters, meshgrid, automeshgrid, diag
__author__ = "Moritz Lotze"
__copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration."
__credits__ = []
__license__ = "MIT"
__maintainer__ = "Moritz Lotze"
__email__ = "mlotze@km3net.de"
__status__ = "Development"
class TestBins(TestCase):
def test_binlims(self):
bins = np.linspace(0, 20, 21)
assert bincenters(bins).shape[0] == bins.shape[0] - 1
class TestMeshStuff(TestCase):
def test_meshgrid(self):
xx, yy = meshgrid(-1, 1, 0.8)
assert np.allclose([[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6]], xx)
assert np.allclose([[-1.0, -1.0, -1.0],
[-0.2, -0.2, -0.2],
[0.6, 0.6, 0.6]], yy)
def test_meshgrid_with_y_specs(self):
xx, yy = meshgrid(-1, 1, 0.8, -10, 10, 8)
assert np.allclose([[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6]], xx)
assert np.allclose([[-10, -10, -10],
[-2, -2, -2],
[6, 6, 6]], yy)
class TestDiag(TestCase):
def test_call(self):
diag()
|
# Filename: test_plot.py
# pylint: disable=locally-disabled,C0111,R0904,C0103
import numpy as np
from km3pipe.testing import TestCase
from km3pipe.plot import bincenters
__author__ = "Moritz Lotze"
__copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration."
__credits__ = []
__license__ = "MIT"
__maintainer__ = "Moritz Lotze"
__email__ = "mlotze@km3net.de"
__status__ = "Development"
class TestBins(TestCase):
def test_binlims(self):
bins = np.linspace(0, 20, 21)
assert bincenters(bins).shape[0] == bins.shape[0] - 1
Add tests for plot functions# Filename: test_plot.py
# pylint: disable=locally-disabled,C0111,R0904,C0103
import numpy as np
from km3pipe.testing import TestCase, patch
from km3pipe.plot import bincenters, meshgrid, automeshgrid, diag
__author__ = "Moritz Lotze"
__copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration."
__credits__ = []
__license__ = "MIT"
__maintainer__ = "Moritz Lotze"
__email__ = "mlotze@km3net.de"
__status__ = "Development"
class TestBins(TestCase):
def test_binlims(self):
bins = np.linspace(0, 20, 21)
assert bincenters(bins).shape[0] == bins.shape[0] - 1
class TestMeshStuff(TestCase):
def test_meshgrid(self):
xx, yy = meshgrid(-1, 1, 0.8)
assert np.allclose([[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6]], xx)
assert np.allclose([[-1.0, -1.0, -1.0],
[-0.2, -0.2, -0.2],
[0.6, 0.6, 0.6]], yy)
def test_meshgrid_with_y_specs(self):
xx, yy = meshgrid(-1, 1, 0.8, -10, 10, 8)
assert np.allclose([[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6]], xx)
assert np.allclose([[-10, -10, -10],
[-2, -2, -2],
[6, 6, 6]], yy)
class TestDiag(TestCase):
def test_call(self):
diag()
|
<commit_before># Filename: test_plot.py
# pylint: disable=locally-disabled,C0111,R0904,C0103
import numpy as np
from km3pipe.testing import TestCase
from km3pipe.plot import bincenters
__author__ = "Moritz Lotze"
__copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration."
__credits__ = []
__license__ = "MIT"
__maintainer__ = "Moritz Lotze"
__email__ = "mlotze@km3net.de"
__status__ = "Development"
class TestBins(TestCase):
def test_binlims(self):
bins = np.linspace(0, 20, 21)
assert bincenters(bins).shape[0] == bins.shape[0] - 1
<commit_msg>Add tests for plot functions<commit_after># Filename: test_plot.py
# pylint: disable=locally-disabled,C0111,R0904,C0103
import numpy as np
from km3pipe.testing import TestCase, patch
from km3pipe.plot import bincenters, meshgrid, automeshgrid, diag
__author__ = "Moritz Lotze"
__copyright__ = "Copyright 2016, Tamas Gal and the KM3NeT collaboration."
__credits__ = []
__license__ = "MIT"
__maintainer__ = "Moritz Lotze"
__email__ = "mlotze@km3net.de"
__status__ = "Development"
class TestBins(TestCase):
def test_binlims(self):
bins = np.linspace(0, 20, 21)
assert bincenters(bins).shape[0] == bins.shape[0] - 1
class TestMeshStuff(TestCase):
def test_meshgrid(self):
xx, yy = meshgrid(-1, 1, 0.8)
assert np.allclose([[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6]], xx)
assert np.allclose([[-1.0, -1.0, -1.0],
[-0.2, -0.2, -0.2],
[0.6, 0.6, 0.6]], yy)
def test_meshgrid_with_y_specs(self):
xx, yy = meshgrid(-1, 1, 0.8, -10, 10, 8)
assert np.allclose([[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6],
[-1.0, -0.2, 0.6]], xx)
assert np.allclose([[-10, -10, -10],
[-2, -2, -2],
[6, 6, 6]], yy)
class TestDiag(TestCase):
def test_call(self):
diag()
|
ef4c9f6a2e6fc1db01d93d937d24e444b0bb0ede
|
tests/memory_profiling.py
|
tests/memory_profiling.py
|
"""
Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector.
"""
import inspect
import sys
import time
import memory_profiler
import vector_test
try:
from pvectorc import pvector
except ImportError:
print("No C implementation of PVector available, terminating")
sys.exit()
PROFILING_DURATION = 2.0
def run_function(fn):
stop = time.time() + PROFILING_DURATION
while time.time() < stop:
fn(pvector)
def detect_memory_leak(samples):
# Skip the first half to get rid of the build up period and the last since it seems
# a little less precise
samples = samples[int(len(samples)/2):len(samples)-1]
return not samples.count(samples[0]) > len(samples) - 2
def profile_tests():
test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction)
if fn[0].startswith('test_')]
for name, fn in test_functions:
# There are a couple of tests that are not run for the C implementation, skip those
fn_args = inspect.getargspec(fn)[0]
if 'pvector' in fn_args:
print('Executing %s' % name)
result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1)
assert not detect_memory_leak(result), (name, result)
if __name__ == "__main__":
profile_tests()
|
"""
Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector.
"""
import inspect
import sys
import time
import memory_profiler
import vector_test
try:
from pvectorc import pvector
except ImportError:
print("No C implementation of PVector available, terminating")
sys.exit()
PROFILING_DURATION = 2.0
def run_function(fn):
stop = time.time() + PROFILING_DURATION
while time.time() < stop:
fn(pvector)
def detect_memory_leak(samples):
# Skip the first samples to get rid of the build up period and the last sample since it seems
# a little less precise
rising = 0
for i in range(5, len(samples)-1):
if samples[i] < samples[i+1]:
rising += 1
return (rising / float(len(samples) - 6)) > 0.2
def profile_tests():
test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction)
if fn[0].startswith('test_')]
for name, fn in test_functions:
# There are a couple of tests that are not run for the C implementation, skip those
fn_args = inspect.getargspec(fn)[0]
if 'pvector' in fn_args:
print('Executing %s' % name)
result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1)
assert not detect_memory_leak(result), (name, result)
if __name__ == "__main__":
profile_tests()
|
Improve memory error detection for less false positives
|
Improve memory error detection for less false positives
|
Python
|
mit
|
tobgu/pyrsistent,jkbjh/pyrsistent,Futrell/pyrsistent,tobgu/pyrsistent,jml/pyrsistent,jml/pyrsistent,tobgu/pyrsistent,jkbjh/pyrsistent,Futrell/pyrsistent,jkbjh/pyrsistent,Futrell/pyrsistent,jml/pyrsistent
|
"""
Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector.
"""
import inspect
import sys
import time
import memory_profiler
import vector_test
try:
from pvectorc import pvector
except ImportError:
print("No C implementation of PVector available, terminating")
sys.exit()
PROFILING_DURATION = 2.0
def run_function(fn):
stop = time.time() + PROFILING_DURATION
while time.time() < stop:
fn(pvector)
def detect_memory_leak(samples):
# Skip the first half to get rid of the build up period and the last since it seems
# a little less precise
samples = samples[int(len(samples)/2):len(samples)-1]
return not samples.count(samples[0]) > len(samples) - 2
def profile_tests():
test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction)
if fn[0].startswith('test_')]
for name, fn in test_functions:
# There are a couple of tests that are not run for the C implementation, skip those
fn_args = inspect.getargspec(fn)[0]
if 'pvector' in fn_args:
print('Executing %s' % name)
result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1)
assert not detect_memory_leak(result), (name, result)
if __name__ == "__main__":
profile_tests()Improve memory error detection for less false positives
|
"""
Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector.
"""
import inspect
import sys
import time
import memory_profiler
import vector_test
try:
from pvectorc import pvector
except ImportError:
print("No C implementation of PVector available, terminating")
sys.exit()
PROFILING_DURATION = 2.0
def run_function(fn):
stop = time.time() + PROFILING_DURATION
while time.time() < stop:
fn(pvector)
def detect_memory_leak(samples):
# Skip the first samples to get rid of the build up period and the last sample since it seems
# a little less precise
rising = 0
for i in range(5, len(samples)-1):
if samples[i] < samples[i+1]:
rising += 1
return (rising / float(len(samples) - 6)) > 0.2
def profile_tests():
test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction)
if fn[0].startswith('test_')]
for name, fn in test_functions:
# There are a couple of tests that are not run for the C implementation, skip those
fn_args = inspect.getargspec(fn)[0]
if 'pvector' in fn_args:
print('Executing %s' % name)
result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1)
assert not detect_memory_leak(result), (name, result)
if __name__ == "__main__":
profile_tests()
|
<commit_before>"""
Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector.
"""
import inspect
import sys
import time
import memory_profiler
import vector_test
try:
from pvectorc import pvector
except ImportError:
print("No C implementation of PVector available, terminating")
sys.exit()
PROFILING_DURATION = 2.0
def run_function(fn):
stop = time.time() + PROFILING_DURATION
while time.time() < stop:
fn(pvector)
def detect_memory_leak(samples):
# Skip the first half to get rid of the build up period and the last since it seems
# a little less precise
samples = samples[int(len(samples)/2):len(samples)-1]
return not samples.count(samples[0]) > len(samples) - 2
def profile_tests():
test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction)
if fn[0].startswith('test_')]
for name, fn in test_functions:
# There are a couple of tests that are not run for the C implementation, skip those
fn_args = inspect.getargspec(fn)[0]
if 'pvector' in fn_args:
print('Executing %s' % name)
result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1)
assert not detect_memory_leak(result), (name, result)
if __name__ == "__main__":
profile_tests()<commit_msg>Improve memory error detection for less false positives<commit_after>
|
"""
Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector.
"""
import inspect
import sys
import time
import memory_profiler
import vector_test
try:
from pvectorc import pvector
except ImportError:
print("No C implementation of PVector available, terminating")
sys.exit()
PROFILING_DURATION = 2.0
def run_function(fn):
stop = time.time() + PROFILING_DURATION
while time.time() < stop:
fn(pvector)
def detect_memory_leak(samples):
# Skip the first samples to get rid of the build up period and the last sample since it seems
# a little less precise
rising = 0
for i in range(5, len(samples)-1):
if samples[i] < samples[i+1]:
rising += 1
return (rising / float(len(samples) - 6)) > 0.2
def profile_tests():
test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction)
if fn[0].startswith('test_')]
for name, fn in test_functions:
# There are a couple of tests that are not run for the C implementation, skip those
fn_args = inspect.getargspec(fn)[0]
if 'pvector' in fn_args:
print('Executing %s' % name)
result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1)
assert not detect_memory_leak(result), (name, result)
if __name__ == "__main__":
profile_tests()
|
"""
Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector.
"""
import inspect
import sys
import time
import memory_profiler
import vector_test
try:
from pvectorc import pvector
except ImportError:
print("No C implementation of PVector available, terminating")
sys.exit()
PROFILING_DURATION = 2.0
def run_function(fn):
stop = time.time() + PROFILING_DURATION
while time.time() < stop:
fn(pvector)
def detect_memory_leak(samples):
# Skip the first half to get rid of the build up period and the last since it seems
# a little less precise
samples = samples[int(len(samples)/2):len(samples)-1]
return not samples.count(samples[0]) > len(samples) - 2
def profile_tests():
test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction)
if fn[0].startswith('test_')]
for name, fn in test_functions:
# There are a couple of tests that are not run for the C implementation, skip those
fn_args = inspect.getargspec(fn)[0]
if 'pvector' in fn_args:
print('Executing %s' % name)
result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1)
assert not detect_memory_leak(result), (name, result)
if __name__ == "__main__":
profile_tests()Improve memory error detection for less false positives"""
Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector.
"""
import inspect
import sys
import time
import memory_profiler
import vector_test
try:
from pvectorc import pvector
except ImportError:
print("No C implementation of PVector available, terminating")
sys.exit()
PROFILING_DURATION = 2.0
def run_function(fn):
stop = time.time() + PROFILING_DURATION
while time.time() < stop:
fn(pvector)
def detect_memory_leak(samples):
# Skip the first samples to get rid of the build up period and the last sample since it seems
# a little less precise
rising = 0
for i in range(5, len(samples)-1):
if samples[i] < samples[i+1]:
rising += 1
return (rising / float(len(samples) - 6)) > 0.2
def profile_tests():
test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction)
if fn[0].startswith('test_')]
for name, fn in test_functions:
# There are a couple of tests that are not run for the C implementation, skip those
fn_args = inspect.getargspec(fn)[0]
if 'pvector' in fn_args:
print('Executing %s' % name)
result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1)
assert not detect_memory_leak(result), (name, result)
if __name__ == "__main__":
profile_tests()
|
<commit_before>"""
Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector.
"""
import inspect
import sys
import time
import memory_profiler
import vector_test
try:
from pvectorc import pvector
except ImportError:
print("No C implementation of PVector available, terminating")
sys.exit()
PROFILING_DURATION = 2.0
def run_function(fn):
stop = time.time() + PROFILING_DURATION
while time.time() < stop:
fn(pvector)
def detect_memory_leak(samples):
# Skip the first half to get rid of the build up period and the last since it seems
# a little less precise
samples = samples[int(len(samples)/2):len(samples)-1]
return not samples.count(samples[0]) > len(samples) - 2
def profile_tests():
test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction)
if fn[0].startswith('test_')]
for name, fn in test_functions:
# There are a couple of tests that are not run for the C implementation, skip those
fn_args = inspect.getargspec(fn)[0]
if 'pvector' in fn_args:
print('Executing %s' % name)
result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1)
assert not detect_memory_leak(result), (name, result)
if __name__ == "__main__":
profile_tests()<commit_msg>Improve memory error detection for less false positives<commit_after>"""
Script to try do detect any memory leaks that may be lurking in the C implementation of the PVector.
"""
import inspect
import sys
import time
import memory_profiler
import vector_test
try:
from pvectorc import pvector
except ImportError:
print("No C implementation of PVector available, terminating")
sys.exit()
PROFILING_DURATION = 2.0
def run_function(fn):
stop = time.time() + PROFILING_DURATION
while time.time() < stop:
fn(pvector)
def detect_memory_leak(samples):
# Skip the first samples to get rid of the build up period and the last sample since it seems
# a little less precise
rising = 0
for i in range(5, len(samples)-1):
if samples[i] < samples[i+1]:
rising += 1
return (rising / float(len(samples) - 6)) > 0.2
def profile_tests():
test_functions = [fn for fn in inspect.getmembers(vector_test, inspect.isfunction)
if fn[0].startswith('test_')]
for name, fn in test_functions:
# There are a couple of tests that are not run for the C implementation, skip those
fn_args = inspect.getargspec(fn)[0]
if 'pvector' in fn_args:
print('Executing %s' % name)
result = memory_profiler.memory_usage((run_function, (fn,), {}), interval=.1)
assert not detect_memory_leak(result), (name, result)
if __name__ == "__main__":
profile_tests()
|
ee2db892b4dafa33115779166773e248c17a1b43
|
kyoto/tests/test_client.py
|
kyoto/tests/test_client.py
|
import unittest
import kyoto.server
import kyoto.tests.dummy
import kyoto.client
class ServiceTestCase(unittest.TestCase):
def setUp(self):
self.address = ('localhost', 1337)
self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy])
self.server.start()
self.service = kyoto.client.Service(self.address, ":dummy")
def test_invalid_service_name_type(self):
with self.assertRaises(ValueError):
service = kyoto.client.Service(self.address, "dummy")
def test_sync_request(self):
response = self.service.call(":echo", ["hello"])
self.assertEqual(response, "hello?")
def test_async_request(self):
response = self.service.cast(":echo", ["hello"])
self.assertEqual(response, None)
def tearDown(self):
self.server.stop()
|
import unittest
import kyoto.server
import kyoto.tests.dummy
import kyoto.client
class ServiceTestCase(unittest.TestCase):
def setUp(self):
self.address = ('localhost', 1337)
self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy])
self.server.start()
self.service = kyoto.client.Service(self.address, ":dummy")
def test_invalid_module_name_type(self):
with self.assertRaises(ValueError):
service = kyoto.client.Service(self.address, "dummy")
service = kyoto.client.Service(self.address, ":dummy")
def test_sync_request(self):
response = self.service.call(":echo", ["hello"])
self.assertEqual(response, "hello?")
def test_async_request(self):
response = self.service.cast(":echo", ["hello"])
self.assertEqual(response, None)
def tearDown(self):
self.server.stop()
|
Add valid module name test case
|
Add valid module name test case
|
Python
|
mit
|
kyoto-project/kyoto
|
import unittest
import kyoto.server
import kyoto.tests.dummy
import kyoto.client
class ServiceTestCase(unittest.TestCase):
def setUp(self):
self.address = ('localhost', 1337)
self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy])
self.server.start()
self.service = kyoto.client.Service(self.address, ":dummy")
def test_invalid_service_name_type(self):
with self.assertRaises(ValueError):
service = kyoto.client.Service(self.address, "dummy")
def test_sync_request(self):
response = self.service.call(":echo", ["hello"])
self.assertEqual(response, "hello?")
def test_async_request(self):
response = self.service.cast(":echo", ["hello"])
self.assertEqual(response, None)
def tearDown(self):
self.server.stop()
Add valid module name test case
|
import unittest
import kyoto.server
import kyoto.tests.dummy
import kyoto.client
class ServiceTestCase(unittest.TestCase):
def setUp(self):
self.address = ('localhost', 1337)
self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy])
self.server.start()
self.service = kyoto.client.Service(self.address, ":dummy")
def test_invalid_module_name_type(self):
with self.assertRaises(ValueError):
service = kyoto.client.Service(self.address, "dummy")
service = kyoto.client.Service(self.address, ":dummy")
def test_sync_request(self):
response = self.service.call(":echo", ["hello"])
self.assertEqual(response, "hello?")
def test_async_request(self):
response = self.service.cast(":echo", ["hello"])
self.assertEqual(response, None)
def tearDown(self):
self.server.stop()
|
<commit_before>import unittest
import kyoto.server
import kyoto.tests.dummy
import kyoto.client
class ServiceTestCase(unittest.TestCase):
def setUp(self):
self.address = ('localhost', 1337)
self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy])
self.server.start()
self.service = kyoto.client.Service(self.address, ":dummy")
def test_invalid_service_name_type(self):
with self.assertRaises(ValueError):
service = kyoto.client.Service(self.address, "dummy")
def test_sync_request(self):
response = self.service.call(":echo", ["hello"])
self.assertEqual(response, "hello?")
def test_async_request(self):
response = self.service.cast(":echo", ["hello"])
self.assertEqual(response, None)
def tearDown(self):
self.server.stop()
<commit_msg>Add valid module name test case<commit_after>
|
import unittest
import kyoto.server
import kyoto.tests.dummy
import kyoto.client
class ServiceTestCase(unittest.TestCase):
def setUp(self):
self.address = ('localhost', 1337)
self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy])
self.server.start()
self.service = kyoto.client.Service(self.address, ":dummy")
def test_invalid_module_name_type(self):
with self.assertRaises(ValueError):
service = kyoto.client.Service(self.address, "dummy")
service = kyoto.client.Service(self.address, ":dummy")
def test_sync_request(self):
response = self.service.call(":echo", ["hello"])
self.assertEqual(response, "hello?")
def test_async_request(self):
response = self.service.cast(":echo", ["hello"])
self.assertEqual(response, None)
def tearDown(self):
self.server.stop()
|
import unittest
import kyoto.server
import kyoto.tests.dummy
import kyoto.client
class ServiceTestCase(unittest.TestCase):
def setUp(self):
self.address = ('localhost', 1337)
self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy])
self.server.start()
self.service = kyoto.client.Service(self.address, ":dummy")
def test_invalid_service_name_type(self):
with self.assertRaises(ValueError):
service = kyoto.client.Service(self.address, "dummy")
def test_sync_request(self):
response = self.service.call(":echo", ["hello"])
self.assertEqual(response, "hello?")
def test_async_request(self):
response = self.service.cast(":echo", ["hello"])
self.assertEqual(response, None)
def tearDown(self):
self.server.stop()
Add valid module name test caseimport unittest
import kyoto.server
import kyoto.tests.dummy
import kyoto.client
class ServiceTestCase(unittest.TestCase):
def setUp(self):
self.address = ('localhost', 1337)
self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy])
self.server.start()
self.service = kyoto.client.Service(self.address, ":dummy")
def test_invalid_module_name_type(self):
with self.assertRaises(ValueError):
service = kyoto.client.Service(self.address, "dummy")
service = kyoto.client.Service(self.address, ":dummy")
def test_sync_request(self):
response = self.service.call(":echo", ["hello"])
self.assertEqual(response, "hello?")
def test_async_request(self):
response = self.service.cast(":echo", ["hello"])
self.assertEqual(response, None)
def tearDown(self):
self.server.stop()
|
<commit_before>import unittest
import kyoto.server
import kyoto.tests.dummy
import kyoto.client
class ServiceTestCase(unittest.TestCase):
def setUp(self):
self.address = ('localhost', 1337)
self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy])
self.server.start()
self.service = kyoto.client.Service(self.address, ":dummy")
def test_invalid_service_name_type(self):
with self.assertRaises(ValueError):
service = kyoto.client.Service(self.address, "dummy")
def test_sync_request(self):
response = self.service.call(":echo", ["hello"])
self.assertEqual(response, "hello?")
def test_async_request(self):
response = self.service.cast(":echo", ["hello"])
self.assertEqual(response, None)
def tearDown(self):
self.server.stop()
<commit_msg>Add valid module name test case<commit_after>import unittest
import kyoto.server
import kyoto.tests.dummy
import kyoto.client
class ServiceTestCase(unittest.TestCase):
def setUp(self):
self.address = ('localhost', 1337)
self.server = kyoto.server.BertRPCServer([kyoto.tests.dummy])
self.server.start()
self.service = kyoto.client.Service(self.address, ":dummy")
def test_invalid_module_name_type(self):
with self.assertRaises(ValueError):
service = kyoto.client.Service(self.address, "dummy")
service = kyoto.client.Service(self.address, ":dummy")
def test_sync_request(self):
response = self.service.call(":echo", ["hello"])
self.assertEqual(response, "hello?")
def test_async_request(self):
response = self.service.cast(":echo", ["hello"])
self.assertEqual(response, None)
def tearDown(self):
self.server.stop()
|
84304d8c04f59421a76b7c070eb9bdcf58a72567
|
callbackLoader.py
|
callbackLoader.py
|
# -*- coding: latin-1 -*-
'''
Created on 16.10.2012
@author: Teemu Pkknen
'''
import imp
import sys
import os
from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError
def get_callback_module( name ):
scriptDir = os.path.dirname(os.path.realpath(__file__))
# Already loaded?
try:
return sys.modules[name]
except KeyError:
pass
fp = pathname = description = None
try:
fp, pathname, description = imp.find_module(name, [os.getcwdu(), scriptDir])
return imp.load_module(name, fp, pathname, description)
except:
return None
finally:
if fp:
fp.close()
|
# -*- coding: latin-1 -*-
'''
Created on 16.10.2012
@author: Teemu Pkknen
'''
import imp
import sys
import os
import ntpath
from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError
def path_leaf(path):
head, tail = ntpath.split(path)
return tail or ntpath.basename(head)
def get_callback_module( name ):
scriptDir = os.path.dirname(os.path.realpath(__file__))
callback_module_dir = scriptDir + '/' + ntpath.dirname( name )
callback_module_name = path_leaf( name )
# Already loaded?
try:
return sys.modules[name]
except KeyError:
pass
fp = pathname = description = None
try:
fp, pathname, description = imp.find_module(callback_module_name, [callback_module_dir, os.getcwdu(), scriptDir])
return imp.load_module(name, fp, pathname, description)
except:
return None
finally:
if fp:
fp.close()
|
Make callback loader take into account directory names in loadable module name
|
Make callback loader take into account directory names in loadable module name
|
Python
|
mit
|
fire-uta/ir-simulation,fire-uta/ir-simulation
|
# -*- coding: latin-1 -*-
'''
Created on 16.10.2012
@author: Teemu Pkknen
'''
import imp
import sys
import os
from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError
def get_callback_module( name ):
scriptDir = os.path.dirname(os.path.realpath(__file__))
# Already loaded?
try:
return sys.modules[name]
except KeyError:
pass
fp = pathname = description = None
try:
fp, pathname, description = imp.find_module(name, [os.getcwdu(), scriptDir])
return imp.load_module(name, fp, pathname, description)
except:
return None
finally:
if fp:
fp.close()
Make callback loader take into account directory names in loadable module name
|
# -*- coding: latin-1 -*-
'''
Created on 16.10.2012
@author: Teemu Pkknen
'''
import imp
import sys
import os
import ntpath
from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError
def path_leaf(path):
head, tail = ntpath.split(path)
return tail or ntpath.basename(head)
def get_callback_module( name ):
scriptDir = os.path.dirname(os.path.realpath(__file__))
callback_module_dir = scriptDir + '/' + ntpath.dirname( name )
callback_module_name = path_leaf( name )
# Already loaded?
try:
return sys.modules[name]
except KeyError:
pass
fp = pathname = description = None
try:
fp, pathname, description = imp.find_module(callback_module_name, [callback_module_dir, os.getcwdu(), scriptDir])
return imp.load_module(name, fp, pathname, description)
except:
return None
finally:
if fp:
fp.close()
|
<commit_before># -*- coding: latin-1 -*-
'''
Created on 16.10.2012
@author: Teemu Pkknen
'''
import imp
import sys
import os
from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError
def get_callback_module( name ):
scriptDir = os.path.dirname(os.path.realpath(__file__))
# Already loaded?
try:
return sys.modules[name]
except KeyError:
pass
fp = pathname = description = None
try:
fp, pathname, description = imp.find_module(name, [os.getcwdu(), scriptDir])
return imp.load_module(name, fp, pathname, description)
except:
return None
finally:
if fp:
fp.close()
<commit_msg>Make callback loader take into account directory names in loadable module name<commit_after>
|
# -*- coding: latin-1 -*-
'''
Created on 16.10.2012
@author: Teemu Pkknen
'''
import imp
import sys
import os
import ntpath
from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError
def path_leaf(path):
head, tail = ntpath.split(path)
return tail or ntpath.basename(head)
def get_callback_module( name ):
scriptDir = os.path.dirname(os.path.realpath(__file__))
callback_module_dir = scriptDir + '/' + ntpath.dirname( name )
callback_module_name = path_leaf( name )
# Already loaded?
try:
return sys.modules[name]
except KeyError:
pass
fp = pathname = description = None
try:
fp, pathname, description = imp.find_module(callback_module_name, [callback_module_dir, os.getcwdu(), scriptDir])
return imp.load_module(name, fp, pathname, description)
except:
return None
finally:
if fp:
fp.close()
|
# -*- coding: latin-1 -*-
'''
Created on 16.10.2012
@author: Teemu Pkknen
'''
import imp
import sys
import os
from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError
def get_callback_module( name ):
scriptDir = os.path.dirname(os.path.realpath(__file__))
# Already loaded?
try:
return sys.modules[name]
except KeyError:
pass
fp = pathname = description = None
try:
fp, pathname, description = imp.find_module(name, [os.getcwdu(), scriptDir])
return imp.load_module(name, fp, pathname, description)
except:
return None
finally:
if fp:
fp.close()
Make callback loader take into account directory names in loadable module name# -*- coding: latin-1 -*-
'''
Created on 16.10.2012
@author: Teemu Pkknen
'''
import imp
import sys
import os
import ntpath
from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError
def path_leaf(path):
head, tail = ntpath.split(path)
return tail or ntpath.basename(head)
def get_callback_module( name ):
scriptDir = os.path.dirname(os.path.realpath(__file__))
callback_module_dir = scriptDir + '/' + ntpath.dirname( name )
callback_module_name = path_leaf( name )
# Already loaded?
try:
return sys.modules[name]
except KeyError:
pass
fp = pathname = description = None
try:
fp, pathname, description = imp.find_module(callback_module_name, [callback_module_dir, os.getcwdu(), scriptDir])
return imp.load_module(name, fp, pathname, description)
except:
return None
finally:
if fp:
fp.close()
|
<commit_before># -*- coding: latin-1 -*-
'''
Created on 16.10.2012
@author: Teemu Pkknen
'''
import imp
import sys
import os
from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError
def get_callback_module( name ):
scriptDir = os.path.dirname(os.path.realpath(__file__))
# Already loaded?
try:
return sys.modules[name]
except KeyError:
pass
fp = pathname = description = None
try:
fp, pathname, description = imp.find_module(name, [os.getcwdu(), scriptDir])
return imp.load_module(name, fp, pathname, description)
except:
return None
finally:
if fp:
fp.close()
<commit_msg>Make callback loader take into account directory names in loadable module name<commit_after># -*- coding: latin-1 -*-
'''
Created on 16.10.2012
@author: Teemu Pkknen
'''
import imp
import sys
import os
import ntpath
from qsdl.simulator.errors.ConfigurationInvalidError import ConfigurationInvalidError
def path_leaf(path):
head, tail = ntpath.split(path)
return tail or ntpath.basename(head)
def get_callback_module( name ):
scriptDir = os.path.dirname(os.path.realpath(__file__))
callback_module_dir = scriptDir + '/' + ntpath.dirname( name )
callback_module_name = path_leaf( name )
# Already loaded?
try:
return sys.modules[name]
except KeyError:
pass
fp = pathname = description = None
try:
fp, pathname, description = imp.find_module(callback_module_name, [callback_module_dir, os.getcwdu(), scriptDir])
return imp.load_module(name, fp, pathname, description)
except:
return None
finally:
if fp:
fp.close()
|
4a509970cb48b64046f88193efc141344437b151
|
tests/test_list_struct.py
|
tests/test_list_struct.py
|
import pytest
from hypothesis import given
from hypothesis.strategies import lists, integers, floats, one_of, composite
from datatyping.datatyping import validate
def test_empty():
assert validate([], []) is None
@given(li=lists(integers()))
def test_plain(li):
assert validate([int], li) is None
@given(lst=lists(floats(), min_size=1))
def test_plain_type_error(lst):
with pytest.raises(TypeError):
validate([int], lst)
@given(lst=one_of(lists(integers(), min_size=5),
lists(integers(), max_size=3)))
def test_list_lengths(lst):
with pytest.raises(ValueError):
validate([int, int, int, str], lst)
@given(lst=lists(lists(integers())))
def test_nested(lst):
assert validate([[int]], lst)
with pytest.raises(TypeError):
validate([int], lst)
@composite
def heavy_nested_data(draw):
return [draw(lists(integers)), draw(floats()), lists(lists(floats()))]
@given(lst=heavy_nested_data())
def test_heavy_nested(lst):
assert validate([[int], float, [[float]]], lst) is None
with pytest.raises(TypeError):
assert validate([[str], int, int], lst)
with pytest.raises(ValueError):
validate([[[float]]], lst)
|
import pytest
from hypothesis import given
from hypothesis.strategies import lists, integers, floats, one_of, composite
from datatyping.datatyping import validate
def test_empty():
assert validate([], []) is None
@given(li=lists(integers()))
def test_plain(li):
assert validate([int], li) is None
@given(lst=lists(floats(), min_size=1))
def test_plain_type_error(lst):
with pytest.raises(TypeError):
validate([int], lst)
@given(lst=one_of(lists(integers(), min_size=5),
lists(integers(), max_size=3)))
def test_list_lengths(lst):
with pytest.raises(ValueError):
validate([int, int, int, str], lst)
@given(lst=lists(lists(integers(), min_size=1), min_size=1))
def test_nested(lst):
assert validate([[int]], lst) is None
with pytest.raises(TypeError):
validate([int], lst)
@composite
def heavy_nested_data(draw):
return [draw(lists(integers(), min_size=1, max_size=3)),
draw(floats()),
draw(lists(lists(floats(), min_size=1, max_size=3), min_size=1, max_size=3))]
@given(lst=heavy_nested_data())
def test_heavy_nested(lst):
assert validate([[int], float, [[float]]], lst) is None
with pytest.raises(TypeError):
validate([[str], int, int], lst)
with pytest.raises(TypeError):
validate([[[float]]], lst)
|
Fix up mistakes in tests
|
Fix up mistakes in tests
|
Python
|
mit
|
Zaab1t/datatyping
|
import pytest
from hypothesis import given
from hypothesis.strategies import lists, integers, floats, one_of, composite
from datatyping.datatyping import validate
def test_empty():
assert validate([], []) is None
@given(li=lists(integers()))
def test_plain(li):
assert validate([int], li) is None
@given(lst=lists(floats(), min_size=1))
def test_plain_type_error(lst):
with pytest.raises(TypeError):
validate([int], lst)
@given(lst=one_of(lists(integers(), min_size=5),
lists(integers(), max_size=3)))
def test_list_lengths(lst):
with pytest.raises(ValueError):
validate([int, int, int, str], lst)
@given(lst=lists(lists(integers())))
def test_nested(lst):
assert validate([[int]], lst)
with pytest.raises(TypeError):
validate([int], lst)
@composite
def heavy_nested_data(draw):
return [draw(lists(integers)), draw(floats()), lists(lists(floats()))]
@given(lst=heavy_nested_data())
def test_heavy_nested(lst):
assert validate([[int], float, [[float]]], lst) is None
with pytest.raises(TypeError):
assert validate([[str], int, int], lst)
with pytest.raises(ValueError):
validate([[[float]]], lst)
Fix up mistakes in tests
|
import pytest
from hypothesis import given
from hypothesis.strategies import lists, integers, floats, one_of, composite
from datatyping.datatyping import validate
def test_empty():
assert validate([], []) is None
@given(li=lists(integers()))
def test_plain(li):
assert validate([int], li) is None
@given(lst=lists(floats(), min_size=1))
def test_plain_type_error(lst):
with pytest.raises(TypeError):
validate([int], lst)
@given(lst=one_of(lists(integers(), min_size=5),
lists(integers(), max_size=3)))
def test_list_lengths(lst):
with pytest.raises(ValueError):
validate([int, int, int, str], lst)
@given(lst=lists(lists(integers(), min_size=1), min_size=1))
def test_nested(lst):
assert validate([[int]], lst) is None
with pytest.raises(TypeError):
validate([int], lst)
@composite
def heavy_nested_data(draw):
return [draw(lists(integers(), min_size=1, max_size=3)),
draw(floats()),
draw(lists(lists(floats(), min_size=1, max_size=3), min_size=1, max_size=3))]
@given(lst=heavy_nested_data())
def test_heavy_nested(lst):
assert validate([[int], float, [[float]]], lst) is None
with pytest.raises(TypeError):
validate([[str], int, int], lst)
with pytest.raises(TypeError):
validate([[[float]]], lst)
|
<commit_before>import pytest
from hypothesis import given
from hypothesis.strategies import lists, integers, floats, one_of, composite
from datatyping.datatyping import validate
def test_empty():
assert validate([], []) is None
@given(li=lists(integers()))
def test_plain(li):
assert validate([int], li) is None
@given(lst=lists(floats(), min_size=1))
def test_plain_type_error(lst):
with pytest.raises(TypeError):
validate([int], lst)
@given(lst=one_of(lists(integers(), min_size=5),
lists(integers(), max_size=3)))
def test_list_lengths(lst):
with pytest.raises(ValueError):
validate([int, int, int, str], lst)
@given(lst=lists(lists(integers())))
def test_nested(lst):
assert validate([[int]], lst)
with pytest.raises(TypeError):
validate([int], lst)
@composite
def heavy_nested_data(draw):
return [draw(lists(integers)), draw(floats()), lists(lists(floats()))]
@given(lst=heavy_nested_data())
def test_heavy_nested(lst):
assert validate([[int], float, [[float]]], lst) is None
with pytest.raises(TypeError):
assert validate([[str], int, int], lst)
with pytest.raises(ValueError):
validate([[[float]]], lst)
<commit_msg>Fix up mistakes in tests<commit_after>
|
import pytest
from hypothesis import given
from hypothesis.strategies import lists, integers, floats, one_of, composite
from datatyping.datatyping import validate
def test_empty():
assert validate([], []) is None
@given(li=lists(integers()))
def test_plain(li):
assert validate([int], li) is None
@given(lst=lists(floats(), min_size=1))
def test_plain_type_error(lst):
with pytest.raises(TypeError):
validate([int], lst)
@given(lst=one_of(lists(integers(), min_size=5),
lists(integers(), max_size=3)))
def test_list_lengths(lst):
with pytest.raises(ValueError):
validate([int, int, int, str], lst)
@given(lst=lists(lists(integers(), min_size=1), min_size=1))
def test_nested(lst):
assert validate([[int]], lst) is None
with pytest.raises(TypeError):
validate([int], lst)
@composite
def heavy_nested_data(draw):
return [draw(lists(integers(), min_size=1, max_size=3)),
draw(floats()),
draw(lists(lists(floats(), min_size=1, max_size=3), min_size=1, max_size=3))]
@given(lst=heavy_nested_data())
def test_heavy_nested(lst):
assert validate([[int], float, [[float]]], lst) is None
with pytest.raises(TypeError):
validate([[str], int, int], lst)
with pytest.raises(TypeError):
validate([[[float]]], lst)
|
import pytest
from hypothesis import given
from hypothesis.strategies import lists, integers, floats, one_of, composite
from datatyping.datatyping import validate
def test_empty():
assert validate([], []) is None
@given(li=lists(integers()))
def test_plain(li):
assert validate([int], li) is None
@given(lst=lists(floats(), min_size=1))
def test_plain_type_error(lst):
with pytest.raises(TypeError):
validate([int], lst)
@given(lst=one_of(lists(integers(), min_size=5),
lists(integers(), max_size=3)))
def test_list_lengths(lst):
with pytest.raises(ValueError):
validate([int, int, int, str], lst)
@given(lst=lists(lists(integers())))
def test_nested(lst):
assert validate([[int]], lst)
with pytest.raises(TypeError):
validate([int], lst)
@composite
def heavy_nested_data(draw):
return [draw(lists(integers)), draw(floats()), lists(lists(floats()))]
@given(lst=heavy_nested_data())
def test_heavy_nested(lst):
assert validate([[int], float, [[float]]], lst) is None
with pytest.raises(TypeError):
assert validate([[str], int, int], lst)
with pytest.raises(ValueError):
validate([[[float]]], lst)
Fix up mistakes in testsimport pytest
from hypothesis import given
from hypothesis.strategies import lists, integers, floats, one_of, composite
from datatyping.datatyping import validate
def test_empty():
assert validate([], []) is None
@given(li=lists(integers()))
def test_plain(li):
assert validate([int], li) is None
@given(lst=lists(floats(), min_size=1))
def test_plain_type_error(lst):
with pytest.raises(TypeError):
validate([int], lst)
@given(lst=one_of(lists(integers(), min_size=5),
lists(integers(), max_size=3)))
def test_list_lengths(lst):
with pytest.raises(ValueError):
validate([int, int, int, str], lst)
@given(lst=lists(lists(integers(), min_size=1), min_size=1))
def test_nested(lst):
assert validate([[int]], lst) is None
with pytest.raises(TypeError):
validate([int], lst)
@composite
def heavy_nested_data(draw):
return [draw(lists(integers(), min_size=1, max_size=3)),
draw(floats()),
draw(lists(lists(floats(), min_size=1, max_size=3), min_size=1, max_size=3))]
@given(lst=heavy_nested_data())
def test_heavy_nested(lst):
assert validate([[int], float, [[float]]], lst) is None
with pytest.raises(TypeError):
validate([[str], int, int], lst)
with pytest.raises(TypeError):
validate([[[float]]], lst)
|
<commit_before>import pytest
from hypothesis import given
from hypothesis.strategies import lists, integers, floats, one_of, composite
from datatyping.datatyping import validate
def test_empty():
assert validate([], []) is None
@given(li=lists(integers()))
def test_plain(li):
assert validate([int], li) is None
@given(lst=lists(floats(), min_size=1))
def test_plain_type_error(lst):
with pytest.raises(TypeError):
validate([int], lst)
@given(lst=one_of(lists(integers(), min_size=5),
lists(integers(), max_size=3)))
def test_list_lengths(lst):
with pytest.raises(ValueError):
validate([int, int, int, str], lst)
@given(lst=lists(lists(integers())))
def test_nested(lst):
assert validate([[int]], lst)
with pytest.raises(TypeError):
validate([int], lst)
@composite
def heavy_nested_data(draw):
return [draw(lists(integers)), draw(floats()), lists(lists(floats()))]
@given(lst=heavy_nested_data())
def test_heavy_nested(lst):
assert validate([[int], float, [[float]]], lst) is None
with pytest.raises(TypeError):
assert validate([[str], int, int], lst)
with pytest.raises(ValueError):
validate([[[float]]], lst)
<commit_msg>Fix up mistakes in tests<commit_after>import pytest
from hypothesis import given
from hypothesis.strategies import lists, integers, floats, one_of, composite
from datatyping.datatyping import validate
def test_empty():
assert validate([], []) is None
@given(li=lists(integers()))
def test_plain(li):
assert validate([int], li) is None
@given(lst=lists(floats(), min_size=1))
def test_plain_type_error(lst):
with pytest.raises(TypeError):
validate([int], lst)
@given(lst=one_of(lists(integers(), min_size=5),
lists(integers(), max_size=3)))
def test_list_lengths(lst):
with pytest.raises(ValueError):
validate([int, int, int, str], lst)
@given(lst=lists(lists(integers(), min_size=1), min_size=1))
def test_nested(lst):
assert validate([[int]], lst) is None
with pytest.raises(TypeError):
validate([int], lst)
@composite
def heavy_nested_data(draw):
return [draw(lists(integers(), min_size=1, max_size=3)),
draw(floats()),
draw(lists(lists(floats(), min_size=1, max_size=3), min_size=1, max_size=3))]
@given(lst=heavy_nested_data())
def test_heavy_nested(lst):
assert validate([[int], float, [[float]]], lst) is None
with pytest.raises(TypeError):
validate([[str], int, int], lst)
with pytest.raises(TypeError):
validate([[[float]]], lst)
|
3ce9f6d8537c6b6d0ec5a5e09c5f1f6b7b34699c
|
troposphere/eventschemas.py
|
troposphere/eventschemas.py
|
# Copyright (c) 2012-2019, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
#
# *** Do not modify - this file is autogenerated ***
# Resource specification version: 14.1.0
from troposphere import Tags
from . import AWSObject
class Discoverer(AWSObject):
resource_type = "AWS::EventSchemas::Discoverer"
props = {
"Description": (str, False),
"SourceArn": (str, True),
"Tags": (Tags, False),
}
class Registry(AWSObject):
resource_type = "AWS::EventSchemas::Registry"
props = {
"Description": (str, False),
"RegistryName": (str, False),
"Tags": (Tags, False),
}
class RegistryPolicy(AWSObject):
resource_type = "AWS::EventSchemas::RegistryPolicy"
props = {
"Policy": (dict, True),
"RegistryName": (str, True),
"RevisionId": (str, False),
}
class Schema(AWSObject):
resource_type = "AWS::EventSchemas::Schema"
props = {
"Content": (str, True),
"Description": (str, False),
"RegistryName": (str, True),
"SchemaName": (str, False),
"Tags": (Tags, False),
"Type": (str, True),
}
|
# Copyright (c) 2012-2021, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
#
# *** Do not modify - this file is autogenerated ***
# Resource specification version: 41.0.0
from troposphere import Tags
from . import AWSObject
from .validators import boolean
class Discoverer(AWSObject):
resource_type = "AWS::EventSchemas::Discoverer"
props = {
"CrossAccount": (boolean, False),
"Description": (str, False),
"SourceArn": (str, True),
"Tags": (Tags, False),
}
class Registry(AWSObject):
resource_type = "AWS::EventSchemas::Registry"
props = {
"Description": (str, False),
"RegistryName": (str, False),
"Tags": (Tags, False),
}
class RegistryPolicy(AWSObject):
resource_type = "AWS::EventSchemas::RegistryPolicy"
props = {
"Policy": (dict, True),
"RegistryName": (str, True),
"RevisionId": (str, False),
}
class Schema(AWSObject):
resource_type = "AWS::EventSchemas::Schema"
props = {
"Content": (str, True),
"Description": (str, False),
"RegistryName": (str, True),
"SchemaName": (str, False),
"Tags": (Tags, False),
"Type": (str, True),
}
|
Update EventSchemas per 2021-09-02 changes
|
Update EventSchemas per 2021-09-02 changes
|
Python
|
bsd-2-clause
|
cloudtools/troposphere,cloudtools/troposphere
|
# Copyright (c) 2012-2019, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
#
# *** Do not modify - this file is autogenerated ***
# Resource specification version: 14.1.0
from troposphere import Tags
from . import AWSObject
class Discoverer(AWSObject):
resource_type = "AWS::EventSchemas::Discoverer"
props = {
"Description": (str, False),
"SourceArn": (str, True),
"Tags": (Tags, False),
}
class Registry(AWSObject):
resource_type = "AWS::EventSchemas::Registry"
props = {
"Description": (str, False),
"RegistryName": (str, False),
"Tags": (Tags, False),
}
class RegistryPolicy(AWSObject):
resource_type = "AWS::EventSchemas::RegistryPolicy"
props = {
"Policy": (dict, True),
"RegistryName": (str, True),
"RevisionId": (str, False),
}
class Schema(AWSObject):
resource_type = "AWS::EventSchemas::Schema"
props = {
"Content": (str, True),
"Description": (str, False),
"RegistryName": (str, True),
"SchemaName": (str, False),
"Tags": (Tags, False),
"Type": (str, True),
}
Update EventSchemas per 2021-09-02 changes
|
# Copyright (c) 2012-2021, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
#
# *** Do not modify - this file is autogenerated ***
# Resource specification version: 41.0.0
from troposphere import Tags
from . import AWSObject
from .validators import boolean
class Discoverer(AWSObject):
resource_type = "AWS::EventSchemas::Discoverer"
props = {
"CrossAccount": (boolean, False),
"Description": (str, False),
"SourceArn": (str, True),
"Tags": (Tags, False),
}
class Registry(AWSObject):
resource_type = "AWS::EventSchemas::Registry"
props = {
"Description": (str, False),
"RegistryName": (str, False),
"Tags": (Tags, False),
}
class RegistryPolicy(AWSObject):
resource_type = "AWS::EventSchemas::RegistryPolicy"
props = {
"Policy": (dict, True),
"RegistryName": (str, True),
"RevisionId": (str, False),
}
class Schema(AWSObject):
resource_type = "AWS::EventSchemas::Schema"
props = {
"Content": (str, True),
"Description": (str, False),
"RegistryName": (str, True),
"SchemaName": (str, False),
"Tags": (Tags, False),
"Type": (str, True),
}
|
<commit_before># Copyright (c) 2012-2019, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
#
# *** Do not modify - this file is autogenerated ***
# Resource specification version: 14.1.0
from troposphere import Tags
from . import AWSObject
class Discoverer(AWSObject):
resource_type = "AWS::EventSchemas::Discoverer"
props = {
"Description": (str, False),
"SourceArn": (str, True),
"Tags": (Tags, False),
}
class Registry(AWSObject):
resource_type = "AWS::EventSchemas::Registry"
props = {
"Description": (str, False),
"RegistryName": (str, False),
"Tags": (Tags, False),
}
class RegistryPolicy(AWSObject):
resource_type = "AWS::EventSchemas::RegistryPolicy"
props = {
"Policy": (dict, True),
"RegistryName": (str, True),
"RevisionId": (str, False),
}
class Schema(AWSObject):
resource_type = "AWS::EventSchemas::Schema"
props = {
"Content": (str, True),
"Description": (str, False),
"RegistryName": (str, True),
"SchemaName": (str, False),
"Tags": (Tags, False),
"Type": (str, True),
}
<commit_msg>Update EventSchemas per 2021-09-02 changes<commit_after>
|
# Copyright (c) 2012-2021, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
#
# *** Do not modify - this file is autogenerated ***
# Resource specification version: 41.0.0
from troposphere import Tags
from . import AWSObject
from .validators import boolean
class Discoverer(AWSObject):
resource_type = "AWS::EventSchemas::Discoverer"
props = {
"CrossAccount": (boolean, False),
"Description": (str, False),
"SourceArn": (str, True),
"Tags": (Tags, False),
}
class Registry(AWSObject):
resource_type = "AWS::EventSchemas::Registry"
props = {
"Description": (str, False),
"RegistryName": (str, False),
"Tags": (Tags, False),
}
class RegistryPolicy(AWSObject):
resource_type = "AWS::EventSchemas::RegistryPolicy"
props = {
"Policy": (dict, True),
"RegistryName": (str, True),
"RevisionId": (str, False),
}
class Schema(AWSObject):
resource_type = "AWS::EventSchemas::Schema"
props = {
"Content": (str, True),
"Description": (str, False),
"RegistryName": (str, True),
"SchemaName": (str, False),
"Tags": (Tags, False),
"Type": (str, True),
}
|
# Copyright (c) 2012-2019, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
#
# *** Do not modify - this file is autogenerated ***
# Resource specification version: 14.1.0
from troposphere import Tags
from . import AWSObject
class Discoverer(AWSObject):
resource_type = "AWS::EventSchemas::Discoverer"
props = {
"Description": (str, False),
"SourceArn": (str, True),
"Tags": (Tags, False),
}
class Registry(AWSObject):
resource_type = "AWS::EventSchemas::Registry"
props = {
"Description": (str, False),
"RegistryName": (str, False),
"Tags": (Tags, False),
}
class RegistryPolicy(AWSObject):
resource_type = "AWS::EventSchemas::RegistryPolicy"
props = {
"Policy": (dict, True),
"RegistryName": (str, True),
"RevisionId": (str, False),
}
class Schema(AWSObject):
resource_type = "AWS::EventSchemas::Schema"
props = {
"Content": (str, True),
"Description": (str, False),
"RegistryName": (str, True),
"SchemaName": (str, False),
"Tags": (Tags, False),
"Type": (str, True),
}
Update EventSchemas per 2021-09-02 changes# Copyright (c) 2012-2021, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
#
# *** Do not modify - this file is autogenerated ***
# Resource specification version: 41.0.0
from troposphere import Tags
from . import AWSObject
from .validators import boolean
class Discoverer(AWSObject):
resource_type = "AWS::EventSchemas::Discoverer"
props = {
"CrossAccount": (boolean, False),
"Description": (str, False),
"SourceArn": (str, True),
"Tags": (Tags, False),
}
class Registry(AWSObject):
resource_type = "AWS::EventSchemas::Registry"
props = {
"Description": (str, False),
"RegistryName": (str, False),
"Tags": (Tags, False),
}
class RegistryPolicy(AWSObject):
resource_type = "AWS::EventSchemas::RegistryPolicy"
props = {
"Policy": (dict, True),
"RegistryName": (str, True),
"RevisionId": (str, False),
}
class Schema(AWSObject):
resource_type = "AWS::EventSchemas::Schema"
props = {
"Content": (str, True),
"Description": (str, False),
"RegistryName": (str, True),
"SchemaName": (str, False),
"Tags": (Tags, False),
"Type": (str, True),
}
|
<commit_before># Copyright (c) 2012-2019, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
#
# *** Do not modify - this file is autogenerated ***
# Resource specification version: 14.1.0
from troposphere import Tags
from . import AWSObject
class Discoverer(AWSObject):
resource_type = "AWS::EventSchemas::Discoverer"
props = {
"Description": (str, False),
"SourceArn": (str, True),
"Tags": (Tags, False),
}
class Registry(AWSObject):
resource_type = "AWS::EventSchemas::Registry"
props = {
"Description": (str, False),
"RegistryName": (str, False),
"Tags": (Tags, False),
}
class RegistryPolicy(AWSObject):
resource_type = "AWS::EventSchemas::RegistryPolicy"
props = {
"Policy": (dict, True),
"RegistryName": (str, True),
"RevisionId": (str, False),
}
class Schema(AWSObject):
resource_type = "AWS::EventSchemas::Schema"
props = {
"Content": (str, True),
"Description": (str, False),
"RegistryName": (str, True),
"SchemaName": (str, False),
"Tags": (Tags, False),
"Type": (str, True),
}
<commit_msg>Update EventSchemas per 2021-09-02 changes<commit_after># Copyright (c) 2012-2021, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
#
# *** Do not modify - this file is autogenerated ***
# Resource specification version: 41.0.0
from troposphere import Tags
from . import AWSObject
from .validators import boolean
class Discoverer(AWSObject):
resource_type = "AWS::EventSchemas::Discoverer"
props = {
"CrossAccount": (boolean, False),
"Description": (str, False),
"SourceArn": (str, True),
"Tags": (Tags, False),
}
class Registry(AWSObject):
resource_type = "AWS::EventSchemas::Registry"
props = {
"Description": (str, False),
"RegistryName": (str, False),
"Tags": (Tags, False),
}
class RegistryPolicy(AWSObject):
resource_type = "AWS::EventSchemas::RegistryPolicy"
props = {
"Policy": (dict, True),
"RegistryName": (str, True),
"RevisionId": (str, False),
}
class Schema(AWSObject):
resource_type = "AWS::EventSchemas::Schema"
props = {
"Content": (str, True),
"Description": (str, False),
"RegistryName": (str, True),
"SchemaName": (str, False),
"Tags": (Tags, False),
"Type": (str, True),
}
|
76ed79593a832c1cf85615d21b31f18f2c7adebf
|
yanico/session/__init__.py
|
yanico/session/__init__.py
|
# Copyright 2015-2016 Masayuki Yamamoto
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handle nicovideo.jp user_session."""
import pkg_resources
class LoaderNotFoundError(Exception):
"""Session loader is not found."""
class UserSessionNotFoundError(Exception):
"""Profile exists, but user_session is not found."""
def load(ltype, profile):
for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype):
load_func = entry.load()
return load_func(profile)
raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
|
# Copyright 2015-2016 Masayuki Yamamoto
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handle nicovideo.jp user_session."""
import pkg_resources
class LoaderNotFoundError(Exception):
"""Session loader is not found."""
class UserSessionNotFoundError(Exception):
"""Profile exists, but user_session is not found."""
def load(ltype, profile):
"""Return nicovideo.jp user session string.
Args:
ltype (str): loader type
profile (str): file path for profile
Returns:
str: user session
Raises:
LoaderNotFoundError
Error from loader
"""
for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype):
load_func = entry.load()
return load_func(profile)
raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
|
Add docstring into load function
|
Add docstring into load function
Follow to Google style.
|
Python
|
apache-2.0
|
ma8ma/yanico
|
# Copyright 2015-2016 Masayuki Yamamoto
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handle nicovideo.jp user_session."""
import pkg_resources
class LoaderNotFoundError(Exception):
"""Session loader is not found."""
class UserSessionNotFoundError(Exception):
"""Profile exists, but user_session is not found."""
def load(ltype, profile):
for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype):
load_func = entry.load()
return load_func(profile)
raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
Add docstring into load function
Follow to Google style.
|
# Copyright 2015-2016 Masayuki Yamamoto
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handle nicovideo.jp user_session."""
import pkg_resources
class LoaderNotFoundError(Exception):
"""Session loader is not found."""
class UserSessionNotFoundError(Exception):
"""Profile exists, but user_session is not found."""
def load(ltype, profile):
"""Return nicovideo.jp user session string.
Args:
ltype (str): loader type
profile (str): file path for profile
Returns:
str: user session
Raises:
LoaderNotFoundError
Error from loader
"""
for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype):
load_func = entry.load()
return load_func(profile)
raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
|
<commit_before># Copyright 2015-2016 Masayuki Yamamoto
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handle nicovideo.jp user_session."""
import pkg_resources
class LoaderNotFoundError(Exception):
"""Session loader is not found."""
class UserSessionNotFoundError(Exception):
"""Profile exists, but user_session is not found."""
def load(ltype, profile):
for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype):
load_func = entry.load()
return load_func(profile)
raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
<commit_msg>Add docstring into load function
Follow to Google style.<commit_after>
|
# Copyright 2015-2016 Masayuki Yamamoto
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handle nicovideo.jp user_session."""
import pkg_resources
class LoaderNotFoundError(Exception):
"""Session loader is not found."""
class UserSessionNotFoundError(Exception):
"""Profile exists, but user_session is not found."""
def load(ltype, profile):
"""Return nicovideo.jp user session string.
Args:
ltype (str): loader type
profile (str): file path for profile
Returns:
str: user session
Raises:
LoaderNotFoundError
Error from loader
"""
for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype):
load_func = entry.load()
return load_func(profile)
raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
|
# Copyright 2015-2016 Masayuki Yamamoto
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handle nicovideo.jp user_session."""
import pkg_resources
class LoaderNotFoundError(Exception):
"""Session loader is not found."""
class UserSessionNotFoundError(Exception):
"""Profile exists, but user_session is not found."""
def load(ltype, profile):
for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype):
load_func = entry.load()
return load_func(profile)
raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
Add docstring into load function
Follow to Google style.# Copyright 2015-2016 Masayuki Yamamoto
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handle nicovideo.jp user_session."""
import pkg_resources
class LoaderNotFoundError(Exception):
"""Session loader is not found."""
class UserSessionNotFoundError(Exception):
"""Profile exists, but user_session is not found."""
def load(ltype, profile):
"""Return nicovideo.jp user session string.
Args:
ltype (str): loader type
profile (str): file path for profile
Returns:
str: user session
Raises:
LoaderNotFoundError
Error from loader
"""
for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype):
load_func = entry.load()
return load_func(profile)
raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
|
<commit_before># Copyright 2015-2016 Masayuki Yamamoto
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handle nicovideo.jp user_session."""
import pkg_resources
class LoaderNotFoundError(Exception):
"""Session loader is not found."""
class UserSessionNotFoundError(Exception):
"""Profile exists, but user_session is not found."""
def load(ltype, profile):
for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype):
load_func = entry.load()
return load_func(profile)
raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
<commit_msg>Add docstring into load function
Follow to Google style.<commit_after># Copyright 2015-2016 Masayuki Yamamoto
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Handle nicovideo.jp user_session."""
import pkg_resources
class LoaderNotFoundError(Exception):
"""Session loader is not found."""
class UserSessionNotFoundError(Exception):
"""Profile exists, but user_session is not found."""
def load(ltype, profile):
"""Return nicovideo.jp user session string.
Args:
ltype (str): loader type
profile (str): file path for profile
Returns:
str: user session
Raises:
LoaderNotFoundError
Error from loader
"""
for entry in pkg_resources.iter_entry_points('yanico.sessions', ltype):
load_func = entry.load()
return load_func(profile)
raise LoaderNotFoundError('{} loader is not found.'.format(ltype))
|
086e2bb85d0076c55dff886154664dc7179561fa
|
utils/summary_downloader.py
|
utils/summary_downloader.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
class SummaryDownloader():
# base url for official schedule json page
SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule"
# url template for official json gamefeed page
JSON_GAME_FEED_URL_TEMPLATE = (
"http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from dateutil.parser import parse
from dateutil.relativedelta import DAILY
from dateutil.rrule import rrule
class SummaryDownloader():
# base url for official schedule json page
SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule"
# url template for official json gamefeed page
JSON_GAME_FEED_URL_TEMPLATE = (
"http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
MAX_DOWNLOAD_WORKERS = 8
def __init__(self, tgt_dir, date, to_date='', threads=0):
self.date = parse(date)
if to_date:
self.to_date = parse(to_date)
else:
self.to_date = self.date
# preparing list of dates to download summary data for
self.game_dates = list(
rrule(DAILY, dtstart=self.date, until=self.to_date))
print(self.game_dates)
if __name__ == '__main__':
date = "1997/04/20"
d = SummaryDownloader(r"d:\tmp", date)
|
Add constructor to summary downloader
|
Add constructor to summary downloader
|
Python
|
mit
|
leaffan/pynhldb
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
class SummaryDownloader():
# base url for official schedule json page
SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule"
# url template for official json gamefeed page
JSON_GAME_FEED_URL_TEMPLATE = (
"http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
Add constructor to summary downloader
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from dateutil.parser import parse
from dateutil.relativedelta import DAILY
from dateutil.rrule import rrule
class SummaryDownloader():
# base url for official schedule json page
SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule"
# url template for official json gamefeed page
JSON_GAME_FEED_URL_TEMPLATE = (
"http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
MAX_DOWNLOAD_WORKERS = 8
def __init__(self, tgt_dir, date, to_date='', threads=0):
self.date = parse(date)
if to_date:
self.to_date = parse(to_date)
else:
self.to_date = self.date
# preparing list of dates to download summary data for
self.game_dates = list(
rrule(DAILY, dtstart=self.date, until=self.to_date))
print(self.game_dates)
if __name__ == '__main__':
date = "1997/04/20"
d = SummaryDownloader(r"d:\tmp", date)
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
class SummaryDownloader():
# base url for official schedule json page
SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule"
# url template for official json gamefeed page
JSON_GAME_FEED_URL_TEMPLATE = (
"http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
<commit_msg>Add constructor to summary downloader<commit_after>
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from dateutil.parser import parse
from dateutil.relativedelta import DAILY
from dateutil.rrule import rrule
class SummaryDownloader():
# base url for official schedule json page
SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule"
# url template for official json gamefeed page
JSON_GAME_FEED_URL_TEMPLATE = (
"http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
MAX_DOWNLOAD_WORKERS = 8
def __init__(self, tgt_dir, date, to_date='', threads=0):
self.date = parse(date)
if to_date:
self.to_date = parse(to_date)
else:
self.to_date = self.date
# preparing list of dates to download summary data for
self.game_dates = list(
rrule(DAILY, dtstart=self.date, until=self.to_date))
print(self.game_dates)
if __name__ == '__main__':
date = "1997/04/20"
d = SummaryDownloader(r"d:\tmp", date)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
class SummaryDownloader():
# base url for official schedule json page
SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule"
# url template for official json gamefeed page
JSON_GAME_FEED_URL_TEMPLATE = (
"http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
Add constructor to summary downloader#!/usr/bin/env python
# -*- coding: utf-8 -*-
from dateutil.parser import parse
from dateutil.relativedelta import DAILY
from dateutil.rrule import rrule
class SummaryDownloader():
# base url for official schedule json page
SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule"
# url template for official json gamefeed page
JSON_GAME_FEED_URL_TEMPLATE = (
"http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
MAX_DOWNLOAD_WORKERS = 8
def __init__(self, tgt_dir, date, to_date='', threads=0):
self.date = parse(date)
if to_date:
self.to_date = parse(to_date)
else:
self.to_date = self.date
# preparing list of dates to download summary data for
self.game_dates = list(
rrule(DAILY, dtstart=self.date, until=self.to_date))
print(self.game_dates)
if __name__ == '__main__':
date = "1997/04/20"
d = SummaryDownloader(r"d:\tmp", date)
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
class SummaryDownloader():
# base url for official schedule json page
SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule"
# url template for official json gamefeed page
JSON_GAME_FEED_URL_TEMPLATE = (
"http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
<commit_msg>Add constructor to summary downloader<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from dateutil.parser import parse
from dateutil.relativedelta import DAILY
from dateutil.rrule import rrule
class SummaryDownloader():
# base url for official schedule json page
SCHEDULE_URL_BASE = "http://statsapi.web.nhl.com/api/v1/schedule"
# url template for official json gamefeed page
JSON_GAME_FEED_URL_TEMPLATE = (
"http://statsapi.web.nhl.com/api/v1/game/%s/feed/live")
MAX_DOWNLOAD_WORKERS = 8
def __init__(self, tgt_dir, date, to_date='', threads=0):
self.date = parse(date)
if to_date:
self.to_date = parse(to_date)
else:
self.to_date = self.date
# preparing list of dates to download summary data for
self.game_dates = list(
rrule(DAILY, dtstart=self.date, until=self.to_date))
print(self.game_dates)
if __name__ == '__main__':
date = "1997/04/20"
d = SummaryDownloader(r"d:\tmp", date)
|
10a241938d5469f9da3d7d8a695963ac7cff87b2
|
website/mosaic/settings_gondor.py
|
website/mosaic/settings_gondor.py
|
import os
import urlparse
from .settings import *
DEBUG = True
TEMPLATE_DEBUG = DEBUG
ADMINS = (
# ('Your Name', 'your_email@example.com'),
)
MANAGERS = ADMINS
if "GONDOR_DATABASE_URL" in os.environ:
urlparse.uses_netloc.append("postgres")
url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"])
DATABASES = {
"default": {
"ENGINE": {
"postgres": "django.db.backends.postgresql_psycopg2"
}[url.scheme],
"NAME": url.path[1:],
"USER": url.username,
"PASSWORD": url.password,
"HOST": url.hostname,
"PORT": url.port
}
}
MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media")
STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static")
MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url
STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url
ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/"
FILE_UPLOAD_PERMISSIONS = 0640
SECRET_KEY = '8f89aej9fa89898__fd9er38948934fjsfoijdaf.'
|
import os
import urlparse
from .settings import *
DEBUG = True
TEMPLATE_DEBUG = DEBUG
ADMINS = (
# ('Your Name', 'your_email@example.com'),
)
MANAGERS = ADMINS
if "GONDOR_DATABASE_URL" in os.environ:
urlparse.uses_netloc.append("postgres")
url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"])
DATABASES = {
"default": {
"ENGINE": {
"postgres": "django.db.backends.postgresql_psycopg2"
}[url.scheme],
"NAME": url.path[1:],
"USER": url.username,
"PASSWORD": url.password,
"HOST": url.hostname,
"PORT": url.port
}
}
MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media")
STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static")
MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url
STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url
ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/"
FILE_UPLOAD_PERMISSIONS = 0640
SECRET_KEY = os.environ.get('DJANGO_SECRET_KEY')
|
Remove secret key from the gondor settings file
|
Remove secret key from the gondor settings file
|
Python
|
mit
|
sema/django-2012,sema/django-2012
|
import os
import urlparse
from .settings import *
DEBUG = True
TEMPLATE_DEBUG = DEBUG
ADMINS = (
# ('Your Name', 'your_email@example.com'),
)
MANAGERS = ADMINS
if "GONDOR_DATABASE_URL" in os.environ:
urlparse.uses_netloc.append("postgres")
url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"])
DATABASES = {
"default": {
"ENGINE": {
"postgres": "django.db.backends.postgresql_psycopg2"
}[url.scheme],
"NAME": url.path[1:],
"USER": url.username,
"PASSWORD": url.password,
"HOST": url.hostname,
"PORT": url.port
}
}
MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media")
STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static")
MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url
STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url
ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/"
FILE_UPLOAD_PERMISSIONS = 0640
SECRET_KEY = '8f89aej9fa89898__fd9er38948934fjsfoijdaf.'Remove secret key from the gondor settings file
|
import os
import urlparse
from .settings import *
DEBUG = True
TEMPLATE_DEBUG = DEBUG
ADMINS = (
# ('Your Name', 'your_email@example.com'),
)
MANAGERS = ADMINS
if "GONDOR_DATABASE_URL" in os.environ:
urlparse.uses_netloc.append("postgres")
url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"])
DATABASES = {
"default": {
"ENGINE": {
"postgres": "django.db.backends.postgresql_psycopg2"
}[url.scheme],
"NAME": url.path[1:],
"USER": url.username,
"PASSWORD": url.password,
"HOST": url.hostname,
"PORT": url.port
}
}
MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media")
STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static")
MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url
STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url
ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/"
FILE_UPLOAD_PERMISSIONS = 0640
SECRET_KEY = os.environ.get('DJANGO_SECRET_KEY')
|
<commit_before>import os
import urlparse
from .settings import *
DEBUG = True
TEMPLATE_DEBUG = DEBUG
ADMINS = (
# ('Your Name', 'your_email@example.com'),
)
MANAGERS = ADMINS
if "GONDOR_DATABASE_URL" in os.environ:
urlparse.uses_netloc.append("postgres")
url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"])
DATABASES = {
"default": {
"ENGINE": {
"postgres": "django.db.backends.postgresql_psycopg2"
}[url.scheme],
"NAME": url.path[1:],
"USER": url.username,
"PASSWORD": url.password,
"HOST": url.hostname,
"PORT": url.port
}
}
MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media")
STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static")
MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url
STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url
ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/"
FILE_UPLOAD_PERMISSIONS = 0640
SECRET_KEY = '8f89aej9fa89898__fd9er38948934fjsfoijdaf.'<commit_msg>Remove secret key from the gondor settings file<commit_after>
|
import os
import urlparse
from .settings import *
DEBUG = True
TEMPLATE_DEBUG = DEBUG
ADMINS = (
# ('Your Name', 'your_email@example.com'),
)
MANAGERS = ADMINS
if "GONDOR_DATABASE_URL" in os.environ:
urlparse.uses_netloc.append("postgres")
url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"])
DATABASES = {
"default": {
"ENGINE": {
"postgres": "django.db.backends.postgresql_psycopg2"
}[url.scheme],
"NAME": url.path[1:],
"USER": url.username,
"PASSWORD": url.password,
"HOST": url.hostname,
"PORT": url.port
}
}
MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media")
STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static")
MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url
STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url
ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/"
FILE_UPLOAD_PERMISSIONS = 0640
SECRET_KEY = os.environ.get('DJANGO_SECRET_KEY')
|
import os
import urlparse
from .settings import *
DEBUG = True
TEMPLATE_DEBUG = DEBUG
ADMINS = (
# ('Your Name', 'your_email@example.com'),
)
MANAGERS = ADMINS
if "GONDOR_DATABASE_URL" in os.environ:
urlparse.uses_netloc.append("postgres")
url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"])
DATABASES = {
"default": {
"ENGINE": {
"postgres": "django.db.backends.postgresql_psycopg2"
}[url.scheme],
"NAME": url.path[1:],
"USER": url.username,
"PASSWORD": url.password,
"HOST": url.hostname,
"PORT": url.port
}
}
MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media")
STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static")
MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url
STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url
ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/"
FILE_UPLOAD_PERMISSIONS = 0640
SECRET_KEY = '8f89aej9fa89898__fd9er38948934fjsfoijdaf.'Remove secret key from the gondor settings fileimport os
import urlparse
from .settings import *
DEBUG = True
TEMPLATE_DEBUG = DEBUG
ADMINS = (
# ('Your Name', 'your_email@example.com'),
)
MANAGERS = ADMINS
if "GONDOR_DATABASE_URL" in os.environ:
urlparse.uses_netloc.append("postgres")
url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"])
DATABASES = {
"default": {
"ENGINE": {
"postgres": "django.db.backends.postgresql_psycopg2"
}[url.scheme],
"NAME": url.path[1:],
"USER": url.username,
"PASSWORD": url.password,
"HOST": url.hostname,
"PORT": url.port
}
}
MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media")
STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static")
MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url
STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url
ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/"
FILE_UPLOAD_PERMISSIONS = 0640
SECRET_KEY = os.environ.get('DJANGO_SECRET_KEY')
|
<commit_before>import os
import urlparse
from .settings import *
DEBUG = True
TEMPLATE_DEBUG = DEBUG
ADMINS = (
# ('Your Name', 'your_email@example.com'),
)
MANAGERS = ADMINS
if "GONDOR_DATABASE_URL" in os.environ:
urlparse.uses_netloc.append("postgres")
url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"])
DATABASES = {
"default": {
"ENGINE": {
"postgres": "django.db.backends.postgresql_psycopg2"
}[url.scheme],
"NAME": url.path[1:],
"USER": url.username,
"PASSWORD": url.password,
"HOST": url.hostname,
"PORT": url.port
}
}
MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media")
STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static")
MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url
STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url
ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/"
FILE_UPLOAD_PERMISSIONS = 0640
SECRET_KEY = '8f89aej9fa89898__fd9er38948934fjsfoijdaf.'<commit_msg>Remove secret key from the gondor settings file<commit_after>import os
import urlparse
from .settings import *
DEBUG = True
TEMPLATE_DEBUG = DEBUG
ADMINS = (
# ('Your Name', 'your_email@example.com'),
)
MANAGERS = ADMINS
if "GONDOR_DATABASE_URL" in os.environ:
urlparse.uses_netloc.append("postgres")
url = urlparse.urlparse(os.environ["GONDOR_DATABASE_URL"])
DATABASES = {
"default": {
"ENGINE": {
"postgres": "django.db.backends.postgresql_psycopg2"
}[url.scheme],
"NAME": url.path[1:],
"USER": url.username,
"PASSWORD": url.password,
"HOST": url.hostname,
"PORT": url.port
}
}
MEDIA_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "media")
STATIC_ROOT = os.path.join(os.environ["GONDOR_DATA_DIR"], "site_media", "static")
MEDIA_URL = "/site_media/media/" # make sure this maps inside of site_media_url
STATIC_URL = "/site_media/static/" # make sure this maps inside of site_media_url
ADMIN_MEDIA_PREFIX = STATIC_URL + "admin/"
FILE_UPLOAD_PERMISSIONS = 0640
SECRET_KEY = os.environ.get('DJANGO_SECRET_KEY')
|
e8e7bb5b7f063cc48b761fc17ef8f2264a17a2ce
|
dthm4kaiako/config/__init__.py
|
dthm4kaiako/config/__init__.py
|
"""Configuration for Django system."""
__version__ = "0.17.2"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
|
"""Configuration for Django system."""
__version__ = "0.17.3"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
|
Increment version number to 0.17.3
|
Increment version number to 0.17.3
|
Python
|
mit
|
uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers
|
"""Configuration for Django system."""
__version__ = "0.17.2"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
Increment version number to 0.17.3
|
"""Configuration for Django system."""
__version__ = "0.17.3"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
|
<commit_before>"""Configuration for Django system."""
__version__ = "0.17.2"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
<commit_msg>Increment version number to 0.17.3<commit_after>
|
"""Configuration for Django system."""
__version__ = "0.17.3"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
|
"""Configuration for Django system."""
__version__ = "0.17.2"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
Increment version number to 0.17.3"""Configuration for Django system."""
__version__ = "0.17.3"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
|
<commit_before>"""Configuration for Django system."""
__version__ = "0.17.2"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
<commit_msg>Increment version number to 0.17.3<commit_after>"""Configuration for Django system."""
__version__ = "0.17.3"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
|
50992031229ea903418935613cd5e1e561b04c91
|
Control/PID.py
|
Control/PID.py
|
class PID:
def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0):
self.Kp = Kp # Proporiional gain
self.Ki = Ki # Integral gain
self.Kd = Kd # Derivative gain
self.integral = 0
def incrementTime(self, error, dt):
self.integral = self.integral + error*self.Ki*dt
def computeCorrection(self, error, derror):
correction = self.Kp*(error) + self.Kd*derror + self.integral
return correction
|
class PID:
def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0):
self.Kp = Kp # Proportional gain
self.Ki = Ki # Integral gain
self.Kd = Kd # Derivative gain
self.integral = 0
def incrementTime(self, error, dt):
self.integral = self.integral + error*self.Ki*dt
def computeCorrection(self, error, derror):
correction = self.Kp*(error) + self.Kd*derror + self.integral
return correction
|
Correct typing error and arrange indentation
|
Correct typing error and arrange indentation
|
Python
|
mit
|
baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite
|
class PID:
def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0):
self.Kp = Kp # Proporiional gain
self.Ki = Ki # Integral gain
self.Kd = Kd # Derivative gain
self.integral = 0
def incrementTime(self, error, dt):
self.integral = self.integral + error*self.Ki*dt
def computeCorrection(self, error, derror):
correction = self.Kp*(error) + self.Kd*derror + self.integral
return correction
Correct typing error and arrange indentation
|
class PID:
def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0):
self.Kp = Kp # Proportional gain
self.Ki = Ki # Integral gain
self.Kd = Kd # Derivative gain
self.integral = 0
def incrementTime(self, error, dt):
self.integral = self.integral + error*self.Ki*dt
def computeCorrection(self, error, derror):
correction = self.Kp*(error) + self.Kd*derror + self.integral
return correction
|
<commit_before>class PID:
def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0):
self.Kp = Kp # Proporiional gain
self.Ki = Ki # Integral gain
self.Kd = Kd # Derivative gain
self.integral = 0
def incrementTime(self, error, dt):
self.integral = self.integral + error*self.Ki*dt
def computeCorrection(self, error, derror):
correction = self.Kp*(error) + self.Kd*derror + self.integral
return correction
<commit_msg>Correct typing error and arrange indentation<commit_after>
|
class PID:
def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0):
self.Kp = Kp # Proportional gain
self.Ki = Ki # Integral gain
self.Kd = Kd # Derivative gain
self.integral = 0
def incrementTime(self, error, dt):
self.integral = self.integral + error*self.Ki*dt
def computeCorrection(self, error, derror):
correction = self.Kp*(error) + self.Kd*derror + self.integral
return correction
|
class PID:
def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0):
self.Kp = Kp # Proporiional gain
self.Ki = Ki # Integral gain
self.Kd = Kd # Derivative gain
self.integral = 0
def incrementTime(self, error, dt):
self.integral = self.integral + error*self.Ki*dt
def computeCorrection(self, error, derror):
correction = self.Kp*(error) + self.Kd*derror + self.integral
return correction
Correct typing error and arrange indentationclass PID:
def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0):
self.Kp = Kp # Proportional gain
self.Ki = Ki # Integral gain
self.Kd = Kd # Derivative gain
self.integral = 0
def incrementTime(self, error, dt):
self.integral = self.integral + error*self.Ki*dt
def computeCorrection(self, error, derror):
correction = self.Kp*(error) + self.Kd*derror + self.integral
return correction
|
<commit_before>class PID:
def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0):
self.Kp = Kp # Proporiional gain
self.Ki = Ki # Integral gain
self.Kd = Kd # Derivative gain
self.integral = 0
def incrementTime(self, error, dt):
self.integral = self.integral + error*self.Ki*dt
def computeCorrection(self, error, derror):
correction = self.Kp*(error) + self.Kd*derror + self.integral
return correction
<commit_msg>Correct typing error and arrange indentation<commit_after>class PID:
def __init__(self, Kp=1, Ki=0.1, Kd=1, maxIntegralCorrection=0, minIntegralCorrection=-0):
self.Kp = Kp # Proportional gain
self.Ki = Ki # Integral gain
self.Kd = Kd # Derivative gain
self.integral = 0
def incrementTime(self, error, dt):
self.integral = self.integral + error*self.Ki*dt
def computeCorrection(self, error, derror):
correction = self.Kp*(error) + self.Kd*derror + self.integral
return correction
|
7d5a259460b4e8b8325fa55793ed4456425bda78
|
xd/tool/log.py
|
xd/tool/log.py
|
import logging
class ConsoleFormatter(logging.Formatter):
"""A logging formatter for use when logging to console.
Log message above logging.INFO will be prefixed with the levelname, fx.:
ERROR: this is wrong
And logging.DEBUG messages will be prefixed with name of the logger, which
should normally be the module name, fx.:
xd.tool.shell: chdir /home/user/my-project
To achive this (logger name being the module name), all modules should
setup logging this way:
import logging
log = logging.getLogger(__name__)
"""
def __init__(self):
"""Initialize the formatter."""
logging.Formatter.__init__(self)
return
def format(self, record):
"""Format the specified record as text."""
record.message = record.getMessage()
fmt = ""
if record.levelno > logging.INFO:
fmt += "%(levelname)s: "
if record.levelno == logging.DEBUG:
fmt += "%(name)s: "
fmt += "%(message)s"
s = fmt % record.__dict__
if record.exc_info:
if not record.exc_text:
record.exc_text = self.formatException(record.exc_info)
if record.exc_text:
if s[-1:] != "\n":
s += "\n\n"
s = s + record.exc_text + "\n"
return s
def init():
"""Initialize logging module for logging to console.
The root_logger will be setup and initialized to output print out
logging.INFO level and above.
"""
console_formatter = ConsoleFormatter()
console_logger = logging.StreamHandler()
console_logger.setFormatter(console_formatter)
root_logger = logging.getLogger()
root_logger.addHandler(console_logger)
root_logger.setLevel(logging.INFO)
|
import logging
class ConsoleFormatter(logging.Formatter):
"""A logging formatter for use when logging to console.
Log message above logging.INFO will be prefixed with the levelname, fx.:
ERROR: this is wrong
And logging.DEBUG messages will be prefixed with name of the logger, which
should normally be the module name, fx.:
xd.tool.shell: chdir /home/user/my-project
To achive this (logger name being the module name), all modules should
setup logging this way:
import logging
log = logging.getLogger(__name__)
"""
def __init__(self):
"""Initialize the formatter."""
logging.Formatter.__init__(self)
return
def formatMessage(self, record):
"""Format the specified record message as text."""
fmt = ""
if record.levelno > logging.INFO:
fmt += "%(levelname)s: "
if record.levelno == logging.DEBUG:
fmt += "%(name)s: "
fmt += "%(message)s"
return fmt % record.__dict__
def init():
"""Initialize logging module for logging to console.
The root_logger will be setup and initialized to output print out
logging.INFO level and above.
"""
console_formatter = ConsoleFormatter()
console_logger = logging.StreamHandler()
console_logger.setFormatter(console_formatter)
root_logger = logging.getLogger()
root_logger.addHandler(console_logger)
root_logger.setLevel(logging.INFO)
|
Refactor ConsoleFormatter.format() method to formatMessage()
|
Refactor ConsoleFormatter.format() method to formatMessage()
Re-use the standard Formatter.format() method, and only override the
Formatter.formatMessage() method.
Signed-off-by: Esben Haabendal <da90c138e4a9573086862393cde34fa33d74f6e5@haabendal.dk>
|
Python
|
mit
|
esben/xd-tool,XD-embedded/xd-tool,esben/xd-tool,XD-embedded/xd-tool
|
import logging
class ConsoleFormatter(logging.Formatter):
"""A logging formatter for use when logging to console.
Log message above logging.INFO will be prefixed with the levelname, fx.:
ERROR: this is wrong
And logging.DEBUG messages will be prefixed with name of the logger, which
should normally be the module name, fx.:
xd.tool.shell: chdir /home/user/my-project
To achive this (logger name being the module name), all modules should
setup logging this way:
import logging
log = logging.getLogger(__name__)
"""
def __init__(self):
"""Initialize the formatter."""
logging.Formatter.__init__(self)
return
def format(self, record):
"""Format the specified record as text."""
record.message = record.getMessage()
fmt = ""
if record.levelno > logging.INFO:
fmt += "%(levelname)s: "
if record.levelno == logging.DEBUG:
fmt += "%(name)s: "
fmt += "%(message)s"
s = fmt % record.__dict__
if record.exc_info:
if not record.exc_text:
record.exc_text = self.formatException(record.exc_info)
if record.exc_text:
if s[-1:] != "\n":
s += "\n\n"
s = s + record.exc_text + "\n"
return s
def init():
"""Initialize logging module for logging to console.
The root_logger will be setup and initialized to output print out
logging.INFO level and above.
"""
console_formatter = ConsoleFormatter()
console_logger = logging.StreamHandler()
console_logger.setFormatter(console_formatter)
root_logger = logging.getLogger()
root_logger.addHandler(console_logger)
root_logger.setLevel(logging.INFO)
Refactor ConsoleFormatter.format() method to formatMessage()
Re-use the standard Formatter.format() method, and only override the
Formatter.formatMessage() method.
Signed-off-by: Esben Haabendal <da90c138e4a9573086862393cde34fa33d74f6e5@haabendal.dk>
|
import logging
class ConsoleFormatter(logging.Formatter):
"""A logging formatter for use when logging to console.
Log message above logging.INFO will be prefixed with the levelname, fx.:
ERROR: this is wrong
And logging.DEBUG messages will be prefixed with name of the logger, which
should normally be the module name, fx.:
xd.tool.shell: chdir /home/user/my-project
To achive this (logger name being the module name), all modules should
setup logging this way:
import logging
log = logging.getLogger(__name__)
"""
def __init__(self):
"""Initialize the formatter."""
logging.Formatter.__init__(self)
return
def formatMessage(self, record):
"""Format the specified record message as text."""
fmt = ""
if record.levelno > logging.INFO:
fmt += "%(levelname)s: "
if record.levelno == logging.DEBUG:
fmt += "%(name)s: "
fmt += "%(message)s"
return fmt % record.__dict__
def init():
"""Initialize logging module for logging to console.
The root_logger will be setup and initialized to output print out
logging.INFO level and above.
"""
console_formatter = ConsoleFormatter()
console_logger = logging.StreamHandler()
console_logger.setFormatter(console_formatter)
root_logger = logging.getLogger()
root_logger.addHandler(console_logger)
root_logger.setLevel(logging.INFO)
|
<commit_before>import logging
class ConsoleFormatter(logging.Formatter):
"""A logging formatter for use when logging to console.
Log message above logging.INFO will be prefixed with the levelname, fx.:
ERROR: this is wrong
And logging.DEBUG messages will be prefixed with name of the logger, which
should normally be the module name, fx.:
xd.tool.shell: chdir /home/user/my-project
To achive this (logger name being the module name), all modules should
setup logging this way:
import logging
log = logging.getLogger(__name__)
"""
def __init__(self):
"""Initialize the formatter."""
logging.Formatter.__init__(self)
return
def format(self, record):
"""Format the specified record as text."""
record.message = record.getMessage()
fmt = ""
if record.levelno > logging.INFO:
fmt += "%(levelname)s: "
if record.levelno == logging.DEBUG:
fmt += "%(name)s: "
fmt += "%(message)s"
s = fmt % record.__dict__
if record.exc_info:
if not record.exc_text:
record.exc_text = self.formatException(record.exc_info)
if record.exc_text:
if s[-1:] != "\n":
s += "\n\n"
s = s + record.exc_text + "\n"
return s
def init():
"""Initialize logging module for logging to console.
The root_logger will be setup and initialized to output print out
logging.INFO level and above.
"""
console_formatter = ConsoleFormatter()
console_logger = logging.StreamHandler()
console_logger.setFormatter(console_formatter)
root_logger = logging.getLogger()
root_logger.addHandler(console_logger)
root_logger.setLevel(logging.INFO)
<commit_msg>Refactor ConsoleFormatter.format() method to formatMessage()
Re-use the standard Formatter.format() method, and only override the
Formatter.formatMessage() method.
Signed-off-by: Esben Haabendal <da90c138e4a9573086862393cde34fa33d74f6e5@haabendal.dk><commit_after>
|
import logging
class ConsoleFormatter(logging.Formatter):
"""A logging formatter for use when logging to console.
Log message above logging.INFO will be prefixed with the levelname, fx.:
ERROR: this is wrong
And logging.DEBUG messages will be prefixed with name of the logger, which
should normally be the module name, fx.:
xd.tool.shell: chdir /home/user/my-project
To achive this (logger name being the module name), all modules should
setup logging this way:
import logging
log = logging.getLogger(__name__)
"""
def __init__(self):
"""Initialize the formatter."""
logging.Formatter.__init__(self)
return
def formatMessage(self, record):
"""Format the specified record message as text."""
fmt = ""
if record.levelno > logging.INFO:
fmt += "%(levelname)s: "
if record.levelno == logging.DEBUG:
fmt += "%(name)s: "
fmt += "%(message)s"
return fmt % record.__dict__
def init():
"""Initialize logging module for logging to console.
The root_logger will be setup and initialized to output print out
logging.INFO level and above.
"""
console_formatter = ConsoleFormatter()
console_logger = logging.StreamHandler()
console_logger.setFormatter(console_formatter)
root_logger = logging.getLogger()
root_logger.addHandler(console_logger)
root_logger.setLevel(logging.INFO)
|
import logging
class ConsoleFormatter(logging.Formatter):
"""A logging formatter for use when logging to console.
Log message above logging.INFO will be prefixed with the levelname, fx.:
ERROR: this is wrong
And logging.DEBUG messages will be prefixed with name of the logger, which
should normally be the module name, fx.:
xd.tool.shell: chdir /home/user/my-project
To achive this (logger name being the module name), all modules should
setup logging this way:
import logging
log = logging.getLogger(__name__)
"""
def __init__(self):
"""Initialize the formatter."""
logging.Formatter.__init__(self)
return
def format(self, record):
"""Format the specified record as text."""
record.message = record.getMessage()
fmt = ""
if record.levelno > logging.INFO:
fmt += "%(levelname)s: "
if record.levelno == logging.DEBUG:
fmt += "%(name)s: "
fmt += "%(message)s"
s = fmt % record.__dict__
if record.exc_info:
if not record.exc_text:
record.exc_text = self.formatException(record.exc_info)
if record.exc_text:
if s[-1:] != "\n":
s += "\n\n"
s = s + record.exc_text + "\n"
return s
def init():
"""Initialize logging module for logging to console.
The root_logger will be setup and initialized to output print out
logging.INFO level and above.
"""
console_formatter = ConsoleFormatter()
console_logger = logging.StreamHandler()
console_logger.setFormatter(console_formatter)
root_logger = logging.getLogger()
root_logger.addHandler(console_logger)
root_logger.setLevel(logging.INFO)
Refactor ConsoleFormatter.format() method to formatMessage()
Re-use the standard Formatter.format() method, and only override the
Formatter.formatMessage() method.
Signed-off-by: Esben Haabendal <da90c138e4a9573086862393cde34fa33d74f6e5@haabendal.dk>import logging
class ConsoleFormatter(logging.Formatter):
"""A logging formatter for use when logging to console.
Log message above logging.INFO will be prefixed with the levelname, fx.:
ERROR: this is wrong
And logging.DEBUG messages will be prefixed with name of the logger, which
should normally be the module name, fx.:
xd.tool.shell: chdir /home/user/my-project
To achive this (logger name being the module name), all modules should
setup logging this way:
import logging
log = logging.getLogger(__name__)
"""
def __init__(self):
"""Initialize the formatter."""
logging.Formatter.__init__(self)
return
def formatMessage(self, record):
"""Format the specified record message as text."""
fmt = ""
if record.levelno > logging.INFO:
fmt += "%(levelname)s: "
if record.levelno == logging.DEBUG:
fmt += "%(name)s: "
fmt += "%(message)s"
return fmt % record.__dict__
def init():
"""Initialize logging module for logging to console.
The root_logger will be setup and initialized to output print out
logging.INFO level and above.
"""
console_formatter = ConsoleFormatter()
console_logger = logging.StreamHandler()
console_logger.setFormatter(console_formatter)
root_logger = logging.getLogger()
root_logger.addHandler(console_logger)
root_logger.setLevel(logging.INFO)
|
<commit_before>import logging
class ConsoleFormatter(logging.Formatter):
"""A logging formatter for use when logging to console.
Log message above logging.INFO will be prefixed with the levelname, fx.:
ERROR: this is wrong
And logging.DEBUG messages will be prefixed with name of the logger, which
should normally be the module name, fx.:
xd.tool.shell: chdir /home/user/my-project
To achive this (logger name being the module name), all modules should
setup logging this way:
import logging
log = logging.getLogger(__name__)
"""
def __init__(self):
"""Initialize the formatter."""
logging.Formatter.__init__(self)
return
def format(self, record):
"""Format the specified record as text."""
record.message = record.getMessage()
fmt = ""
if record.levelno > logging.INFO:
fmt += "%(levelname)s: "
if record.levelno == logging.DEBUG:
fmt += "%(name)s: "
fmt += "%(message)s"
s = fmt % record.__dict__
if record.exc_info:
if not record.exc_text:
record.exc_text = self.formatException(record.exc_info)
if record.exc_text:
if s[-1:] != "\n":
s += "\n\n"
s = s + record.exc_text + "\n"
return s
def init():
"""Initialize logging module for logging to console.
The root_logger will be setup and initialized to output print out
logging.INFO level and above.
"""
console_formatter = ConsoleFormatter()
console_logger = logging.StreamHandler()
console_logger.setFormatter(console_formatter)
root_logger = logging.getLogger()
root_logger.addHandler(console_logger)
root_logger.setLevel(logging.INFO)
<commit_msg>Refactor ConsoleFormatter.format() method to formatMessage()
Re-use the standard Formatter.format() method, and only override the
Formatter.formatMessage() method.
Signed-off-by: Esben Haabendal <da90c138e4a9573086862393cde34fa33d74f6e5@haabendal.dk><commit_after>import logging
class ConsoleFormatter(logging.Formatter):
"""A logging formatter for use when logging to console.
Log message above logging.INFO will be prefixed with the levelname, fx.:
ERROR: this is wrong
And logging.DEBUG messages will be prefixed with name of the logger, which
should normally be the module name, fx.:
xd.tool.shell: chdir /home/user/my-project
To achive this (logger name being the module name), all modules should
setup logging this way:
import logging
log = logging.getLogger(__name__)
"""
def __init__(self):
"""Initialize the formatter."""
logging.Formatter.__init__(self)
return
def formatMessage(self, record):
"""Format the specified record message as text."""
fmt = ""
if record.levelno > logging.INFO:
fmt += "%(levelname)s: "
if record.levelno == logging.DEBUG:
fmt += "%(name)s: "
fmt += "%(message)s"
return fmt % record.__dict__
def init():
"""Initialize logging module for logging to console.
The root_logger will be setup and initialized to output print out
logging.INFO level and above.
"""
console_formatter = ConsoleFormatter()
console_logger = logging.StreamHandler()
console_logger.setFormatter(console_formatter)
root_logger = logging.getLogger()
root_logger.addHandler(console_logger)
root_logger.setLevel(logging.INFO)
|
5162275b9b6136f2b97d195384bb9979a0d79bfc
|
script/lib/config.py
|
script/lib/config.py
|
#!/usr/bin/env python
import platform
import sys
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '9f5271d31e0f32eac5a20ef6f543e3f1d43ad645'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
|
#!/usr/bin/env python
import platform
import sys
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '56984fa0e4c3c745652510f342c0fb2724d846c2'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
|
Upgrade libchromiumcontent for dbus headers
|
Upgrade libchromiumcontent for dbus headers
|
Python
|
mit
|
ianscrivener/electron,chriskdon/electron,yalexx/electron,subblue/electron,nekuz0r/electron,systembugtj/electron,trankmichael/electron,posix4e/electron,bitemyapp/electron,beni55/electron,mrwizard82d1/electron,Faiz7412/electron,rajatsingla28/electron,tomashanacek/electron,kokdemo/electron,darwin/electron,vipulroxx/electron,nicobot/electron,tincan24/electron,jcblw/electron,neutrous/electron,iftekeriba/electron,arturts/electron,robinvandernoord/electron,matiasinsaurralde/electron,pandoraui/electron,vaginessa/electron,jaanus/electron,jlord/electron,faizalpribadi/electron,BionicClick/electron,Faiz7412/electron,ianscrivener/electron,sky7sea/electron,brave/electron,jtburke/electron,jaanus/electron,hokein/atom-shell,jhen0409/electron,evgenyzinoviev/electron,faizalpribadi/electron,yan-foto/electron,cos2004/electron,natgolov/electron,astoilkov/electron,cqqccqc/electron,stevemao/electron,dahal/electron,edulan/electron,greyhwndz/electron,wolfflow/electron,simongregory/electron,DivyaKMenon/electron,GoooIce/electron,IonicaBizauKitchen/electron,leftstick/electron,MaxWhere/electron,thomsonreuters/electron,shaundunne/electron,bitemyapp/electron,seanchas116/electron,aichingm/electron,stevekinney/electron,bwiggs/electron,leethomas/electron,eric-seekas/electron,John-Lin/electron,vipulroxx/electron,synaptek/electron,neutrous/electron,jonatasfreitasv/electron,egoist/electron,rreimann/electron,brave/muon,thingsinjars/electron,hokein/atom-shell,ankitaggarwal011/electron,digideskio/electron,adcentury/electron,kcrt/electron,Ivshti/electron,mjaniszew/electron,natgolov/electron,yan-foto/electron,fomojola/electron,RobertJGabriel/electron,JussMee15/electron,jaanus/electron,Faiz7412/electron,jtburke/electron,chriskdon/electron,jjz/electron,thomsonreuters/electron,vaginessa/electron,astoilkov/electron,aecca/electron,miniak/electron,nicholasess/electron,seanchas116/electron,GoooIce/electron,brave/electron,digideskio/electron,mattotodd/electron,Floato/electron,wan-qy/electron,nicobot/electron,maxogden/atom-shell,wan-qy/electron,farmisen/electron,posix4e/electron,thingsinjars/electron,evgenyzinoviev/electron,shockone/electron,leolujuyi/electron,fabien-d/electron,minggo/electron,brave/muon,wan-qy/electron,jhen0409/electron,sircharleswatson/electron,destan/electron,brenca/electron,rhencke/electron,shockone/electron,simongregory/electron,adcentury/electron,fomojola/electron,bwiggs/electron,sshiting/electron,tinydew4/electron,icattlecoder/electron,evgenyzinoviev/electron,xiruibing/electron,bright-sparks/electron,dongjoon-hyun/electron,zhakui/electron,seanchas116/electron,mattotodd/electron,jjz/electron,miniak/electron,benweissmann/electron,Faiz7412/electron,dkfiresky/electron,jannishuebl/electron,sky7sea/electron,JesselJohn/electron,joneit/electron,soulteary/electron,micalan/electron,vaginessa/electron,setzer777/electron,JesselJohn/electron,JesselJohn/electron,adamjgray/electron,shennushi/electron,medixdev/electron,takashi/electron,bbondy/electron,kenmozi/electron,BionicClick/electron,yalexx/electron,baiwyc119/electron,twolfson/electron,aliib/electron,renaesop/electron,fffej/electron,tinydew4/electron,nagyistoce/electron-atom-shell,simongregory/electron,meowlab/electron,gabrielPeart/electron,robinvandernoord/electron,nicholasess/electron,yan-foto/electron,arusakov/electron,shiftkey/electron,roadev/electron,RIAEvangelist/electron,christian-bromann/electron,egoist/electron,trankmichael/electron,d-salas/electron,bobwol/electron,zhakui/electron,d-salas/electron,kikong/electron,minggo/electron,gamedevsam/electron,yalexx/electron,baiwyc119/electron,gabriel/electron,fritx/electron,voidbridge/electron,jsutcodes/electron,Jacobichou/electron,BionicClick/electron,arturts/electron,jcblw/electron,yalexx/electron,JussMee15/electron,thompsonemerson/electron,deepak1556/atom-shell,robinvandernoord/electron,aichingm/electron,shockone/electron,egoist/electron,RIAEvangelist/electron,nekuz0r/electron,gabrielPeart/electron,nagyistoce/electron-atom-shell,maxogden/atom-shell,cqqccqc/electron,MaxGraey/electron,webmechanicx/electron,eric-seekas/electron,DivyaKMenon/electron,electron/electron,joaomoreno/atom-shell,mubassirhayat/electron,destan/electron,arusakov/electron,vipulroxx/electron,baiwyc119/electron,pirafrank/electron,greyhwndz/electron,gabriel/electron,renaesop/electron,RobertJGabriel/electron,setzer777/electron,egoist/electron,thomsonreuters/electron,kokdemo/electron,matiasinsaurralde/electron,Gerhut/electron,tincan24/electron,rsvip/electron,bpasero/electron,maxogden/atom-shell,eriser/electron,joneit/electron,iftekeriba/electron,takashi/electron,lzpfmh/electron,mhkeller/electron,jacksondc/electron,synaptek/electron,DivyaKMenon/electron,jiaz/electron,takashi/electron,icattlecoder/electron,fritx/electron,tincan24/electron,trankmichael/electron,vaginessa/electron,the-ress/electron,aliib/electron,vHanda/electron,rreimann/electron,simongregory/electron,bruce/electron,maxogden/atom-shell,cos2004/electron,beni55/electron,bitemyapp/electron,preco21/electron,ianscrivener/electron,aliib/electron,jjz/electron,systembugtj/electron,roadev/electron,arturts/electron,MaxWhere/electron,coderhaoxin/electron,jonatasfreitasv/electron,bpasero/electron,gstack/infinium-shell,saronwei/electron,subblue/electron,MaxWhere/electron,shaundunne/electron,brenca/electron,dongjoon-hyun/electron,joneit/electron,sky7sea/electron,digideskio/electron,vaginessa/electron,digideskio/electron,jlhbaseball15/electron,greyhwndz/electron,fireball-x/atom-shell,shaundunne/electron,arusakov/electron,yalexx/electron,trigrass2/electron,kostia/electron,bruce/electron,subblue/electron,medixdev/electron,edulan/electron,rhencke/electron,wolfflow/electron,bbondy/electron,chriskdon/electron,stevemao/electron,deed02392/electron,evgenyzinoviev/electron,dkfiresky/electron,leftstick/electron,mattdesl/electron,kenmozi/electron,John-Lin/electron,egoist/electron,aliib/electron,gerhardberger/electron,beni55/electron,beni55/electron,gamedevsam/electron,darwin/electron,kenmozi/electron,stevekinney/electron,xiruibing/electron,LadyNaggaga/electron,medixdev/electron,kostia/electron,mrwizard82d1/electron,pandoraui/electron,saronwei/electron,dahal/electron,gbn972/electron,leolujuyi/electron,nicobot/electron,shaundunne/electron,davazp/electron,voidbridge/electron,rhencke/electron,rajatsingla28/electron,zhakui/electron,sircharleswatson/electron,anko/electron,ervinb/electron,fomojola/electron,darwin/electron,DivyaKMenon/electron,MaxGraey/electron,tonyganch/electron,gstack/infinium-shell,egoist/electron,roadev/electron,jannishuebl/electron,edulan/electron,jsutcodes/electron,noikiy/electron,chrisswk/electron,simonfork/electron,thomsonreuters/electron,mattdesl/electron,stevemao/electron,IonicaBizauKitchen/electron,nicholasess/electron,minggo/electron,rprichard/electron,sshiting/electron,felixrieseberg/electron,bwiggs/electron,christian-bromann/electron,micalan/electron,Neron-X5/electron,michaelchiche/electron,leethomas/electron,tomashanacek/electron,benweissmann/electron,rajatsingla28/electron,d-salas/electron,kokdemo/electron,shiftkey/electron,noikiy/electron,ianscrivener/electron,sky7sea/electron,sky7sea/electron,chrisswk/electron,smczk/electron,the-ress/electron,RobertJGabriel/electron,sshiting/electron,twolfson/electron,cos2004/electron,kcrt/electron,Zagorakiss/electron,simonfork/electron,tinydew4/electron,jcblw/electron,cos2004/electron,gerhardberger/electron,carsonmcdonald/electron,kcrt/electron,timruffles/electron,bwiggs/electron,jacksondc/electron,kazupon/electron,pirafrank/electron,preco21/electron,nicobot/electron,miniak/electron,shennushi/electron,fabien-d/electron,Jacobichou/electron,fffej/electron,ankitaggarwal011/electron,arturts/electron,chriskdon/electron,mattdesl/electron,setzer777/electron,micalan/electron,nicholasess/electron,Gerhut/electron,edulan/electron,Zagorakiss/electron,Jonekee/electron,Rokt33r/electron,jtburke/electron,pandoraui/electron,joaomoreno/atom-shell,jacksondc/electron,vHanda/electron,soulteary/electron,d-salas/electron,rsvip/electron,xfstudio/electron,fomojola/electron,etiktin/electron,eriser/electron,darwin/electron,rsvip/electron,vipulroxx/electron,jsutcodes/electron,tinydew4/electron,preco21/electron,MaxGraey/electron,gabriel/electron,saronwei/electron,brave/muon,pombredanne/electron,jaanus/electron,Zagorakiss/electron,aaron-goshine/electron,pirafrank/electron,soulteary/electron,roadev/electron,davazp/electron,Rokt33r/electron,coderhaoxin/electron,systembugtj/electron,ervinb/electron,faizalpribadi/electron,astoilkov/electron,thompsonemerson/electron,carsonmcdonald/electron,rhencke/electron,miniak/electron,carsonmcdonald/electron,bright-sparks/electron,aichingm/electron,takashi/electron,jlord/electron,natgolov/electron,leethomas/electron,LadyNaggaga/electron,lzpfmh/electron,tincan24/electron,jsutcodes/electron,preco21/electron,kenmozi/electron,Ivshti/electron,jiaz/electron,christian-bromann/electron,zhakui/electron,fomojola/electron,Andrey-Pavlov/electron,dahal/electron,JussMee15/electron,jlhbaseball15/electron,zhakui/electron,trankmichael/electron,lrlna/electron,fomojola/electron,setzer777/electron,adamjgray/electron,mjaniszew/electron,jtburke/electron,noikiy/electron,smczk/electron,RobertJGabriel/electron,adamjgray/electron,wolfflow/electron,joneit/electron,bruce/electron,Evercoder/electron,Jacobichou/electron,thingsinjars/electron,coderhaoxin/electron,the-ress/electron,hokein/atom-shell,electron/electron,eriser/electron,jonatasfreitasv/electron,ervinb/electron,bruce/electron,jannishuebl/electron,cqqccqc/electron,roadev/electron,tomashanacek/electron,fritx/electron,synaptek/electron,mubassirhayat/electron,jannishuebl/electron,evgenyzinoviev/electron,mattdesl/electron,DivyaKMenon/electron,gamedevsam/electron,biblerule/UMCTelnetHub,eric-seekas/electron,michaelchiche/electron,michaelchiche/electron,sshiting/electron,Evercoder/electron,destan/electron,gbn972/electron,LadyNaggaga/electron,icattlecoder/electron,greyhwndz/electron,roadev/electron,oiledCode/electron,gabriel/electron,kostia/electron,jannishuebl/electron,RobertJGabriel/electron,gerhardberger/electron,Jacobichou/electron,aliib/electron,simonfork/electron,coderhaoxin/electron,posix4e/electron,kazupon/electron,jacksondc/electron,stevemao/electron,thompsonemerson/electron,nicobot/electron,John-Lin/electron,etiktin/electron,cos2004/electron,hokein/atom-shell,biblerule/UMCTelnetHub,iftekeriba/electron,eriser/electron,medixdev/electron,kcrt/electron,shaundunne/electron,Neron-X5/electron,vHanda/electron,arusakov/electron,aaron-goshine/electron,stevekinney/electron,howmuchcomputer/electron,deed02392/electron,nicobot/electron,soulteary/electron,jhen0409/electron,Ivshti/electron,jlhbaseball15/electron,xfstudio/electron,fireball-x/atom-shell,JussMee15/electron,Gerhut/electron,mubassirhayat/electron,sircharleswatson/electron,preco21/electron,gamedevsam/electron,nicholasess/electron,bpasero/electron,JesselJohn/electron,jhen0409/electron,jhen0409/electron,astoilkov/electron,fireball-x/atom-shell,gerhardberger/electron,dkfiresky/electron,bbondy/electron,GoooIce/electron,digideskio/electron,meowlab/electron,John-Lin/electron,icattlecoder/electron,tonyganch/electron,sky7sea/electron,LadyNaggaga/electron,minggo/electron,tinydew4/electron,mrwizard82d1/electron,dongjoon-hyun/electron,carsonmcdonald/electron,Andrey-Pavlov/electron,leftstick/electron,SufianHassan/electron,deed02392/electron,smczk/electron,gstack/infinium-shell,Neron-X5/electron,anko/electron,bobwol/electron,aecca/electron,leethomas/electron,subblue/electron,pandoraui/electron,davazp/electron,adcentury/electron,stevekinney/electron,jiaz/electron,minggo/electron,lzpfmh/electron,Rokt33r/electron,bobwol/electron,bbondy/electron,timruffles/electron,robinvandernoord/electron,aaron-goshine/electron,mirrh/electron,deepak1556/atom-shell,pirafrank/electron,Neron-X5/electron,oiledCode/electron,abhishekgahlot/electron,voidbridge/electron,lzpfmh/electron,rreimann/electron,davazp/electron,BionicClick/electron,micalan/electron,rajatsingla28/electron,xiruibing/electron,destan/electron,arusakov/electron,tomashanacek/electron,greyhwndz/electron,fireball-x/atom-shell,pandoraui/electron,biblerule/UMCTelnetHub,shaundunne/electron,eric-seekas/electron,timruffles/electron,darwin/electron,gabrielPeart/electron,minggo/electron,Andrey-Pavlov/electron,takashi/electron,nekuz0r/electron,mirrh/electron,eriser/electron,rprichard/electron,xfstudio/electron,bwiggs/electron,Floato/electron,nekuz0r/electron,shiftkey/electron,subblue/electron,edulan/electron,wan-qy/electron,matiasinsaurralde/electron,oiledCode/electron,abhishekgahlot/electron,MaxWhere/electron,howmuchcomputer/electron,jcblw/electron,the-ress/electron,sshiting/electron,tomashanacek/electron,trigrass2/electron,brave/electron,benweissmann/electron,maxogden/atom-shell,noikiy/electron,medixdev/electron,brave/muon,smczk/electron,kikong/electron,aichingm/electron,carsonmcdonald/electron,twolfson/electron,natgolov/electron,kokdemo/electron,RIAEvangelist/electron,arusakov/electron,abhishekgahlot/electron,seanchas116/electron,xfstudio/electron,shockone/electron,mattdesl/electron,gabrielPeart/electron,micalan/electron,jlord/electron,aecca/electron,Andrey-Pavlov/electron,setzer777/electron,bruce/electron,pombredanne/electron,chrisswk/electron,dongjoon-hyun/electron,simongregory/electron,bitemyapp/electron,mjaniszew/electron,evgenyzinoviev/electron,ankitaggarwal011/electron,jacksondc/electron,baiwyc119/electron,tylergibson/electron,Zagorakiss/electron,mubassirhayat/electron,electron/electron,tylergibson/electron,chrisswk/electron,gbn972/electron,zhakui/electron,cqqccqc/electron,shennushi/electron,jtburke/electron,deepak1556/atom-shell,tylergibson/electron,jiaz/electron,stevemao/electron,fabien-d/electron,rsvip/electron,bbondy/electron,stevemao/electron,shockone/electron,rsvip/electron,GoooIce/electron,leolujuyi/electron,thomsonreuters/electron,GoooIce/electron,iftekeriba/electron,DivyaKMenon/electron,bobwol/electron,wan-qy/electron,systembugtj/electron,wan-qy/electron,Gerhut/electron,biblerule/UMCTelnetHub,adcentury/electron,pirafrank/electron,trankmichael/electron,shennushi/electron,nagyistoce/electron-atom-shell,faizalpribadi/electron,tylergibson/electron,chrisswk/electron,gbn972/electron,joneit/electron,kikong/electron,gbn972/electron,mirrh/electron,leolujuyi/electron,MaxWhere/electron,IonicaBizauKitchen/electron,abhishekgahlot/electron,thomsonreuters/electron,joaomoreno/atom-shell,vHanda/electron,greyhwndz/electron,eric-seekas/electron,aaron-goshine/electron,mattotodd/electron,deed02392/electron,twolfson/electron,beni55/electron,kikong/electron,jlord/electron,saronwei/electron,ianscrivener/electron,Rokt33r/electron,Jacobichou/electron,preco21/electron,aecca/electron,miniak/electron,SufianHassan/electron,rreimann/electron,iftekeriba/electron,sircharleswatson/electron,trigrass2/electron,cos2004/electron,anko/electron,Floato/electron,mirrh/electron,mjaniszew/electron,JussMee15/electron,thingsinjars/electron,michaelchiche/electron,tinydew4/electron,mattotodd/electron,nekuz0r/electron,systembugtj/electron,mjaniszew/electron,aecca/electron,smczk/electron,felixrieseberg/electron,anko/electron,MaxGraey/electron,dongjoon-hyun/electron,wolfflow/electron,felixrieseberg/electron,xiruibing/electron,farmisen/electron,tylergibson/electron,brenca/electron,rajatsingla28/electron,adamjgray/electron,mhkeller/electron,mhkeller/electron,eriser/electron,JesselJohn/electron,mjaniszew/electron,jjz/electron,christian-bromann/electron,pandoraui/electron,gbn972/electron,matiasinsaurralde/electron,mirrh/electron,noikiy/electron,LadyNaggaga/electron,carsonmcdonald/electron,jjz/electron,fffej/electron,JussMee15/electron,Zagorakiss/electron,tomashanacek/electron,fabien-d/electron,jiaz/electron,gamedevsam/electron,saronwei/electron,voidbridge/electron,michaelchiche/electron,adcentury/electron,icattlecoder/electron,shennushi/electron,pombredanne/electron,ankitaggarwal011/electron,jlhbaseball15/electron,John-Lin/electron,Evercoder/electron,biblerule/UMCTelnetHub,electron/electron,ervinb/electron,jlhbaseball15/electron,tincan24/electron,thompsonemerson/electron,xiruibing/electron,electron/electron,bpasero/electron,simonfork/electron,adcentury/electron,tylergibson/electron,bpasero/electron,simonfork/electron,etiktin/electron,bobwol/electron,stevekinney/electron,aaron-goshine/electron,electron/electron,wolfflow/electron,bpasero/electron,ervinb/electron,rhencke/electron,seanchas116/electron,synaptek/electron,iftekeriba/electron,Gerhut/electron,bobwol/electron,mhkeller/electron,shiftkey/electron,kikong/electron,tincan24/electron,kenmozi/electron,fffej/electron,pombredanne/electron,Evercoder/electron,systembugtj/electron,Floato/electron,vipulroxx/electron,electron/electron,oiledCode/electron,fritx/electron,joneit/electron,neutrous/electron,mubassirhayat/electron,kcrt/electron,bright-sparks/electron,etiktin/electron,jonatasfreitasv/electron,aaron-goshine/electron,trigrass2/electron,eric-seekas/electron,Rokt33r/electron,micalan/electron,leolujuyi/electron,brenca/electron,JesselJohn/electron,leftstick/electron,bwiggs/electron,jaanus/electron,LadyNaggaga/electron,webmechanicx/electron,bitemyapp/electron,SufianHassan/electron,davazp/electron,Gerhut/electron,renaesop/electron,mhkeller/electron,synaptek/electron,chriskdon/electron,jaanus/electron,John-Lin/electron,mirrh/electron,trigrass2/electron,oiledCode/electron,aecca/electron,jjz/electron,xfstudio/electron,Jonekee/electron,deepak1556/atom-shell,trigrass2/electron,gerhardberger/electron,SufianHassan/electron,shiftkey/electron,xfstudio/electron,aichingm/electron,dahal/electron,Zagorakiss/electron,hokein/atom-shell,vipulroxx/electron,jsutcodes/electron,thingsinjars/electron,voidbridge/electron,mattotodd/electron,adamjgray/electron,howmuchcomputer/electron,soulteary/electron,beni55/electron,Ivshti/electron,baiwyc119/electron,mrwizard82d1/electron,ianscrivener/electron,kazupon/electron,pirafrank/electron,timruffles/electron,faizalpribadi/electron,abhishekgahlot/electron,farmisen/electron,Jacobichou/electron,gerhardberger/electron,lrlna/electron,vHanda/electron,gabrielPeart/electron,cqqccqc/electron,sircharleswatson/electron,mrwizard82d1/electron,brenca/electron,pombredanne/electron,icattlecoder/electron,simonfork/electron,deepak1556/atom-shell,fritx/electron,tonyganch/electron,RIAEvangelist/electron,brave/electron,trankmichael/electron,chriskdon/electron,digideskio/electron,gabriel/electron,mattdesl/electron,thingsinjars/electron,SufianHassan/electron,rajatsingla28/electron,dahal/electron,miniak/electron,gstack/infinium-shell,kazupon/electron,farmisen/electron,nekuz0r/electron,fffej/electron,vaginessa/electron,arturts/electron,faizalpribadi/electron,jiaz/electron,jcblw/electron,nagyistoce/electron-atom-shell,biblerule/UMCTelnetHub,nagyistoce/electron-atom-shell,thompsonemerson/electron,GoooIce/electron,lrlna/electron,bbondy/electron,rprichard/electron,MaxWhere/electron,cqqccqc/electron,webmechanicx/electron,posix4e/electron,posix4e/electron,yan-foto/electron,jonatasfreitasv/electron,twolfson/electron,tonyganch/electron,ankitaggarwal011/electron,neutrous/electron,benweissmann/electron,jlord/electron,mattotodd/electron,leftstick/electron,dongjoon-hyun/electron,fireball-x/atom-shell,d-salas/electron,howmuchcomputer/electron,kcrt/electron,IonicaBizauKitchen/electron,Andrey-Pavlov/electron,lzpfmh/electron,gabriel/electron,Evercoder/electron,kokdemo/electron,fffej/electron,howmuchcomputer/electron,aliib/electron,shockone/electron,gabrielPeart/electron,wolfflow/electron,jsutcodes/electron,aichingm/electron,bruce/electron,natgolov/electron,brave/electron,the-ress/electron,joaomoreno/atom-shell,anko/electron,matiasinsaurralde/electron,renaesop/electron,lzpfmh/electron,oiledCode/electron,rprichard/electron,bright-sparks/electron,lrlna/electron,BionicClick/electron,dahal/electron,Neron-X5/electron,adamjgray/electron,subblue/electron,edulan/electron,stevekinney/electron,rhencke/electron,robinvandernoord/electron,coderhaoxin/electron,ervinb/electron,pombredanne/electron,twolfson/electron,astoilkov/electron,jcblw/electron,the-ress/electron,farmisen/electron,leethomas/electron,leethomas/electron,Floato/electron,matiasinsaurralde/electron,felixrieseberg/electron,sircharleswatson/electron,voidbridge/electron,gerhardberger/electron,anko/electron,d-salas/electron,arturts/electron,Jonekee/electron,SufianHassan/electron,dkfiresky/electron,coderhaoxin/electron,mhkeller/electron,neutrous/electron,IonicaBizauKitchen/electron,lrlna/electron,seanchas116/electron,takashi/electron,the-ress/electron,etiktin/electron,Andrey-Pavlov/electron,posix4e/electron,fabien-d/electron,IonicaBizauKitchen/electron,bpasero/electron,noikiy/electron,farmisen/electron,kostia/electron,deed02392/electron,saronwei/electron,yalexx/electron,tonyganch/electron,MaxGraey/electron,christian-bromann/electron,sshiting/electron,webmechanicx/electron,gamedevsam/electron,destan/electron,felixrieseberg/electron,benweissmann/electron,Rokt33r/electron,ankitaggarwal011/electron,dkfiresky/electron,deed02392/electron,RIAEvangelist/electron,neutrous/electron,Jonekee/electron,brave/muon,Floato/electron,timruffles/electron,meowlab/electron,kenmozi/electron,howmuchcomputer/electron,kostia/electron,xiruibing/electron,Faiz7412/electron,robinvandernoord/electron,renaesop/electron,Neron-X5/electron,webmechanicx/electron,Ivshti/electron,meowlab/electron,brave/electron,yan-foto/electron,nicholasess/electron,Jonekee/electron,meowlab/electron,soulteary/electron,tonyganch/electron,smczk/electron,natgolov/electron,jacksondc/electron,lrlna/electron,leftstick/electron,jlhbaseball15/electron,kostia/electron,abhishekgahlot/electron,dkfiresky/electron,mrwizard82d1/electron,medixdev/electron,synaptek/electron,fritx/electron,rreimann/electron,joaomoreno/atom-shell,benweissmann/electron,RobertJGabriel/electron,Evercoder/electron,jonatasfreitasv/electron,bitemyapp/electron,Jonekee/electron,joaomoreno/atom-shell,christian-bromann/electron,kokdemo/electron,etiktin/electron,renaesop/electron,kazupon/electron,simongregory/electron,brave/muon,felixrieseberg/electron,jhen0409/electron,BionicClick/electron,vHanda/electron,RIAEvangelist/electron,michaelchiche/electron,destan/electron,davazp/electron,brenca/electron,astoilkov/electron,meowlab/electron,setzer777/electron,baiwyc119/electron,thompsonemerson/electron,yan-foto/electron,shiftkey/electron,kazupon/electron,jannishuebl/electron,webmechanicx/electron,jtburke/electron,bright-sparks/electron,bright-sparks/electron,rreimann/electron,shennushi/electron,gstack/infinium-shell,leolujuyi/electron
|
#!/usr/bin/env python
import platform
import sys
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '9f5271d31e0f32eac5a20ef6f543e3f1d43ad645'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
Upgrade libchromiumcontent for dbus headers
|
#!/usr/bin/env python
import platform
import sys
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '56984fa0e4c3c745652510f342c0fb2724d846c2'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
|
<commit_before>#!/usr/bin/env python
import platform
import sys
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '9f5271d31e0f32eac5a20ef6f543e3f1d43ad645'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
<commit_msg>Upgrade libchromiumcontent for dbus headers<commit_after>
|
#!/usr/bin/env python
import platform
import sys
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '56984fa0e4c3c745652510f342c0fb2724d846c2'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
|
#!/usr/bin/env python
import platform
import sys
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '9f5271d31e0f32eac5a20ef6f543e3f1d43ad645'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
Upgrade libchromiumcontent for dbus headers#!/usr/bin/env python
import platform
import sys
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '56984fa0e4c3c745652510f342c0fb2724d846c2'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
|
<commit_before>#!/usr/bin/env python
import platform
import sys
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '9f5271d31e0f32eac5a20ef6f543e3f1d43ad645'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
<commit_msg>Upgrade libchromiumcontent for dbus headers<commit_after>#!/usr/bin/env python
import platform
import sys
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '56984fa0e4c3c745652510f342c0fb2724d846c2'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
|
5a2c03b9369ccd00cc8c5c7bca4b2fc40bb18a7f
|
passpie/credential.py
|
passpie/credential.py
|
import re
def split_fullname(fullname):
rgx = re.compile(r"(?P<login>.*)@(?P<name>.*)")
try:
name = rgx.match(fullname).group("name")
login = rgx.match(fullname).group("login")
except AttributeError:
raise ValueError("Not a valid name")
return login if login else "_", name
def make_fullname(login, name):
return "{}@{}".format("_" if login is None else login, name)
|
import re
def split_fullname(fullname):
rgx = re.compile(r"(?P<login>.*)?@(?P<name>.*)")
try:
name = rgx.match(fullname).group("name")
login = rgx.match(fullname).group("login")
except AttributeError:
raise ValueError("Not a valid name")
return login if login else "_", name
def make_fullname(login, name):
return "{}@{}".format("_" if login is None else login, name)
|
Fix regex for spliting fullnames
|
Fix regex for spliting fullnames
|
Python
|
mit
|
marcwebbie/passpie,scorphus/passpie,marcwebbie/passpie,eiginn/passpie,scorphus/passpie,eiginn/passpie
|
import re
def split_fullname(fullname):
rgx = re.compile(r"(?P<login>.*)@(?P<name>.*)")
try:
name = rgx.match(fullname).group("name")
login = rgx.match(fullname).group("login")
except AttributeError:
raise ValueError("Not a valid name")
return login if login else "_", name
def make_fullname(login, name):
return "{}@{}".format("_" if login is None else login, name)
Fix regex for spliting fullnames
|
import re
def split_fullname(fullname):
rgx = re.compile(r"(?P<login>.*)?@(?P<name>.*)")
try:
name = rgx.match(fullname).group("name")
login = rgx.match(fullname).group("login")
except AttributeError:
raise ValueError("Not a valid name")
return login if login else "_", name
def make_fullname(login, name):
return "{}@{}".format("_" if login is None else login, name)
|
<commit_before>import re
def split_fullname(fullname):
rgx = re.compile(r"(?P<login>.*)@(?P<name>.*)")
try:
name = rgx.match(fullname).group("name")
login = rgx.match(fullname).group("login")
except AttributeError:
raise ValueError("Not a valid name")
return login if login else "_", name
def make_fullname(login, name):
return "{}@{}".format("_" if login is None else login, name)
<commit_msg>Fix regex for spliting fullnames<commit_after>
|
import re
def split_fullname(fullname):
rgx = re.compile(r"(?P<login>.*)?@(?P<name>.*)")
try:
name = rgx.match(fullname).group("name")
login = rgx.match(fullname).group("login")
except AttributeError:
raise ValueError("Not a valid name")
return login if login else "_", name
def make_fullname(login, name):
return "{}@{}".format("_" if login is None else login, name)
|
import re
def split_fullname(fullname):
rgx = re.compile(r"(?P<login>.*)@(?P<name>.*)")
try:
name = rgx.match(fullname).group("name")
login = rgx.match(fullname).group("login")
except AttributeError:
raise ValueError("Not a valid name")
return login if login else "_", name
def make_fullname(login, name):
return "{}@{}".format("_" if login is None else login, name)
Fix regex for spliting fullnamesimport re
def split_fullname(fullname):
rgx = re.compile(r"(?P<login>.*)?@(?P<name>.*)")
try:
name = rgx.match(fullname).group("name")
login = rgx.match(fullname).group("login")
except AttributeError:
raise ValueError("Not a valid name")
return login if login else "_", name
def make_fullname(login, name):
return "{}@{}".format("_" if login is None else login, name)
|
<commit_before>import re
def split_fullname(fullname):
rgx = re.compile(r"(?P<login>.*)@(?P<name>.*)")
try:
name = rgx.match(fullname).group("name")
login = rgx.match(fullname).group("login")
except AttributeError:
raise ValueError("Not a valid name")
return login if login else "_", name
def make_fullname(login, name):
return "{}@{}".format("_" if login is None else login, name)
<commit_msg>Fix regex for spliting fullnames<commit_after>import re
def split_fullname(fullname):
rgx = re.compile(r"(?P<login>.*)?@(?P<name>.*)")
try:
name = rgx.match(fullname).group("name")
login = rgx.match(fullname).group("login")
except AttributeError:
raise ValueError("Not a valid name")
return login if login else "_", name
def make_fullname(login, name):
return "{}@{}".format("_" if login is None else login, name)
|
335abda444cbd5651af0d9a298570144627c7022
|
passwordless/utils.py
|
passwordless/utils.py
|
import os
import random
import uuid
from django.contrib.auth.hashers import make_password,is_password_usable
from django.utils import timezone
from datetime import timedelta
WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt')
def make_token():
"""
Generate a random token suitable for activation/confirmation via email
A hex-encoded random UUID has plenty of entropy to be secure enough for our
needs.
"""
return uuid.uuid4().hex
def expiration_date():
"""
AuthToken objects expire 1 hour after creation by default
"""
return timezone.now() + timedelta(hours=1)
def new_app_password(size=6):
f = open(WORDLIST_FILE, 'r')
words = []
for i in range(size):
words.append(next(f).strip())
for num,line in enumerate(f):
j = random.randrange(size+num)
if j < size:
words[j] = line.strip()
return words
|
import os
import random
import uuid
from django.utils import timezone
from datetime import timedelta
WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt')
def make_token():
"""
Generate a random token suitable for activation/confirmation via email
A hex-encoded random UUID has plenty of entropy to be secure enough for our
needs.
"""
return uuid.uuid4().hex
def expiration_date():
"""
AuthToken objects expire 1 hour after creation by default
"""
return timezone.now() + timedelta(hours=1)
def new_app_password(size=6):
f = open(WORDLIST_FILE, 'r')
words = []
for i in range(size):
words.append(next(f).strip())
for num,line in enumerate(f):
j = random.randrange(size+num)
if j < size:
words[j] = line.strip()
return ' '.join(words)
|
Return app passwords as string
|
Return app passwords as string
|
Python
|
mit
|
Kromey/fbxnano,Kromey/akwriters,Kromey/fbxnano,Kromey/fbxnano,Kromey/akwriters,Kromey/fbxnano,Kromey/akwriters,Kromey/akwriters
|
import os
import random
import uuid
from django.contrib.auth.hashers import make_password,is_password_usable
from django.utils import timezone
from datetime import timedelta
WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt')
def make_token():
"""
Generate a random token suitable for activation/confirmation via email
A hex-encoded random UUID has plenty of entropy to be secure enough for our
needs.
"""
return uuid.uuid4().hex
def expiration_date():
"""
AuthToken objects expire 1 hour after creation by default
"""
return timezone.now() + timedelta(hours=1)
def new_app_password(size=6):
f = open(WORDLIST_FILE, 'r')
words = []
for i in range(size):
words.append(next(f).strip())
for num,line in enumerate(f):
j = random.randrange(size+num)
if j < size:
words[j] = line.strip()
return words
Return app passwords as string
|
import os
import random
import uuid
from django.utils import timezone
from datetime import timedelta
WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt')
def make_token():
"""
Generate a random token suitable for activation/confirmation via email
A hex-encoded random UUID has plenty of entropy to be secure enough for our
needs.
"""
return uuid.uuid4().hex
def expiration_date():
"""
AuthToken objects expire 1 hour after creation by default
"""
return timezone.now() + timedelta(hours=1)
def new_app_password(size=6):
f = open(WORDLIST_FILE, 'r')
words = []
for i in range(size):
words.append(next(f).strip())
for num,line in enumerate(f):
j = random.randrange(size+num)
if j < size:
words[j] = line.strip()
return ' '.join(words)
|
<commit_before>import os
import random
import uuid
from django.contrib.auth.hashers import make_password,is_password_usable
from django.utils import timezone
from datetime import timedelta
WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt')
def make_token():
"""
Generate a random token suitable for activation/confirmation via email
A hex-encoded random UUID has plenty of entropy to be secure enough for our
needs.
"""
return uuid.uuid4().hex
def expiration_date():
"""
AuthToken objects expire 1 hour after creation by default
"""
return timezone.now() + timedelta(hours=1)
def new_app_password(size=6):
f = open(WORDLIST_FILE, 'r')
words = []
for i in range(size):
words.append(next(f).strip())
for num,line in enumerate(f):
j = random.randrange(size+num)
if j < size:
words[j] = line.strip()
return words
<commit_msg>Return app passwords as string<commit_after>
|
import os
import random
import uuid
from django.utils import timezone
from datetime import timedelta
WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt')
def make_token():
"""
Generate a random token suitable for activation/confirmation via email
A hex-encoded random UUID has plenty of entropy to be secure enough for our
needs.
"""
return uuid.uuid4().hex
def expiration_date():
"""
AuthToken objects expire 1 hour after creation by default
"""
return timezone.now() + timedelta(hours=1)
def new_app_password(size=6):
f = open(WORDLIST_FILE, 'r')
words = []
for i in range(size):
words.append(next(f).strip())
for num,line in enumerate(f):
j = random.randrange(size+num)
if j < size:
words[j] = line.strip()
return ' '.join(words)
|
import os
import random
import uuid
from django.contrib.auth.hashers import make_password,is_password_usable
from django.utils import timezone
from datetime import timedelta
WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt')
def make_token():
"""
Generate a random token suitable for activation/confirmation via email
A hex-encoded random UUID has plenty of entropy to be secure enough for our
needs.
"""
return uuid.uuid4().hex
def expiration_date():
"""
AuthToken objects expire 1 hour after creation by default
"""
return timezone.now() + timedelta(hours=1)
def new_app_password(size=6):
f = open(WORDLIST_FILE, 'r')
words = []
for i in range(size):
words.append(next(f).strip())
for num,line in enumerate(f):
j = random.randrange(size+num)
if j < size:
words[j] = line.strip()
return words
Return app passwords as stringimport os
import random
import uuid
from django.utils import timezone
from datetime import timedelta
WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt')
def make_token():
"""
Generate a random token suitable for activation/confirmation via email
A hex-encoded random UUID has plenty of entropy to be secure enough for our
needs.
"""
return uuid.uuid4().hex
def expiration_date():
"""
AuthToken objects expire 1 hour after creation by default
"""
return timezone.now() + timedelta(hours=1)
def new_app_password(size=6):
f = open(WORDLIST_FILE, 'r')
words = []
for i in range(size):
words.append(next(f).strip())
for num,line in enumerate(f):
j = random.randrange(size+num)
if j < size:
words[j] = line.strip()
return ' '.join(words)
|
<commit_before>import os
import random
import uuid
from django.contrib.auth.hashers import make_password,is_password_usable
from django.utils import timezone
from datetime import timedelta
WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt')
def make_token():
"""
Generate a random token suitable for activation/confirmation via email
A hex-encoded random UUID has plenty of entropy to be secure enough for our
needs.
"""
return uuid.uuid4().hex
def expiration_date():
"""
AuthToken objects expire 1 hour after creation by default
"""
return timezone.now() + timedelta(hours=1)
def new_app_password(size=6):
f = open(WORDLIST_FILE, 'r')
words = []
for i in range(size):
words.append(next(f).strip())
for num,line in enumerate(f):
j = random.randrange(size+num)
if j < size:
words[j] = line.strip()
return words
<commit_msg>Return app passwords as string<commit_after>import os
import random
import uuid
from django.utils import timezone
from datetime import timedelta
WORDLIST_FILE = os.path.join(os.path.dirname(__file__), 'wordlist.txt')
def make_token():
"""
Generate a random token suitable for activation/confirmation via email
A hex-encoded random UUID has plenty of entropy to be secure enough for our
needs.
"""
return uuid.uuid4().hex
def expiration_date():
"""
AuthToken objects expire 1 hour after creation by default
"""
return timezone.now() + timedelta(hours=1)
def new_app_password(size=6):
f = open(WORDLIST_FILE, 'r')
words = []
for i in range(size):
words.append(next(f).strip())
for num,line in enumerate(f):
j = random.randrange(size+num)
if j < size:
words[j] = line.strip()
return ' '.join(words)
|
551d86f64e1dadf54a4c63b633af6523dd5cdc05
|
urbansim/utils/logutil.py
|
urbansim/utils/logutil.py
|
import contextlib
import logging
@contextlib.contextmanager
def log_start_finish(msg, logger, level=logging.DEBUG):
"""
A context manager to log messages with "start: " and "finish: "
prefixes before and after a block.
Parameters
----------
msg : str
Will be prefixed with "start: " and "finish: ".
logger : logging.Logger
level : int, optional
Level at which to log, passed to ``logger.log``.
"""
logger.log(level, 'start: ' + msg)
yield
logger.log(level, 'finish: ' + msg)
|
import contextlib
import logging
US_LOG_FMT = ('%(asctime)s|%(levelname)s|%(name)s|'
'%(funcName)s|%(filename)s|%(lineno)s|%(message)s')
US_LOG_DATE_FMT = '%Y-%m-%d %H:%M:%S'
US_FMT = logging.Formatter(fmt=US_LOG_FMT, datefmt=US_LOG_DATE_FMT)
@contextlib.contextmanager
def log_start_finish(msg, logger, level=logging.DEBUG):
"""
A context manager to log messages with "start: " and "finish: "
prefixes before and after a block.
Parameters
----------
msg : str
Will be prefixed with "start: " and "finish: ".
logger : logging.Logger
level : int, optional
Level at which to log, passed to ``logger.log``.
"""
logger.log(level, 'start: ' + msg)
yield
logger.log(level, 'finish: ' + msg)
def set_log_level(level):
"""
Set the logging level for urbansim.
Parameters
----------
level : int
A supporting logging level. Use logging constants like logging.DEBUG.
"""
logging.getLogger('urbansim').setLevel(level)
def log_to_stream(level=None):
"""
Send log messages to the console.
"""
handler = logging.StreamHandler()
handler.setFormatter(US_FMT)
if level is not None:
handler.setLevel(level)
logger = logging.getLogger()
logger.addHandler(handler)
def log_to_file(filename, level=None):
"""
Send log output to the given file.
Parameters
----------
filename : str
level : int, optional
Optional logging level for the file handler.
"""
handler = logging.FileHandler(filename)
handler.setFormatter(US_FMT)
if level is not None:
handler.setLevel(level)
logger = logging.getLogger()
logger.addHandler(handler)
|
Add utilities for controlling urbansim logging.
|
Add utilities for controlling urbansim logging.
|
Python
|
bsd-3-clause
|
UDST/urbansim,waddell/urbansim,waddell/urbansim,ual/urbansim,UDST/urbansim,VladimirTyrin/urbansim,waddell/urbansim,ual/urbansim,AZMAG/urbansim,SANDAG/urbansim,synthicity/urbansim,SANDAG/urbansim,bricegnichols/urbansim,synthicity/urbansim,SANDAG/urbansim,AZMAG/urbansim,synthicity/urbansim,apdjustino/urbansim,ual/urbansim,bricegnichols/urbansim,AZMAG/urbansim,apdjustino/urbansim,apdjustino/urbansim,ual/urbansim,SANDAG/urbansim,synthicity/urbansim,bricegnichols/urbansim,VladimirTyrin/urbansim,AZMAG/urbansim,VladimirTyrin/urbansim,waddell/urbansim,VladimirTyrin/urbansim,bricegnichols/urbansim,UDST/urbansim,apdjustino/urbansim,UDST/urbansim
|
import contextlib
import logging
@contextlib.contextmanager
def log_start_finish(msg, logger, level=logging.DEBUG):
"""
A context manager to log messages with "start: " and "finish: "
prefixes before and after a block.
Parameters
----------
msg : str
Will be prefixed with "start: " and "finish: ".
logger : logging.Logger
level : int, optional
Level at which to log, passed to ``logger.log``.
"""
logger.log(level, 'start: ' + msg)
yield
logger.log(level, 'finish: ' + msg)
Add utilities for controlling urbansim logging.
|
import contextlib
import logging
US_LOG_FMT = ('%(asctime)s|%(levelname)s|%(name)s|'
'%(funcName)s|%(filename)s|%(lineno)s|%(message)s')
US_LOG_DATE_FMT = '%Y-%m-%d %H:%M:%S'
US_FMT = logging.Formatter(fmt=US_LOG_FMT, datefmt=US_LOG_DATE_FMT)
@contextlib.contextmanager
def log_start_finish(msg, logger, level=logging.DEBUG):
"""
A context manager to log messages with "start: " and "finish: "
prefixes before and after a block.
Parameters
----------
msg : str
Will be prefixed with "start: " and "finish: ".
logger : logging.Logger
level : int, optional
Level at which to log, passed to ``logger.log``.
"""
logger.log(level, 'start: ' + msg)
yield
logger.log(level, 'finish: ' + msg)
def set_log_level(level):
"""
Set the logging level for urbansim.
Parameters
----------
level : int
A supporting logging level. Use logging constants like logging.DEBUG.
"""
logging.getLogger('urbansim').setLevel(level)
def log_to_stream(level=None):
"""
Send log messages to the console.
"""
handler = logging.StreamHandler()
handler.setFormatter(US_FMT)
if level is not None:
handler.setLevel(level)
logger = logging.getLogger()
logger.addHandler(handler)
def log_to_file(filename, level=None):
"""
Send log output to the given file.
Parameters
----------
filename : str
level : int, optional
Optional logging level for the file handler.
"""
handler = logging.FileHandler(filename)
handler.setFormatter(US_FMT)
if level is not None:
handler.setLevel(level)
logger = logging.getLogger()
logger.addHandler(handler)
|
<commit_before>import contextlib
import logging
@contextlib.contextmanager
def log_start_finish(msg, logger, level=logging.DEBUG):
"""
A context manager to log messages with "start: " and "finish: "
prefixes before and after a block.
Parameters
----------
msg : str
Will be prefixed with "start: " and "finish: ".
logger : logging.Logger
level : int, optional
Level at which to log, passed to ``logger.log``.
"""
logger.log(level, 'start: ' + msg)
yield
logger.log(level, 'finish: ' + msg)
<commit_msg>Add utilities for controlling urbansim logging.<commit_after>
|
import contextlib
import logging
US_LOG_FMT = ('%(asctime)s|%(levelname)s|%(name)s|'
'%(funcName)s|%(filename)s|%(lineno)s|%(message)s')
US_LOG_DATE_FMT = '%Y-%m-%d %H:%M:%S'
US_FMT = logging.Formatter(fmt=US_LOG_FMT, datefmt=US_LOG_DATE_FMT)
@contextlib.contextmanager
def log_start_finish(msg, logger, level=logging.DEBUG):
"""
A context manager to log messages with "start: " and "finish: "
prefixes before and after a block.
Parameters
----------
msg : str
Will be prefixed with "start: " and "finish: ".
logger : logging.Logger
level : int, optional
Level at which to log, passed to ``logger.log``.
"""
logger.log(level, 'start: ' + msg)
yield
logger.log(level, 'finish: ' + msg)
def set_log_level(level):
"""
Set the logging level for urbansim.
Parameters
----------
level : int
A supporting logging level. Use logging constants like logging.DEBUG.
"""
logging.getLogger('urbansim').setLevel(level)
def log_to_stream(level=None):
"""
Send log messages to the console.
"""
handler = logging.StreamHandler()
handler.setFormatter(US_FMT)
if level is not None:
handler.setLevel(level)
logger = logging.getLogger()
logger.addHandler(handler)
def log_to_file(filename, level=None):
"""
Send log output to the given file.
Parameters
----------
filename : str
level : int, optional
Optional logging level for the file handler.
"""
handler = logging.FileHandler(filename)
handler.setFormatter(US_FMT)
if level is not None:
handler.setLevel(level)
logger = logging.getLogger()
logger.addHandler(handler)
|
import contextlib
import logging
@contextlib.contextmanager
def log_start_finish(msg, logger, level=logging.DEBUG):
"""
A context manager to log messages with "start: " and "finish: "
prefixes before and after a block.
Parameters
----------
msg : str
Will be prefixed with "start: " and "finish: ".
logger : logging.Logger
level : int, optional
Level at which to log, passed to ``logger.log``.
"""
logger.log(level, 'start: ' + msg)
yield
logger.log(level, 'finish: ' + msg)
Add utilities for controlling urbansim logging.import contextlib
import logging
US_LOG_FMT = ('%(asctime)s|%(levelname)s|%(name)s|'
'%(funcName)s|%(filename)s|%(lineno)s|%(message)s')
US_LOG_DATE_FMT = '%Y-%m-%d %H:%M:%S'
US_FMT = logging.Formatter(fmt=US_LOG_FMT, datefmt=US_LOG_DATE_FMT)
@contextlib.contextmanager
def log_start_finish(msg, logger, level=logging.DEBUG):
"""
A context manager to log messages with "start: " and "finish: "
prefixes before and after a block.
Parameters
----------
msg : str
Will be prefixed with "start: " and "finish: ".
logger : logging.Logger
level : int, optional
Level at which to log, passed to ``logger.log``.
"""
logger.log(level, 'start: ' + msg)
yield
logger.log(level, 'finish: ' + msg)
def set_log_level(level):
"""
Set the logging level for urbansim.
Parameters
----------
level : int
A supporting logging level. Use logging constants like logging.DEBUG.
"""
logging.getLogger('urbansim').setLevel(level)
def log_to_stream(level=None):
"""
Send log messages to the console.
"""
handler = logging.StreamHandler()
handler.setFormatter(US_FMT)
if level is not None:
handler.setLevel(level)
logger = logging.getLogger()
logger.addHandler(handler)
def log_to_file(filename, level=None):
"""
Send log output to the given file.
Parameters
----------
filename : str
level : int, optional
Optional logging level for the file handler.
"""
handler = logging.FileHandler(filename)
handler.setFormatter(US_FMT)
if level is not None:
handler.setLevel(level)
logger = logging.getLogger()
logger.addHandler(handler)
|
<commit_before>import contextlib
import logging
@contextlib.contextmanager
def log_start_finish(msg, logger, level=logging.DEBUG):
"""
A context manager to log messages with "start: " and "finish: "
prefixes before and after a block.
Parameters
----------
msg : str
Will be prefixed with "start: " and "finish: ".
logger : logging.Logger
level : int, optional
Level at which to log, passed to ``logger.log``.
"""
logger.log(level, 'start: ' + msg)
yield
logger.log(level, 'finish: ' + msg)
<commit_msg>Add utilities for controlling urbansim logging.<commit_after>import contextlib
import logging
US_LOG_FMT = ('%(asctime)s|%(levelname)s|%(name)s|'
'%(funcName)s|%(filename)s|%(lineno)s|%(message)s')
US_LOG_DATE_FMT = '%Y-%m-%d %H:%M:%S'
US_FMT = logging.Formatter(fmt=US_LOG_FMT, datefmt=US_LOG_DATE_FMT)
@contextlib.contextmanager
def log_start_finish(msg, logger, level=logging.DEBUG):
"""
A context manager to log messages with "start: " and "finish: "
prefixes before and after a block.
Parameters
----------
msg : str
Will be prefixed with "start: " and "finish: ".
logger : logging.Logger
level : int, optional
Level at which to log, passed to ``logger.log``.
"""
logger.log(level, 'start: ' + msg)
yield
logger.log(level, 'finish: ' + msg)
def set_log_level(level):
"""
Set the logging level for urbansim.
Parameters
----------
level : int
A supporting logging level. Use logging constants like logging.DEBUG.
"""
logging.getLogger('urbansim').setLevel(level)
def log_to_stream(level=None):
"""
Send log messages to the console.
"""
handler = logging.StreamHandler()
handler.setFormatter(US_FMT)
if level is not None:
handler.setLevel(level)
logger = logging.getLogger()
logger.addHandler(handler)
def log_to_file(filename, level=None):
"""
Send log output to the given file.
Parameters
----------
filename : str
level : int, optional
Optional logging level for the file handler.
"""
handler = logging.FileHandler(filename)
handler.setFormatter(US_FMT)
if level is not None:
handler.setLevel(level)
logger = logging.getLogger()
logger.addHandler(handler)
|
57b375d7bab3b88137b2ef5d6b0c38056b758a48
|
Mscthesis/IO/municipios_parser.py
|
Mscthesis/IO/municipios_parser.py
|
"""
municipios_parser
-----------------
Module which contains the process of parsing data of municipios.
TODO
----
"""
import pandas as pd
from pythonUtils.ProcessTools import Processer
class Municipios_Parser(Processer):
"""This class is the one which controls the parsing process of municipios
information.
"""
indices = None
files = ''
def __init__(self, logfile, bool_inform=False):
"Instantiation of the class remembering it is a subclass of Processer."
self.proc_name = "Municipios parser"
self.proc_desc = "Parser the standarize data from file"
self.subproc_desc = []
self.t_expended_subproc = []
self.logfile = logfile
def parse(self, filepath):
"Parse the data from the file given."
data = pd.read_csv(filepath, sep=';', index_col=0)
typevars = {}
typevars['pop_vars'] = ['Poblacion', "Superficie", "Densidad"]
typevars['loc_vars'] = ["longitud", "latitud"]
return data, typevars
|
"""
municipios_parser
-----------------
Module which contains the process of parsing data of municipios.
TODO
----
"""
import pandas as pd
from pythonUtils.ProcessTools import Processer
class Municipios_Parser(Processer):
"""This class is the one which controls the parsing process of municipios
information.
"""
indices = None
files = ''
def __init__(self, logfile, bool_inform=False):
"Instantiation of the class remembering it is a subclass of Processer."
self.proc_name = "Municipios parser"
self.proc_desc = "Parser the standarize data from file"
self.subproc_desc = []
self.t_expended_subproc = []
self.logfile = logfile
def parse(self, filepath):
"Parse the data from the file given."
data = pd.read_csv(filepath, sep=';', index_col=0)
typevars = {}
typevars['feat_vars'] = ['Poblacion', "Superficie", "Densidad"]
typevars['loc_vars'] = ["longitud", "latitud"]
return data, typevars
|
Change in the typ output.
|
Change in the typ output.
|
Python
|
mit
|
tgquintela/Mscthesis
|
"""
municipios_parser
-----------------
Module which contains the process of parsing data of municipios.
TODO
----
"""
import pandas as pd
from pythonUtils.ProcessTools import Processer
class Municipios_Parser(Processer):
"""This class is the one which controls the parsing process of municipios
information.
"""
indices = None
files = ''
def __init__(self, logfile, bool_inform=False):
"Instantiation of the class remembering it is a subclass of Processer."
self.proc_name = "Municipios parser"
self.proc_desc = "Parser the standarize data from file"
self.subproc_desc = []
self.t_expended_subproc = []
self.logfile = logfile
def parse(self, filepath):
"Parse the data from the file given."
data = pd.read_csv(filepath, sep=';', index_col=0)
typevars = {}
typevars['pop_vars'] = ['Poblacion', "Superficie", "Densidad"]
typevars['loc_vars'] = ["longitud", "latitud"]
return data, typevars
Change in the typ output.
|
"""
municipios_parser
-----------------
Module which contains the process of parsing data of municipios.
TODO
----
"""
import pandas as pd
from pythonUtils.ProcessTools import Processer
class Municipios_Parser(Processer):
"""This class is the one which controls the parsing process of municipios
information.
"""
indices = None
files = ''
def __init__(self, logfile, bool_inform=False):
"Instantiation of the class remembering it is a subclass of Processer."
self.proc_name = "Municipios parser"
self.proc_desc = "Parser the standarize data from file"
self.subproc_desc = []
self.t_expended_subproc = []
self.logfile = logfile
def parse(self, filepath):
"Parse the data from the file given."
data = pd.read_csv(filepath, sep=';', index_col=0)
typevars = {}
typevars['feat_vars'] = ['Poblacion', "Superficie", "Densidad"]
typevars['loc_vars'] = ["longitud", "latitud"]
return data, typevars
|
<commit_before>
"""
municipios_parser
-----------------
Module which contains the process of parsing data of municipios.
TODO
----
"""
import pandas as pd
from pythonUtils.ProcessTools import Processer
class Municipios_Parser(Processer):
"""This class is the one which controls the parsing process of municipios
information.
"""
indices = None
files = ''
def __init__(self, logfile, bool_inform=False):
"Instantiation of the class remembering it is a subclass of Processer."
self.proc_name = "Municipios parser"
self.proc_desc = "Parser the standarize data from file"
self.subproc_desc = []
self.t_expended_subproc = []
self.logfile = logfile
def parse(self, filepath):
"Parse the data from the file given."
data = pd.read_csv(filepath, sep=';', index_col=0)
typevars = {}
typevars['pop_vars'] = ['Poblacion', "Superficie", "Densidad"]
typevars['loc_vars'] = ["longitud", "latitud"]
return data, typevars
<commit_msg>Change in the typ output.<commit_after>
|
"""
municipios_parser
-----------------
Module which contains the process of parsing data of municipios.
TODO
----
"""
import pandas as pd
from pythonUtils.ProcessTools import Processer
class Municipios_Parser(Processer):
"""This class is the one which controls the parsing process of municipios
information.
"""
indices = None
files = ''
def __init__(self, logfile, bool_inform=False):
"Instantiation of the class remembering it is a subclass of Processer."
self.proc_name = "Municipios parser"
self.proc_desc = "Parser the standarize data from file"
self.subproc_desc = []
self.t_expended_subproc = []
self.logfile = logfile
def parse(self, filepath):
"Parse the data from the file given."
data = pd.read_csv(filepath, sep=';', index_col=0)
typevars = {}
typevars['feat_vars'] = ['Poblacion', "Superficie", "Densidad"]
typevars['loc_vars'] = ["longitud", "latitud"]
return data, typevars
|
"""
municipios_parser
-----------------
Module which contains the process of parsing data of municipios.
TODO
----
"""
import pandas as pd
from pythonUtils.ProcessTools import Processer
class Municipios_Parser(Processer):
"""This class is the one which controls the parsing process of municipios
information.
"""
indices = None
files = ''
def __init__(self, logfile, bool_inform=False):
"Instantiation of the class remembering it is a subclass of Processer."
self.proc_name = "Municipios parser"
self.proc_desc = "Parser the standarize data from file"
self.subproc_desc = []
self.t_expended_subproc = []
self.logfile = logfile
def parse(self, filepath):
"Parse the data from the file given."
data = pd.read_csv(filepath, sep=';', index_col=0)
typevars = {}
typevars['pop_vars'] = ['Poblacion', "Superficie", "Densidad"]
typevars['loc_vars'] = ["longitud", "latitud"]
return data, typevars
Change in the typ output.
"""
municipios_parser
-----------------
Module which contains the process of parsing data of municipios.
TODO
----
"""
import pandas as pd
from pythonUtils.ProcessTools import Processer
class Municipios_Parser(Processer):
"""This class is the one which controls the parsing process of municipios
information.
"""
indices = None
files = ''
def __init__(self, logfile, bool_inform=False):
"Instantiation of the class remembering it is a subclass of Processer."
self.proc_name = "Municipios parser"
self.proc_desc = "Parser the standarize data from file"
self.subproc_desc = []
self.t_expended_subproc = []
self.logfile = logfile
def parse(self, filepath):
"Parse the data from the file given."
data = pd.read_csv(filepath, sep=';', index_col=0)
typevars = {}
typevars['feat_vars'] = ['Poblacion', "Superficie", "Densidad"]
typevars['loc_vars'] = ["longitud", "latitud"]
return data, typevars
|
<commit_before>
"""
municipios_parser
-----------------
Module which contains the process of parsing data of municipios.
TODO
----
"""
import pandas as pd
from pythonUtils.ProcessTools import Processer
class Municipios_Parser(Processer):
"""This class is the one which controls the parsing process of municipios
information.
"""
indices = None
files = ''
def __init__(self, logfile, bool_inform=False):
"Instantiation of the class remembering it is a subclass of Processer."
self.proc_name = "Municipios parser"
self.proc_desc = "Parser the standarize data from file"
self.subproc_desc = []
self.t_expended_subproc = []
self.logfile = logfile
def parse(self, filepath):
"Parse the data from the file given."
data = pd.read_csv(filepath, sep=';', index_col=0)
typevars = {}
typevars['pop_vars'] = ['Poblacion', "Superficie", "Densidad"]
typevars['loc_vars'] = ["longitud", "latitud"]
return data, typevars
<commit_msg>Change in the typ output.<commit_after>
"""
municipios_parser
-----------------
Module which contains the process of parsing data of municipios.
TODO
----
"""
import pandas as pd
from pythonUtils.ProcessTools import Processer
class Municipios_Parser(Processer):
"""This class is the one which controls the parsing process of municipios
information.
"""
indices = None
files = ''
def __init__(self, logfile, bool_inform=False):
"Instantiation of the class remembering it is a subclass of Processer."
self.proc_name = "Municipios parser"
self.proc_desc = "Parser the standarize data from file"
self.subproc_desc = []
self.t_expended_subproc = []
self.logfile = logfile
def parse(self, filepath):
"Parse the data from the file given."
data = pd.read_csv(filepath, sep=';', index_col=0)
typevars = {}
typevars['feat_vars'] = ['Poblacion', "Superficie", "Densidad"]
typevars['loc_vars'] = ["longitud", "latitud"]
return data, typevars
|
4854015a61f0b582065b0d5561df231314abcce1
|
django_redux_generator/management/commands/redux_generator.py
|
django_redux_generator/management/commands/redux_generator.py
|
from django.core.management.base import BaseCommand
from django.template.loader import render_to_string
class Command(BaseCommand):
help = 'Generate redux boilerplate'
def add_arguments(self, parser):
parser.add_argument('action_name', type=str)
parser.add_argument('--thunk',
action='store_true',
dest='thunk',
default=False,
help='Generate a redux thunk')
def handle(self, *args, **options):
if options['thunk']:
template_name = 'django_redux_generator/thunk_fetch.js'
else:
template_name = 'django_redux_generator/action_creator.js'
print(render_to_string(template_name, {
'action_name': options['action_name'],
}))
|
from django.core.management.base import BaseCommand
from django.template.loader import render_to_string
class Command(BaseCommand):
help = 'Generate redux boilerplate'
def add_arguments(self, parser):
parser.add_argument('action_name', type=str)
parser.add_argument('--thunk',
action='store_true',
dest='thunk',
default=False,
help='Generate a redux thunk')
def handle(self, *args, **options):
if options['thunk']:
template_name = 'django_redux_generator/thunk_fetch.js'
else:
template_name = 'django_redux_generator/action_creator.js'
return render_to_string(template_name, {
'action_name': options['action_name'],
})
|
Return the output rather than print
|
Return the output rather than print
|
Python
|
mit
|
rapilabs/django-redux-generator,rapilabs/django-redux-generator
|
from django.core.management.base import BaseCommand
from django.template.loader import render_to_string
class Command(BaseCommand):
help = 'Generate redux boilerplate'
def add_arguments(self, parser):
parser.add_argument('action_name', type=str)
parser.add_argument('--thunk',
action='store_true',
dest='thunk',
default=False,
help='Generate a redux thunk')
def handle(self, *args, **options):
if options['thunk']:
template_name = 'django_redux_generator/thunk_fetch.js'
else:
template_name = 'django_redux_generator/action_creator.js'
print(render_to_string(template_name, {
'action_name': options['action_name'],
}))
Return the output rather than print
|
from django.core.management.base import BaseCommand
from django.template.loader import render_to_string
class Command(BaseCommand):
help = 'Generate redux boilerplate'
def add_arguments(self, parser):
parser.add_argument('action_name', type=str)
parser.add_argument('--thunk',
action='store_true',
dest='thunk',
default=False,
help='Generate a redux thunk')
def handle(self, *args, **options):
if options['thunk']:
template_name = 'django_redux_generator/thunk_fetch.js'
else:
template_name = 'django_redux_generator/action_creator.js'
return render_to_string(template_name, {
'action_name': options['action_name'],
})
|
<commit_before>from django.core.management.base import BaseCommand
from django.template.loader import render_to_string
class Command(BaseCommand):
help = 'Generate redux boilerplate'
def add_arguments(self, parser):
parser.add_argument('action_name', type=str)
parser.add_argument('--thunk',
action='store_true',
dest='thunk',
default=False,
help='Generate a redux thunk')
def handle(self, *args, **options):
if options['thunk']:
template_name = 'django_redux_generator/thunk_fetch.js'
else:
template_name = 'django_redux_generator/action_creator.js'
print(render_to_string(template_name, {
'action_name': options['action_name'],
}))
<commit_msg>Return the output rather than print<commit_after>
|
from django.core.management.base import BaseCommand
from django.template.loader import render_to_string
class Command(BaseCommand):
help = 'Generate redux boilerplate'
def add_arguments(self, parser):
parser.add_argument('action_name', type=str)
parser.add_argument('--thunk',
action='store_true',
dest='thunk',
default=False,
help='Generate a redux thunk')
def handle(self, *args, **options):
if options['thunk']:
template_name = 'django_redux_generator/thunk_fetch.js'
else:
template_name = 'django_redux_generator/action_creator.js'
return render_to_string(template_name, {
'action_name': options['action_name'],
})
|
from django.core.management.base import BaseCommand
from django.template.loader import render_to_string
class Command(BaseCommand):
help = 'Generate redux boilerplate'
def add_arguments(self, parser):
parser.add_argument('action_name', type=str)
parser.add_argument('--thunk',
action='store_true',
dest='thunk',
default=False,
help='Generate a redux thunk')
def handle(self, *args, **options):
if options['thunk']:
template_name = 'django_redux_generator/thunk_fetch.js'
else:
template_name = 'django_redux_generator/action_creator.js'
print(render_to_string(template_name, {
'action_name': options['action_name'],
}))
Return the output rather than printfrom django.core.management.base import BaseCommand
from django.template.loader import render_to_string
class Command(BaseCommand):
help = 'Generate redux boilerplate'
def add_arguments(self, parser):
parser.add_argument('action_name', type=str)
parser.add_argument('--thunk',
action='store_true',
dest='thunk',
default=False,
help='Generate a redux thunk')
def handle(self, *args, **options):
if options['thunk']:
template_name = 'django_redux_generator/thunk_fetch.js'
else:
template_name = 'django_redux_generator/action_creator.js'
return render_to_string(template_name, {
'action_name': options['action_name'],
})
|
<commit_before>from django.core.management.base import BaseCommand
from django.template.loader import render_to_string
class Command(BaseCommand):
help = 'Generate redux boilerplate'
def add_arguments(self, parser):
parser.add_argument('action_name', type=str)
parser.add_argument('--thunk',
action='store_true',
dest='thunk',
default=False,
help='Generate a redux thunk')
def handle(self, *args, **options):
if options['thunk']:
template_name = 'django_redux_generator/thunk_fetch.js'
else:
template_name = 'django_redux_generator/action_creator.js'
print(render_to_string(template_name, {
'action_name': options['action_name'],
}))
<commit_msg>Return the output rather than print<commit_after>from django.core.management.base import BaseCommand
from django.template.loader import render_to_string
class Command(BaseCommand):
help = 'Generate redux boilerplate'
def add_arguments(self, parser):
parser.add_argument('action_name', type=str)
parser.add_argument('--thunk',
action='store_true',
dest='thunk',
default=False,
help='Generate a redux thunk')
def handle(self, *args, **options):
if options['thunk']:
template_name = 'django_redux_generator/thunk_fetch.js'
else:
template_name = 'django_redux_generator/action_creator.js'
return render_to_string(template_name, {
'action_name': options['action_name'],
})
|
d8ba1531b2e0faa71c57e8970af471ec2caa4a18
|
en-2014-06-21-unit-testing-with-unittest-mock-patch/chdir2.py
|
en-2014-06-21-unit-testing-with-unittest-mock-patch/chdir2.py
|
"""
chdir2
~~~~~~
An alternative implementation of :func:`chdir.chdir`.
:copyright: © 2014 by Petr Zemek <s3rvac@gmail.com>
:license: BSD, see LICENSE for more details
"""
import os
class chdir2():
"""An alternative implementation of :func:`chdir.chdir`."""
def __init__(self, dir):
self.dir = dir
def __enter__(self):
self.orig_cwd = os.getcwd()
os.chdir(self.dir)
def __exit__(self, *exc_info):
os.chdir(self.orig_cwd)
|
"""
chdir2
~~~~~~
An alternative implementation of :func:`chdir.chdir()`.
:copyright: © 2014 by Petr Zemek <s3rvac@gmail.com>
:license: BSD, see LICENSE for more details
"""
import os
class chdir2():
"""An alternative implementation of :func:`chdir.chdir()`."""
def __init__(self, dir):
self.dir = dir
def __enter__(self):
self.orig_cwd = os.getcwd()
os.chdir(self.dir)
def __exit__(self, *exc_info):
os.chdir(self.orig_cwd)
|
Add missing parentheses after 'chdir'.
|
blog/en-2014-06-21: Add missing parentheses after 'chdir'.
It is a function, so we better add parentheses to make this clearer.
|
Python
|
bsd-3-clause
|
s3rvac/blog,s3rvac/blog,s3rvac/blog,s3rvac/blog
|
"""
chdir2
~~~~~~
An alternative implementation of :func:`chdir.chdir`.
:copyright: © 2014 by Petr Zemek <s3rvac@gmail.com>
:license: BSD, see LICENSE for more details
"""
import os
class chdir2():
"""An alternative implementation of :func:`chdir.chdir`."""
def __init__(self, dir):
self.dir = dir
def __enter__(self):
self.orig_cwd = os.getcwd()
os.chdir(self.dir)
def __exit__(self, *exc_info):
os.chdir(self.orig_cwd)
blog/en-2014-06-21: Add missing parentheses after 'chdir'.
It is a function, so we better add parentheses to make this clearer.
|
"""
chdir2
~~~~~~
An alternative implementation of :func:`chdir.chdir()`.
:copyright: © 2014 by Petr Zemek <s3rvac@gmail.com>
:license: BSD, see LICENSE for more details
"""
import os
class chdir2():
"""An alternative implementation of :func:`chdir.chdir()`."""
def __init__(self, dir):
self.dir = dir
def __enter__(self):
self.orig_cwd = os.getcwd()
os.chdir(self.dir)
def __exit__(self, *exc_info):
os.chdir(self.orig_cwd)
|
<commit_before>"""
chdir2
~~~~~~
An alternative implementation of :func:`chdir.chdir`.
:copyright: © 2014 by Petr Zemek <s3rvac@gmail.com>
:license: BSD, see LICENSE for more details
"""
import os
class chdir2():
"""An alternative implementation of :func:`chdir.chdir`."""
def __init__(self, dir):
self.dir = dir
def __enter__(self):
self.orig_cwd = os.getcwd()
os.chdir(self.dir)
def __exit__(self, *exc_info):
os.chdir(self.orig_cwd)
<commit_msg>blog/en-2014-06-21: Add missing parentheses after 'chdir'.
It is a function, so we better add parentheses to make this clearer.<commit_after>
|
"""
chdir2
~~~~~~
An alternative implementation of :func:`chdir.chdir()`.
:copyright: © 2014 by Petr Zemek <s3rvac@gmail.com>
:license: BSD, see LICENSE for more details
"""
import os
class chdir2():
"""An alternative implementation of :func:`chdir.chdir()`."""
def __init__(self, dir):
self.dir = dir
def __enter__(self):
self.orig_cwd = os.getcwd()
os.chdir(self.dir)
def __exit__(self, *exc_info):
os.chdir(self.orig_cwd)
|
"""
chdir2
~~~~~~
An alternative implementation of :func:`chdir.chdir`.
:copyright: © 2014 by Petr Zemek <s3rvac@gmail.com>
:license: BSD, see LICENSE for more details
"""
import os
class chdir2():
"""An alternative implementation of :func:`chdir.chdir`."""
def __init__(self, dir):
self.dir = dir
def __enter__(self):
self.orig_cwd = os.getcwd()
os.chdir(self.dir)
def __exit__(self, *exc_info):
os.chdir(self.orig_cwd)
blog/en-2014-06-21: Add missing parentheses after 'chdir'.
It is a function, so we better add parentheses to make this clearer."""
chdir2
~~~~~~
An alternative implementation of :func:`chdir.chdir()`.
:copyright: © 2014 by Petr Zemek <s3rvac@gmail.com>
:license: BSD, see LICENSE for more details
"""
import os
class chdir2():
"""An alternative implementation of :func:`chdir.chdir()`."""
def __init__(self, dir):
self.dir = dir
def __enter__(self):
self.orig_cwd = os.getcwd()
os.chdir(self.dir)
def __exit__(self, *exc_info):
os.chdir(self.orig_cwd)
|
<commit_before>"""
chdir2
~~~~~~
An alternative implementation of :func:`chdir.chdir`.
:copyright: © 2014 by Petr Zemek <s3rvac@gmail.com>
:license: BSD, see LICENSE for more details
"""
import os
class chdir2():
"""An alternative implementation of :func:`chdir.chdir`."""
def __init__(self, dir):
self.dir = dir
def __enter__(self):
self.orig_cwd = os.getcwd()
os.chdir(self.dir)
def __exit__(self, *exc_info):
os.chdir(self.orig_cwd)
<commit_msg>blog/en-2014-06-21: Add missing parentheses after 'chdir'.
It is a function, so we better add parentheses to make this clearer.<commit_after>"""
chdir2
~~~~~~
An alternative implementation of :func:`chdir.chdir()`.
:copyright: © 2014 by Petr Zemek <s3rvac@gmail.com>
:license: BSD, see LICENSE for more details
"""
import os
class chdir2():
"""An alternative implementation of :func:`chdir.chdir()`."""
def __init__(self, dir):
self.dir = dir
def __enter__(self):
self.orig_cwd = os.getcwd()
os.chdir(self.dir)
def __exit__(self, *exc_info):
os.chdir(self.orig_cwd)
|
f59919efefb78fffff564ec17c55f6df644e8d7e
|
server/lib/python/cartodb_services/cartodb_services/here/__init__.py
|
server/lib/python/cartodb_services/cartodb_services/here/__init__.py
|
from cartodb_services.here.geocoder import HereMapsGeocoder
from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder
from cartodb_services.here.routing import HereMapsRoutingIsoline
|
from cartodb_services.here.geocoder import HereMapsGeocoder, HereMapsGeocoderV7
from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder, HereMapsBulkGeocoderV7
from cartodb_services.here.service_factory import get_geocoder, get_bulk_geocoder, get_routing_isoline
from cartodb_services.here.routing import HereMapsRoutingIsoline
|
Add new imports for apikey parameter support
|
Add new imports for apikey parameter support
|
Python
|
bsd-3-clause
|
CartoDB/dataservices-api,CartoDB/dataservices-api,CartoDB/dataservices-api,CartoDB/dataservices-api,CartoDB/geocoder-api,CartoDB/geocoder-api,CartoDB/geocoder-api,CartoDB/geocoder-api
|
from cartodb_services.here.geocoder import HereMapsGeocoder
from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder
from cartodb_services.here.routing import HereMapsRoutingIsoline
Add new imports for apikey parameter support
|
from cartodb_services.here.geocoder import HereMapsGeocoder, HereMapsGeocoderV7
from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder, HereMapsBulkGeocoderV7
from cartodb_services.here.service_factory import get_geocoder, get_bulk_geocoder, get_routing_isoline
from cartodb_services.here.routing import HereMapsRoutingIsoline
|
<commit_before>from cartodb_services.here.geocoder import HereMapsGeocoder
from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder
from cartodb_services.here.routing import HereMapsRoutingIsoline
<commit_msg>Add new imports for apikey parameter support<commit_after>
|
from cartodb_services.here.geocoder import HereMapsGeocoder, HereMapsGeocoderV7
from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder, HereMapsBulkGeocoderV7
from cartodb_services.here.service_factory import get_geocoder, get_bulk_geocoder, get_routing_isoline
from cartodb_services.here.routing import HereMapsRoutingIsoline
|
from cartodb_services.here.geocoder import HereMapsGeocoder
from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder
from cartodb_services.here.routing import HereMapsRoutingIsoline
Add new imports for apikey parameter supportfrom cartodb_services.here.geocoder import HereMapsGeocoder, HereMapsGeocoderV7
from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder, HereMapsBulkGeocoderV7
from cartodb_services.here.service_factory import get_geocoder, get_bulk_geocoder, get_routing_isoline
from cartodb_services.here.routing import HereMapsRoutingIsoline
|
<commit_before>from cartodb_services.here.geocoder import HereMapsGeocoder
from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder
from cartodb_services.here.routing import HereMapsRoutingIsoline
<commit_msg>Add new imports for apikey parameter support<commit_after>from cartodb_services.here.geocoder import HereMapsGeocoder, HereMapsGeocoderV7
from cartodb_services.here.bulk_geocoder import HereMapsBulkGeocoder, HereMapsBulkGeocoderV7
from cartodb_services.here.service_factory import get_geocoder, get_bulk_geocoder, get_routing_isoline
from cartodb_services.here.routing import HereMapsRoutingIsoline
|
a6405ccfc7f53f601088206c216c5167fd86359f
|
symposion/teams/backends.py
|
symposion/teams/backends.py
|
from django.db.models import Q
from .models import Team
class TeamPermissionsBackend(object):
def authenticate(self, username=None, password=None):
return None
def get_team_permissions(self, user_obj, obj=None):
"""
Returns a set of permission strings that this user has through his/her
team memberships.
"""
if user_obj.is_anonymous() or obj is not None:
return set()
if not hasattr(user_obj, "_team_perm_cache"):
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="manager") | Q(memberships__state="member"),
)
perms = memberships.values_list(
"permissions__content_type__app_label",
"permissions__codename"
).order_by()
user_obj._team_perm_cache = set(["%s.%s" % (ct, name) for ct, name in perms])
return user_obj._team_perm_cache
def has_perm(self, user_obj, perm, obj=None):
if not user_obj.is_active:
return False
return perm in self.get_team_permissions(user_obj, obj)
|
from django.db.models import Q
from .models import Team
class TeamPermissionsBackend(object):
def authenticate(self, username=None, password=None):
return None
def get_team_permissions(self, user_obj, obj=None):
"""
Returns a set of permission strings that this user has through his/her
team memberships.
"""
if user_obj.is_anonymous() or obj is not None:
return set()
if not hasattr(user_obj, "_team_perm_cache"):
# Member permissions
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="member"),
)
perms = memberships.values_list(
"permissions__content_type__app_label",
"permissions__codename"
).order_by()
permissions = ["%s.%s" % (ct, name) for ct, name in perms]
# Manager permissions
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="manager"),
)
perms = memberships.values_list(
"manager_permissions__content_type__app_label",
"manager_permissions__codename"
).order_by()
permissions += ["%s.%s" % (ct, name) for ct, name in perms]
user_obj._team_perm_cache = set(permissions)
return user_obj._team_perm_cache
def has_perm(self, user_obj, perm, obj=None):
if not user_obj.is_active:
return False
return perm in self.get_team_permissions(user_obj, obj)
|
Fix team permissions backend not pulling out manager_permissions
|
Fix team permissions backend not pulling out manager_permissions
Something like
request.user.has_perm('reviews.can_manage_%s' % proposal.kind.section.slug)
Will aways return false as the backend does a lookup of team membership
(member or manager) but only grabs the 'permissions' and not the
'manager_permissions' field
|
Python
|
bsd-3-clause
|
pyconau2017/symposion,pyconau2017/symposion
|
from django.db.models import Q
from .models import Team
class TeamPermissionsBackend(object):
def authenticate(self, username=None, password=None):
return None
def get_team_permissions(self, user_obj, obj=None):
"""
Returns a set of permission strings that this user has through his/her
team memberships.
"""
if user_obj.is_anonymous() or obj is not None:
return set()
if not hasattr(user_obj, "_team_perm_cache"):
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="manager") | Q(memberships__state="member"),
)
perms = memberships.values_list(
"permissions__content_type__app_label",
"permissions__codename"
).order_by()
user_obj._team_perm_cache = set(["%s.%s" % (ct, name) for ct, name in perms])
return user_obj._team_perm_cache
def has_perm(self, user_obj, perm, obj=None):
if not user_obj.is_active:
return False
return perm in self.get_team_permissions(user_obj, obj)
Fix team permissions backend not pulling out manager_permissions
Something like
request.user.has_perm('reviews.can_manage_%s' % proposal.kind.section.slug)
Will aways return false as the backend does a lookup of team membership
(member or manager) but only grabs the 'permissions' and not the
'manager_permissions' field
|
from django.db.models import Q
from .models import Team
class TeamPermissionsBackend(object):
def authenticate(self, username=None, password=None):
return None
def get_team_permissions(self, user_obj, obj=None):
"""
Returns a set of permission strings that this user has through his/her
team memberships.
"""
if user_obj.is_anonymous() or obj is not None:
return set()
if not hasattr(user_obj, "_team_perm_cache"):
# Member permissions
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="member"),
)
perms = memberships.values_list(
"permissions__content_type__app_label",
"permissions__codename"
).order_by()
permissions = ["%s.%s" % (ct, name) for ct, name in perms]
# Manager permissions
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="manager"),
)
perms = memberships.values_list(
"manager_permissions__content_type__app_label",
"manager_permissions__codename"
).order_by()
permissions += ["%s.%s" % (ct, name) for ct, name in perms]
user_obj._team_perm_cache = set(permissions)
return user_obj._team_perm_cache
def has_perm(self, user_obj, perm, obj=None):
if not user_obj.is_active:
return False
return perm in self.get_team_permissions(user_obj, obj)
|
<commit_before>from django.db.models import Q
from .models import Team
class TeamPermissionsBackend(object):
def authenticate(self, username=None, password=None):
return None
def get_team_permissions(self, user_obj, obj=None):
"""
Returns a set of permission strings that this user has through his/her
team memberships.
"""
if user_obj.is_anonymous() or obj is not None:
return set()
if not hasattr(user_obj, "_team_perm_cache"):
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="manager") | Q(memberships__state="member"),
)
perms = memberships.values_list(
"permissions__content_type__app_label",
"permissions__codename"
).order_by()
user_obj._team_perm_cache = set(["%s.%s" % (ct, name) for ct, name in perms])
return user_obj._team_perm_cache
def has_perm(self, user_obj, perm, obj=None):
if not user_obj.is_active:
return False
return perm in self.get_team_permissions(user_obj, obj)
<commit_msg>Fix team permissions backend not pulling out manager_permissions
Something like
request.user.has_perm('reviews.can_manage_%s' % proposal.kind.section.slug)
Will aways return false as the backend does a lookup of team membership
(member or manager) but only grabs the 'permissions' and not the
'manager_permissions' field<commit_after>
|
from django.db.models import Q
from .models import Team
class TeamPermissionsBackend(object):
def authenticate(self, username=None, password=None):
return None
def get_team_permissions(self, user_obj, obj=None):
"""
Returns a set of permission strings that this user has through his/her
team memberships.
"""
if user_obj.is_anonymous() or obj is not None:
return set()
if not hasattr(user_obj, "_team_perm_cache"):
# Member permissions
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="member"),
)
perms = memberships.values_list(
"permissions__content_type__app_label",
"permissions__codename"
).order_by()
permissions = ["%s.%s" % (ct, name) for ct, name in perms]
# Manager permissions
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="manager"),
)
perms = memberships.values_list(
"manager_permissions__content_type__app_label",
"manager_permissions__codename"
).order_by()
permissions += ["%s.%s" % (ct, name) for ct, name in perms]
user_obj._team_perm_cache = set(permissions)
return user_obj._team_perm_cache
def has_perm(self, user_obj, perm, obj=None):
if not user_obj.is_active:
return False
return perm in self.get_team_permissions(user_obj, obj)
|
from django.db.models import Q
from .models import Team
class TeamPermissionsBackend(object):
def authenticate(self, username=None, password=None):
return None
def get_team_permissions(self, user_obj, obj=None):
"""
Returns a set of permission strings that this user has through his/her
team memberships.
"""
if user_obj.is_anonymous() or obj is not None:
return set()
if not hasattr(user_obj, "_team_perm_cache"):
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="manager") | Q(memberships__state="member"),
)
perms = memberships.values_list(
"permissions__content_type__app_label",
"permissions__codename"
).order_by()
user_obj._team_perm_cache = set(["%s.%s" % (ct, name) for ct, name in perms])
return user_obj._team_perm_cache
def has_perm(self, user_obj, perm, obj=None):
if not user_obj.is_active:
return False
return perm in self.get_team_permissions(user_obj, obj)
Fix team permissions backend not pulling out manager_permissions
Something like
request.user.has_perm('reviews.can_manage_%s' % proposal.kind.section.slug)
Will aways return false as the backend does a lookup of team membership
(member or manager) but only grabs the 'permissions' and not the
'manager_permissions' fieldfrom django.db.models import Q
from .models import Team
class TeamPermissionsBackend(object):
def authenticate(self, username=None, password=None):
return None
def get_team_permissions(self, user_obj, obj=None):
"""
Returns a set of permission strings that this user has through his/her
team memberships.
"""
if user_obj.is_anonymous() or obj is not None:
return set()
if not hasattr(user_obj, "_team_perm_cache"):
# Member permissions
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="member"),
)
perms = memberships.values_list(
"permissions__content_type__app_label",
"permissions__codename"
).order_by()
permissions = ["%s.%s" % (ct, name) for ct, name in perms]
# Manager permissions
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="manager"),
)
perms = memberships.values_list(
"manager_permissions__content_type__app_label",
"manager_permissions__codename"
).order_by()
permissions += ["%s.%s" % (ct, name) for ct, name in perms]
user_obj._team_perm_cache = set(permissions)
return user_obj._team_perm_cache
def has_perm(self, user_obj, perm, obj=None):
if not user_obj.is_active:
return False
return perm in self.get_team_permissions(user_obj, obj)
|
<commit_before>from django.db.models import Q
from .models import Team
class TeamPermissionsBackend(object):
def authenticate(self, username=None, password=None):
return None
def get_team_permissions(self, user_obj, obj=None):
"""
Returns a set of permission strings that this user has through his/her
team memberships.
"""
if user_obj.is_anonymous() or obj is not None:
return set()
if not hasattr(user_obj, "_team_perm_cache"):
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="manager") | Q(memberships__state="member"),
)
perms = memberships.values_list(
"permissions__content_type__app_label",
"permissions__codename"
).order_by()
user_obj._team_perm_cache = set(["%s.%s" % (ct, name) for ct, name in perms])
return user_obj._team_perm_cache
def has_perm(self, user_obj, perm, obj=None):
if not user_obj.is_active:
return False
return perm in self.get_team_permissions(user_obj, obj)
<commit_msg>Fix team permissions backend not pulling out manager_permissions
Something like
request.user.has_perm('reviews.can_manage_%s' % proposal.kind.section.slug)
Will aways return false as the backend does a lookup of team membership
(member or manager) but only grabs the 'permissions' and not the
'manager_permissions' field<commit_after>from django.db.models import Q
from .models import Team
class TeamPermissionsBackend(object):
def authenticate(self, username=None, password=None):
return None
def get_team_permissions(self, user_obj, obj=None):
"""
Returns a set of permission strings that this user has through his/her
team memberships.
"""
if user_obj.is_anonymous() or obj is not None:
return set()
if not hasattr(user_obj, "_team_perm_cache"):
# Member permissions
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="member"),
)
perms = memberships.values_list(
"permissions__content_type__app_label",
"permissions__codename"
).order_by()
permissions = ["%s.%s" % (ct, name) for ct, name in perms]
# Manager permissions
memberships = Team.objects.filter(
Q(memberships__user=user_obj),
Q(memberships__state="manager"),
)
perms = memberships.values_list(
"manager_permissions__content_type__app_label",
"manager_permissions__codename"
).order_by()
permissions += ["%s.%s" % (ct, name) for ct, name in perms]
user_obj._team_perm_cache = set(permissions)
return user_obj._team_perm_cache
def has_perm(self, user_obj, perm, obj=None):
if not user_obj.is_active:
return False
return perm in self.get_team_permissions(user_obj, obj)
|
23072e882edb6da55cb12ef0591a786235249670
|
ome/__main__.py
|
ome/__main__.py
|
# ome - Object Message Expressions
# Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved.
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
if command_args.verbose:
print('ome: using target {}'.format(target.name))
print('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
if command_args.verbose:
print('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
|
# ome - Object Message Expressions
# Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved.
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def print_verbose(*args, **kwargs):
if command_args.verbose:
print(*args, **kwargs)
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
print_verbose('ome: using target {}'.format(target.name))
print_verbose('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
print_verbose('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
|
Use print_verbose for conditional printing.
|
Use print_verbose for conditional printing.
|
Python
|
mit
|
shaurz/ome,shaurz/ome
|
# ome - Object Message Expressions
# Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved.
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
if command_args.verbose:
print('ome: using target {}'.format(target.name))
print('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
if command_args.verbose:
print('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
Use print_verbose for conditional printing.
|
# ome - Object Message Expressions
# Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved.
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def print_verbose(*args, **kwargs):
if command_args.verbose:
print(*args, **kwargs)
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
print_verbose('ome: using target {}'.format(target.name))
print_verbose('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
print_verbose('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
|
<commit_before># ome - Object Message Expressions
# Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved.
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
if command_args.verbose:
print('ome: using target {}'.format(target.name))
print('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
if command_args.verbose:
print('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
<commit_msg>Use print_verbose for conditional printing.<commit_after>
|
# ome - Object Message Expressions
# Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved.
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def print_verbose(*args, **kwargs):
if command_args.verbose:
print(*args, **kwargs)
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
print_verbose('ome: using target {}'.format(target.name))
print_verbose('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
print_verbose('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
|
# ome - Object Message Expressions
# Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved.
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
if command_args.verbose:
print('ome: using target {}'.format(target.name))
print('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
if command_args.verbose:
print('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
Use print_verbose for conditional printing.# ome - Object Message Expressions
# Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved.
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def print_verbose(*args, **kwargs):
if command_args.verbose:
print(*args, **kwargs)
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
print_verbose('ome: using target {}'.format(target.name))
print_verbose('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
print_verbose('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
|
<commit_before># ome - Object Message Expressions
# Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved.
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
if command_args.verbose:
print('ome: using target {}'.format(target.name))
print('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
if command_args.verbose:
print('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
<commit_msg>Use print_verbose for conditional printing.<commit_after># ome - Object Message Expressions
# Copyright (c) 2015-2016 Luke McCarthy <luke@iogopro.co.uk>. All rights reserved.
import sys
from .command import command_args
from .error import OmeError
from .terminal import stderr
def print_verbose(*args, **kwargs):
if command_args.verbose:
print(*args, **kwargs)
def main():
stderr.reset()
try:
from . import compiler
target = compiler.get_target(command_args.target)
build_options = compiler.BuildOptions(target)
backend = compiler.get_backend(target, command_args.backend)
print_verbose('ome: using target {}'.format(target.name))
print_verbose('ome: using backend {} {}'.format(backend.name, backend.version))
for filename in command_args.filename:
print_verbose('ome: compiling {}'.format(filename))
if command_args.print_code:
print(compiler.compile_file(filename, target).decode(target.encoding))
else:
build_options.make_executable(filename, backend)
except OmeError as error:
error.write_ansi(stderr)
stderr.reset()
sys.exit(1)
if __name__ == '__main__':
if sys.version_info[0] < 3:
sys.exit('ome: error: please use python 3.x')
main()
|
36d2b9843160d9c3d439bc36c0188840fcdfa8b5
|
examples/rmg/minimal_sensitivity/input.py
|
examples/rmg/minimal_sensitivity/input.py
|
# Data sources
database(
thermoLibraries = ['primaryThermoLibrary'],
reactionLibraries = [],
seedMechanisms = [],
kineticsDepositories = ['training'],
kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'],
kineticsEstimator = 'rate rules',
)
# Constraints on generated species
generatedSpeciesConstraints(
maximumRadicalElectrons = 2,
)
# List of species
species(
label='ethane',
reactive=True,
structure=SMILES("CC"),
)
# Reaction systems
simpleReactor(
temperature=(1350,'K'),
pressure=(1.0,'bar'),
initialMoleFractions={
"ethane": 1.0,
},
terminationConversion={
'ethane': 0.9,
},
terminationTime=(1e6,'s'),
sensitivity=['ethane'],
sensitivityThreshold=0.01,
)
simulator(
atol=1e-16,
rtol=1e-8,
sens_atol=1e-6,
sens_rtol=1e-4,
)
model(
toleranceKeepInEdge=0.0,
toleranceMoveToCore=0.1,
toleranceInterruptSimulation=0.1,
maximumEdgeSpecies=100000
)
options(
units='si',
saveRestartPeriod=None,
saveSimulationProfiles=True,
generateOutputHTML=False,
generatePlots=False,
)
|
# Data sources
database(
thermoLibraries = ['primaryThermoLibrary'],
reactionLibraries = [],
seedMechanisms = [],
kineticsDepositories = ['training'],
kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'],
kineticsEstimator = 'rate rules',
)
# Constraints on generated species
generatedSpeciesConstraints(
maximumRadicalElectrons = 2,
)
# List of species
species(
label='ethane',
reactive=True,
structure=SMILES("CC"),
)
# Reaction systems
simpleReactor(
temperature=(1350,'K'),
pressure=(1.0,'bar'),
initialMoleFractions={
"ethane": 1.0,
},
terminationConversion={
'ethane': 0.9,
},
terminationTime=(1e6,'s'),
sensitivity=['ethane'],
sensitivityThreshold=0.01,
)
simulator(
atol=1e-16,
rtol=1e-8,
sens_atol=1e-6,
sens_rtol=1e-4,
)
model(
toleranceKeepInEdge=0.0,
toleranceMoveToCore=0.1,
toleranceInterruptSimulation=0.1,
maximumEdgeSpecies=100000
)
options(
units='si',
saveRestartPeriod=None,
saveSimulationProfiles=False,
generateOutputHTML=False,
generatePlots=False,
)
|
Change saveSimulationProfiles to False in minimal_sensitivity
|
Change saveSimulationProfiles to False in minimal_sensitivity
just to test a diff parameter in this job
|
Python
|
mit
|
chatelak/RMG-Py,nyee/RMG-Py,chatelak/RMG-Py,pierrelb/RMG-Py,nickvandewiele/RMG-Py,pierrelb/RMG-Py,nyee/RMG-Py,nickvandewiele/RMG-Py
|
# Data sources
database(
thermoLibraries = ['primaryThermoLibrary'],
reactionLibraries = [],
seedMechanisms = [],
kineticsDepositories = ['training'],
kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'],
kineticsEstimator = 'rate rules',
)
# Constraints on generated species
generatedSpeciesConstraints(
maximumRadicalElectrons = 2,
)
# List of species
species(
label='ethane',
reactive=True,
structure=SMILES("CC"),
)
# Reaction systems
simpleReactor(
temperature=(1350,'K'),
pressure=(1.0,'bar'),
initialMoleFractions={
"ethane": 1.0,
},
terminationConversion={
'ethane': 0.9,
},
terminationTime=(1e6,'s'),
sensitivity=['ethane'],
sensitivityThreshold=0.01,
)
simulator(
atol=1e-16,
rtol=1e-8,
sens_atol=1e-6,
sens_rtol=1e-4,
)
model(
toleranceKeepInEdge=0.0,
toleranceMoveToCore=0.1,
toleranceInterruptSimulation=0.1,
maximumEdgeSpecies=100000
)
options(
units='si',
saveRestartPeriod=None,
saveSimulationProfiles=True,
generateOutputHTML=False,
generatePlots=False,
)
Change saveSimulationProfiles to False in minimal_sensitivity
just to test a diff parameter in this job
|
# Data sources
database(
thermoLibraries = ['primaryThermoLibrary'],
reactionLibraries = [],
seedMechanisms = [],
kineticsDepositories = ['training'],
kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'],
kineticsEstimator = 'rate rules',
)
# Constraints on generated species
generatedSpeciesConstraints(
maximumRadicalElectrons = 2,
)
# List of species
species(
label='ethane',
reactive=True,
structure=SMILES("CC"),
)
# Reaction systems
simpleReactor(
temperature=(1350,'K'),
pressure=(1.0,'bar'),
initialMoleFractions={
"ethane": 1.0,
},
terminationConversion={
'ethane': 0.9,
},
terminationTime=(1e6,'s'),
sensitivity=['ethane'],
sensitivityThreshold=0.01,
)
simulator(
atol=1e-16,
rtol=1e-8,
sens_atol=1e-6,
sens_rtol=1e-4,
)
model(
toleranceKeepInEdge=0.0,
toleranceMoveToCore=0.1,
toleranceInterruptSimulation=0.1,
maximumEdgeSpecies=100000
)
options(
units='si',
saveRestartPeriod=None,
saveSimulationProfiles=False,
generateOutputHTML=False,
generatePlots=False,
)
|
<commit_before># Data sources
database(
thermoLibraries = ['primaryThermoLibrary'],
reactionLibraries = [],
seedMechanisms = [],
kineticsDepositories = ['training'],
kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'],
kineticsEstimator = 'rate rules',
)
# Constraints on generated species
generatedSpeciesConstraints(
maximumRadicalElectrons = 2,
)
# List of species
species(
label='ethane',
reactive=True,
structure=SMILES("CC"),
)
# Reaction systems
simpleReactor(
temperature=(1350,'K'),
pressure=(1.0,'bar'),
initialMoleFractions={
"ethane": 1.0,
},
terminationConversion={
'ethane': 0.9,
},
terminationTime=(1e6,'s'),
sensitivity=['ethane'],
sensitivityThreshold=0.01,
)
simulator(
atol=1e-16,
rtol=1e-8,
sens_atol=1e-6,
sens_rtol=1e-4,
)
model(
toleranceKeepInEdge=0.0,
toleranceMoveToCore=0.1,
toleranceInterruptSimulation=0.1,
maximumEdgeSpecies=100000
)
options(
units='si',
saveRestartPeriod=None,
saveSimulationProfiles=True,
generateOutputHTML=False,
generatePlots=False,
)
<commit_msg>Change saveSimulationProfiles to False in minimal_sensitivity
just to test a diff parameter in this job<commit_after>
|
# Data sources
database(
thermoLibraries = ['primaryThermoLibrary'],
reactionLibraries = [],
seedMechanisms = [],
kineticsDepositories = ['training'],
kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'],
kineticsEstimator = 'rate rules',
)
# Constraints on generated species
generatedSpeciesConstraints(
maximumRadicalElectrons = 2,
)
# List of species
species(
label='ethane',
reactive=True,
structure=SMILES("CC"),
)
# Reaction systems
simpleReactor(
temperature=(1350,'K'),
pressure=(1.0,'bar'),
initialMoleFractions={
"ethane": 1.0,
},
terminationConversion={
'ethane': 0.9,
},
terminationTime=(1e6,'s'),
sensitivity=['ethane'],
sensitivityThreshold=0.01,
)
simulator(
atol=1e-16,
rtol=1e-8,
sens_atol=1e-6,
sens_rtol=1e-4,
)
model(
toleranceKeepInEdge=0.0,
toleranceMoveToCore=0.1,
toleranceInterruptSimulation=0.1,
maximumEdgeSpecies=100000
)
options(
units='si',
saveRestartPeriod=None,
saveSimulationProfiles=False,
generateOutputHTML=False,
generatePlots=False,
)
|
# Data sources
database(
thermoLibraries = ['primaryThermoLibrary'],
reactionLibraries = [],
seedMechanisms = [],
kineticsDepositories = ['training'],
kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'],
kineticsEstimator = 'rate rules',
)
# Constraints on generated species
generatedSpeciesConstraints(
maximumRadicalElectrons = 2,
)
# List of species
species(
label='ethane',
reactive=True,
structure=SMILES("CC"),
)
# Reaction systems
simpleReactor(
temperature=(1350,'K'),
pressure=(1.0,'bar'),
initialMoleFractions={
"ethane": 1.0,
},
terminationConversion={
'ethane': 0.9,
},
terminationTime=(1e6,'s'),
sensitivity=['ethane'],
sensitivityThreshold=0.01,
)
simulator(
atol=1e-16,
rtol=1e-8,
sens_atol=1e-6,
sens_rtol=1e-4,
)
model(
toleranceKeepInEdge=0.0,
toleranceMoveToCore=0.1,
toleranceInterruptSimulation=0.1,
maximumEdgeSpecies=100000
)
options(
units='si',
saveRestartPeriod=None,
saveSimulationProfiles=True,
generateOutputHTML=False,
generatePlots=False,
)
Change saveSimulationProfiles to False in minimal_sensitivity
just to test a diff parameter in this job# Data sources
database(
thermoLibraries = ['primaryThermoLibrary'],
reactionLibraries = [],
seedMechanisms = [],
kineticsDepositories = ['training'],
kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'],
kineticsEstimator = 'rate rules',
)
# Constraints on generated species
generatedSpeciesConstraints(
maximumRadicalElectrons = 2,
)
# List of species
species(
label='ethane',
reactive=True,
structure=SMILES("CC"),
)
# Reaction systems
simpleReactor(
temperature=(1350,'K'),
pressure=(1.0,'bar'),
initialMoleFractions={
"ethane": 1.0,
},
terminationConversion={
'ethane': 0.9,
},
terminationTime=(1e6,'s'),
sensitivity=['ethane'],
sensitivityThreshold=0.01,
)
simulator(
atol=1e-16,
rtol=1e-8,
sens_atol=1e-6,
sens_rtol=1e-4,
)
model(
toleranceKeepInEdge=0.0,
toleranceMoveToCore=0.1,
toleranceInterruptSimulation=0.1,
maximumEdgeSpecies=100000
)
options(
units='si',
saveRestartPeriod=None,
saveSimulationProfiles=False,
generateOutputHTML=False,
generatePlots=False,
)
|
<commit_before># Data sources
database(
thermoLibraries = ['primaryThermoLibrary'],
reactionLibraries = [],
seedMechanisms = [],
kineticsDepositories = ['training'],
kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'],
kineticsEstimator = 'rate rules',
)
# Constraints on generated species
generatedSpeciesConstraints(
maximumRadicalElectrons = 2,
)
# List of species
species(
label='ethane',
reactive=True,
structure=SMILES("CC"),
)
# Reaction systems
simpleReactor(
temperature=(1350,'K'),
pressure=(1.0,'bar'),
initialMoleFractions={
"ethane": 1.0,
},
terminationConversion={
'ethane': 0.9,
},
terminationTime=(1e6,'s'),
sensitivity=['ethane'],
sensitivityThreshold=0.01,
)
simulator(
atol=1e-16,
rtol=1e-8,
sens_atol=1e-6,
sens_rtol=1e-4,
)
model(
toleranceKeepInEdge=0.0,
toleranceMoveToCore=0.1,
toleranceInterruptSimulation=0.1,
maximumEdgeSpecies=100000
)
options(
units='si',
saveRestartPeriod=None,
saveSimulationProfiles=True,
generateOutputHTML=False,
generatePlots=False,
)
<commit_msg>Change saveSimulationProfiles to False in minimal_sensitivity
just to test a diff parameter in this job<commit_after># Data sources
database(
thermoLibraries = ['primaryThermoLibrary'],
reactionLibraries = [],
seedMechanisms = [],
kineticsDepositories = ['training'],
kineticsFamilies = ['!Intra_Disproportionation','!Substitution_O'],
kineticsEstimator = 'rate rules',
)
# Constraints on generated species
generatedSpeciesConstraints(
maximumRadicalElectrons = 2,
)
# List of species
species(
label='ethane',
reactive=True,
structure=SMILES("CC"),
)
# Reaction systems
simpleReactor(
temperature=(1350,'K'),
pressure=(1.0,'bar'),
initialMoleFractions={
"ethane": 1.0,
},
terminationConversion={
'ethane': 0.9,
},
terminationTime=(1e6,'s'),
sensitivity=['ethane'],
sensitivityThreshold=0.01,
)
simulator(
atol=1e-16,
rtol=1e-8,
sens_atol=1e-6,
sens_rtol=1e-4,
)
model(
toleranceKeepInEdge=0.0,
toleranceMoveToCore=0.1,
toleranceInterruptSimulation=0.1,
maximumEdgeSpecies=100000
)
options(
units='si',
saveRestartPeriod=None,
saveSimulationProfiles=False,
generateOutputHTML=False,
generatePlots=False,
)
|
1bf4116bbd449769d209c4ff98b609b72bd312aa
|
api/views.py
|
api/views.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from rest_framework import viewsets
from core.models import Timesheet, Task, Entry
from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer,
EntrySerializer)
class UserViewSet(viewsets.ModelViewSet):
queryset = User.objects.all()
serializer_class = UserSerializer
class TimesheetViewSet(viewsets.ModelViewSet):
queryset = Timesheet.objects.all()
serializer_class = TimesheetSerializer
filter_fields = ('id',)
class TaskViewSet(viewsets.ModelViewSet):
queryset = Task.objects.all()
serializer_class = TaskSerializer
filter_fields = ('id', 'timesheet',)
class EntryViewSet(viewsets.ModelViewSet):
queryset = Entry.objects.all()
serializer_class = EntrySerializer
filter_fields = ('id', 'user', 'task', 'task__timesheet',)
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from rest_framework import viewsets
import django_filters
from core.models import Timesheet, Task, Entry
from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer,
EntrySerializer)
class UserViewSet(viewsets.ModelViewSet):
queryset = User.objects.all()
serializer_class = UserSerializer
class TimesheetViewSet(viewsets.ModelViewSet):
queryset = Timesheet.objects.all()
serializer_class = TimesheetSerializer
filter_fields = ('id',)
class TaskViewSet(viewsets.ModelViewSet):
queryset = Task.objects.all()
serializer_class = TaskSerializer
filter_fields = ('id', 'timesheet',)
class EntryFilter(django_filters.rest_framework.FilterSet):
min_date = django_filters.DateFilter(name="date", lookup_expr="gte")
max_date = django_filters.DateFilter(name="date", lookup_expr="lte")
class Meta:
model = Entry
fields = ('id', 'date', 'user', 'task', 'task__timesheet',)
class EntryViewSet(viewsets.ModelViewSet):
queryset = Entry.objects.all()
serializer_class = EntrySerializer
filter_class = EntryFilter
|
Add date min-max filtering to API
|
Add date min-max filtering to API
|
Python
|
bsd-2-clause
|
Leahelisabeth/timestrap,muhleder/timestrap,cdubz/timestrap,Leahelisabeth/timestrap,muhleder/timestrap,overshard/timestrap,overshard/timestrap,Leahelisabeth/timestrap,muhleder/timestrap,cdubz/timestrap,Leahelisabeth/timestrap,overshard/timestrap,cdubz/timestrap
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from rest_framework import viewsets
from core.models import Timesheet, Task, Entry
from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer,
EntrySerializer)
class UserViewSet(viewsets.ModelViewSet):
queryset = User.objects.all()
serializer_class = UserSerializer
class TimesheetViewSet(viewsets.ModelViewSet):
queryset = Timesheet.objects.all()
serializer_class = TimesheetSerializer
filter_fields = ('id',)
class TaskViewSet(viewsets.ModelViewSet):
queryset = Task.objects.all()
serializer_class = TaskSerializer
filter_fields = ('id', 'timesheet',)
class EntryViewSet(viewsets.ModelViewSet):
queryset = Entry.objects.all()
serializer_class = EntrySerializer
filter_fields = ('id', 'user', 'task', 'task__timesheet',)
Add date min-max filtering to API
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from rest_framework import viewsets
import django_filters
from core.models import Timesheet, Task, Entry
from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer,
EntrySerializer)
class UserViewSet(viewsets.ModelViewSet):
queryset = User.objects.all()
serializer_class = UserSerializer
class TimesheetViewSet(viewsets.ModelViewSet):
queryset = Timesheet.objects.all()
serializer_class = TimesheetSerializer
filter_fields = ('id',)
class TaskViewSet(viewsets.ModelViewSet):
queryset = Task.objects.all()
serializer_class = TaskSerializer
filter_fields = ('id', 'timesheet',)
class EntryFilter(django_filters.rest_framework.FilterSet):
min_date = django_filters.DateFilter(name="date", lookup_expr="gte")
max_date = django_filters.DateFilter(name="date", lookup_expr="lte")
class Meta:
model = Entry
fields = ('id', 'date', 'user', 'task', 'task__timesheet',)
class EntryViewSet(viewsets.ModelViewSet):
queryset = Entry.objects.all()
serializer_class = EntrySerializer
filter_class = EntryFilter
|
<commit_before># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from rest_framework import viewsets
from core.models import Timesheet, Task, Entry
from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer,
EntrySerializer)
class UserViewSet(viewsets.ModelViewSet):
queryset = User.objects.all()
serializer_class = UserSerializer
class TimesheetViewSet(viewsets.ModelViewSet):
queryset = Timesheet.objects.all()
serializer_class = TimesheetSerializer
filter_fields = ('id',)
class TaskViewSet(viewsets.ModelViewSet):
queryset = Task.objects.all()
serializer_class = TaskSerializer
filter_fields = ('id', 'timesheet',)
class EntryViewSet(viewsets.ModelViewSet):
queryset = Entry.objects.all()
serializer_class = EntrySerializer
filter_fields = ('id', 'user', 'task', 'task__timesheet',)
<commit_msg>Add date min-max filtering to API<commit_after>
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from rest_framework import viewsets
import django_filters
from core.models import Timesheet, Task, Entry
from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer,
EntrySerializer)
class UserViewSet(viewsets.ModelViewSet):
queryset = User.objects.all()
serializer_class = UserSerializer
class TimesheetViewSet(viewsets.ModelViewSet):
queryset = Timesheet.objects.all()
serializer_class = TimesheetSerializer
filter_fields = ('id',)
class TaskViewSet(viewsets.ModelViewSet):
queryset = Task.objects.all()
serializer_class = TaskSerializer
filter_fields = ('id', 'timesheet',)
class EntryFilter(django_filters.rest_framework.FilterSet):
min_date = django_filters.DateFilter(name="date", lookup_expr="gte")
max_date = django_filters.DateFilter(name="date", lookup_expr="lte")
class Meta:
model = Entry
fields = ('id', 'date', 'user', 'task', 'task__timesheet',)
class EntryViewSet(viewsets.ModelViewSet):
queryset = Entry.objects.all()
serializer_class = EntrySerializer
filter_class = EntryFilter
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from rest_framework import viewsets
from core.models import Timesheet, Task, Entry
from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer,
EntrySerializer)
class UserViewSet(viewsets.ModelViewSet):
queryset = User.objects.all()
serializer_class = UserSerializer
class TimesheetViewSet(viewsets.ModelViewSet):
queryset = Timesheet.objects.all()
serializer_class = TimesheetSerializer
filter_fields = ('id',)
class TaskViewSet(viewsets.ModelViewSet):
queryset = Task.objects.all()
serializer_class = TaskSerializer
filter_fields = ('id', 'timesheet',)
class EntryViewSet(viewsets.ModelViewSet):
queryset = Entry.objects.all()
serializer_class = EntrySerializer
filter_fields = ('id', 'user', 'task', 'task__timesheet',)
Add date min-max filtering to API# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from rest_framework import viewsets
import django_filters
from core.models import Timesheet, Task, Entry
from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer,
EntrySerializer)
class UserViewSet(viewsets.ModelViewSet):
queryset = User.objects.all()
serializer_class = UserSerializer
class TimesheetViewSet(viewsets.ModelViewSet):
queryset = Timesheet.objects.all()
serializer_class = TimesheetSerializer
filter_fields = ('id',)
class TaskViewSet(viewsets.ModelViewSet):
queryset = Task.objects.all()
serializer_class = TaskSerializer
filter_fields = ('id', 'timesheet',)
class EntryFilter(django_filters.rest_framework.FilterSet):
min_date = django_filters.DateFilter(name="date", lookup_expr="gte")
max_date = django_filters.DateFilter(name="date", lookup_expr="lte")
class Meta:
model = Entry
fields = ('id', 'date', 'user', 'task', 'task__timesheet',)
class EntryViewSet(viewsets.ModelViewSet):
queryset = Entry.objects.all()
serializer_class = EntrySerializer
filter_class = EntryFilter
|
<commit_before># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from rest_framework import viewsets
from core.models import Timesheet, Task, Entry
from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer,
EntrySerializer)
class UserViewSet(viewsets.ModelViewSet):
queryset = User.objects.all()
serializer_class = UserSerializer
class TimesheetViewSet(viewsets.ModelViewSet):
queryset = Timesheet.objects.all()
serializer_class = TimesheetSerializer
filter_fields = ('id',)
class TaskViewSet(viewsets.ModelViewSet):
queryset = Task.objects.all()
serializer_class = TaskSerializer
filter_fields = ('id', 'timesheet',)
class EntryViewSet(viewsets.ModelViewSet):
queryset = Entry.objects.all()
serializer_class = EntrySerializer
filter_fields = ('id', 'user', 'task', 'task__timesheet',)
<commit_msg>Add date min-max filtering to API<commit_after># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from rest_framework import viewsets
import django_filters
from core.models import Timesheet, Task, Entry
from .serializers import (UserSerializer, TimesheetSerializer, TaskSerializer,
EntrySerializer)
class UserViewSet(viewsets.ModelViewSet):
queryset = User.objects.all()
serializer_class = UserSerializer
class TimesheetViewSet(viewsets.ModelViewSet):
queryset = Timesheet.objects.all()
serializer_class = TimesheetSerializer
filter_fields = ('id',)
class TaskViewSet(viewsets.ModelViewSet):
queryset = Task.objects.all()
serializer_class = TaskSerializer
filter_fields = ('id', 'timesheet',)
class EntryFilter(django_filters.rest_framework.FilterSet):
min_date = django_filters.DateFilter(name="date", lookup_expr="gte")
max_date = django_filters.DateFilter(name="date", lookup_expr="lte")
class Meta:
model = Entry
fields = ('id', 'date', 'user', 'task', 'task__timesheet',)
class EntryViewSet(viewsets.ModelViewSet):
queryset = Entry.objects.all()
serializer_class = EntrySerializer
filter_class = EntryFilter
|
b5e11827929f37da8d18616f1fb3fc2d62591515
|
djangocms_spa/decorators.py
|
djangocms_spa/decorators.py
|
from functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.template.response import ContentNotRenderedError
from django.utils.decorators import available_attrs
def cache_view(view_func):
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view_func(view, *args, **kwargs):
cache_key = view.request.get_full_path()
cached_response = cache.get(cache_key)
if cached_response and not view.request.user.is_authenticated():
return cached_response
response = view_func(view, *args, **kwargs)
if response.status_code == 200 and not view.request.user.is_authenticated():
try:
set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
except ContentNotRenderedError:
response.add_post_render_callback(
lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
)
return response
return _wrapped_view_func
def set_cache_after_rendering(cache_key, response, timeout):
cache.set(cache_key, response, timeout)
|
from functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.template.response import ContentNotRenderedError
from django.utils.decorators import available_attrs
def cache_view(view_func):
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view_func(view, *args, **kwargs):
request = view.request
language_code = request.LANGUAGE_CODE
cache_key = '{path}:{lang}'.format(path=request.get_full_path(), lang=language_code)
cached_response = cache.get(cache_key)
if cached_response and not request.user.is_authenticated():
return cached_response
response = view_func(view, *args, **kwargs)
if response.status_code == 200 and not request.user.is_authenticated():
try:
set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
except ContentNotRenderedError:
response.add_post_render_callback(
lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
)
return response
return _wrapped_view_func
def set_cache_after_rendering(cache_key, response, timeout):
cache.set(cache_key, response, timeout)
|
Add language code to cache key explicitly
|
[language_activation] Add language code to cache key explicitly
|
Python
|
mit
|
dreipol/djangocms-spa,dreipol/djangocms-spa
|
from functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.template.response import ContentNotRenderedError
from django.utils.decorators import available_attrs
def cache_view(view_func):
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view_func(view, *args, **kwargs):
cache_key = view.request.get_full_path()
cached_response = cache.get(cache_key)
if cached_response and not view.request.user.is_authenticated():
return cached_response
response = view_func(view, *args, **kwargs)
if response.status_code == 200 and not view.request.user.is_authenticated():
try:
set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
except ContentNotRenderedError:
response.add_post_render_callback(
lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
)
return response
return _wrapped_view_func
def set_cache_after_rendering(cache_key, response, timeout):
cache.set(cache_key, response, timeout)
[language_activation] Add language code to cache key explicitly
|
from functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.template.response import ContentNotRenderedError
from django.utils.decorators import available_attrs
def cache_view(view_func):
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view_func(view, *args, **kwargs):
request = view.request
language_code = request.LANGUAGE_CODE
cache_key = '{path}:{lang}'.format(path=request.get_full_path(), lang=language_code)
cached_response = cache.get(cache_key)
if cached_response and not request.user.is_authenticated():
return cached_response
response = view_func(view, *args, **kwargs)
if response.status_code == 200 and not request.user.is_authenticated():
try:
set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
except ContentNotRenderedError:
response.add_post_render_callback(
lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
)
return response
return _wrapped_view_func
def set_cache_after_rendering(cache_key, response, timeout):
cache.set(cache_key, response, timeout)
|
<commit_before>from functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.template.response import ContentNotRenderedError
from django.utils.decorators import available_attrs
def cache_view(view_func):
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view_func(view, *args, **kwargs):
cache_key = view.request.get_full_path()
cached_response = cache.get(cache_key)
if cached_response and not view.request.user.is_authenticated():
return cached_response
response = view_func(view, *args, **kwargs)
if response.status_code == 200 and not view.request.user.is_authenticated():
try:
set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
except ContentNotRenderedError:
response.add_post_render_callback(
lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
)
return response
return _wrapped_view_func
def set_cache_after_rendering(cache_key, response, timeout):
cache.set(cache_key, response, timeout)
<commit_msg>[language_activation] Add language code to cache key explicitly<commit_after>
|
from functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.template.response import ContentNotRenderedError
from django.utils.decorators import available_attrs
def cache_view(view_func):
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view_func(view, *args, **kwargs):
request = view.request
language_code = request.LANGUAGE_CODE
cache_key = '{path}:{lang}'.format(path=request.get_full_path(), lang=language_code)
cached_response = cache.get(cache_key)
if cached_response and not request.user.is_authenticated():
return cached_response
response = view_func(view, *args, **kwargs)
if response.status_code == 200 and not request.user.is_authenticated():
try:
set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
except ContentNotRenderedError:
response.add_post_render_callback(
lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
)
return response
return _wrapped_view_func
def set_cache_after_rendering(cache_key, response, timeout):
cache.set(cache_key, response, timeout)
|
from functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.template.response import ContentNotRenderedError
from django.utils.decorators import available_attrs
def cache_view(view_func):
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view_func(view, *args, **kwargs):
cache_key = view.request.get_full_path()
cached_response = cache.get(cache_key)
if cached_response and not view.request.user.is_authenticated():
return cached_response
response = view_func(view, *args, **kwargs)
if response.status_code == 200 and not view.request.user.is_authenticated():
try:
set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
except ContentNotRenderedError:
response.add_post_render_callback(
lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
)
return response
return _wrapped_view_func
def set_cache_after_rendering(cache_key, response, timeout):
cache.set(cache_key, response, timeout)
[language_activation] Add language code to cache key explicitlyfrom functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.template.response import ContentNotRenderedError
from django.utils.decorators import available_attrs
def cache_view(view_func):
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view_func(view, *args, **kwargs):
request = view.request
language_code = request.LANGUAGE_CODE
cache_key = '{path}:{lang}'.format(path=request.get_full_path(), lang=language_code)
cached_response = cache.get(cache_key)
if cached_response and not request.user.is_authenticated():
return cached_response
response = view_func(view, *args, **kwargs)
if response.status_code == 200 and not request.user.is_authenticated():
try:
set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
except ContentNotRenderedError:
response.add_post_render_callback(
lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
)
return response
return _wrapped_view_func
def set_cache_after_rendering(cache_key, response, timeout):
cache.set(cache_key, response, timeout)
|
<commit_before>from functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.template.response import ContentNotRenderedError
from django.utils.decorators import available_attrs
def cache_view(view_func):
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view_func(view, *args, **kwargs):
cache_key = view.request.get_full_path()
cached_response = cache.get(cache_key)
if cached_response and not view.request.user.is_authenticated():
return cached_response
response = view_func(view, *args, **kwargs)
if response.status_code == 200 and not view.request.user.is_authenticated():
try:
set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
except ContentNotRenderedError:
response.add_post_render_callback(
lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
)
return response
return _wrapped_view_func
def set_cache_after_rendering(cache_key, response, timeout):
cache.set(cache_key, response, timeout)
<commit_msg>[language_activation] Add language code to cache key explicitly<commit_after>from functools import wraps
from django.conf import settings
from django.core.cache import cache
from django.template.response import ContentNotRenderedError
from django.utils.decorators import available_attrs
def cache_view(view_func):
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view_func(view, *args, **kwargs):
request = view.request
language_code = request.LANGUAGE_CODE
cache_key = '{path}:{lang}'.format(path=request.get_full_path(), lang=language_code)
cached_response = cache.get(cache_key)
if cached_response and not request.user.is_authenticated():
return cached_response
response = view_func(view, *args, **kwargs)
if response.status_code == 200 and not request.user.is_authenticated():
try:
set_cache_after_rendering(cache_key, response, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
except ContentNotRenderedError:
response.add_post_render_callback(
lambda r: set_cache_after_rendering(cache_key, r, settings.DJANGOCMS_SPA_CACHE_TIMEOUT)
)
return response
return _wrapped_view_func
def set_cache_after_rendering(cache_key, response, timeout):
cache.set(cache_key, response, timeout)
|
e0aab62f2a693ca20a81c9e55c4220f379ac9eb1
|
socialregistration/templatetags/socialregistration_tags.py
|
socialregistration/templatetags/socialregistration_tags.py
|
from django import template
register = template.Library()
@register.tag
def social_csrf_token():
"""
Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration
work with both Django v1.2 and Django < v1.2
"""
return CsrfNode()
class CsrfNode(template.Node):
def render(self, context):
try:
from django.template.defaulttags import CsrfTokenNode
return CsrfTokenNode().render(context)
except ImportError:
return u''
|
from django import template
register = template.Library()
@register.tag
def social_csrf_token(parser, token):
"""
Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration
work with both Django v1.2 and Django < v1.2
"""
return CsrfNode()
class CsrfNode(template.Node):
def render(self, context):
try:
from django.template.defaulttags import CsrfTokenNode
return CsrfTokenNode().render(context)
except ImportError:
return u''
|
Add necessary arguments to the social_csrf_token tag.
|
Add necessary arguments to the social_csrf_token tag.
|
Python
|
mit
|
praekelt/django-socialregistration,aditweb/django-socialregistration,minlex/django-socialregistration,kapt/django-socialregistration,coxmediagroup/django-socialregistration,aditweb/django-socialregistration,flashingpumpkin/django-socialregistration,mark-adams/django-socialregistration,mark-adams/django-socialregistration,minlex/django-socialregistration,brodie/django-socialregistration,minlex/django-socialregistration,lgapontes/django-socialregistration,bopo/django-socialregistration,0101/django-socialregistration,aditweb/django-socialregistration,lgapontes/django-socialregistration,amakhnach/django-socialregistration,bopo/django-socialregistration,flashingpumpkin/django-socialregistration,coxmediagroup/django-socialregistration,itmustbejj/django-socialregistration,lgapontes/django-socialregistration,Soovox/django-socialregistration,bopo/django-socialregistration,mark-adams/django-socialregistration,kapt/django-socialregistration,brodie/django-socialregistration
|
from django import template
register = template.Library()
@register.tag
def social_csrf_token():
"""
Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration
work with both Django v1.2 and Django < v1.2
"""
return CsrfNode()
class CsrfNode(template.Node):
def render(self, context):
try:
from django.template.defaulttags import CsrfTokenNode
return CsrfTokenNode().render(context)
except ImportError:
return u''Add necessary arguments to the social_csrf_token tag.
|
from django import template
register = template.Library()
@register.tag
def social_csrf_token(parser, token):
"""
Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration
work with both Django v1.2 and Django < v1.2
"""
return CsrfNode()
class CsrfNode(template.Node):
def render(self, context):
try:
from django.template.defaulttags import CsrfTokenNode
return CsrfTokenNode().render(context)
except ImportError:
return u''
|
<commit_before>from django import template
register = template.Library()
@register.tag
def social_csrf_token():
"""
Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration
work with both Django v1.2 and Django < v1.2
"""
return CsrfNode()
class CsrfNode(template.Node):
def render(self, context):
try:
from django.template.defaulttags import CsrfTokenNode
return CsrfTokenNode().render(context)
except ImportError:
return u''<commit_msg>Add necessary arguments to the social_csrf_token tag.<commit_after>
|
from django import template
register = template.Library()
@register.tag
def social_csrf_token(parser, token):
"""
Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration
work with both Django v1.2 and Django < v1.2
"""
return CsrfNode()
class CsrfNode(template.Node):
def render(self, context):
try:
from django.template.defaulttags import CsrfTokenNode
return CsrfTokenNode().render(context)
except ImportError:
return u''
|
from django import template
register = template.Library()
@register.tag
def social_csrf_token():
"""
Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration
work with both Django v1.2 and Django < v1.2
"""
return CsrfNode()
class CsrfNode(template.Node):
def render(self, context):
try:
from django.template.defaulttags import CsrfTokenNode
return CsrfTokenNode().render(context)
except ImportError:
return u''Add necessary arguments to the social_csrf_token tag.from django import template
register = template.Library()
@register.tag
def social_csrf_token(parser, token):
"""
Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration
work with both Django v1.2 and Django < v1.2
"""
return CsrfNode()
class CsrfNode(template.Node):
def render(self, context):
try:
from django.template.defaulttags import CsrfTokenNode
return CsrfTokenNode().render(context)
except ImportError:
return u''
|
<commit_before>from django import template
register = template.Library()
@register.tag
def social_csrf_token():
"""
Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration
work with both Django v1.2 and Django < v1.2
"""
return CsrfNode()
class CsrfNode(template.Node):
def render(self, context):
try:
from django.template.defaulttags import CsrfTokenNode
return CsrfTokenNode().render(context)
except ImportError:
return u''<commit_msg>Add necessary arguments to the social_csrf_token tag.<commit_after>from django import template
register = template.Library()
@register.tag
def social_csrf_token(parser, token):
"""
Wrapper around the ``{% csrf_token %}`` template tag to make socialregistration
work with both Django v1.2 and Django < v1.2
"""
return CsrfNode()
class CsrfNode(template.Node):
def render(self, context):
try:
from django.template.defaulttags import CsrfTokenNode
return CsrfTokenNode().render(context)
except ImportError:
return u''
|
7ea8420e9653765d960938340124b8c2274c69fc
|
Site/Settings.py
|
Site/Settings.py
|
import os
settings = {}
domain = 'localhost:8080'
framework = 'flask'
debug = False
clearOnLoad = False
emailOnRegister = False
registrationDisabled = False
cacheDirectory = "./Site/Cache"
siteDown = False
database = 'osf20120530' # Mongo
cookieDomain = '.openscienceframework.org' # Beaker
static = '%s/static' % os.path.dirname(os.path.abspath(__file__))
local = True
|
import os
settings = {}
domain = 'localhost:8080'
framework = 'flask'
debug = False
clearOnLoad = False
emailOnRegister = False
registrationDisabled = False
cacheDirectory = "./Site/Cache"
siteDown = False
database = 'osf20120530' # Mongo
cookieDomain = '.openscienceframework.org' # Beaker
static = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'static')
local = True
|
Improve OS compatibility for settings
|
Improve OS compatibility for settings
|
Python
|
apache-2.0
|
cslzchen/osf.io,mattclark/osf.io,kwierman/osf.io,SSJohns/osf.io,caseyrollins/osf.io,HarryRybacki/osf.io,monikagrabowska/osf.io,HarryRybacki/osf.io,Nesiehr/osf.io,himanshuo/osf.io,SSJohns/osf.io,jnayak1/osf.io,billyhunt/osf.io,fabianvf/osf.io,CenterForOpenScience/osf.io,TomHeatwole/osf.io,chennan47/osf.io,HarryRybacki/osf.io,jinluyuan/osf.io,CenterForOpenScience/osf.io,brandonPurvis/osf.io,jnayak1/osf.io,cosenal/osf.io,samanehsan/osf.io,Nesiehr/osf.io,GageGaskins/osf.io,bdyetton/prettychart,jinluyuan/osf.io,arpitar/osf.io,jmcarp/osf.io,kch8qx/osf.io,Nesiehr/osf.io,monikagrabowska/osf.io,Johnetordoff/osf.io,cwisecarver/osf.io,reinaH/osf.io,jeffreyliu3230/osf.io,RomanZWang/osf.io,barbour-em/osf.io,lamdnhan/osf.io,revanthkolli/osf.io,amyshi188/osf.io,danielneis/osf.io,ckc6cz/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,njantrania/osf.io,GaryKriebel/osf.io,doublebits/osf.io,barbour-em/osf.io,revanthkolli/osf.io,aaxelb/osf.io,mluo613/osf.io,brianjgeiger/osf.io,kushG/osf.io,icereval/osf.io,jolene-esposito/osf.io,jolene-esposito/osf.io,wearpants/osf.io,felliott/osf.io,mluo613/osf.io,zachjanicki/osf.io,arpitar/osf.io,mattclark/osf.io,brandonPurvis/osf.io,rdhyee/osf.io,jeffreyliu3230/osf.io,cslzchen/osf.io,ticklemepierce/osf.io,zachjanicki/osf.io,kch8qx/osf.io,rdhyee/osf.io,AndrewSallans/osf.io,danielneis/osf.io,KAsante95/osf.io,KAsante95/osf.io,GageGaskins/osf.io,kch8qx/osf.io,pattisdr/osf.io,barbour-em/osf.io,amyshi188/osf.io,asanfilippo7/osf.io,alexschiller/osf.io,KAsante95/osf.io,jmcarp/osf.io,ticklemepierce/osf.io,lyndsysimon/osf.io,leb2dg/osf.io,Ghalko/osf.io,fabianvf/osf.io,samchrisinger/osf.io,GageGaskins/osf.io,wearpants/osf.io,sbt9uc/osf.io,AndrewSallans/osf.io,felliott/osf.io,samanehsan/osf.io,cosenal/osf.io,bdyetton/prettychart,bdyetton/prettychart,binoculars/osf.io,mluo613/osf.io,MerlinZhang/osf.io,caneruguz/osf.io,mfraezz/osf.io,amyshi188/osf.io,crcresearch/osf.io,dplorimer/osf,kushG/osf.io,Johnetordoff/osf.io,leb2dg/osf.io,TomBaxter/osf.io,alexschiller/osf.io,acshi/osf.io,arpitar/osf.io,GageGaskins/osf.io,amyshi188/osf.io,erinspace/osf.io,haoyuchen1992/osf.io,lamdnhan/osf.io,himanshuo/osf.io,Ghalko/osf.io,caneruguz/osf.io,HalcyonChimera/osf.io,felliott/osf.io,zachjanicki/osf.io,billyhunt/osf.io,caseyrygt/osf.io,danielneis/osf.io,mattclark/osf.io,cldershem/osf.io,billyhunt/osf.io,monikagrabowska/osf.io,zkraime/osf.io,acshi/osf.io,mfraezz/osf.io,leb2dg/osf.io,dplorimer/osf,sloria/osf.io,binoculars/osf.io,sbt9uc/osf.io,leb2dg/osf.io,laurenrevere/osf.io,acshi/osf.io,dplorimer/osf,acshi/osf.io,crcresearch/osf.io,jnayak1/osf.io,chrisseto/osf.io,alexschiller/osf.io,SSJohns/osf.io,baylee-d/osf.io,aaxelb/osf.io,sbt9uc/osf.io,Nesiehr/osf.io,kushG/osf.io,saradbowman/osf.io,samchrisinger/osf.io,cldershem/osf.io,zamattiac/osf.io,hmoco/osf.io,caseyrollins/osf.io,GaryKriebel/osf.io,TomBaxter/osf.io,adlius/osf.io,hmoco/osf.io,petermalcolm/osf.io,jnayak1/osf.io,ZobairAlijan/osf.io,barbour-em/osf.io,RomanZWang/osf.io,abought/osf.io,felliott/osf.io,chrisseto/osf.io,erinspace/osf.io,mluke93/osf.io,wearpants/osf.io,cldershem/osf.io,Ghalko/osf.io,zamattiac/osf.io,cwisecarver/osf.io,mfraezz/osf.io,caneruguz/osf.io,ticklemepierce/osf.io,GaryKriebel/osf.io,monikagrabowska/osf.io,emetsger/osf.io,brianjgeiger/osf.io,icereval/osf.io,danielneis/osf.io,cldershem/osf.io,laurenrevere/osf.io,GaryKriebel/osf.io,monikagrabowska/osf.io,DanielSBrown/osf.io,doublebits/osf.io,chrisseto/osf.io,zkraime/osf.io,brianjgeiger/osf.io,brandonPurvis/osf.io,kushG/osf.io,lyndsysimon/osf.io,petermalcolm/osf.io,samchrisinger/osf.io,caseyrollins/osf.io,himanshuo/osf.io,aaxelb/osf.io,jmcarp/osf.io,erinspace/osf.io,ticklemepierce/osf.io,RomanZWang/osf.io,billyhunt/osf.io,kch8qx/osf.io,reinaH/osf.io,jolene-esposito/osf.io,mfraezz/osf.io,emetsger/osf.io,doublebits/osf.io,samanehsan/osf.io,RomanZWang/osf.io,haoyuchen1992/osf.io,himanshuo/osf.io,dplorimer/osf,cosenal/osf.io,njantrania/osf.io,chennan47/osf.io,billyhunt/osf.io,MerlinZhang/osf.io,jeffreyliu3230/osf.io,HarryRybacki/osf.io,cslzchen/osf.io,caseyrygt/osf.io,baylee-d/osf.io,ZobairAlijan/osf.io,TomHeatwole/osf.io,GageGaskins/osf.io,njantrania/osf.io,kwierman/osf.io,asanfilippo7/osf.io,DanielSBrown/osf.io,petermalcolm/osf.io,asanfilippo7/osf.io,alexschiller/osf.io,ckc6cz/osf.io,alexschiller/osf.io,adlius/osf.io,kwierman/osf.io,chennan47/osf.io,mluke93/osf.io,revanthkolli/osf.io,njantrania/osf.io,petermalcolm/osf.io,samchrisinger/osf.io,rdhyee/osf.io,sbt9uc/osf.io,ckc6cz/osf.io,hmoco/osf.io,TomHeatwole/osf.io,caneruguz/osf.io,mluke93/osf.io,caseyrygt/osf.io,saradbowman/osf.io,cwisecarver/osf.io,lamdnhan/osf.io,hmoco/osf.io,zachjanicki/osf.io,MerlinZhang/osf.io,wearpants/osf.io,abought/osf.io,Johnetordoff/osf.io,ZobairAlijan/osf.io,reinaH/osf.io,abought/osf.io,cslzchen/osf.io,zamattiac/osf.io,abought/osf.io,samanehsan/osf.io,SSJohns/osf.io,CenterForOpenScience/osf.io,adlius/osf.io,jinluyuan/osf.io,arpitar/osf.io,jolene-esposito/osf.io,zamattiac/osf.io,fabianvf/osf.io,aaxelb/osf.io,doublebits/osf.io,haoyuchen1992/osf.io,KAsante95/osf.io,caseyrygt/osf.io,brandonPurvis/osf.io,icereval/osf.io,sloria/osf.io,HalcyonChimera/osf.io,ckc6cz/osf.io,lamdnhan/osf.io,lyndsysimon/osf.io,DanielSBrown/osf.io,rdhyee/osf.io,lyndsysimon/osf.io,bdyetton/prettychart,zkraime/osf.io,emetsger/osf.io,doublebits/osf.io,mluo613/osf.io,kch8qx/osf.io,baylee-d/osf.io,pattisdr/osf.io,laurenrevere/osf.io,jmcarp/osf.io,HalcyonChimera/osf.io,sloria/osf.io,ZobairAlijan/osf.io,crcresearch/osf.io,KAsante95/osf.io,HalcyonChimera/osf.io,zkraime/osf.io,mluke93/osf.io,kwierman/osf.io,chrisseto/osf.io,Ghalko/osf.io,jeffreyliu3230/osf.io,fabianvf/osf.io,haoyuchen1992/osf.io,RomanZWang/osf.io,MerlinZhang/osf.io,TomHeatwole/osf.io,binoculars/osf.io,asanfilippo7/osf.io,pattisdr/osf.io,emetsger/osf.io,cosenal/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,mluo613/osf.io,revanthkolli/osf.io,cwisecarver/osf.io,jinluyuan/osf.io,reinaH/osf.io,DanielSBrown/osf.io,adlius/osf.io,acshi/osf.io,TomBaxter/osf.io
|
import os
settings = {}
domain = 'localhost:8080'
framework = 'flask'
debug = False
clearOnLoad = False
emailOnRegister = False
registrationDisabled = False
cacheDirectory = "./Site/Cache"
siteDown = False
database = 'osf20120530' # Mongo
cookieDomain = '.openscienceframework.org' # Beaker
static = '%s/static' % os.path.dirname(os.path.abspath(__file__))
local = TrueImprove OS compatibility for settings
|
import os
settings = {}
domain = 'localhost:8080'
framework = 'flask'
debug = False
clearOnLoad = False
emailOnRegister = False
registrationDisabled = False
cacheDirectory = "./Site/Cache"
siteDown = False
database = 'osf20120530' # Mongo
cookieDomain = '.openscienceframework.org' # Beaker
static = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'static')
local = True
|
<commit_before>import os
settings = {}
domain = 'localhost:8080'
framework = 'flask'
debug = False
clearOnLoad = False
emailOnRegister = False
registrationDisabled = False
cacheDirectory = "./Site/Cache"
siteDown = False
database = 'osf20120530' # Mongo
cookieDomain = '.openscienceframework.org' # Beaker
static = '%s/static' % os.path.dirname(os.path.abspath(__file__))
local = True<commit_msg>Improve OS compatibility for settings<commit_after>
|
import os
settings = {}
domain = 'localhost:8080'
framework = 'flask'
debug = False
clearOnLoad = False
emailOnRegister = False
registrationDisabled = False
cacheDirectory = "./Site/Cache"
siteDown = False
database = 'osf20120530' # Mongo
cookieDomain = '.openscienceframework.org' # Beaker
static = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'static')
local = True
|
import os
settings = {}
domain = 'localhost:8080'
framework = 'flask'
debug = False
clearOnLoad = False
emailOnRegister = False
registrationDisabled = False
cacheDirectory = "./Site/Cache"
siteDown = False
database = 'osf20120530' # Mongo
cookieDomain = '.openscienceframework.org' # Beaker
static = '%s/static' % os.path.dirname(os.path.abspath(__file__))
local = TrueImprove OS compatibility for settingsimport os
settings = {}
domain = 'localhost:8080'
framework = 'flask'
debug = False
clearOnLoad = False
emailOnRegister = False
registrationDisabled = False
cacheDirectory = "./Site/Cache"
siteDown = False
database = 'osf20120530' # Mongo
cookieDomain = '.openscienceframework.org' # Beaker
static = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'static')
local = True
|
<commit_before>import os
settings = {}
domain = 'localhost:8080'
framework = 'flask'
debug = False
clearOnLoad = False
emailOnRegister = False
registrationDisabled = False
cacheDirectory = "./Site/Cache"
siteDown = False
database = 'osf20120530' # Mongo
cookieDomain = '.openscienceframework.org' # Beaker
static = '%s/static' % os.path.dirname(os.path.abspath(__file__))
local = True<commit_msg>Improve OS compatibility for settings<commit_after>import os
settings = {}
domain = 'localhost:8080'
framework = 'flask'
debug = False
clearOnLoad = False
emailOnRegister = False
registrationDisabled = False
cacheDirectory = "./Site/Cache"
siteDown = False
database = 'osf20120530' # Mongo
cookieDomain = '.openscienceframework.org' # Beaker
static = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'static')
local = True
|
09ec6e4611a763e823a5e3d15fb233a0132fd06b
|
imagersite/imagersite/tests.py
|
imagersite/imagersite/tests.py
|
from django.test import Client, TestCase
from django.contrib.auth.models import User
HOME = '/'
REGISTER = '/accounts/register/'
LOGIN = '/login'
LOGOUT = '/logout'
class UnauthenticatedUser(TestCase):
"""Create unauth user for testing."""
def setUp(self):
"""Setup unauth user."""
client = Client()
self.home = client.get(HOME)
self.login = client.get(LOGIN)
self.logout = client.get(LOGOUT)
self.register = client.get(REGISTER)
def test_no_user_in_db(self):
"""No user i db."""
self.assertFalse(User.objects.count())
def test_homepage(self):
"""Test homepage can be reached."""
self.assertEqual(self.home.status_code, 200)
def test_login(self):
"""Test login cna be reached."""
self.assertEqual(self.login.status_code, 200)
def test_logout(self):
"""Test logout can be reached."""
self.assertEqual(self.logout.status_code, 200)
def test_register(self):
"""Test register can be reached."""
self.assertEqual(self.register.status_code, 200)
|
"""Tests for project level urls and views."""
from __future__ import unicode_literals
from django.contrib.staticfiles import finders
from django.test import Client, TestCase
from django.contrib.auth.models import User
HOME = '/'
REGISTER = '/accounts/register/'
LOGIN = '/login'
LOGOUT = '/logout'
DEFAULT_IMAGE = finders.find('static/imagersite/images/default-image.jpg')
class UnauthenticatedUser(TestCase):
"""Create unauth user for testing."""
def setUp(self):
"""Setup unauth user."""
client = Client()
self.home = client.get(HOME)
self.login = client.get(LOGIN)
self.logout = client.get(LOGOUT)
self.register = client.get(REGISTER)
def test_no_user_in_db(self):
"""No user i db."""
self.assertFalse(User.objects.count())
def test_homepage(self):
"""Test homepage can be reached."""
self.assertEqual(self.home.status_code, 200)
def test_login(self):
"""Test login cna be reached."""
self.assertEqual(self.login.status_code, 200)
def test_logout(self):
"""Test logout can be reached."""
self.assertEqual(self.logout.status_code, 200)
def test_register(self):
"""Test register can be reached."""
self.assertEqual(self.register.status_code, 200)
def test_default_image(self):
"""Test default image shows up."""
img_path = self.home.context['image']
self.assertEqual(img_path, DEFAULT_IMAGE)
|
Add passing test for default image
|
Add passing test for default image
|
Python
|
mit
|
SeleniumK/django-imager,SeleniumK/django-imager,SeleniumK/django-imager
|
from django.test import Client, TestCase
from django.contrib.auth.models import User
HOME = '/'
REGISTER = '/accounts/register/'
LOGIN = '/login'
LOGOUT = '/logout'
class UnauthenticatedUser(TestCase):
"""Create unauth user for testing."""
def setUp(self):
"""Setup unauth user."""
client = Client()
self.home = client.get(HOME)
self.login = client.get(LOGIN)
self.logout = client.get(LOGOUT)
self.register = client.get(REGISTER)
def test_no_user_in_db(self):
"""No user i db."""
self.assertFalse(User.objects.count())
def test_homepage(self):
"""Test homepage can be reached."""
self.assertEqual(self.home.status_code, 200)
def test_login(self):
"""Test login cna be reached."""
self.assertEqual(self.login.status_code, 200)
def test_logout(self):
"""Test logout can be reached."""
self.assertEqual(self.logout.status_code, 200)
def test_register(self):
"""Test register can be reached."""
self.assertEqual(self.register.status_code, 200)
Add passing test for default image
|
"""Tests for project level urls and views."""
from __future__ import unicode_literals
from django.contrib.staticfiles import finders
from django.test import Client, TestCase
from django.contrib.auth.models import User
HOME = '/'
REGISTER = '/accounts/register/'
LOGIN = '/login'
LOGOUT = '/logout'
DEFAULT_IMAGE = finders.find('static/imagersite/images/default-image.jpg')
class UnauthenticatedUser(TestCase):
"""Create unauth user for testing."""
def setUp(self):
"""Setup unauth user."""
client = Client()
self.home = client.get(HOME)
self.login = client.get(LOGIN)
self.logout = client.get(LOGOUT)
self.register = client.get(REGISTER)
def test_no_user_in_db(self):
"""No user i db."""
self.assertFalse(User.objects.count())
def test_homepage(self):
"""Test homepage can be reached."""
self.assertEqual(self.home.status_code, 200)
def test_login(self):
"""Test login cna be reached."""
self.assertEqual(self.login.status_code, 200)
def test_logout(self):
"""Test logout can be reached."""
self.assertEqual(self.logout.status_code, 200)
def test_register(self):
"""Test register can be reached."""
self.assertEqual(self.register.status_code, 200)
def test_default_image(self):
"""Test default image shows up."""
img_path = self.home.context['image']
self.assertEqual(img_path, DEFAULT_IMAGE)
|
<commit_before>from django.test import Client, TestCase
from django.contrib.auth.models import User
HOME = '/'
REGISTER = '/accounts/register/'
LOGIN = '/login'
LOGOUT = '/logout'
class UnauthenticatedUser(TestCase):
"""Create unauth user for testing."""
def setUp(self):
"""Setup unauth user."""
client = Client()
self.home = client.get(HOME)
self.login = client.get(LOGIN)
self.logout = client.get(LOGOUT)
self.register = client.get(REGISTER)
def test_no_user_in_db(self):
"""No user i db."""
self.assertFalse(User.objects.count())
def test_homepage(self):
"""Test homepage can be reached."""
self.assertEqual(self.home.status_code, 200)
def test_login(self):
"""Test login cna be reached."""
self.assertEqual(self.login.status_code, 200)
def test_logout(self):
"""Test logout can be reached."""
self.assertEqual(self.logout.status_code, 200)
def test_register(self):
"""Test register can be reached."""
self.assertEqual(self.register.status_code, 200)
<commit_msg>Add passing test for default image<commit_after>
|
"""Tests for project level urls and views."""
from __future__ import unicode_literals
from django.contrib.staticfiles import finders
from django.test import Client, TestCase
from django.contrib.auth.models import User
HOME = '/'
REGISTER = '/accounts/register/'
LOGIN = '/login'
LOGOUT = '/logout'
DEFAULT_IMAGE = finders.find('static/imagersite/images/default-image.jpg')
class UnauthenticatedUser(TestCase):
"""Create unauth user for testing."""
def setUp(self):
"""Setup unauth user."""
client = Client()
self.home = client.get(HOME)
self.login = client.get(LOGIN)
self.logout = client.get(LOGOUT)
self.register = client.get(REGISTER)
def test_no_user_in_db(self):
"""No user i db."""
self.assertFalse(User.objects.count())
def test_homepage(self):
"""Test homepage can be reached."""
self.assertEqual(self.home.status_code, 200)
def test_login(self):
"""Test login cna be reached."""
self.assertEqual(self.login.status_code, 200)
def test_logout(self):
"""Test logout can be reached."""
self.assertEqual(self.logout.status_code, 200)
def test_register(self):
"""Test register can be reached."""
self.assertEqual(self.register.status_code, 200)
def test_default_image(self):
"""Test default image shows up."""
img_path = self.home.context['image']
self.assertEqual(img_path, DEFAULT_IMAGE)
|
from django.test import Client, TestCase
from django.contrib.auth.models import User
HOME = '/'
REGISTER = '/accounts/register/'
LOGIN = '/login'
LOGOUT = '/logout'
class UnauthenticatedUser(TestCase):
"""Create unauth user for testing."""
def setUp(self):
"""Setup unauth user."""
client = Client()
self.home = client.get(HOME)
self.login = client.get(LOGIN)
self.logout = client.get(LOGOUT)
self.register = client.get(REGISTER)
def test_no_user_in_db(self):
"""No user i db."""
self.assertFalse(User.objects.count())
def test_homepage(self):
"""Test homepage can be reached."""
self.assertEqual(self.home.status_code, 200)
def test_login(self):
"""Test login cna be reached."""
self.assertEqual(self.login.status_code, 200)
def test_logout(self):
"""Test logout can be reached."""
self.assertEqual(self.logout.status_code, 200)
def test_register(self):
"""Test register can be reached."""
self.assertEqual(self.register.status_code, 200)
Add passing test for default image"""Tests for project level urls and views."""
from __future__ import unicode_literals
from django.contrib.staticfiles import finders
from django.test import Client, TestCase
from django.contrib.auth.models import User
HOME = '/'
REGISTER = '/accounts/register/'
LOGIN = '/login'
LOGOUT = '/logout'
DEFAULT_IMAGE = finders.find('static/imagersite/images/default-image.jpg')
class UnauthenticatedUser(TestCase):
"""Create unauth user for testing."""
def setUp(self):
"""Setup unauth user."""
client = Client()
self.home = client.get(HOME)
self.login = client.get(LOGIN)
self.logout = client.get(LOGOUT)
self.register = client.get(REGISTER)
def test_no_user_in_db(self):
"""No user i db."""
self.assertFalse(User.objects.count())
def test_homepage(self):
"""Test homepage can be reached."""
self.assertEqual(self.home.status_code, 200)
def test_login(self):
"""Test login cna be reached."""
self.assertEqual(self.login.status_code, 200)
def test_logout(self):
"""Test logout can be reached."""
self.assertEqual(self.logout.status_code, 200)
def test_register(self):
"""Test register can be reached."""
self.assertEqual(self.register.status_code, 200)
def test_default_image(self):
"""Test default image shows up."""
img_path = self.home.context['image']
self.assertEqual(img_path, DEFAULT_IMAGE)
|
<commit_before>from django.test import Client, TestCase
from django.contrib.auth.models import User
HOME = '/'
REGISTER = '/accounts/register/'
LOGIN = '/login'
LOGOUT = '/logout'
class UnauthenticatedUser(TestCase):
"""Create unauth user for testing."""
def setUp(self):
"""Setup unauth user."""
client = Client()
self.home = client.get(HOME)
self.login = client.get(LOGIN)
self.logout = client.get(LOGOUT)
self.register = client.get(REGISTER)
def test_no_user_in_db(self):
"""No user i db."""
self.assertFalse(User.objects.count())
def test_homepage(self):
"""Test homepage can be reached."""
self.assertEqual(self.home.status_code, 200)
def test_login(self):
"""Test login cna be reached."""
self.assertEqual(self.login.status_code, 200)
def test_logout(self):
"""Test logout can be reached."""
self.assertEqual(self.logout.status_code, 200)
def test_register(self):
"""Test register can be reached."""
self.assertEqual(self.register.status_code, 200)
<commit_msg>Add passing test for default image<commit_after>"""Tests for project level urls and views."""
from __future__ import unicode_literals
from django.contrib.staticfiles import finders
from django.test import Client, TestCase
from django.contrib.auth.models import User
HOME = '/'
REGISTER = '/accounts/register/'
LOGIN = '/login'
LOGOUT = '/logout'
DEFAULT_IMAGE = finders.find('static/imagersite/images/default-image.jpg')
class UnauthenticatedUser(TestCase):
"""Create unauth user for testing."""
def setUp(self):
"""Setup unauth user."""
client = Client()
self.home = client.get(HOME)
self.login = client.get(LOGIN)
self.logout = client.get(LOGOUT)
self.register = client.get(REGISTER)
def test_no_user_in_db(self):
"""No user i db."""
self.assertFalse(User.objects.count())
def test_homepage(self):
"""Test homepage can be reached."""
self.assertEqual(self.home.status_code, 200)
def test_login(self):
"""Test login cna be reached."""
self.assertEqual(self.login.status_code, 200)
def test_logout(self):
"""Test logout can be reached."""
self.assertEqual(self.logout.status_code, 200)
def test_register(self):
"""Test register can be reached."""
self.assertEqual(self.register.status_code, 200)
def test_default_image(self):
"""Test default image shows up."""
img_path = self.home.context['image']
self.assertEqual(img_path, DEFAULT_IMAGE)
|
24c24ab8a6c662079c397a8d91228b3b8d45f033
|
testing/test_integration.py
|
testing/test_integration.py
|
import sys
from setuptools_scm.utils import do
def test_pyproject_support(tmpdir, monkeypatch):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.ensure("package", dir=42)
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]
fallback_version = "12.34"
"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "12.34"
|
import sys
import pytest
from setuptools_scm.utils import do
@pytest.fixture
def wd(wd):
wd("git init")
wd("git config user.email test@example.com")
wd('git config user.name "a test"')
wd.add_command = "git add ."
wd.commit_command = "git commit -m test-{reason}"
return wd
def test_pyproject_support(tmpdir, monkeypatch):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.ensure("package", dir=42)
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]
fallback_version = "12.34"
"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "12.34"
@pytest.mark.xfail
def test_pyproject_support_with_git(tmpdir, monkeypatch, wd):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.join("wd")
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "0.1.dev0"
|
Test pyproject.toml integration using git
|
Test pyproject.toml integration using git
See #374.
|
Python
|
mit
|
pypa/setuptools_scm,pypa/setuptools_scm,RonnyPfannschmidt/setuptools_scm,RonnyPfannschmidt/setuptools_scm
|
import sys
from setuptools_scm.utils import do
def test_pyproject_support(tmpdir, monkeypatch):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.ensure("package", dir=42)
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]
fallback_version = "12.34"
"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "12.34"
Test pyproject.toml integration using git
See #374.
|
import sys
import pytest
from setuptools_scm.utils import do
@pytest.fixture
def wd(wd):
wd("git init")
wd("git config user.email test@example.com")
wd('git config user.name "a test"')
wd.add_command = "git add ."
wd.commit_command = "git commit -m test-{reason}"
return wd
def test_pyproject_support(tmpdir, monkeypatch):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.ensure("package", dir=42)
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]
fallback_version = "12.34"
"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "12.34"
@pytest.mark.xfail
def test_pyproject_support_with_git(tmpdir, monkeypatch, wd):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.join("wd")
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "0.1.dev0"
|
<commit_before>import sys
from setuptools_scm.utils import do
def test_pyproject_support(tmpdir, monkeypatch):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.ensure("package", dir=42)
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]
fallback_version = "12.34"
"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "12.34"
<commit_msg>Test pyproject.toml integration using git
See #374.<commit_after>
|
import sys
import pytest
from setuptools_scm.utils import do
@pytest.fixture
def wd(wd):
wd("git init")
wd("git config user.email test@example.com")
wd('git config user.name "a test"')
wd.add_command = "git add ."
wd.commit_command = "git commit -m test-{reason}"
return wd
def test_pyproject_support(tmpdir, monkeypatch):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.ensure("package", dir=42)
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]
fallback_version = "12.34"
"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "12.34"
@pytest.mark.xfail
def test_pyproject_support_with_git(tmpdir, monkeypatch, wd):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.join("wd")
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "0.1.dev0"
|
import sys
from setuptools_scm.utils import do
def test_pyproject_support(tmpdir, monkeypatch):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.ensure("package", dir=42)
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]
fallback_version = "12.34"
"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "12.34"
Test pyproject.toml integration using git
See #374.import sys
import pytest
from setuptools_scm.utils import do
@pytest.fixture
def wd(wd):
wd("git init")
wd("git config user.email test@example.com")
wd('git config user.name "a test"')
wd.add_command = "git add ."
wd.commit_command = "git commit -m test-{reason}"
return wd
def test_pyproject_support(tmpdir, monkeypatch):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.ensure("package", dir=42)
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]
fallback_version = "12.34"
"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "12.34"
@pytest.mark.xfail
def test_pyproject_support_with_git(tmpdir, monkeypatch, wd):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.join("wd")
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "0.1.dev0"
|
<commit_before>import sys
from setuptools_scm.utils import do
def test_pyproject_support(tmpdir, monkeypatch):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.ensure("package", dir=42)
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]
fallback_version = "12.34"
"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "12.34"
<commit_msg>Test pyproject.toml integration using git
See #374.<commit_after>import sys
import pytest
from setuptools_scm.utils import do
@pytest.fixture
def wd(wd):
wd("git init")
wd("git config user.email test@example.com")
wd('git config user.name "a test"')
wd.add_command = "git add ."
wd.commit_command = "git commit -m test-{reason}"
return wd
def test_pyproject_support(tmpdir, monkeypatch):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.ensure("package", dir=42)
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]
fallback_version = "12.34"
"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "12.34"
@pytest.mark.xfail
def test_pyproject_support_with_git(tmpdir, monkeypatch, wd):
monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG")
pkg = tmpdir.join("wd")
pkg.join("pyproject.toml").write(
"""[tool.setuptools_scm]"""
)
pkg.join("setup.py").write("__import__('setuptools').setup()")
res = do((sys.executable, "setup.py", "--version"), pkg)
assert res == "0.1.dev0"
|
81936bfbac9254fcd90d294c299ad635504cd93c
|
police_api/service.py
|
police_api/service.py
|
import logging
import requests
from .exceptions import APIError
from .version import __version__
logger = logging.getLogger(__name__)
class BaseService(object):
def __init__(self, api, **config):
self.api = api
self.config = {
'base_url': 'http://data.police.uk/api/',
'user_agent': 'police-api-client-python/%s' % __version__,
}
self.config.update(config)
def raise_for_status(self, request):
try:
request.raise_for_status()
except requests.models.HTTPError as e:
raise APIError(e)
def request(self, verb, method, **kwargs):
verb = verb.upper()
request_kwargs = {
'headers': {
'User-Agent': self.config['user_agent'],
},
'timeout': self.config.get('timeout', 30),
}
if 'username' in self.config:
request_kwargs['auth'] = (self.config.get('username', ''),
self.config.get('password', ''))
if verb == 'GET':
request_kwargs['params'] = kwargs
else:
request_kwargs['data'] = kwargs
url = self.config['base_url'] + method
logger.debug('%s %s' % (verb, url))
r = requests.request(verb, url, **request_kwargs)
self.raise_for_status(r)
return r.json()
|
import logging
import requests
from .exceptions import APIError
from .version import __version__
logger = logging.getLogger(__name__)
class BaseService(object):
def __init__(self, api, **config):
self.api = api
self.config = {
'base_url': 'http://data.police.uk/api/',
'user_agent': 'police-api-client-python/%s' % __version__,
}
self.config.update(config)
def raise_for_status(self, request):
try:
request.raise_for_status()
except requests.models.HTTPError as e:
raise APIError(e)
def _make_request(self, verb, url, params={}):
request_kwargs = {
'headers': {
'User-Agent': self.config['user_agent'],
},
'timeout': self.config.get('timeout', 30),
}
if 'username' in self.config:
request_kwargs['auth'] = (self.config.get('username', ''),
self.config.get('password', ''))
if verb == 'GET':
request_kwargs['params'] = params
else:
request_kwargs['data'] = params
logger.debug('%s %s' % (verb, url))
r = requests.request(verb, url, **request_kwargs)
self.raise_for_status(r)
return r.json()
def request(self, verb, method, **kwargs):
url = self.config['base_url'] + method
return self._make_request(verb.upper(), url, kwargs)
|
Refactor request mechanics into an internal method on BaseService
|
Refactor request mechanics into an internal method on BaseService
|
Python
|
mit
|
rkhleics/police-api-client-python
|
import logging
import requests
from .exceptions import APIError
from .version import __version__
logger = logging.getLogger(__name__)
class BaseService(object):
def __init__(self, api, **config):
self.api = api
self.config = {
'base_url': 'http://data.police.uk/api/',
'user_agent': 'police-api-client-python/%s' % __version__,
}
self.config.update(config)
def raise_for_status(self, request):
try:
request.raise_for_status()
except requests.models.HTTPError as e:
raise APIError(e)
def request(self, verb, method, **kwargs):
verb = verb.upper()
request_kwargs = {
'headers': {
'User-Agent': self.config['user_agent'],
},
'timeout': self.config.get('timeout', 30),
}
if 'username' in self.config:
request_kwargs['auth'] = (self.config.get('username', ''),
self.config.get('password', ''))
if verb == 'GET':
request_kwargs['params'] = kwargs
else:
request_kwargs['data'] = kwargs
url = self.config['base_url'] + method
logger.debug('%s %s' % (verb, url))
r = requests.request(verb, url, **request_kwargs)
self.raise_for_status(r)
return r.json()
Refactor request mechanics into an internal method on BaseService
|
import logging
import requests
from .exceptions import APIError
from .version import __version__
logger = logging.getLogger(__name__)
class BaseService(object):
def __init__(self, api, **config):
self.api = api
self.config = {
'base_url': 'http://data.police.uk/api/',
'user_agent': 'police-api-client-python/%s' % __version__,
}
self.config.update(config)
def raise_for_status(self, request):
try:
request.raise_for_status()
except requests.models.HTTPError as e:
raise APIError(e)
def _make_request(self, verb, url, params={}):
request_kwargs = {
'headers': {
'User-Agent': self.config['user_agent'],
},
'timeout': self.config.get('timeout', 30),
}
if 'username' in self.config:
request_kwargs['auth'] = (self.config.get('username', ''),
self.config.get('password', ''))
if verb == 'GET':
request_kwargs['params'] = params
else:
request_kwargs['data'] = params
logger.debug('%s %s' % (verb, url))
r = requests.request(verb, url, **request_kwargs)
self.raise_for_status(r)
return r.json()
def request(self, verb, method, **kwargs):
url = self.config['base_url'] + method
return self._make_request(verb.upper(), url, kwargs)
|
<commit_before>import logging
import requests
from .exceptions import APIError
from .version import __version__
logger = logging.getLogger(__name__)
class BaseService(object):
def __init__(self, api, **config):
self.api = api
self.config = {
'base_url': 'http://data.police.uk/api/',
'user_agent': 'police-api-client-python/%s' % __version__,
}
self.config.update(config)
def raise_for_status(self, request):
try:
request.raise_for_status()
except requests.models.HTTPError as e:
raise APIError(e)
def request(self, verb, method, **kwargs):
verb = verb.upper()
request_kwargs = {
'headers': {
'User-Agent': self.config['user_agent'],
},
'timeout': self.config.get('timeout', 30),
}
if 'username' in self.config:
request_kwargs['auth'] = (self.config.get('username', ''),
self.config.get('password', ''))
if verb == 'GET':
request_kwargs['params'] = kwargs
else:
request_kwargs['data'] = kwargs
url = self.config['base_url'] + method
logger.debug('%s %s' % (verb, url))
r = requests.request(verb, url, **request_kwargs)
self.raise_for_status(r)
return r.json()
<commit_msg>Refactor request mechanics into an internal method on BaseService<commit_after>
|
import logging
import requests
from .exceptions import APIError
from .version import __version__
logger = logging.getLogger(__name__)
class BaseService(object):
def __init__(self, api, **config):
self.api = api
self.config = {
'base_url': 'http://data.police.uk/api/',
'user_agent': 'police-api-client-python/%s' % __version__,
}
self.config.update(config)
def raise_for_status(self, request):
try:
request.raise_for_status()
except requests.models.HTTPError as e:
raise APIError(e)
def _make_request(self, verb, url, params={}):
request_kwargs = {
'headers': {
'User-Agent': self.config['user_agent'],
},
'timeout': self.config.get('timeout', 30),
}
if 'username' in self.config:
request_kwargs['auth'] = (self.config.get('username', ''),
self.config.get('password', ''))
if verb == 'GET':
request_kwargs['params'] = params
else:
request_kwargs['data'] = params
logger.debug('%s %s' % (verb, url))
r = requests.request(verb, url, **request_kwargs)
self.raise_for_status(r)
return r.json()
def request(self, verb, method, **kwargs):
url = self.config['base_url'] + method
return self._make_request(verb.upper(), url, kwargs)
|
import logging
import requests
from .exceptions import APIError
from .version import __version__
logger = logging.getLogger(__name__)
class BaseService(object):
def __init__(self, api, **config):
self.api = api
self.config = {
'base_url': 'http://data.police.uk/api/',
'user_agent': 'police-api-client-python/%s' % __version__,
}
self.config.update(config)
def raise_for_status(self, request):
try:
request.raise_for_status()
except requests.models.HTTPError as e:
raise APIError(e)
def request(self, verb, method, **kwargs):
verb = verb.upper()
request_kwargs = {
'headers': {
'User-Agent': self.config['user_agent'],
},
'timeout': self.config.get('timeout', 30),
}
if 'username' in self.config:
request_kwargs['auth'] = (self.config.get('username', ''),
self.config.get('password', ''))
if verb == 'GET':
request_kwargs['params'] = kwargs
else:
request_kwargs['data'] = kwargs
url = self.config['base_url'] + method
logger.debug('%s %s' % (verb, url))
r = requests.request(verb, url, **request_kwargs)
self.raise_for_status(r)
return r.json()
Refactor request mechanics into an internal method on BaseServiceimport logging
import requests
from .exceptions import APIError
from .version import __version__
logger = logging.getLogger(__name__)
class BaseService(object):
def __init__(self, api, **config):
self.api = api
self.config = {
'base_url': 'http://data.police.uk/api/',
'user_agent': 'police-api-client-python/%s' % __version__,
}
self.config.update(config)
def raise_for_status(self, request):
try:
request.raise_for_status()
except requests.models.HTTPError as e:
raise APIError(e)
def _make_request(self, verb, url, params={}):
request_kwargs = {
'headers': {
'User-Agent': self.config['user_agent'],
},
'timeout': self.config.get('timeout', 30),
}
if 'username' in self.config:
request_kwargs['auth'] = (self.config.get('username', ''),
self.config.get('password', ''))
if verb == 'GET':
request_kwargs['params'] = params
else:
request_kwargs['data'] = params
logger.debug('%s %s' % (verb, url))
r = requests.request(verb, url, **request_kwargs)
self.raise_for_status(r)
return r.json()
def request(self, verb, method, **kwargs):
url = self.config['base_url'] + method
return self._make_request(verb.upper(), url, kwargs)
|
<commit_before>import logging
import requests
from .exceptions import APIError
from .version import __version__
logger = logging.getLogger(__name__)
class BaseService(object):
def __init__(self, api, **config):
self.api = api
self.config = {
'base_url': 'http://data.police.uk/api/',
'user_agent': 'police-api-client-python/%s' % __version__,
}
self.config.update(config)
def raise_for_status(self, request):
try:
request.raise_for_status()
except requests.models.HTTPError as e:
raise APIError(e)
def request(self, verb, method, **kwargs):
verb = verb.upper()
request_kwargs = {
'headers': {
'User-Agent': self.config['user_agent'],
},
'timeout': self.config.get('timeout', 30),
}
if 'username' in self.config:
request_kwargs['auth'] = (self.config.get('username', ''),
self.config.get('password', ''))
if verb == 'GET':
request_kwargs['params'] = kwargs
else:
request_kwargs['data'] = kwargs
url = self.config['base_url'] + method
logger.debug('%s %s' % (verb, url))
r = requests.request(verb, url, **request_kwargs)
self.raise_for_status(r)
return r.json()
<commit_msg>Refactor request mechanics into an internal method on BaseService<commit_after>import logging
import requests
from .exceptions import APIError
from .version import __version__
logger = logging.getLogger(__name__)
class BaseService(object):
def __init__(self, api, **config):
self.api = api
self.config = {
'base_url': 'http://data.police.uk/api/',
'user_agent': 'police-api-client-python/%s' % __version__,
}
self.config.update(config)
def raise_for_status(self, request):
try:
request.raise_for_status()
except requests.models.HTTPError as e:
raise APIError(e)
def _make_request(self, verb, url, params={}):
request_kwargs = {
'headers': {
'User-Agent': self.config['user_agent'],
},
'timeout': self.config.get('timeout', 30),
}
if 'username' in self.config:
request_kwargs['auth'] = (self.config.get('username', ''),
self.config.get('password', ''))
if verb == 'GET':
request_kwargs['params'] = params
else:
request_kwargs['data'] = params
logger.debug('%s %s' % (verb, url))
r = requests.request(verb, url, **request_kwargs)
self.raise_for_status(r)
return r.json()
def request(self, verb, method, **kwargs):
url = self.config['base_url'] + method
return self._make_request(verb.upper(), url, kwargs)
|
178c25714aaae056c115f1580f19d833486a54ac
|
datapipe/targets/objects.py
|
datapipe/targets/objects.py
|
from ..target import Target
import hashlib
import dill
import joblib
class PyTarget(Target):
def __init__(self, name, obj=None):
self._name = name
self._obj = obj
super(PyTarget, self).__init__()
if not obj is None:
self.set(obj)
def identifier(self):
return self._name
def get(self):
return self._obj
def set(self, obj):
self._obj = obj
def is_damaged(self):
stored = self.stored()
if stored:
if self._obj is None:
self._obj = stored._obj
return stored._obj is None
else:
return joblib.hash(self._obj) != joblib.hash(stored._obj)
else:
return self._obj is None
|
from ..target import Target
import hashlib
import dill
import joblib
class PyTarget(Target):
def __init__(self, name, obj=None):
self._name = name
self._obj = obj
super(PyTarget, self).__init__()
if not obj is None:
self.set(obj)
def identifier(self):
return self._name
def get(self):
return self._obj
def set(self, obj):
self._obj = obj
self._memory['obj'] = dill.dumps(obj).encode('base64')
def is_damaged(self):
mem = self.stored()
if mem:
if self._obj is None:
self._obj = dill.loads(mem['obj'].decode('base64'))
return stored._obj is None
else:
return joblib.hash(self._obj) != joblib.hash(dill.loads(mem['obj'].decode('base64')))
else:
return self._obj is None
|
Make PyTarget object work again
|
Make PyTarget object work again
We now save a base64 encoded pickled version of the object.
|
Python
|
mit
|
ibab/datapipe
|
from ..target import Target
import hashlib
import dill
import joblib
class PyTarget(Target):
def __init__(self, name, obj=None):
self._name = name
self._obj = obj
super(PyTarget, self).__init__()
if not obj is None:
self.set(obj)
def identifier(self):
return self._name
def get(self):
return self._obj
def set(self, obj):
self._obj = obj
def is_damaged(self):
stored = self.stored()
if stored:
if self._obj is None:
self._obj = stored._obj
return stored._obj is None
else:
return joblib.hash(self._obj) != joblib.hash(stored._obj)
else:
return self._obj is None
Make PyTarget object work again
We now save a base64 encoded pickled version of the object.
|
from ..target import Target
import hashlib
import dill
import joblib
class PyTarget(Target):
def __init__(self, name, obj=None):
self._name = name
self._obj = obj
super(PyTarget, self).__init__()
if not obj is None:
self.set(obj)
def identifier(self):
return self._name
def get(self):
return self._obj
def set(self, obj):
self._obj = obj
self._memory['obj'] = dill.dumps(obj).encode('base64')
def is_damaged(self):
mem = self.stored()
if mem:
if self._obj is None:
self._obj = dill.loads(mem['obj'].decode('base64'))
return stored._obj is None
else:
return joblib.hash(self._obj) != joblib.hash(dill.loads(mem['obj'].decode('base64')))
else:
return self._obj is None
|
<commit_before>from ..target import Target
import hashlib
import dill
import joblib
class PyTarget(Target):
def __init__(self, name, obj=None):
self._name = name
self._obj = obj
super(PyTarget, self).__init__()
if not obj is None:
self.set(obj)
def identifier(self):
return self._name
def get(self):
return self._obj
def set(self, obj):
self._obj = obj
def is_damaged(self):
stored = self.stored()
if stored:
if self._obj is None:
self._obj = stored._obj
return stored._obj is None
else:
return joblib.hash(self._obj) != joblib.hash(stored._obj)
else:
return self._obj is None
<commit_msg>Make PyTarget object work again
We now save a base64 encoded pickled version of the object.<commit_after>
|
from ..target import Target
import hashlib
import dill
import joblib
class PyTarget(Target):
def __init__(self, name, obj=None):
self._name = name
self._obj = obj
super(PyTarget, self).__init__()
if not obj is None:
self.set(obj)
def identifier(self):
return self._name
def get(self):
return self._obj
def set(self, obj):
self._obj = obj
self._memory['obj'] = dill.dumps(obj).encode('base64')
def is_damaged(self):
mem = self.stored()
if mem:
if self._obj is None:
self._obj = dill.loads(mem['obj'].decode('base64'))
return stored._obj is None
else:
return joblib.hash(self._obj) != joblib.hash(dill.loads(mem['obj'].decode('base64')))
else:
return self._obj is None
|
from ..target import Target
import hashlib
import dill
import joblib
class PyTarget(Target):
def __init__(self, name, obj=None):
self._name = name
self._obj = obj
super(PyTarget, self).__init__()
if not obj is None:
self.set(obj)
def identifier(self):
return self._name
def get(self):
return self._obj
def set(self, obj):
self._obj = obj
def is_damaged(self):
stored = self.stored()
if stored:
if self._obj is None:
self._obj = stored._obj
return stored._obj is None
else:
return joblib.hash(self._obj) != joblib.hash(stored._obj)
else:
return self._obj is None
Make PyTarget object work again
We now save a base64 encoded pickled version of the object.from ..target import Target
import hashlib
import dill
import joblib
class PyTarget(Target):
def __init__(self, name, obj=None):
self._name = name
self._obj = obj
super(PyTarget, self).__init__()
if not obj is None:
self.set(obj)
def identifier(self):
return self._name
def get(self):
return self._obj
def set(self, obj):
self._obj = obj
self._memory['obj'] = dill.dumps(obj).encode('base64')
def is_damaged(self):
mem = self.stored()
if mem:
if self._obj is None:
self._obj = dill.loads(mem['obj'].decode('base64'))
return stored._obj is None
else:
return joblib.hash(self._obj) != joblib.hash(dill.loads(mem['obj'].decode('base64')))
else:
return self._obj is None
|
<commit_before>from ..target import Target
import hashlib
import dill
import joblib
class PyTarget(Target):
def __init__(self, name, obj=None):
self._name = name
self._obj = obj
super(PyTarget, self).__init__()
if not obj is None:
self.set(obj)
def identifier(self):
return self._name
def get(self):
return self._obj
def set(self, obj):
self._obj = obj
def is_damaged(self):
stored = self.stored()
if stored:
if self._obj is None:
self._obj = stored._obj
return stored._obj is None
else:
return joblib.hash(self._obj) != joblib.hash(stored._obj)
else:
return self._obj is None
<commit_msg>Make PyTarget object work again
We now save a base64 encoded pickled version of the object.<commit_after>from ..target import Target
import hashlib
import dill
import joblib
class PyTarget(Target):
def __init__(self, name, obj=None):
self._name = name
self._obj = obj
super(PyTarget, self).__init__()
if not obj is None:
self.set(obj)
def identifier(self):
return self._name
def get(self):
return self._obj
def set(self, obj):
self._obj = obj
self._memory['obj'] = dill.dumps(obj).encode('base64')
def is_damaged(self):
mem = self.stored()
if mem:
if self._obj is None:
self._obj = dill.loads(mem['obj'].decode('base64'))
return stored._obj is None
else:
return joblib.hash(self._obj) != joblib.hash(dill.loads(mem['obj'].decode('base64')))
else:
return self._obj is None
|
4d3f809ba5e1b5109e6f2e73d9c9630371660210
|
Bookie/bookie/lib/access.py
|
Bookie/bookie/lib/access.py
|
"""Handle auth and authz activities in bookie"""
from pyramid.httpexceptions import HTTPForbidden
class Authorize(object):
"""Context manager to check if the user is authorized
use:
with Authorize(some_key):
# do work
Will return NotAuthorized if it fails
"""
def __init__(self, submitted_key, config_key):
"""Create the context manager"""
self.api_key = config_key
self.check_key = submitted_key
def __enter__(self):
"""Verify api key set in constructor"""
if self.api_key != self.check_key:
raise HTTPForbidden('Invalid Authorization')
def __exit__(self, exc_type, exc_value, traceback):
"""No cleanup work to do after usage"""
pass
|
"""Handle auth and authz activities in bookie"""
import logging
from pyramid.httpexceptions import HTTPForbidden
LOG = logging.getLogger(__name__)
class Authorize(object):
"""Context manager to check if the user is authorized
use:
with Authorize(some_key):
# do work
Will return NotAuthorized if it fails
"""
def __init__(self, submitted_key, config_key):
"""Create the context manager"""
self.api_key = config_key
self.check_key = submitted_key
def __enter__(self):
"""Verify api key set in constructor"""
if self.api_key != self.check_key:
LOG.error('Invalid API Key! {0} v {1}'.format(self.api_key,
self.check_key))
raise HTTPForbidden('Invalid Authorization')
def __exit__(self, exc_type, exc_value, traceback):
"""No cleanup work to do after usage"""
pass
|
Update to make sure we log an error with an invalid key
|
Update to make sure we log an error with an invalid key
|
Python
|
agpl-3.0
|
GreenLunar/Bookie,adamlincoln/Bookie,wangjun/Bookie,bookieio/Bookie,GreenLunar/Bookie,teodesson/Bookie,adamlincoln/Bookie,pombredanne/Bookie,bookieio/Bookie,teodesson/Bookie,adamlincoln/Bookie,GreenLunar/Bookie,wangjun/Bookie,wangjun/Bookie,skmezanul/Bookie,wangjun/Bookie,GreenLunar/Bookie,skmezanul/Bookie,adamlincoln/Bookie,pombredanne/Bookie,skmezanul/Bookie,bookieio/Bookie,charany1/Bookie,teodesson/Bookie,skmezanul/Bookie,charany1/Bookie,pombredanne/Bookie,charany1/Bookie,teodesson/Bookie,bookieio/Bookie
|
"""Handle auth and authz activities in bookie"""
from pyramid.httpexceptions import HTTPForbidden
class Authorize(object):
"""Context manager to check if the user is authorized
use:
with Authorize(some_key):
# do work
Will return NotAuthorized if it fails
"""
def __init__(self, submitted_key, config_key):
"""Create the context manager"""
self.api_key = config_key
self.check_key = submitted_key
def __enter__(self):
"""Verify api key set in constructor"""
if self.api_key != self.check_key:
raise HTTPForbidden('Invalid Authorization')
def __exit__(self, exc_type, exc_value, traceback):
"""No cleanup work to do after usage"""
pass
Update to make sure we log an error with an invalid key
|
"""Handle auth and authz activities in bookie"""
import logging
from pyramid.httpexceptions import HTTPForbidden
LOG = logging.getLogger(__name__)
class Authorize(object):
"""Context manager to check if the user is authorized
use:
with Authorize(some_key):
# do work
Will return NotAuthorized if it fails
"""
def __init__(self, submitted_key, config_key):
"""Create the context manager"""
self.api_key = config_key
self.check_key = submitted_key
def __enter__(self):
"""Verify api key set in constructor"""
if self.api_key != self.check_key:
LOG.error('Invalid API Key! {0} v {1}'.format(self.api_key,
self.check_key))
raise HTTPForbidden('Invalid Authorization')
def __exit__(self, exc_type, exc_value, traceback):
"""No cleanup work to do after usage"""
pass
|
<commit_before>"""Handle auth and authz activities in bookie"""
from pyramid.httpexceptions import HTTPForbidden
class Authorize(object):
"""Context manager to check if the user is authorized
use:
with Authorize(some_key):
# do work
Will return NotAuthorized if it fails
"""
def __init__(self, submitted_key, config_key):
"""Create the context manager"""
self.api_key = config_key
self.check_key = submitted_key
def __enter__(self):
"""Verify api key set in constructor"""
if self.api_key != self.check_key:
raise HTTPForbidden('Invalid Authorization')
def __exit__(self, exc_type, exc_value, traceback):
"""No cleanup work to do after usage"""
pass
<commit_msg>Update to make sure we log an error with an invalid key<commit_after>
|
"""Handle auth and authz activities in bookie"""
import logging
from pyramid.httpexceptions import HTTPForbidden
LOG = logging.getLogger(__name__)
class Authorize(object):
"""Context manager to check if the user is authorized
use:
with Authorize(some_key):
# do work
Will return NotAuthorized if it fails
"""
def __init__(self, submitted_key, config_key):
"""Create the context manager"""
self.api_key = config_key
self.check_key = submitted_key
def __enter__(self):
"""Verify api key set in constructor"""
if self.api_key != self.check_key:
LOG.error('Invalid API Key! {0} v {1}'.format(self.api_key,
self.check_key))
raise HTTPForbidden('Invalid Authorization')
def __exit__(self, exc_type, exc_value, traceback):
"""No cleanup work to do after usage"""
pass
|
"""Handle auth and authz activities in bookie"""
from pyramid.httpexceptions import HTTPForbidden
class Authorize(object):
"""Context manager to check if the user is authorized
use:
with Authorize(some_key):
# do work
Will return NotAuthorized if it fails
"""
def __init__(self, submitted_key, config_key):
"""Create the context manager"""
self.api_key = config_key
self.check_key = submitted_key
def __enter__(self):
"""Verify api key set in constructor"""
if self.api_key != self.check_key:
raise HTTPForbidden('Invalid Authorization')
def __exit__(self, exc_type, exc_value, traceback):
"""No cleanup work to do after usage"""
pass
Update to make sure we log an error with an invalid key"""Handle auth and authz activities in bookie"""
import logging
from pyramid.httpexceptions import HTTPForbidden
LOG = logging.getLogger(__name__)
class Authorize(object):
"""Context manager to check if the user is authorized
use:
with Authorize(some_key):
# do work
Will return NotAuthorized if it fails
"""
def __init__(self, submitted_key, config_key):
"""Create the context manager"""
self.api_key = config_key
self.check_key = submitted_key
def __enter__(self):
"""Verify api key set in constructor"""
if self.api_key != self.check_key:
LOG.error('Invalid API Key! {0} v {1}'.format(self.api_key,
self.check_key))
raise HTTPForbidden('Invalid Authorization')
def __exit__(self, exc_type, exc_value, traceback):
"""No cleanup work to do after usage"""
pass
|
<commit_before>"""Handle auth and authz activities in bookie"""
from pyramid.httpexceptions import HTTPForbidden
class Authorize(object):
"""Context manager to check if the user is authorized
use:
with Authorize(some_key):
# do work
Will return NotAuthorized if it fails
"""
def __init__(self, submitted_key, config_key):
"""Create the context manager"""
self.api_key = config_key
self.check_key = submitted_key
def __enter__(self):
"""Verify api key set in constructor"""
if self.api_key != self.check_key:
raise HTTPForbidden('Invalid Authorization')
def __exit__(self, exc_type, exc_value, traceback):
"""No cleanup work to do after usage"""
pass
<commit_msg>Update to make sure we log an error with an invalid key<commit_after>"""Handle auth and authz activities in bookie"""
import logging
from pyramid.httpexceptions import HTTPForbidden
LOG = logging.getLogger(__name__)
class Authorize(object):
"""Context manager to check if the user is authorized
use:
with Authorize(some_key):
# do work
Will return NotAuthorized if it fails
"""
def __init__(self, submitted_key, config_key):
"""Create the context manager"""
self.api_key = config_key
self.check_key = submitted_key
def __enter__(self):
"""Verify api key set in constructor"""
if self.api_key != self.check_key:
LOG.error('Invalid API Key! {0} v {1}'.format(self.api_key,
self.check_key))
raise HTTPForbidden('Invalid Authorization')
def __exit__(self, exc_type, exc_value, traceback):
"""No cleanup work to do after usage"""
pass
|
8b4ea06ae8c61f0745a13e4c0118d6f499a31738
|
app.py
|
app.py
|
from flask import Flask, jsonify, request
from dotenv import load_dotenv, find_dotenv
from twilio import twiml
from Queue import Queue
from threading import Thread
from time import sleep
load_dotenv(find_dotenv())
directions = ['forward', 'backward']
task_q = Queue()
def send_rasp(task_q):
while True:
sleep(2)
if not task_q.empty():
message = task_q.get()
print(message)
rasp_signal = Thread(target=send_rasp, args=(task_q, ))
rasp_signal.setDaemon(True)
rasp_signal.start()
app = Flask(__name__)
@app.route('/message', methods=['POST'])
def roomba_command():
# twilio text message
body = request.form['Body']
resp = handle_twilio_message(body)
twilio_resp = twiml.Response()
twilio_resp.message(resp)
return str(twilio_resp)
def handle_twilio_message(message):
if message.lower() in directions:
task_q.put(message.lower())
return 'Message sent'
try:
degree = float(message)
except ValueError as e:
return 'Invalid command'
task_q.put(str(degree))
return 'Message sent'
if __name__ == '__main__':
app.run(debug=True)
|
from flask import Flask, jsonify, request
from dotenv import load_dotenv, find_dotenv
from twilio import twiml
from Queue import Queue
from threading import Thread
from time import sleep
load_dotenv(find_dotenv())
directions = ['forward', 'backward']
task_q = Queue()
def send_rasp(task_q):
while True:
sleep(2)
if task_q.empty():
continue
message = task_q.get()
print(message)
rasp_signal = Thread(target=send_rasp, args=(task_q, ))
rasp_signal.setDaemon(True)
rasp_signal.start()
app = Flask(__name__)
@app.route('/message', methods=['POST'])
def roomba_command():
# twilio text message
body = request.form['Body']
resp = handle_twilio_message(body)
twilio_resp = twiml.Response()
twilio_resp.message(resp)
return str(twilio_resp)
def handle_twilio_message(message):
if message.lower() in directions:
task_q.put(message.lower())
return 'Message sent'
try:
degree = float(message)
except ValueError as e:
return 'Invalid command'
task_q.put(str(degree))
return 'Message sent'
if __name__ == '__main__':
app.run(debug=True)
|
Use continue in task loop
|
Use continue in task loop
|
Python
|
mit
|
tforrest/twilio-plays-roomba-flask
|
from flask import Flask, jsonify, request
from dotenv import load_dotenv, find_dotenv
from twilio import twiml
from Queue import Queue
from threading import Thread
from time import sleep
load_dotenv(find_dotenv())
directions = ['forward', 'backward']
task_q = Queue()
def send_rasp(task_q):
while True:
sleep(2)
if not task_q.empty():
message = task_q.get()
print(message)
rasp_signal = Thread(target=send_rasp, args=(task_q, ))
rasp_signal.setDaemon(True)
rasp_signal.start()
app = Flask(__name__)
@app.route('/message', methods=['POST'])
def roomba_command():
# twilio text message
body = request.form['Body']
resp = handle_twilio_message(body)
twilio_resp = twiml.Response()
twilio_resp.message(resp)
return str(twilio_resp)
def handle_twilio_message(message):
if message.lower() in directions:
task_q.put(message.lower())
return 'Message sent'
try:
degree = float(message)
except ValueError as e:
return 'Invalid command'
task_q.put(str(degree))
return 'Message sent'
if __name__ == '__main__':
app.run(debug=True)Use continue in task loop
|
from flask import Flask, jsonify, request
from dotenv import load_dotenv, find_dotenv
from twilio import twiml
from Queue import Queue
from threading import Thread
from time import sleep
load_dotenv(find_dotenv())
directions = ['forward', 'backward']
task_q = Queue()
def send_rasp(task_q):
while True:
sleep(2)
if task_q.empty():
continue
message = task_q.get()
print(message)
rasp_signal = Thread(target=send_rasp, args=(task_q, ))
rasp_signal.setDaemon(True)
rasp_signal.start()
app = Flask(__name__)
@app.route('/message', methods=['POST'])
def roomba_command():
# twilio text message
body = request.form['Body']
resp = handle_twilio_message(body)
twilio_resp = twiml.Response()
twilio_resp.message(resp)
return str(twilio_resp)
def handle_twilio_message(message):
if message.lower() in directions:
task_q.put(message.lower())
return 'Message sent'
try:
degree = float(message)
except ValueError as e:
return 'Invalid command'
task_q.put(str(degree))
return 'Message sent'
if __name__ == '__main__':
app.run(debug=True)
|
<commit_before>from flask import Flask, jsonify, request
from dotenv import load_dotenv, find_dotenv
from twilio import twiml
from Queue import Queue
from threading import Thread
from time import sleep
load_dotenv(find_dotenv())
directions = ['forward', 'backward']
task_q = Queue()
def send_rasp(task_q):
while True:
sleep(2)
if not task_q.empty():
message = task_q.get()
print(message)
rasp_signal = Thread(target=send_rasp, args=(task_q, ))
rasp_signal.setDaemon(True)
rasp_signal.start()
app = Flask(__name__)
@app.route('/message', methods=['POST'])
def roomba_command():
# twilio text message
body = request.form['Body']
resp = handle_twilio_message(body)
twilio_resp = twiml.Response()
twilio_resp.message(resp)
return str(twilio_resp)
def handle_twilio_message(message):
if message.lower() in directions:
task_q.put(message.lower())
return 'Message sent'
try:
degree = float(message)
except ValueError as e:
return 'Invalid command'
task_q.put(str(degree))
return 'Message sent'
if __name__ == '__main__':
app.run(debug=True)<commit_msg>Use continue in task loop<commit_after>
|
from flask import Flask, jsonify, request
from dotenv import load_dotenv, find_dotenv
from twilio import twiml
from Queue import Queue
from threading import Thread
from time import sleep
load_dotenv(find_dotenv())
directions = ['forward', 'backward']
task_q = Queue()
def send_rasp(task_q):
while True:
sleep(2)
if task_q.empty():
continue
message = task_q.get()
print(message)
rasp_signal = Thread(target=send_rasp, args=(task_q, ))
rasp_signal.setDaemon(True)
rasp_signal.start()
app = Flask(__name__)
@app.route('/message', methods=['POST'])
def roomba_command():
# twilio text message
body = request.form['Body']
resp = handle_twilio_message(body)
twilio_resp = twiml.Response()
twilio_resp.message(resp)
return str(twilio_resp)
def handle_twilio_message(message):
if message.lower() in directions:
task_q.put(message.lower())
return 'Message sent'
try:
degree = float(message)
except ValueError as e:
return 'Invalid command'
task_q.put(str(degree))
return 'Message sent'
if __name__ == '__main__':
app.run(debug=True)
|
from flask import Flask, jsonify, request
from dotenv import load_dotenv, find_dotenv
from twilio import twiml
from Queue import Queue
from threading import Thread
from time import sleep
load_dotenv(find_dotenv())
directions = ['forward', 'backward']
task_q = Queue()
def send_rasp(task_q):
while True:
sleep(2)
if not task_q.empty():
message = task_q.get()
print(message)
rasp_signal = Thread(target=send_rasp, args=(task_q, ))
rasp_signal.setDaemon(True)
rasp_signal.start()
app = Flask(__name__)
@app.route('/message', methods=['POST'])
def roomba_command():
# twilio text message
body = request.form['Body']
resp = handle_twilio_message(body)
twilio_resp = twiml.Response()
twilio_resp.message(resp)
return str(twilio_resp)
def handle_twilio_message(message):
if message.lower() in directions:
task_q.put(message.lower())
return 'Message sent'
try:
degree = float(message)
except ValueError as e:
return 'Invalid command'
task_q.put(str(degree))
return 'Message sent'
if __name__ == '__main__':
app.run(debug=True)Use continue in task loopfrom flask import Flask, jsonify, request
from dotenv import load_dotenv, find_dotenv
from twilio import twiml
from Queue import Queue
from threading import Thread
from time import sleep
load_dotenv(find_dotenv())
directions = ['forward', 'backward']
task_q = Queue()
def send_rasp(task_q):
while True:
sleep(2)
if task_q.empty():
continue
message = task_q.get()
print(message)
rasp_signal = Thread(target=send_rasp, args=(task_q, ))
rasp_signal.setDaemon(True)
rasp_signal.start()
app = Flask(__name__)
@app.route('/message', methods=['POST'])
def roomba_command():
# twilio text message
body = request.form['Body']
resp = handle_twilio_message(body)
twilio_resp = twiml.Response()
twilio_resp.message(resp)
return str(twilio_resp)
def handle_twilio_message(message):
if message.lower() in directions:
task_q.put(message.lower())
return 'Message sent'
try:
degree = float(message)
except ValueError as e:
return 'Invalid command'
task_q.put(str(degree))
return 'Message sent'
if __name__ == '__main__':
app.run(debug=True)
|
<commit_before>from flask import Flask, jsonify, request
from dotenv import load_dotenv, find_dotenv
from twilio import twiml
from Queue import Queue
from threading import Thread
from time import sleep
load_dotenv(find_dotenv())
directions = ['forward', 'backward']
task_q = Queue()
def send_rasp(task_q):
while True:
sleep(2)
if not task_q.empty():
message = task_q.get()
print(message)
rasp_signal = Thread(target=send_rasp, args=(task_q, ))
rasp_signal.setDaemon(True)
rasp_signal.start()
app = Flask(__name__)
@app.route('/message', methods=['POST'])
def roomba_command():
# twilio text message
body = request.form['Body']
resp = handle_twilio_message(body)
twilio_resp = twiml.Response()
twilio_resp.message(resp)
return str(twilio_resp)
def handle_twilio_message(message):
if message.lower() in directions:
task_q.put(message.lower())
return 'Message sent'
try:
degree = float(message)
except ValueError as e:
return 'Invalid command'
task_q.put(str(degree))
return 'Message sent'
if __name__ == '__main__':
app.run(debug=True)<commit_msg>Use continue in task loop<commit_after>from flask import Flask, jsonify, request
from dotenv import load_dotenv, find_dotenv
from twilio import twiml
from Queue import Queue
from threading import Thread
from time import sleep
load_dotenv(find_dotenv())
directions = ['forward', 'backward']
task_q = Queue()
def send_rasp(task_q):
while True:
sleep(2)
if task_q.empty():
continue
message = task_q.get()
print(message)
rasp_signal = Thread(target=send_rasp, args=(task_q, ))
rasp_signal.setDaemon(True)
rasp_signal.start()
app = Flask(__name__)
@app.route('/message', methods=['POST'])
def roomba_command():
# twilio text message
body = request.form['Body']
resp = handle_twilio_message(body)
twilio_resp = twiml.Response()
twilio_resp.message(resp)
return str(twilio_resp)
def handle_twilio_message(message):
if message.lower() in directions:
task_q.put(message.lower())
return 'Message sent'
try:
degree = float(message)
except ValueError as e:
return 'Invalid command'
task_q.put(str(degree))
return 'Message sent'
if __name__ == '__main__':
app.run(debug=True)
|
0389fabdb0343b189b153cc909b05e88d3830b94
|
downstream_node/lib/node.py
|
downstream_node/lib/node.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from downstream_node.config import config
from downstream_node.models import Challenges
from heartbeat import Heartbeat
from downstream_node.startup import db
__all__ = ['create_token', 'delete_token', 'add_file', 'remove_file',
'gen_challenges', 'update_challenges']
def create_token(*args, **kwargs):
raise NotImplementedError
def delete_token(*args, **kwargs):
raise NotImplementedError
def add_file(*args, **kwargs):
raise NotImplementedError
def remove_file(*args, **kwargs):
raise NotImplementedError
def gen_challenges(filepath, root_seed):
secret = getattr(config, 'HEARTBEAT_SECRET')
hb = Heartbeat(filepath, secret=secret)
hb.generate_challenges(1000, root_seed)
for challenge in hb.challenges:
chal = Challenges(
filepath=filepath,
root_seed=root_seed,
block=challenge.block,
seed=challenge.seed,
response=challenge.response,
)
db.session.add(chal)
db.session.commit()
def update_challenges(*args, **kwargs):
raise NotImplementedError
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from downstream_node.config import config
from downstream_node.models import Challenges, Files
from heartbeat import Heartbeat
from downstream_node.startup import db
__all__ = ['create_token', 'delete_token', 'add_file', 'remove_file',
'gen_challenges', 'update_challenges']
def create_token(*args, **kwargs):
raise NotImplementedError
def delete_token(*args, **kwargs):
raise NotImplementedError
def add_file(*args, **kwargs):
raise NotImplementedError
def remove_file(*args, **kwargs):
raise NotImplementedError
def gen_challenges(filepath, root_seed):
secret = getattr(config, 'HEARTBEAT_SECRET')
hb = Heartbeat(filepath, secret=secret)
hb.generate_challenges(1000, root_seed)
files = Files(name=filepath)
db.session.add(files)
for challenge in hb.challenges:
chal = Challenges(
filename=filepath,
rootseed=root_seed,
block=challenge.block,
seed=challenge.seed,
response=challenge.response,
)
db.session.add(chal)
db.session.commit()
def update_challenges(*args, **kwargs):
raise NotImplementedError
|
Fix for new column names
|
Fix for new column names
|
Python
|
mit
|
Storj/downstream-node,Storj/downstream-node
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from downstream_node.config import config
from downstream_node.models import Challenges
from heartbeat import Heartbeat
from downstream_node.startup import db
__all__ = ['create_token', 'delete_token', 'add_file', 'remove_file',
'gen_challenges', 'update_challenges']
def create_token(*args, **kwargs):
raise NotImplementedError
def delete_token(*args, **kwargs):
raise NotImplementedError
def add_file(*args, **kwargs):
raise NotImplementedError
def remove_file(*args, **kwargs):
raise NotImplementedError
def gen_challenges(filepath, root_seed):
secret = getattr(config, 'HEARTBEAT_SECRET')
hb = Heartbeat(filepath, secret=secret)
hb.generate_challenges(1000, root_seed)
for challenge in hb.challenges:
chal = Challenges(
filepath=filepath,
root_seed=root_seed,
block=challenge.block,
seed=challenge.seed,
response=challenge.response,
)
db.session.add(chal)
db.session.commit()
def update_challenges(*args, **kwargs):
raise NotImplementedError
Fix for new column names
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from downstream_node.config import config
from downstream_node.models import Challenges, Files
from heartbeat import Heartbeat
from downstream_node.startup import db
__all__ = ['create_token', 'delete_token', 'add_file', 'remove_file',
'gen_challenges', 'update_challenges']
def create_token(*args, **kwargs):
raise NotImplementedError
def delete_token(*args, **kwargs):
raise NotImplementedError
def add_file(*args, **kwargs):
raise NotImplementedError
def remove_file(*args, **kwargs):
raise NotImplementedError
def gen_challenges(filepath, root_seed):
secret = getattr(config, 'HEARTBEAT_SECRET')
hb = Heartbeat(filepath, secret=secret)
hb.generate_challenges(1000, root_seed)
files = Files(name=filepath)
db.session.add(files)
for challenge in hb.challenges:
chal = Challenges(
filename=filepath,
rootseed=root_seed,
block=challenge.block,
seed=challenge.seed,
response=challenge.response,
)
db.session.add(chal)
db.session.commit()
def update_challenges(*args, **kwargs):
raise NotImplementedError
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from downstream_node.config import config
from downstream_node.models import Challenges
from heartbeat import Heartbeat
from downstream_node.startup import db
__all__ = ['create_token', 'delete_token', 'add_file', 'remove_file',
'gen_challenges', 'update_challenges']
def create_token(*args, **kwargs):
raise NotImplementedError
def delete_token(*args, **kwargs):
raise NotImplementedError
def add_file(*args, **kwargs):
raise NotImplementedError
def remove_file(*args, **kwargs):
raise NotImplementedError
def gen_challenges(filepath, root_seed):
secret = getattr(config, 'HEARTBEAT_SECRET')
hb = Heartbeat(filepath, secret=secret)
hb.generate_challenges(1000, root_seed)
for challenge in hb.challenges:
chal = Challenges(
filepath=filepath,
root_seed=root_seed,
block=challenge.block,
seed=challenge.seed,
response=challenge.response,
)
db.session.add(chal)
db.session.commit()
def update_challenges(*args, **kwargs):
raise NotImplementedError
<commit_msg>Fix for new column names<commit_after>
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from downstream_node.config import config
from downstream_node.models import Challenges, Files
from heartbeat import Heartbeat
from downstream_node.startup import db
__all__ = ['create_token', 'delete_token', 'add_file', 'remove_file',
'gen_challenges', 'update_challenges']
def create_token(*args, **kwargs):
raise NotImplementedError
def delete_token(*args, **kwargs):
raise NotImplementedError
def add_file(*args, **kwargs):
raise NotImplementedError
def remove_file(*args, **kwargs):
raise NotImplementedError
def gen_challenges(filepath, root_seed):
secret = getattr(config, 'HEARTBEAT_SECRET')
hb = Heartbeat(filepath, secret=secret)
hb.generate_challenges(1000, root_seed)
files = Files(name=filepath)
db.session.add(files)
for challenge in hb.challenges:
chal = Challenges(
filename=filepath,
rootseed=root_seed,
block=challenge.block,
seed=challenge.seed,
response=challenge.response,
)
db.session.add(chal)
db.session.commit()
def update_challenges(*args, **kwargs):
raise NotImplementedError
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from downstream_node.config import config
from downstream_node.models import Challenges
from heartbeat import Heartbeat
from downstream_node.startup import db
__all__ = ['create_token', 'delete_token', 'add_file', 'remove_file',
'gen_challenges', 'update_challenges']
def create_token(*args, **kwargs):
raise NotImplementedError
def delete_token(*args, **kwargs):
raise NotImplementedError
def add_file(*args, **kwargs):
raise NotImplementedError
def remove_file(*args, **kwargs):
raise NotImplementedError
def gen_challenges(filepath, root_seed):
secret = getattr(config, 'HEARTBEAT_SECRET')
hb = Heartbeat(filepath, secret=secret)
hb.generate_challenges(1000, root_seed)
for challenge in hb.challenges:
chal = Challenges(
filepath=filepath,
root_seed=root_seed,
block=challenge.block,
seed=challenge.seed,
response=challenge.response,
)
db.session.add(chal)
db.session.commit()
def update_challenges(*args, **kwargs):
raise NotImplementedError
Fix for new column names#!/usr/bin/env python
# -*- coding: utf-8 -*-
from downstream_node.config import config
from downstream_node.models import Challenges, Files
from heartbeat import Heartbeat
from downstream_node.startup import db
__all__ = ['create_token', 'delete_token', 'add_file', 'remove_file',
'gen_challenges', 'update_challenges']
def create_token(*args, **kwargs):
raise NotImplementedError
def delete_token(*args, **kwargs):
raise NotImplementedError
def add_file(*args, **kwargs):
raise NotImplementedError
def remove_file(*args, **kwargs):
raise NotImplementedError
def gen_challenges(filepath, root_seed):
secret = getattr(config, 'HEARTBEAT_SECRET')
hb = Heartbeat(filepath, secret=secret)
hb.generate_challenges(1000, root_seed)
files = Files(name=filepath)
db.session.add(files)
for challenge in hb.challenges:
chal = Challenges(
filename=filepath,
rootseed=root_seed,
block=challenge.block,
seed=challenge.seed,
response=challenge.response,
)
db.session.add(chal)
db.session.commit()
def update_challenges(*args, **kwargs):
raise NotImplementedError
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from downstream_node.config import config
from downstream_node.models import Challenges
from heartbeat import Heartbeat
from downstream_node.startup import db
__all__ = ['create_token', 'delete_token', 'add_file', 'remove_file',
'gen_challenges', 'update_challenges']
def create_token(*args, **kwargs):
raise NotImplementedError
def delete_token(*args, **kwargs):
raise NotImplementedError
def add_file(*args, **kwargs):
raise NotImplementedError
def remove_file(*args, **kwargs):
raise NotImplementedError
def gen_challenges(filepath, root_seed):
secret = getattr(config, 'HEARTBEAT_SECRET')
hb = Heartbeat(filepath, secret=secret)
hb.generate_challenges(1000, root_seed)
for challenge in hb.challenges:
chal = Challenges(
filepath=filepath,
root_seed=root_seed,
block=challenge.block,
seed=challenge.seed,
response=challenge.response,
)
db.session.add(chal)
db.session.commit()
def update_challenges(*args, **kwargs):
raise NotImplementedError
<commit_msg>Fix for new column names<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from downstream_node.config import config
from downstream_node.models import Challenges, Files
from heartbeat import Heartbeat
from downstream_node.startup import db
__all__ = ['create_token', 'delete_token', 'add_file', 'remove_file',
'gen_challenges', 'update_challenges']
def create_token(*args, **kwargs):
raise NotImplementedError
def delete_token(*args, **kwargs):
raise NotImplementedError
def add_file(*args, **kwargs):
raise NotImplementedError
def remove_file(*args, **kwargs):
raise NotImplementedError
def gen_challenges(filepath, root_seed):
secret = getattr(config, 'HEARTBEAT_SECRET')
hb = Heartbeat(filepath, secret=secret)
hb.generate_challenges(1000, root_seed)
files = Files(name=filepath)
db.session.add(files)
for challenge in hb.challenges:
chal = Challenges(
filename=filepath,
rootseed=root_seed,
block=challenge.block,
seed=challenge.seed,
response=challenge.response,
)
db.session.add(chal)
db.session.commit()
def update_challenges(*args, **kwargs):
raise NotImplementedError
|
efc1988d704a7a1231046dea8af65dcdba7897fd
|
py/fbx_write.py
|
py/fbx_write.py
|
# !/usr/bin/env python
# Blender has moved to Python 3!
import sys
import os
import bpy
for sysarg in sys.argv:
print(sysarg)
py_args = sys.argv[sys.argv.index('--') + 1]
py_args = py_args.split(' ')
for arg in py_args:
if (arg.startswith('basedir:')):
basedir = arg.split('basedir:')[1]
else:
# can supply filename(s) with or without extension
pdb_code = os.path.splitext(arg)[0]
abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl')
# This is the base directory, used for saving files
molecule = bpy.ops.import_scene.x3d(
filepath = abs_file_in
)
abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx')
bpy.ops.export_scene.fbx(
filepath = abs_file_out
)
|
# !/usr/bin/env python
# Blender has moved to Python 3!
import sys
import os
import bpy
for sysarg in sys.argv:
print(sysarg)
py_args = sys.argv[sys.argv.index('--') + 1]
py_args = py_args.split(' ')
for arg in py_args:
if (arg.startswith('basedir:')):
basedir = arg.split('basedir:')[1]
else:
# can supply filename(s) with or without extension
pdb_code = os.path.splitext(arg)[0]
abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl')
# This is the base directory, used for saving files
molecule = bpy.ops.import_scene.x3d(
filepath = abs_file_in
)
abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx')
bpy.ops.export_scene.fbx(
filepath = abs_file_out
)
bpy.ops.wm.quit_blender()
|
Quit Blender after writing FBX
|
Quit Blender after writing FBX
|
Python
|
mit
|
hackmcr15-code-a-la-mode/mol-vis-hack,hackmcr15-code-a-la-mode/mol-vis-hack
|
# !/usr/bin/env python
# Blender has moved to Python 3!
import sys
import os
import bpy
for sysarg in sys.argv:
print(sysarg)
py_args = sys.argv[sys.argv.index('--') + 1]
py_args = py_args.split(' ')
for arg in py_args:
if (arg.startswith('basedir:')):
basedir = arg.split('basedir:')[1]
else:
# can supply filename(s) with or without extension
pdb_code = os.path.splitext(arg)[0]
abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl')
# This is the base directory, used for saving files
molecule = bpy.ops.import_scene.x3d(
filepath = abs_file_in
)
abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx')
bpy.ops.export_scene.fbx(
filepath = abs_file_out
)
Quit Blender after writing FBX
|
# !/usr/bin/env python
# Blender has moved to Python 3!
import sys
import os
import bpy
for sysarg in sys.argv:
print(sysarg)
py_args = sys.argv[sys.argv.index('--') + 1]
py_args = py_args.split(' ')
for arg in py_args:
if (arg.startswith('basedir:')):
basedir = arg.split('basedir:')[1]
else:
# can supply filename(s) with or without extension
pdb_code = os.path.splitext(arg)[0]
abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl')
# This is the base directory, used for saving files
molecule = bpy.ops.import_scene.x3d(
filepath = abs_file_in
)
abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx')
bpy.ops.export_scene.fbx(
filepath = abs_file_out
)
bpy.ops.wm.quit_blender()
|
<commit_before># !/usr/bin/env python
# Blender has moved to Python 3!
import sys
import os
import bpy
for sysarg in sys.argv:
print(sysarg)
py_args = sys.argv[sys.argv.index('--') + 1]
py_args = py_args.split(' ')
for arg in py_args:
if (arg.startswith('basedir:')):
basedir = arg.split('basedir:')[1]
else:
# can supply filename(s) with or without extension
pdb_code = os.path.splitext(arg)[0]
abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl')
# This is the base directory, used for saving files
molecule = bpy.ops.import_scene.x3d(
filepath = abs_file_in
)
abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx')
bpy.ops.export_scene.fbx(
filepath = abs_file_out
)
<commit_msg>Quit Blender after writing FBX<commit_after>
|
# !/usr/bin/env python
# Blender has moved to Python 3!
import sys
import os
import bpy
for sysarg in sys.argv:
print(sysarg)
py_args = sys.argv[sys.argv.index('--') + 1]
py_args = py_args.split(' ')
for arg in py_args:
if (arg.startswith('basedir:')):
basedir = arg.split('basedir:')[1]
else:
# can supply filename(s) with or without extension
pdb_code = os.path.splitext(arg)[0]
abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl')
# This is the base directory, used for saving files
molecule = bpy.ops.import_scene.x3d(
filepath = abs_file_in
)
abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx')
bpy.ops.export_scene.fbx(
filepath = abs_file_out
)
bpy.ops.wm.quit_blender()
|
# !/usr/bin/env python
# Blender has moved to Python 3!
import sys
import os
import bpy
for sysarg in sys.argv:
print(sysarg)
py_args = sys.argv[sys.argv.index('--') + 1]
py_args = py_args.split(' ')
for arg in py_args:
if (arg.startswith('basedir:')):
basedir = arg.split('basedir:')[1]
else:
# can supply filename(s) with or without extension
pdb_code = os.path.splitext(arg)[0]
abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl')
# This is the base directory, used for saving files
molecule = bpy.ops.import_scene.x3d(
filepath = abs_file_in
)
abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx')
bpy.ops.export_scene.fbx(
filepath = abs_file_out
)
Quit Blender after writing FBX# !/usr/bin/env python
# Blender has moved to Python 3!
import sys
import os
import bpy
for sysarg in sys.argv:
print(sysarg)
py_args = sys.argv[sys.argv.index('--') + 1]
py_args = py_args.split(' ')
for arg in py_args:
if (arg.startswith('basedir:')):
basedir = arg.split('basedir:')[1]
else:
# can supply filename(s) with or without extension
pdb_code = os.path.splitext(arg)[0]
abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl')
# This is the base directory, used for saving files
molecule = bpy.ops.import_scene.x3d(
filepath = abs_file_in
)
abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx')
bpy.ops.export_scene.fbx(
filepath = abs_file_out
)
bpy.ops.wm.quit_blender()
|
<commit_before># !/usr/bin/env python
# Blender has moved to Python 3!
import sys
import os
import bpy
for sysarg in sys.argv:
print(sysarg)
py_args = sys.argv[sys.argv.index('--') + 1]
py_args = py_args.split(' ')
for arg in py_args:
if (arg.startswith('basedir:')):
basedir = arg.split('basedir:')[1]
else:
# can supply filename(s) with or without extension
pdb_code = os.path.splitext(arg)[0]
abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl')
# This is the base directory, used for saving files
molecule = bpy.ops.import_scene.x3d(
filepath = abs_file_in
)
abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx')
bpy.ops.export_scene.fbx(
filepath = abs_file_out
)
<commit_msg>Quit Blender after writing FBX<commit_after># !/usr/bin/env python
# Blender has moved to Python 3!
import sys
import os
import bpy
for sysarg in sys.argv:
print(sysarg)
py_args = sys.argv[sys.argv.index('--') + 1]
py_args = py_args.split(' ')
for arg in py_args:
if (arg.startswith('basedir:')):
basedir = arg.split('basedir:')[1]
else:
# can supply filename(s) with or without extension
pdb_code = os.path.splitext(arg)[0]
abs_file_in = os.path.join(basedir, 'structures/wrl', pdb_code+'.wrl')
# This is the base directory, used for saving files
molecule = bpy.ops.import_scene.x3d(
filepath = abs_file_in
)
abs_file_out = os.path.join(basedir,'structures/fbx',pdb_code+'.fbx')
bpy.ops.export_scene.fbx(
filepath = abs_file_out
)
bpy.ops.wm.quit_blender()
|
1f3fce7cb415e739bdb745295807cceaf853d176
|
easy_thumbnails/__init__.py
|
easy_thumbnails/__init__.py
|
VERSION = (1, 0, 'alpha', 11)
def get_version(join=' ', short=False):
"""
Return the version of this package as a string.
The version number is built from a ``VERSION`` tuple, which should consist
of integers, or trailing version information (such as 'alpha', 'beta' or
'final'). For example:
>>> VERSION = (2, 0, 6)
>>> get_version()
'2.0.6'
>>> VERSION = (1, 0, 'beta', 2)
>>> get_version()
'1.0 beta 2'
Use the ``join`` argument to join the version elements by an alternate
character to the default ``' '``. This is useful when building a distutils
setup module::
from this_package import get_version
setup(
version=get_version(join='-'),
# ...
)
Use the ``short`` argument to get the version number without trailing
version information.
"""
version = []
number = []
remainder = []
for i, bit in enumerate(VERSION):
if isinstance(bit, int):
number.append(str(bit))
else:
remainder = [str(bit) for bit in VERSION[i:]]
break
if number:
version.append('.'.join(number))
if not short:
if remainder == ['alpha', 0]:
version.append('pre-alpha')
elif 'final' not in remainder:
version.extend(remainder)
return join.join(version)
|
VERSION = (1, 0, 'alpha', 12)
def get_version(join=' ', short=False):
"""
Return the version of this package as a string.
The version number is built from a ``VERSION`` tuple, which should consist
of integers, or trailing version information (such as 'alpha', 'beta' or
'final'). For example:
>>> VERSION = (2, 0, 6)
>>> get_version()
'2.0.6'
>>> VERSION = (1, 0, 'beta', 2)
>>> get_version()
'1.0 beta 2'
Use the ``join`` argument to join the version elements by an alternate
character to the default ``' '``. This is useful when building a distutils
setup module::
from this_package import get_version
setup(
version=get_version(join='-'),
# ...
)
Use the ``short`` argument to get the version number without trailing
version information.
"""
version = []
number = []
remainder = []
for i, bit in enumerate(VERSION):
if isinstance(bit, int):
number.append(str(bit))
else:
remainder = [str(bit) for bit in VERSION[i:]]
break
if number:
version.append('.'.join(number))
if not short:
if remainder == ['alpha', 0]:
version.append('pre-alpha')
elif 'final' not in remainder:
version.extend(remainder)
return join.join(version)
|
Bump version number for another release with ducktyping in it
|
Bump version number for another release with ducktyping in it
|
Python
|
bsd-3-clause
|
sandow-digital/easy-thumbnails-cropman,jrief/easy-thumbnails,Mactory/easy-thumbnails,emschorsch/easy-thumbnails,sandow-digital/easy-thumbnails-cropman,siovene/easy-thumbnails,jrief/easy-thumbnails,emschorsch/easy-thumbnails,SmileyChris/easy-thumbnails,jaddison/easy-thumbnails
|
VERSION = (1, 0, 'alpha', 11)
def get_version(join=' ', short=False):
"""
Return the version of this package as a string.
The version number is built from a ``VERSION`` tuple, which should consist
of integers, or trailing version information (such as 'alpha', 'beta' or
'final'). For example:
>>> VERSION = (2, 0, 6)
>>> get_version()
'2.0.6'
>>> VERSION = (1, 0, 'beta', 2)
>>> get_version()
'1.0 beta 2'
Use the ``join`` argument to join the version elements by an alternate
character to the default ``' '``. This is useful when building a distutils
setup module::
from this_package import get_version
setup(
version=get_version(join='-'),
# ...
)
Use the ``short`` argument to get the version number without trailing
version information.
"""
version = []
number = []
remainder = []
for i, bit in enumerate(VERSION):
if isinstance(bit, int):
number.append(str(bit))
else:
remainder = [str(bit) for bit in VERSION[i:]]
break
if number:
version.append('.'.join(number))
if not short:
if remainder == ['alpha', 0]:
version.append('pre-alpha')
elif 'final' not in remainder:
version.extend(remainder)
return join.join(version)
Bump version number for another release with ducktyping in it
|
VERSION = (1, 0, 'alpha', 12)
def get_version(join=' ', short=False):
"""
Return the version of this package as a string.
The version number is built from a ``VERSION`` tuple, which should consist
of integers, or trailing version information (such as 'alpha', 'beta' or
'final'). For example:
>>> VERSION = (2, 0, 6)
>>> get_version()
'2.0.6'
>>> VERSION = (1, 0, 'beta', 2)
>>> get_version()
'1.0 beta 2'
Use the ``join`` argument to join the version elements by an alternate
character to the default ``' '``. This is useful when building a distutils
setup module::
from this_package import get_version
setup(
version=get_version(join='-'),
# ...
)
Use the ``short`` argument to get the version number without trailing
version information.
"""
version = []
number = []
remainder = []
for i, bit in enumerate(VERSION):
if isinstance(bit, int):
number.append(str(bit))
else:
remainder = [str(bit) for bit in VERSION[i:]]
break
if number:
version.append('.'.join(number))
if not short:
if remainder == ['alpha', 0]:
version.append('pre-alpha')
elif 'final' not in remainder:
version.extend(remainder)
return join.join(version)
|
<commit_before>VERSION = (1, 0, 'alpha', 11)
def get_version(join=' ', short=False):
"""
Return the version of this package as a string.
The version number is built from a ``VERSION`` tuple, which should consist
of integers, or trailing version information (such as 'alpha', 'beta' or
'final'). For example:
>>> VERSION = (2, 0, 6)
>>> get_version()
'2.0.6'
>>> VERSION = (1, 0, 'beta', 2)
>>> get_version()
'1.0 beta 2'
Use the ``join`` argument to join the version elements by an alternate
character to the default ``' '``. This is useful when building a distutils
setup module::
from this_package import get_version
setup(
version=get_version(join='-'),
# ...
)
Use the ``short`` argument to get the version number without trailing
version information.
"""
version = []
number = []
remainder = []
for i, bit in enumerate(VERSION):
if isinstance(bit, int):
number.append(str(bit))
else:
remainder = [str(bit) for bit in VERSION[i:]]
break
if number:
version.append('.'.join(number))
if not short:
if remainder == ['alpha', 0]:
version.append('pre-alpha')
elif 'final' not in remainder:
version.extend(remainder)
return join.join(version)
<commit_msg>Bump version number for another release with ducktyping in it<commit_after>
|
VERSION = (1, 0, 'alpha', 12)
def get_version(join=' ', short=False):
"""
Return the version of this package as a string.
The version number is built from a ``VERSION`` tuple, which should consist
of integers, or trailing version information (such as 'alpha', 'beta' or
'final'). For example:
>>> VERSION = (2, 0, 6)
>>> get_version()
'2.0.6'
>>> VERSION = (1, 0, 'beta', 2)
>>> get_version()
'1.0 beta 2'
Use the ``join`` argument to join the version elements by an alternate
character to the default ``' '``. This is useful when building a distutils
setup module::
from this_package import get_version
setup(
version=get_version(join='-'),
# ...
)
Use the ``short`` argument to get the version number without trailing
version information.
"""
version = []
number = []
remainder = []
for i, bit in enumerate(VERSION):
if isinstance(bit, int):
number.append(str(bit))
else:
remainder = [str(bit) for bit in VERSION[i:]]
break
if number:
version.append('.'.join(number))
if not short:
if remainder == ['alpha', 0]:
version.append('pre-alpha')
elif 'final' not in remainder:
version.extend(remainder)
return join.join(version)
|
VERSION = (1, 0, 'alpha', 11)
def get_version(join=' ', short=False):
"""
Return the version of this package as a string.
The version number is built from a ``VERSION`` tuple, which should consist
of integers, or trailing version information (such as 'alpha', 'beta' or
'final'). For example:
>>> VERSION = (2, 0, 6)
>>> get_version()
'2.0.6'
>>> VERSION = (1, 0, 'beta', 2)
>>> get_version()
'1.0 beta 2'
Use the ``join`` argument to join the version elements by an alternate
character to the default ``' '``. This is useful when building a distutils
setup module::
from this_package import get_version
setup(
version=get_version(join='-'),
# ...
)
Use the ``short`` argument to get the version number without trailing
version information.
"""
version = []
number = []
remainder = []
for i, bit in enumerate(VERSION):
if isinstance(bit, int):
number.append(str(bit))
else:
remainder = [str(bit) for bit in VERSION[i:]]
break
if number:
version.append('.'.join(number))
if not short:
if remainder == ['alpha', 0]:
version.append('pre-alpha')
elif 'final' not in remainder:
version.extend(remainder)
return join.join(version)
Bump version number for another release with ducktyping in itVERSION = (1, 0, 'alpha', 12)
def get_version(join=' ', short=False):
"""
Return the version of this package as a string.
The version number is built from a ``VERSION`` tuple, which should consist
of integers, or trailing version information (such as 'alpha', 'beta' or
'final'). For example:
>>> VERSION = (2, 0, 6)
>>> get_version()
'2.0.6'
>>> VERSION = (1, 0, 'beta', 2)
>>> get_version()
'1.0 beta 2'
Use the ``join`` argument to join the version elements by an alternate
character to the default ``' '``. This is useful when building a distutils
setup module::
from this_package import get_version
setup(
version=get_version(join='-'),
# ...
)
Use the ``short`` argument to get the version number without trailing
version information.
"""
version = []
number = []
remainder = []
for i, bit in enumerate(VERSION):
if isinstance(bit, int):
number.append(str(bit))
else:
remainder = [str(bit) for bit in VERSION[i:]]
break
if number:
version.append('.'.join(number))
if not short:
if remainder == ['alpha', 0]:
version.append('pre-alpha')
elif 'final' not in remainder:
version.extend(remainder)
return join.join(version)
|
<commit_before>VERSION = (1, 0, 'alpha', 11)
def get_version(join=' ', short=False):
"""
Return the version of this package as a string.
The version number is built from a ``VERSION`` tuple, which should consist
of integers, or trailing version information (such as 'alpha', 'beta' or
'final'). For example:
>>> VERSION = (2, 0, 6)
>>> get_version()
'2.0.6'
>>> VERSION = (1, 0, 'beta', 2)
>>> get_version()
'1.0 beta 2'
Use the ``join`` argument to join the version elements by an alternate
character to the default ``' '``. This is useful when building a distutils
setup module::
from this_package import get_version
setup(
version=get_version(join='-'),
# ...
)
Use the ``short`` argument to get the version number without trailing
version information.
"""
version = []
number = []
remainder = []
for i, bit in enumerate(VERSION):
if isinstance(bit, int):
number.append(str(bit))
else:
remainder = [str(bit) for bit in VERSION[i:]]
break
if number:
version.append('.'.join(number))
if not short:
if remainder == ['alpha', 0]:
version.append('pre-alpha')
elif 'final' not in remainder:
version.extend(remainder)
return join.join(version)
<commit_msg>Bump version number for another release with ducktyping in it<commit_after>VERSION = (1, 0, 'alpha', 12)
def get_version(join=' ', short=False):
"""
Return the version of this package as a string.
The version number is built from a ``VERSION`` tuple, which should consist
of integers, or trailing version information (such as 'alpha', 'beta' or
'final'). For example:
>>> VERSION = (2, 0, 6)
>>> get_version()
'2.0.6'
>>> VERSION = (1, 0, 'beta', 2)
>>> get_version()
'1.0 beta 2'
Use the ``join`` argument to join the version elements by an alternate
character to the default ``' '``. This is useful when building a distutils
setup module::
from this_package import get_version
setup(
version=get_version(join='-'),
# ...
)
Use the ``short`` argument to get the version number without trailing
version information.
"""
version = []
number = []
remainder = []
for i, bit in enumerate(VERSION):
if isinstance(bit, int):
number.append(str(bit))
else:
remainder = [str(bit) for bit in VERSION[i:]]
break
if number:
version.append('.'.join(number))
if not short:
if remainder == ['alpha', 0]:
version.append('pre-alpha')
elif 'final' not in remainder:
version.extend(remainder)
return join.join(version)
|
aca031267748358c49eac96fe158ba0a2ec3a2e8
|
tota/drawers/json_replay.py
|
tota/drawers/json_replay.py
|
import json
from os import path
from tota.game import Drawer
class JsonReplayDrawer(Drawer):
def __init__(self, replay_dir):
self.replay_dir = replay_dir
def draw(self, game):
"""Draw the world with 'ascii'-art ."""
things_data = []
tick_data = {
't': game.world.t,
'things': things_data,
'effects': [{
'position': position,
'effect': effect,
}
for position, effect in game.world.effects.items()]
}
for thing in game.world.things.values():
thing_data = {
'id': id(thing),
'type': thing.__class__.__name__,
'position': thing.position,
}
if thing_data['type'] != 'Tree':
thing_data.update({
'life': thing.life,
'name': thing.name,
'team': thing.team,
'level': getattr(thing, 'level', None),
'xp': getattr(thing, 'xp', None),
'action': thing.last_action,
'target': thing.last_target,
})
things_data.append(thing_data)
tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t)
with open(tick_path, 'w') as tick_file:
json.dump(tick_data,
tick_file,
indent=2 if game.debug else None)
|
import json
from os import path
from tota.game import Drawer
class JsonReplayDrawer(Drawer):
def __init__(self, replay_dir):
self.replay_dir = replay_dir
def draw(self, game):
"""Draw the world with 'ascii'-art ."""
things_data = []
tick_data = {
't': game.world.t,
'things': things_data,
'effects': [{
'position': position,
'effect': effect,
}
for position, effect in game.world.effects.items()]
}
for thing in game.world.things.values():
thing_data = {
'id': id(thing),
'type': thing.__class__.__name__,
'position': thing.position,
}
if thing_data['type'] != 'Tree':
thing_data.update({
'life': thing.life,
'name': thing.name,
'team': thing.team,
'level': getattr(thing, 'level', None),
'xp': getattr(thing, 'xp', None),
'action': thing.last_action,
'target': thing.last_target,
'action_done': thing.last_action_done,
})
things_data.append(thing_data)
tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t)
with open(tick_path, 'w') as tick_file:
json.dump(tick_data,
tick_file,
indent=2 if game.debug else None)
|
Add done result to replay
|
Add done result to replay
|
Python
|
mit
|
dmoisset/tota,matuu/tota,fisadev/tota
|
import json
from os import path
from tota.game import Drawer
class JsonReplayDrawer(Drawer):
def __init__(self, replay_dir):
self.replay_dir = replay_dir
def draw(self, game):
"""Draw the world with 'ascii'-art ."""
things_data = []
tick_data = {
't': game.world.t,
'things': things_data,
'effects': [{
'position': position,
'effect': effect,
}
for position, effect in game.world.effects.items()]
}
for thing in game.world.things.values():
thing_data = {
'id': id(thing),
'type': thing.__class__.__name__,
'position': thing.position,
}
if thing_data['type'] != 'Tree':
thing_data.update({
'life': thing.life,
'name': thing.name,
'team': thing.team,
'level': getattr(thing, 'level', None),
'xp': getattr(thing, 'xp', None),
'action': thing.last_action,
'target': thing.last_target,
})
things_data.append(thing_data)
tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t)
with open(tick_path, 'w') as tick_file:
json.dump(tick_data,
tick_file,
indent=2 if game.debug else None)
Add done result to replay
|
import json
from os import path
from tota.game import Drawer
class JsonReplayDrawer(Drawer):
def __init__(self, replay_dir):
self.replay_dir = replay_dir
def draw(self, game):
"""Draw the world with 'ascii'-art ."""
things_data = []
tick_data = {
't': game.world.t,
'things': things_data,
'effects': [{
'position': position,
'effect': effect,
}
for position, effect in game.world.effects.items()]
}
for thing in game.world.things.values():
thing_data = {
'id': id(thing),
'type': thing.__class__.__name__,
'position': thing.position,
}
if thing_data['type'] != 'Tree':
thing_data.update({
'life': thing.life,
'name': thing.name,
'team': thing.team,
'level': getattr(thing, 'level', None),
'xp': getattr(thing, 'xp', None),
'action': thing.last_action,
'target': thing.last_target,
'action_done': thing.last_action_done,
})
things_data.append(thing_data)
tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t)
with open(tick_path, 'w') as tick_file:
json.dump(tick_data,
tick_file,
indent=2 if game.debug else None)
|
<commit_before>import json
from os import path
from tota.game import Drawer
class JsonReplayDrawer(Drawer):
def __init__(self, replay_dir):
self.replay_dir = replay_dir
def draw(self, game):
"""Draw the world with 'ascii'-art ."""
things_data = []
tick_data = {
't': game.world.t,
'things': things_data,
'effects': [{
'position': position,
'effect': effect,
}
for position, effect in game.world.effects.items()]
}
for thing in game.world.things.values():
thing_data = {
'id': id(thing),
'type': thing.__class__.__name__,
'position': thing.position,
}
if thing_data['type'] != 'Tree':
thing_data.update({
'life': thing.life,
'name': thing.name,
'team': thing.team,
'level': getattr(thing, 'level', None),
'xp': getattr(thing, 'xp', None),
'action': thing.last_action,
'target': thing.last_target,
})
things_data.append(thing_data)
tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t)
with open(tick_path, 'w') as tick_file:
json.dump(tick_data,
tick_file,
indent=2 if game.debug else None)
<commit_msg>Add done result to replay<commit_after>
|
import json
from os import path
from tota.game import Drawer
class JsonReplayDrawer(Drawer):
def __init__(self, replay_dir):
self.replay_dir = replay_dir
def draw(self, game):
"""Draw the world with 'ascii'-art ."""
things_data = []
tick_data = {
't': game.world.t,
'things': things_data,
'effects': [{
'position': position,
'effect': effect,
}
for position, effect in game.world.effects.items()]
}
for thing in game.world.things.values():
thing_data = {
'id': id(thing),
'type': thing.__class__.__name__,
'position': thing.position,
}
if thing_data['type'] != 'Tree':
thing_data.update({
'life': thing.life,
'name': thing.name,
'team': thing.team,
'level': getattr(thing, 'level', None),
'xp': getattr(thing, 'xp', None),
'action': thing.last_action,
'target': thing.last_target,
'action_done': thing.last_action_done,
})
things_data.append(thing_data)
tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t)
with open(tick_path, 'w') as tick_file:
json.dump(tick_data,
tick_file,
indent=2 if game.debug else None)
|
import json
from os import path
from tota.game import Drawer
class JsonReplayDrawer(Drawer):
def __init__(self, replay_dir):
self.replay_dir = replay_dir
def draw(self, game):
"""Draw the world with 'ascii'-art ."""
things_data = []
tick_data = {
't': game.world.t,
'things': things_data,
'effects': [{
'position': position,
'effect': effect,
}
for position, effect in game.world.effects.items()]
}
for thing in game.world.things.values():
thing_data = {
'id': id(thing),
'type': thing.__class__.__name__,
'position': thing.position,
}
if thing_data['type'] != 'Tree':
thing_data.update({
'life': thing.life,
'name': thing.name,
'team': thing.team,
'level': getattr(thing, 'level', None),
'xp': getattr(thing, 'xp', None),
'action': thing.last_action,
'target': thing.last_target,
})
things_data.append(thing_data)
tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t)
with open(tick_path, 'w') as tick_file:
json.dump(tick_data,
tick_file,
indent=2 if game.debug else None)
Add done result to replayimport json
from os import path
from tota.game import Drawer
class JsonReplayDrawer(Drawer):
def __init__(self, replay_dir):
self.replay_dir = replay_dir
def draw(self, game):
"""Draw the world with 'ascii'-art ."""
things_data = []
tick_data = {
't': game.world.t,
'things': things_data,
'effects': [{
'position': position,
'effect': effect,
}
for position, effect in game.world.effects.items()]
}
for thing in game.world.things.values():
thing_data = {
'id': id(thing),
'type': thing.__class__.__name__,
'position': thing.position,
}
if thing_data['type'] != 'Tree':
thing_data.update({
'life': thing.life,
'name': thing.name,
'team': thing.team,
'level': getattr(thing, 'level', None),
'xp': getattr(thing, 'xp', None),
'action': thing.last_action,
'target': thing.last_target,
'action_done': thing.last_action_done,
})
things_data.append(thing_data)
tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t)
with open(tick_path, 'w') as tick_file:
json.dump(tick_data,
tick_file,
indent=2 if game.debug else None)
|
<commit_before>import json
from os import path
from tota.game import Drawer
class JsonReplayDrawer(Drawer):
def __init__(self, replay_dir):
self.replay_dir = replay_dir
def draw(self, game):
"""Draw the world with 'ascii'-art ."""
things_data = []
tick_data = {
't': game.world.t,
'things': things_data,
'effects': [{
'position': position,
'effect': effect,
}
for position, effect in game.world.effects.items()]
}
for thing in game.world.things.values():
thing_data = {
'id': id(thing),
'type': thing.__class__.__name__,
'position': thing.position,
}
if thing_data['type'] != 'Tree':
thing_data.update({
'life': thing.life,
'name': thing.name,
'team': thing.team,
'level': getattr(thing, 'level', None),
'xp': getattr(thing, 'xp', None),
'action': thing.last_action,
'target': thing.last_target,
})
things_data.append(thing_data)
tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t)
with open(tick_path, 'w') as tick_file:
json.dump(tick_data,
tick_file,
indent=2 if game.debug else None)
<commit_msg>Add done result to replay<commit_after>import json
from os import path
from tota.game import Drawer
class JsonReplayDrawer(Drawer):
def __init__(self, replay_dir):
self.replay_dir = replay_dir
def draw(self, game):
"""Draw the world with 'ascii'-art ."""
things_data = []
tick_data = {
't': game.world.t,
'things': things_data,
'effects': [{
'position': position,
'effect': effect,
}
for position, effect in game.world.effects.items()]
}
for thing in game.world.things.values():
thing_data = {
'id': id(thing),
'type': thing.__class__.__name__,
'position': thing.position,
}
if thing_data['type'] != 'Tree':
thing_data.update({
'life': thing.life,
'name': thing.name,
'team': thing.team,
'level': getattr(thing, 'level', None),
'xp': getattr(thing, 'xp', None),
'action': thing.last_action,
'target': thing.last_target,
'action_done': thing.last_action_done,
})
things_data.append(thing_data)
tick_path = path.join(self.replay_dir, '%08d.json' % game.world.t)
with open(tick_path, 'w') as tick_file:
json.dump(tick_data,
tick_file,
indent=2 if game.debug else None)
|
c3c703b411d05e6f2d52a0b3695b9dc22bc907d8
|
test/test_main.py
|
test/test_main.py
|
from mpf.main import main
def test_main():
main()
|
import matplotlib
matplotlib.use('Agg') # Not to use X server. For TravisCI.
from mpf.main import main
def test_main():
main()
|
Make matplotlib work with TravisCI
|
Make matplotlib work with TravisCI
|
Python
|
mit
|
Vayel/MPF,tartopum/MPF
|
from mpf.main import main
def test_main():
main()
Make matplotlib work with TravisCI
|
import matplotlib
matplotlib.use('Agg') # Not to use X server. For TravisCI.
from mpf.main import main
def test_main():
main()
|
<commit_before>from mpf.main import main
def test_main():
main()
<commit_msg>Make matplotlib work with TravisCI<commit_after>
|
import matplotlib
matplotlib.use('Agg') # Not to use X server. For TravisCI.
from mpf.main import main
def test_main():
main()
|
from mpf.main import main
def test_main():
main()
Make matplotlib work with TravisCIimport matplotlib
matplotlib.use('Agg') # Not to use X server. For TravisCI.
from mpf.main import main
def test_main():
main()
|
<commit_before>from mpf.main import main
def test_main():
main()
<commit_msg>Make matplotlib work with TravisCI<commit_after>import matplotlib
matplotlib.use('Agg') # Not to use X server. For TravisCI.
from mpf.main import main
def test_main():
main()
|
1a5aeabcdfae02125e167e8a221de4151819f5b5
|
test.py
|
test.py
|
import unittest
from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen
class RotorTestCase(unittest.TestCase):
def test_rotor_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('E', rotor.encode('A'))
def run_tests():
runner = unittest.TextTestRunner()
suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)
runner.run(suite)
if __name__ == '__main__': # pragma: no cover
run_tests()
|
import unittest
from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen
class RotorTestCase(unittest.TestCase):
def test_rotor_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('E', rotor.encode('A'))
def test_rotor_reverse_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('U', rotor.encode_reverse('A'))
def run_tests():
runner = unittest.TextTestRunner()
suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)
runner.run(suite)
if __name__ == '__main__': # pragma: no cover
run_tests()
|
Test if default rotor encodes backwards properly
|
Test if default rotor encodes backwards properly
|
Python
|
mit
|
ranisalt/enigma
|
import unittest
from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen
class RotorTestCase(unittest.TestCase):
def test_rotor_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('E', rotor.encode('A'))
def run_tests():
runner = unittest.TextTestRunner()
suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)
runner.run(suite)
if __name__ == '__main__': # pragma: no cover
run_tests()Test if default rotor encodes backwards properly
|
import unittest
from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen
class RotorTestCase(unittest.TestCase):
def test_rotor_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('E', rotor.encode('A'))
def test_rotor_reverse_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('U', rotor.encode_reverse('A'))
def run_tests():
runner = unittest.TextTestRunner()
suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)
runner.run(suite)
if __name__ == '__main__': # pragma: no cover
run_tests()
|
<commit_before>import unittest
from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen
class RotorTestCase(unittest.TestCase):
def test_rotor_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('E', rotor.encode('A'))
def run_tests():
runner = unittest.TextTestRunner()
suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)
runner.run(suite)
if __name__ == '__main__': # pragma: no cover
run_tests()<commit_msg>Test if default rotor encodes backwards properly<commit_after>
|
import unittest
from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen
class RotorTestCase(unittest.TestCase):
def test_rotor_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('E', rotor.encode('A'))
def test_rotor_reverse_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('U', rotor.encode_reverse('A'))
def run_tests():
runner = unittest.TextTestRunner()
suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)
runner.run(suite)
if __name__ == '__main__': # pragma: no cover
run_tests()
|
import unittest
from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen
class RotorTestCase(unittest.TestCase):
def test_rotor_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('E', rotor.encode('A'))
def run_tests():
runner = unittest.TextTestRunner()
suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)
runner.run(suite)
if __name__ == '__main__': # pragma: no cover
run_tests()Test if default rotor encodes backwards properlyimport unittest
from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen
class RotorTestCase(unittest.TestCase):
def test_rotor_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('E', rotor.encode('A'))
def test_rotor_reverse_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('U', rotor.encode_reverse('A'))
def run_tests():
runner = unittest.TextTestRunner()
suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)
runner.run(suite)
if __name__ == '__main__': # pragma: no cover
run_tests()
|
<commit_before>import unittest
from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen
class RotorTestCase(unittest.TestCase):
def test_rotor_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('E', rotor.encode('A'))
def run_tests():
runner = unittest.TextTestRunner()
suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)
runner.run(suite)
if __name__ == '__main__': # pragma: no cover
run_tests()<commit_msg>Test if default rotor encodes backwards properly<commit_after>import unittest
from enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen
class RotorTestCase(unittest.TestCase):
def test_rotor_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('E', rotor.encode('A'))
def test_rotor_reverse_encoding(self):
rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')
self.assertEqual('U', rotor.encode_reverse('A'))
def run_tests():
runner = unittest.TextTestRunner()
suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)
runner.run(suite)
if __name__ == '__main__': # pragma: no cover
run_tests()
|
1a5cc5b69811db2ac63987ab329bd117e61f3f03
|
tests/__init__.py
|
tests/__init__.py
|
import os
from functools import partial
TESTS_DIR = os.path.dirname(os.path.abspath(__file__))
TEST_DATA_DIR = os.path.join(TESTS_DIR, 'test_data')
AppEventTestXml = partial(open, os.path.join(TEST_DATA_DIR, 'app_event.xml'))
|
Make the test data accessible to the tests.
|
Make the test data accessible to the tests.
|
Python
|
bsd-3-clause
|
unt-libraries/django-premis-event-service,unt-libraries/django-premis-event-service,unt-libraries/django-premis-event-service
|
Make the test data accessible to the tests.
|
import os
from functools import partial
TESTS_DIR = os.path.dirname(os.path.abspath(__file__))
TEST_DATA_DIR = os.path.join(TESTS_DIR, 'test_data')
AppEventTestXml = partial(open, os.path.join(TEST_DATA_DIR, 'app_event.xml'))
|
<commit_before><commit_msg>Make the test data accessible to the tests.<commit_after>
|
import os
from functools import partial
TESTS_DIR = os.path.dirname(os.path.abspath(__file__))
TEST_DATA_DIR = os.path.join(TESTS_DIR, 'test_data')
AppEventTestXml = partial(open, os.path.join(TEST_DATA_DIR, 'app_event.xml'))
|
Make the test data accessible to the tests.import os
from functools import partial
TESTS_DIR = os.path.dirname(os.path.abspath(__file__))
TEST_DATA_DIR = os.path.join(TESTS_DIR, 'test_data')
AppEventTestXml = partial(open, os.path.join(TEST_DATA_DIR, 'app_event.xml'))
|
<commit_before><commit_msg>Make the test data accessible to the tests.<commit_after>import os
from functools import partial
TESTS_DIR = os.path.dirname(os.path.abspath(__file__))
TEST_DATA_DIR = os.path.join(TESTS_DIR, 'test_data')
AppEventTestXml = partial(open, os.path.join(TEST_DATA_DIR, 'app_event.xml'))
|
|
6160da958f4b8ecb1553c7bcca0b32bc1a5a1649
|
tests/conftest.py
|
tests/conftest.py
|
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
shutil.rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
shutil.rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
|
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
import sys
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
def rmtree(dirname):
"""Remove a directory, even if it has read-only files (Windows).
Git creates read-only files that must be removed on teardown. See
https://stackoverflow.com/questions/2656322 for more info.
Parameters
----------
dirname : str
Directory to be removed
"""
try:
shutil.rmtree(dirname)
except PermissionError:
if sys.platform == 'win32':
subprocess.check_call(['del', '/F/S/Q', dirname], shell=True)
else:
raise
|
Make sure .git test directory is removed on Windows
|
Make sure .git test directory is removed on Windows
|
Python
|
bsd-3-clause
|
scopatz/rever,ergs/rever
|
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
shutil.rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
shutil.rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
Make sure .git test directory is removed on Windows
|
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
import sys
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
def rmtree(dirname):
"""Remove a directory, even if it has read-only files (Windows).
Git creates read-only files that must be removed on teardown. See
https://stackoverflow.com/questions/2656322 for more info.
Parameters
----------
dirname : str
Directory to be removed
"""
try:
shutil.rmtree(dirname)
except PermissionError:
if sys.platform == 'win32':
subprocess.check_call(['del', '/F/S/Q', dirname], shell=True)
else:
raise
|
<commit_before>import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
shutil.rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
shutil.rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
<commit_msg>Make sure .git test directory is removed on Windows<commit_after>
|
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
import sys
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
def rmtree(dirname):
"""Remove a directory, even if it has read-only files (Windows).
Git creates read-only files that must be removed on teardown. See
https://stackoverflow.com/questions/2656322 for more info.
Parameters
----------
dirname : str
Directory to be removed
"""
try:
shutil.rmtree(dirname)
except PermissionError:
if sys.platform == 'win32':
subprocess.check_call(['del', '/F/S/Q', dirname], shell=True)
else:
raise
|
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
shutil.rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
shutil.rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
Make sure .git test directory is removed on Windowsimport os
import shutil
import tempfile
import builtins
import subprocess
import pytest
import sys
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
def rmtree(dirname):
"""Remove a directory, even if it has read-only files (Windows).
Git creates read-only files that must be removed on teardown. See
https://stackoverflow.com/questions/2656322 for more info.
Parameters
----------
dirname : str
Directory to be removed
"""
try:
shutil.rmtree(dirname)
except PermissionError:
if sys.platform == 'win32':
subprocess.check_call(['del', '/F/S/Q', dirname], shell=True)
else:
raise
|
<commit_before>import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
shutil.rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
shutil.rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
<commit_msg>Make sure .git test directory is removed on Windows<commit_after>import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
import sys
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
def rmtree(dirname):
"""Remove a directory, even if it has read-only files (Windows).
Git creates read-only files that must be removed on teardown. See
https://stackoverflow.com/questions/2656322 for more info.
Parameters
----------
dirname : str
Directory to be removed
"""
try:
shutil.rmtree(dirname)
except PermissionError:
if sys.platform == 'win32':
subprocess.check_call(['del', '/F/S/Q', dirname], shell=True)
else:
raise
|
c4e71b56e74ab8b81a670c690fef6942d4a412b4
|
ocds/storage/backends/fs.py
|
ocds/storage/backends/fs.py
|
import os
import os.path
import logging
import datetime
from .base import Storage
from ocds.storage.errors import InvalidPath
logger = logging.getLogger(__name__)
class FSStorage(Storage):
def __init__(self, base_path):
self.base_path = base_path
if not os.path.exists(self.base_path):
logger.warn('Initial path not exists. Creating')
try:
os.makedirs(self.base_path)
except (IOError, OSError) as e:
logger.error("Couldn't create destination dir."
"Error {}".format(e))
raise InvalidPath('Not destination folder')
def _from_string(self, string):
return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S')
def _path_from_date(self, date):
if isinstance(date, str):
path =
|
import os
import os.path
import logging
import datetime
import simplejson as json
from .base import Storage
from ocds.export.helpers import encoder
from ocds.storage.errors import InvalidPath
join = os.path.join
logger = logging.getLogger(__name__)
class FSStorage(Storage):
def __init__(self, base_path):
self.base_path = base_path
self.path_fmt = '%Y-%m-%d/%H/%M%/%S'
if not os.path.exists(self.base_path):
logger.warn('Initial path not exists. Creating')
try:
os.makedirs(self.base_path)
except (IOError, OSError) as e:
logger.error("Couldn't create destination dir."
"Error {}".format(e))
raise InvalidPath('Not destination folder')
def _walk(self):
for path, _, files in os.walk(self.base_path):
for f in files:
yield join(path, f)
def _write(self, obj):
path = join(self.base_path,
self._path_from_date(obj['date']))
file_path = join(path, '{}.json'.format(obj['id']))
with open(file_path, 'w') as out:
out.write(encoder(obj))
def _load(self, key):
with open(join(self.base_path, key)) as out:
result = json.load(out)
return result
def _from_string(self, string):
return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S')
def _path_from_date(self, date):
if isinstance(date, str):
path = self._from_string(date).srtftime(self.path_fmt)
if isinstance(date, datetime.date):
path = date.strftime(self.path_fmt)
return path
def __contains__(self, key):
try:
fs = open(join(self.base_path, key))
result = True
except (IOError, OSError):
result = False
finally:
fs.close()
return result
def __iter__(self):
for f in self._walk():
yield f
def save(self, obj):
self._write(obj)
def get(self, key):
return self._load(key)
|
Add basic file system storage
|
Add basic file system storage
|
Python
|
apache-2.0
|
yshalenyk/openprocurement.ocds.export,yshalenyk/ocds.storage,yshalenyk/ocds.export,yshalenyk/openprocurement.ocds.export
|
import os
import os.path
import logging
import datetime
from .base import Storage
from ocds.storage.errors import InvalidPath
logger = logging.getLogger(__name__)
class FSStorage(Storage):
def __init__(self, base_path):
self.base_path = base_path
if not os.path.exists(self.base_path):
logger.warn('Initial path not exists. Creating')
try:
os.makedirs(self.base_path)
except (IOError, OSError) as e:
logger.error("Couldn't create destination dir."
"Error {}".format(e))
raise InvalidPath('Not destination folder')
def _from_string(self, string):
return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S')
def _path_from_date(self, date):
if isinstance(date, str):
path =
Add basic file system storage
|
import os
import os.path
import logging
import datetime
import simplejson as json
from .base import Storage
from ocds.export.helpers import encoder
from ocds.storage.errors import InvalidPath
join = os.path.join
logger = logging.getLogger(__name__)
class FSStorage(Storage):
def __init__(self, base_path):
self.base_path = base_path
self.path_fmt = '%Y-%m-%d/%H/%M%/%S'
if not os.path.exists(self.base_path):
logger.warn('Initial path not exists. Creating')
try:
os.makedirs(self.base_path)
except (IOError, OSError) as e:
logger.error("Couldn't create destination dir."
"Error {}".format(e))
raise InvalidPath('Not destination folder')
def _walk(self):
for path, _, files in os.walk(self.base_path):
for f in files:
yield join(path, f)
def _write(self, obj):
path = join(self.base_path,
self._path_from_date(obj['date']))
file_path = join(path, '{}.json'.format(obj['id']))
with open(file_path, 'w') as out:
out.write(encoder(obj))
def _load(self, key):
with open(join(self.base_path, key)) as out:
result = json.load(out)
return result
def _from_string(self, string):
return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S')
def _path_from_date(self, date):
if isinstance(date, str):
path = self._from_string(date).srtftime(self.path_fmt)
if isinstance(date, datetime.date):
path = date.strftime(self.path_fmt)
return path
def __contains__(self, key):
try:
fs = open(join(self.base_path, key))
result = True
except (IOError, OSError):
result = False
finally:
fs.close()
return result
def __iter__(self):
for f in self._walk():
yield f
def save(self, obj):
self._write(obj)
def get(self, key):
return self._load(key)
|
<commit_before>import os
import os.path
import logging
import datetime
from .base import Storage
from ocds.storage.errors import InvalidPath
logger = logging.getLogger(__name__)
class FSStorage(Storage):
def __init__(self, base_path):
self.base_path = base_path
if not os.path.exists(self.base_path):
logger.warn('Initial path not exists. Creating')
try:
os.makedirs(self.base_path)
except (IOError, OSError) as e:
logger.error("Couldn't create destination dir."
"Error {}".format(e))
raise InvalidPath('Not destination folder')
def _from_string(self, string):
return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S')
def _path_from_date(self, date):
if isinstance(date, str):
path =
<commit_msg>Add basic file system storage<commit_after>
|
import os
import os.path
import logging
import datetime
import simplejson as json
from .base import Storage
from ocds.export.helpers import encoder
from ocds.storage.errors import InvalidPath
join = os.path.join
logger = logging.getLogger(__name__)
class FSStorage(Storage):
def __init__(self, base_path):
self.base_path = base_path
self.path_fmt = '%Y-%m-%d/%H/%M%/%S'
if not os.path.exists(self.base_path):
logger.warn('Initial path not exists. Creating')
try:
os.makedirs(self.base_path)
except (IOError, OSError) as e:
logger.error("Couldn't create destination dir."
"Error {}".format(e))
raise InvalidPath('Not destination folder')
def _walk(self):
for path, _, files in os.walk(self.base_path):
for f in files:
yield join(path, f)
def _write(self, obj):
path = join(self.base_path,
self._path_from_date(obj['date']))
file_path = join(path, '{}.json'.format(obj['id']))
with open(file_path, 'w') as out:
out.write(encoder(obj))
def _load(self, key):
with open(join(self.base_path, key)) as out:
result = json.load(out)
return result
def _from_string(self, string):
return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S')
def _path_from_date(self, date):
if isinstance(date, str):
path = self._from_string(date).srtftime(self.path_fmt)
if isinstance(date, datetime.date):
path = date.strftime(self.path_fmt)
return path
def __contains__(self, key):
try:
fs = open(join(self.base_path, key))
result = True
except (IOError, OSError):
result = False
finally:
fs.close()
return result
def __iter__(self):
for f in self._walk():
yield f
def save(self, obj):
self._write(obj)
def get(self, key):
return self._load(key)
|
import os
import os.path
import logging
import datetime
from .base import Storage
from ocds.storage.errors import InvalidPath
logger = logging.getLogger(__name__)
class FSStorage(Storage):
def __init__(self, base_path):
self.base_path = base_path
if not os.path.exists(self.base_path):
logger.warn('Initial path not exists. Creating')
try:
os.makedirs(self.base_path)
except (IOError, OSError) as e:
logger.error("Couldn't create destination dir."
"Error {}".format(e))
raise InvalidPath('Not destination folder')
def _from_string(self, string):
return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S')
def _path_from_date(self, date):
if isinstance(date, str):
path =
Add basic file system storageimport os
import os.path
import logging
import datetime
import simplejson as json
from .base import Storage
from ocds.export.helpers import encoder
from ocds.storage.errors import InvalidPath
join = os.path.join
logger = logging.getLogger(__name__)
class FSStorage(Storage):
def __init__(self, base_path):
self.base_path = base_path
self.path_fmt = '%Y-%m-%d/%H/%M%/%S'
if not os.path.exists(self.base_path):
logger.warn('Initial path not exists. Creating')
try:
os.makedirs(self.base_path)
except (IOError, OSError) as e:
logger.error("Couldn't create destination dir."
"Error {}".format(e))
raise InvalidPath('Not destination folder')
def _walk(self):
for path, _, files in os.walk(self.base_path):
for f in files:
yield join(path, f)
def _write(self, obj):
path = join(self.base_path,
self._path_from_date(obj['date']))
file_path = join(path, '{}.json'.format(obj['id']))
with open(file_path, 'w') as out:
out.write(encoder(obj))
def _load(self, key):
with open(join(self.base_path, key)) as out:
result = json.load(out)
return result
def _from_string(self, string):
return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S')
def _path_from_date(self, date):
if isinstance(date, str):
path = self._from_string(date).srtftime(self.path_fmt)
if isinstance(date, datetime.date):
path = date.strftime(self.path_fmt)
return path
def __contains__(self, key):
try:
fs = open(join(self.base_path, key))
result = True
except (IOError, OSError):
result = False
finally:
fs.close()
return result
def __iter__(self):
for f in self._walk():
yield f
def save(self, obj):
self._write(obj)
def get(self, key):
return self._load(key)
|
<commit_before>import os
import os.path
import logging
import datetime
from .base import Storage
from ocds.storage.errors import InvalidPath
logger = logging.getLogger(__name__)
class FSStorage(Storage):
def __init__(self, base_path):
self.base_path = base_path
if not os.path.exists(self.base_path):
logger.warn('Initial path not exists. Creating')
try:
os.makedirs(self.base_path)
except (IOError, OSError) as e:
logger.error("Couldn't create destination dir."
"Error {}".format(e))
raise InvalidPath('Not destination folder')
def _from_string(self, string):
return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S')
def _path_from_date(self, date):
if isinstance(date, str):
path =
<commit_msg>Add basic file system storage<commit_after>import os
import os.path
import logging
import datetime
import simplejson as json
from .base import Storage
from ocds.export.helpers import encoder
from ocds.storage.errors import InvalidPath
join = os.path.join
logger = logging.getLogger(__name__)
class FSStorage(Storage):
def __init__(self, base_path):
self.base_path = base_path
self.path_fmt = '%Y-%m-%d/%H/%M%/%S'
if not os.path.exists(self.base_path):
logger.warn('Initial path not exists. Creating')
try:
os.makedirs(self.base_path)
except (IOError, OSError) as e:
logger.error("Couldn't create destination dir."
"Error {}".format(e))
raise InvalidPath('Not destination folder')
def _walk(self):
for path, _, files in os.walk(self.base_path):
for f in files:
yield join(path, f)
def _write(self, obj):
path = join(self.base_path,
self._path_from_date(obj['date']))
file_path = join(path, '{}.json'.format(obj['id']))
with open(file_path, 'w') as out:
out.write(encoder(obj))
def _load(self, key):
with open(join(self.base_path, key)) as out:
result = json.load(out)
return result
def _from_string(self, string):
return datetime.datetime.strptime('%Y-%m-%dT%H:%M:$S')
def _path_from_date(self, date):
if isinstance(date, str):
path = self._from_string(date).srtftime(self.path_fmt)
if isinstance(date, datetime.date):
path = date.strftime(self.path_fmt)
return path
def __contains__(self, key):
try:
fs = open(join(self.base_path, key))
result = True
except (IOError, OSError):
result = False
finally:
fs.close()
return result
def __iter__(self):
for f in self._walk():
yield f
def save(self, obj):
self._write(obj)
def get(self, key):
return self._load(key)
|
2ccfb54f493bf0ffa07db910514a8429a2c51d73
|
changes/api/node_job_index.py
|
changes/api/node_job_index.py
|
from __future__ import absolute_import
from sqlalchemy.orm import joinedload
from changes.api.base import APIView
from changes.api.serializer.models.job import JobWithBuildSerializer
from changes.models import Job, JobStep, Node
class NodeJobIndexAPIView(APIView):
def get(self, node_id):
node = Node.query.get(node_id)
if node is None:
return '', 404
jobs = Job.query.join(
JobStep, JobStep.job_id == Job.id,
).options(
joinedload(Job.build, innerjoin=True),
).filter(
JobStep.node_id == node.id,
).order_by(Job.date_created.desc())
return self.paginate(jobs, serializers={
Job: JobWithBuildSerializer(),
})
|
from __future__ import absolute_import
from sqlalchemy.orm import joinedload
from changes.api.base import APIView
from changes.models import Build, Job, JobStep, Node
class NodeJobIndexAPIView(APIView):
def get(self, node_id):
node = Node.query.get(node_id)
if node is None:
return '', 404
jobs = list(Job.query.join(
JobStep, JobStep.job_id == Job.id,
).filter(
JobStep.node_id == node.id,
).order_by(Job.date_created.desc()))
build_list = list(Build.query.options(
joinedload('project'),
joinedload('author'),
joinedload('source').joinedload('revision'),
).filter(
Build.id.in_(j.build_id for j in jobs),
))
build_map = dict(
(b, d) for b, d in zip(build_list, self.serialize(build_list))
)
context = []
for job, data in zip(jobs, self.serialize(jobs)):
print job, data
data['build'] = build_map[job.build]
context.append(data)
return self.paginate(context, serialize=False)
|
Improve query patterns on node job list
|
Improve query patterns on node job list
|
Python
|
apache-2.0
|
wfxiang08/changes,wfxiang08/changes,bowlofstew/changes,wfxiang08/changes,dropbox/changes,dropbox/changes,dropbox/changes,bowlofstew/changes,bowlofstew/changes,wfxiang08/changes,dropbox/changes,bowlofstew/changes
|
from __future__ import absolute_import
from sqlalchemy.orm import joinedload
from changes.api.base import APIView
from changes.api.serializer.models.job import JobWithBuildSerializer
from changes.models import Job, JobStep, Node
class NodeJobIndexAPIView(APIView):
def get(self, node_id):
node = Node.query.get(node_id)
if node is None:
return '', 404
jobs = Job.query.join(
JobStep, JobStep.job_id == Job.id,
).options(
joinedload(Job.build, innerjoin=True),
).filter(
JobStep.node_id == node.id,
).order_by(Job.date_created.desc())
return self.paginate(jobs, serializers={
Job: JobWithBuildSerializer(),
})
Improve query patterns on node job list
|
from __future__ import absolute_import
from sqlalchemy.orm import joinedload
from changes.api.base import APIView
from changes.models import Build, Job, JobStep, Node
class NodeJobIndexAPIView(APIView):
def get(self, node_id):
node = Node.query.get(node_id)
if node is None:
return '', 404
jobs = list(Job.query.join(
JobStep, JobStep.job_id == Job.id,
).filter(
JobStep.node_id == node.id,
).order_by(Job.date_created.desc()))
build_list = list(Build.query.options(
joinedload('project'),
joinedload('author'),
joinedload('source').joinedload('revision'),
).filter(
Build.id.in_(j.build_id for j in jobs),
))
build_map = dict(
(b, d) for b, d in zip(build_list, self.serialize(build_list))
)
context = []
for job, data in zip(jobs, self.serialize(jobs)):
print job, data
data['build'] = build_map[job.build]
context.append(data)
return self.paginate(context, serialize=False)
|
<commit_before>from __future__ import absolute_import
from sqlalchemy.orm import joinedload
from changes.api.base import APIView
from changes.api.serializer.models.job import JobWithBuildSerializer
from changes.models import Job, JobStep, Node
class NodeJobIndexAPIView(APIView):
def get(self, node_id):
node = Node.query.get(node_id)
if node is None:
return '', 404
jobs = Job.query.join(
JobStep, JobStep.job_id == Job.id,
).options(
joinedload(Job.build, innerjoin=True),
).filter(
JobStep.node_id == node.id,
).order_by(Job.date_created.desc())
return self.paginate(jobs, serializers={
Job: JobWithBuildSerializer(),
})
<commit_msg>Improve query patterns on node job list<commit_after>
|
from __future__ import absolute_import
from sqlalchemy.orm import joinedload
from changes.api.base import APIView
from changes.models import Build, Job, JobStep, Node
class NodeJobIndexAPIView(APIView):
def get(self, node_id):
node = Node.query.get(node_id)
if node is None:
return '', 404
jobs = list(Job.query.join(
JobStep, JobStep.job_id == Job.id,
).filter(
JobStep.node_id == node.id,
).order_by(Job.date_created.desc()))
build_list = list(Build.query.options(
joinedload('project'),
joinedload('author'),
joinedload('source').joinedload('revision'),
).filter(
Build.id.in_(j.build_id for j in jobs),
))
build_map = dict(
(b, d) for b, d in zip(build_list, self.serialize(build_list))
)
context = []
for job, data in zip(jobs, self.serialize(jobs)):
print job, data
data['build'] = build_map[job.build]
context.append(data)
return self.paginate(context, serialize=False)
|
from __future__ import absolute_import
from sqlalchemy.orm import joinedload
from changes.api.base import APIView
from changes.api.serializer.models.job import JobWithBuildSerializer
from changes.models import Job, JobStep, Node
class NodeJobIndexAPIView(APIView):
def get(self, node_id):
node = Node.query.get(node_id)
if node is None:
return '', 404
jobs = Job.query.join(
JobStep, JobStep.job_id == Job.id,
).options(
joinedload(Job.build, innerjoin=True),
).filter(
JobStep.node_id == node.id,
).order_by(Job.date_created.desc())
return self.paginate(jobs, serializers={
Job: JobWithBuildSerializer(),
})
Improve query patterns on node job listfrom __future__ import absolute_import
from sqlalchemy.orm import joinedload
from changes.api.base import APIView
from changes.models import Build, Job, JobStep, Node
class NodeJobIndexAPIView(APIView):
def get(self, node_id):
node = Node.query.get(node_id)
if node is None:
return '', 404
jobs = list(Job.query.join(
JobStep, JobStep.job_id == Job.id,
).filter(
JobStep.node_id == node.id,
).order_by(Job.date_created.desc()))
build_list = list(Build.query.options(
joinedload('project'),
joinedload('author'),
joinedload('source').joinedload('revision'),
).filter(
Build.id.in_(j.build_id for j in jobs),
))
build_map = dict(
(b, d) for b, d in zip(build_list, self.serialize(build_list))
)
context = []
for job, data in zip(jobs, self.serialize(jobs)):
print job, data
data['build'] = build_map[job.build]
context.append(data)
return self.paginate(context, serialize=False)
|
<commit_before>from __future__ import absolute_import
from sqlalchemy.orm import joinedload
from changes.api.base import APIView
from changes.api.serializer.models.job import JobWithBuildSerializer
from changes.models import Job, JobStep, Node
class NodeJobIndexAPIView(APIView):
def get(self, node_id):
node = Node.query.get(node_id)
if node is None:
return '', 404
jobs = Job.query.join(
JobStep, JobStep.job_id == Job.id,
).options(
joinedload(Job.build, innerjoin=True),
).filter(
JobStep.node_id == node.id,
).order_by(Job.date_created.desc())
return self.paginate(jobs, serializers={
Job: JobWithBuildSerializer(),
})
<commit_msg>Improve query patterns on node job list<commit_after>from __future__ import absolute_import
from sqlalchemy.orm import joinedload
from changes.api.base import APIView
from changes.models import Build, Job, JobStep, Node
class NodeJobIndexAPIView(APIView):
def get(self, node_id):
node = Node.query.get(node_id)
if node is None:
return '', 404
jobs = list(Job.query.join(
JobStep, JobStep.job_id == Job.id,
).filter(
JobStep.node_id == node.id,
).order_by(Job.date_created.desc()))
build_list = list(Build.query.options(
joinedload('project'),
joinedload('author'),
joinedload('source').joinedload('revision'),
).filter(
Build.id.in_(j.build_id for j in jobs),
))
build_map = dict(
(b, d) for b, d in zip(build_list, self.serialize(build_list))
)
context = []
for job, data in zip(jobs, self.serialize(jobs)):
print job, data
data['build'] = build_map[job.build]
context.append(data)
return self.paginate(context, serialize=False)
|
c12f3e516eb28d306a103582495216253dd98e7e
|
feedreader/tasks/core.py
|
feedreader/tasks/core.py
|
from celery import Celery
class Tasks(object):
def __init__(self, debug=False):
self.app = Celery()
self.app.conf.update(
CELERY_ACCEPT_CONTENT=['json'],
CELERY_ALWAYS_EAGER=True,
CELERY_ENABLE_UTC=True,
CELERY_TASK_SERIALIZER='json',
CELERY_RESULT_SERIALIZER='json',
CELERY_TIMEZONE='America/Vancouver',
)
if not debug:
self.app.conf.update(
BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//',
CELERY_ALWAYS_EAGER=False,
CELERY_RESULT_BACKEND='amqp',
)
# register tasks with celery
self.fetch_feed = self.app.task()(self.fetch_feed)
# celery tasks
def fetch_feed(self, url):
return 'hola bro'
|
from celery import Celery
class Tasks(object):
def __init__(self, debug=False):
self.app = Celery()
self.app.conf.update(
CELERY_ACCEPT_CONTENT=['json'],
CELERY_ALWAYS_EAGER=True,
CELERY_ENABLE_UTC=True,
CELERY_TASK_SERIALIZER='json',
CELERY_RESULT_SERIALIZER='json',
CELERY_TIMEZONE='America/Vancouver',
)
if not debug:
self.app.conf.update(
BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//',
CELERY_ALWAYS_EAGER=False,
CELERY_RESULT_BACKEND='amqp',
)
# register tasks with celery
self.fetch_feed = self.app.task()(self.fetch_feed)
# celery tasks
def fetch_feed(self, feed_url, last_modified=None, etag=None):
"""Fetch and parse the feed at the given URL.
If the given URL is not a feed, this will attempt to find one.
Raises SomeException if an error occurs.
Returns dict containing:
- feed_url: canonical url of the feed resource
- feed: new instance of the Feed model, or None if the feed was
unmodified
- entries: list of new instances of the Entry model, or empty list
if the feed was unmodified
- last_modified: last modified date, if server provides one
- etag: etag, if server provides one
"""
return # TODO
|
Add a stub task for fetch_feed
|
Add a stub task for fetch_feed
|
Python
|
mit
|
tdryer/feeder,tdryer/feeder
|
from celery import Celery
class Tasks(object):
def __init__(self, debug=False):
self.app = Celery()
self.app.conf.update(
CELERY_ACCEPT_CONTENT=['json'],
CELERY_ALWAYS_EAGER=True,
CELERY_ENABLE_UTC=True,
CELERY_TASK_SERIALIZER='json',
CELERY_RESULT_SERIALIZER='json',
CELERY_TIMEZONE='America/Vancouver',
)
if not debug:
self.app.conf.update(
BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//',
CELERY_ALWAYS_EAGER=False,
CELERY_RESULT_BACKEND='amqp',
)
# register tasks with celery
self.fetch_feed = self.app.task()(self.fetch_feed)
# celery tasks
def fetch_feed(self, url):
return 'hola bro'
Add a stub task for fetch_feed
|
from celery import Celery
class Tasks(object):
def __init__(self, debug=False):
self.app = Celery()
self.app.conf.update(
CELERY_ACCEPT_CONTENT=['json'],
CELERY_ALWAYS_EAGER=True,
CELERY_ENABLE_UTC=True,
CELERY_TASK_SERIALIZER='json',
CELERY_RESULT_SERIALIZER='json',
CELERY_TIMEZONE='America/Vancouver',
)
if not debug:
self.app.conf.update(
BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//',
CELERY_ALWAYS_EAGER=False,
CELERY_RESULT_BACKEND='amqp',
)
# register tasks with celery
self.fetch_feed = self.app.task()(self.fetch_feed)
# celery tasks
def fetch_feed(self, feed_url, last_modified=None, etag=None):
"""Fetch and parse the feed at the given URL.
If the given URL is not a feed, this will attempt to find one.
Raises SomeException if an error occurs.
Returns dict containing:
- feed_url: canonical url of the feed resource
- feed: new instance of the Feed model, or None if the feed was
unmodified
- entries: list of new instances of the Entry model, or empty list
if the feed was unmodified
- last_modified: last modified date, if server provides one
- etag: etag, if server provides one
"""
return # TODO
|
<commit_before>from celery import Celery
class Tasks(object):
def __init__(self, debug=False):
self.app = Celery()
self.app.conf.update(
CELERY_ACCEPT_CONTENT=['json'],
CELERY_ALWAYS_EAGER=True,
CELERY_ENABLE_UTC=True,
CELERY_TASK_SERIALIZER='json',
CELERY_RESULT_SERIALIZER='json',
CELERY_TIMEZONE='America/Vancouver',
)
if not debug:
self.app.conf.update(
BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//',
CELERY_ALWAYS_EAGER=False,
CELERY_RESULT_BACKEND='amqp',
)
# register tasks with celery
self.fetch_feed = self.app.task()(self.fetch_feed)
# celery tasks
def fetch_feed(self, url):
return 'hola bro'
<commit_msg>Add a stub task for fetch_feed<commit_after>
|
from celery import Celery
class Tasks(object):
def __init__(self, debug=False):
self.app = Celery()
self.app.conf.update(
CELERY_ACCEPT_CONTENT=['json'],
CELERY_ALWAYS_EAGER=True,
CELERY_ENABLE_UTC=True,
CELERY_TASK_SERIALIZER='json',
CELERY_RESULT_SERIALIZER='json',
CELERY_TIMEZONE='America/Vancouver',
)
if not debug:
self.app.conf.update(
BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//',
CELERY_ALWAYS_EAGER=False,
CELERY_RESULT_BACKEND='amqp',
)
# register tasks with celery
self.fetch_feed = self.app.task()(self.fetch_feed)
# celery tasks
def fetch_feed(self, feed_url, last_modified=None, etag=None):
"""Fetch and parse the feed at the given URL.
If the given URL is not a feed, this will attempt to find one.
Raises SomeException if an error occurs.
Returns dict containing:
- feed_url: canonical url of the feed resource
- feed: new instance of the Feed model, or None if the feed was
unmodified
- entries: list of new instances of the Entry model, or empty list
if the feed was unmodified
- last_modified: last modified date, if server provides one
- etag: etag, if server provides one
"""
return # TODO
|
from celery import Celery
class Tasks(object):
def __init__(self, debug=False):
self.app = Celery()
self.app.conf.update(
CELERY_ACCEPT_CONTENT=['json'],
CELERY_ALWAYS_EAGER=True,
CELERY_ENABLE_UTC=True,
CELERY_TASK_SERIALIZER='json',
CELERY_RESULT_SERIALIZER='json',
CELERY_TIMEZONE='America/Vancouver',
)
if not debug:
self.app.conf.update(
BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//',
CELERY_ALWAYS_EAGER=False,
CELERY_RESULT_BACKEND='amqp',
)
# register tasks with celery
self.fetch_feed = self.app.task()(self.fetch_feed)
# celery tasks
def fetch_feed(self, url):
return 'hola bro'
Add a stub task for fetch_feedfrom celery import Celery
class Tasks(object):
def __init__(self, debug=False):
self.app = Celery()
self.app.conf.update(
CELERY_ACCEPT_CONTENT=['json'],
CELERY_ALWAYS_EAGER=True,
CELERY_ENABLE_UTC=True,
CELERY_TASK_SERIALIZER='json',
CELERY_RESULT_SERIALIZER='json',
CELERY_TIMEZONE='America/Vancouver',
)
if not debug:
self.app.conf.update(
BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//',
CELERY_ALWAYS_EAGER=False,
CELERY_RESULT_BACKEND='amqp',
)
# register tasks with celery
self.fetch_feed = self.app.task()(self.fetch_feed)
# celery tasks
def fetch_feed(self, feed_url, last_modified=None, etag=None):
"""Fetch and parse the feed at the given URL.
If the given URL is not a feed, this will attempt to find one.
Raises SomeException if an error occurs.
Returns dict containing:
- feed_url: canonical url of the feed resource
- feed: new instance of the Feed model, or None if the feed was
unmodified
- entries: list of new instances of the Entry model, or empty list
if the feed was unmodified
- last_modified: last modified date, if server provides one
- etag: etag, if server provides one
"""
return # TODO
|
<commit_before>from celery import Celery
class Tasks(object):
def __init__(self, debug=False):
self.app = Celery()
self.app.conf.update(
CELERY_ACCEPT_CONTENT=['json'],
CELERY_ALWAYS_EAGER=True,
CELERY_ENABLE_UTC=True,
CELERY_TASK_SERIALIZER='json',
CELERY_RESULT_SERIALIZER='json',
CELERY_TIMEZONE='America/Vancouver',
)
if not debug:
self.app.conf.update(
BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//',
CELERY_ALWAYS_EAGER=False,
CELERY_RESULT_BACKEND='amqp',
)
# register tasks with celery
self.fetch_feed = self.app.task()(self.fetch_feed)
# celery tasks
def fetch_feed(self, url):
return 'hola bro'
<commit_msg>Add a stub task for fetch_feed<commit_after>from celery import Celery
class Tasks(object):
def __init__(self, debug=False):
self.app = Celery()
self.app.conf.update(
CELERY_ACCEPT_CONTENT=['json'],
CELERY_ALWAYS_EAGER=True,
CELERY_ENABLE_UTC=True,
CELERY_TASK_SERIALIZER='json',
CELERY_RESULT_SERIALIZER='json',
CELERY_TIMEZONE='America/Vancouver',
)
if not debug:
self.app.conf.update(
BROKER_URL='amqp://guest:guest@novus.mtomwing.com:5673//',
CELERY_ALWAYS_EAGER=False,
CELERY_RESULT_BACKEND='amqp',
)
# register tasks with celery
self.fetch_feed = self.app.task()(self.fetch_feed)
# celery tasks
def fetch_feed(self, feed_url, last_modified=None, etag=None):
"""Fetch and parse the feed at the given URL.
If the given URL is not a feed, this will attempt to find one.
Raises SomeException if an error occurs.
Returns dict containing:
- feed_url: canonical url of the feed resource
- feed: new instance of the Feed model, or None if the feed was
unmodified
- entries: list of new instances of the Entry model, or empty list
if the feed was unmodified
- last_modified: last modified date, if server provides one
- etag: etag, if server provides one
"""
return # TODO
|
8beaab317d5da25edd093be42f57e35ac12408b8
|
feincms3/plugins/html.py
|
feincms3/plugins/html.py
|
"""
Plugin providing a simple textarea where raw HTML, CSS and JS code can be
entered.
Most useful for people wanting to shoot themselves in the foot.
"""
from django.db import models
from django.utils.html import mark_safe
from django.utils.translation import ugettext_lazy as _
from content_editor.admin import ContentEditorInline
__all__ = ("HTML", "HTMLInline", "render_html")
class HTML(models.Model):
"""
Raw HTML plugin
"""
html = models.TextField(
"HTML",
help_text=_(
"The content will be inserted directly into the page."
" It is VERY important that the HTML snippet is well-formed!"
),
)
class Meta:
abstract = True
verbose_name = "HTML"
verbose_name_plural = "HTML"
def __str__(self):
return ""
class HTMLInline(ContentEditorInline):
"""
Just available for consistency, absolutely no difference to a standard
``ContentEditorInline``.
"""
pass
def render_html(plugin, **kwargs):
"""
Return the HTML code as safe string so that it is not escaped. Of course
the contents are not guaranteed to be safe at all
"""
return mark_safe(plugin.html)
|
"""
Plugin providing a simple textarea where raw HTML, CSS and JS code can be
entered.
Most useful for people wanting to shoot themselves in the foot.
"""
from django import forms
from django.db import models
from django.utils.html import mark_safe
from django.utils.translation import ugettext_lazy as _
from content_editor.admin import ContentEditorInline
__all__ = ("HTML", "HTMLInline", "render_html")
class HTML(models.Model):
"""
Raw HTML plugin
"""
html = models.TextField(
"HTML",
help_text=_(
"The content will be inserted directly into the page."
" It is VERY important that the HTML snippet is well-formed!"
),
)
class Meta:
abstract = True
verbose_name = "HTML"
verbose_name_plural = "HTML"
def __str__(self):
return ""
class HTMLInline(ContentEditorInline):
"""
Just available for consistency, absolutely no difference to a standard
``ContentEditorInline``.
"""
formfield_overrides = {
models.TextField: {
"widget": forms.Textarea(
attrs={"rows": 3, "cols": 40, "class": "vLargeTextField"}
)
}
}
def render_html(plugin, **kwargs):
"""
Return the HTML code as safe string so that it is not escaped. Of course
the contents are not guaranteed to be safe at all
"""
return mark_safe(plugin.html)
|
Make the default HTML textarea smaller
|
Make the default HTML textarea smaller
|
Python
|
bsd-3-clause
|
matthiask/feincms3,matthiask/feincms3,matthiask/feincms3
|
"""
Plugin providing a simple textarea where raw HTML, CSS and JS code can be
entered.
Most useful for people wanting to shoot themselves in the foot.
"""
from django.db import models
from django.utils.html import mark_safe
from django.utils.translation import ugettext_lazy as _
from content_editor.admin import ContentEditorInline
__all__ = ("HTML", "HTMLInline", "render_html")
class HTML(models.Model):
"""
Raw HTML plugin
"""
html = models.TextField(
"HTML",
help_text=_(
"The content will be inserted directly into the page."
" It is VERY important that the HTML snippet is well-formed!"
),
)
class Meta:
abstract = True
verbose_name = "HTML"
verbose_name_plural = "HTML"
def __str__(self):
return ""
class HTMLInline(ContentEditorInline):
"""
Just available for consistency, absolutely no difference to a standard
``ContentEditorInline``.
"""
pass
def render_html(plugin, **kwargs):
"""
Return the HTML code as safe string so that it is not escaped. Of course
the contents are not guaranteed to be safe at all
"""
return mark_safe(plugin.html)
Make the default HTML textarea smaller
|
"""
Plugin providing a simple textarea where raw HTML, CSS and JS code can be
entered.
Most useful for people wanting to shoot themselves in the foot.
"""
from django import forms
from django.db import models
from django.utils.html import mark_safe
from django.utils.translation import ugettext_lazy as _
from content_editor.admin import ContentEditorInline
__all__ = ("HTML", "HTMLInline", "render_html")
class HTML(models.Model):
"""
Raw HTML plugin
"""
html = models.TextField(
"HTML",
help_text=_(
"The content will be inserted directly into the page."
" It is VERY important that the HTML snippet is well-formed!"
),
)
class Meta:
abstract = True
verbose_name = "HTML"
verbose_name_plural = "HTML"
def __str__(self):
return ""
class HTMLInline(ContentEditorInline):
"""
Just available for consistency, absolutely no difference to a standard
``ContentEditorInline``.
"""
formfield_overrides = {
models.TextField: {
"widget": forms.Textarea(
attrs={"rows": 3, "cols": 40, "class": "vLargeTextField"}
)
}
}
def render_html(plugin, **kwargs):
"""
Return the HTML code as safe string so that it is not escaped. Of course
the contents are not guaranteed to be safe at all
"""
return mark_safe(plugin.html)
|
<commit_before>"""
Plugin providing a simple textarea where raw HTML, CSS and JS code can be
entered.
Most useful for people wanting to shoot themselves in the foot.
"""
from django.db import models
from django.utils.html import mark_safe
from django.utils.translation import ugettext_lazy as _
from content_editor.admin import ContentEditorInline
__all__ = ("HTML", "HTMLInline", "render_html")
class HTML(models.Model):
"""
Raw HTML plugin
"""
html = models.TextField(
"HTML",
help_text=_(
"The content will be inserted directly into the page."
" It is VERY important that the HTML snippet is well-formed!"
),
)
class Meta:
abstract = True
verbose_name = "HTML"
verbose_name_plural = "HTML"
def __str__(self):
return ""
class HTMLInline(ContentEditorInline):
"""
Just available for consistency, absolutely no difference to a standard
``ContentEditorInline``.
"""
pass
def render_html(plugin, **kwargs):
"""
Return the HTML code as safe string so that it is not escaped. Of course
the contents are not guaranteed to be safe at all
"""
return mark_safe(plugin.html)
<commit_msg>Make the default HTML textarea smaller<commit_after>
|
"""
Plugin providing a simple textarea where raw HTML, CSS and JS code can be
entered.
Most useful for people wanting to shoot themselves in the foot.
"""
from django import forms
from django.db import models
from django.utils.html import mark_safe
from django.utils.translation import ugettext_lazy as _
from content_editor.admin import ContentEditorInline
__all__ = ("HTML", "HTMLInline", "render_html")
class HTML(models.Model):
"""
Raw HTML plugin
"""
html = models.TextField(
"HTML",
help_text=_(
"The content will be inserted directly into the page."
" It is VERY important that the HTML snippet is well-formed!"
),
)
class Meta:
abstract = True
verbose_name = "HTML"
verbose_name_plural = "HTML"
def __str__(self):
return ""
class HTMLInline(ContentEditorInline):
"""
Just available for consistency, absolutely no difference to a standard
``ContentEditorInline``.
"""
formfield_overrides = {
models.TextField: {
"widget": forms.Textarea(
attrs={"rows": 3, "cols": 40, "class": "vLargeTextField"}
)
}
}
def render_html(plugin, **kwargs):
"""
Return the HTML code as safe string so that it is not escaped. Of course
the contents are not guaranteed to be safe at all
"""
return mark_safe(plugin.html)
|
"""
Plugin providing a simple textarea where raw HTML, CSS and JS code can be
entered.
Most useful for people wanting to shoot themselves in the foot.
"""
from django.db import models
from django.utils.html import mark_safe
from django.utils.translation import ugettext_lazy as _
from content_editor.admin import ContentEditorInline
__all__ = ("HTML", "HTMLInline", "render_html")
class HTML(models.Model):
"""
Raw HTML plugin
"""
html = models.TextField(
"HTML",
help_text=_(
"The content will be inserted directly into the page."
" It is VERY important that the HTML snippet is well-formed!"
),
)
class Meta:
abstract = True
verbose_name = "HTML"
verbose_name_plural = "HTML"
def __str__(self):
return ""
class HTMLInline(ContentEditorInline):
"""
Just available for consistency, absolutely no difference to a standard
``ContentEditorInline``.
"""
pass
def render_html(plugin, **kwargs):
"""
Return the HTML code as safe string so that it is not escaped. Of course
the contents are not guaranteed to be safe at all
"""
return mark_safe(plugin.html)
Make the default HTML textarea smaller"""
Plugin providing a simple textarea where raw HTML, CSS and JS code can be
entered.
Most useful for people wanting to shoot themselves in the foot.
"""
from django import forms
from django.db import models
from django.utils.html import mark_safe
from django.utils.translation import ugettext_lazy as _
from content_editor.admin import ContentEditorInline
__all__ = ("HTML", "HTMLInline", "render_html")
class HTML(models.Model):
"""
Raw HTML plugin
"""
html = models.TextField(
"HTML",
help_text=_(
"The content will be inserted directly into the page."
" It is VERY important that the HTML snippet is well-formed!"
),
)
class Meta:
abstract = True
verbose_name = "HTML"
verbose_name_plural = "HTML"
def __str__(self):
return ""
class HTMLInline(ContentEditorInline):
"""
Just available for consistency, absolutely no difference to a standard
``ContentEditorInline``.
"""
formfield_overrides = {
models.TextField: {
"widget": forms.Textarea(
attrs={"rows": 3, "cols": 40, "class": "vLargeTextField"}
)
}
}
def render_html(plugin, **kwargs):
"""
Return the HTML code as safe string so that it is not escaped. Of course
the contents are not guaranteed to be safe at all
"""
return mark_safe(plugin.html)
|
<commit_before>"""
Plugin providing a simple textarea where raw HTML, CSS and JS code can be
entered.
Most useful for people wanting to shoot themselves in the foot.
"""
from django.db import models
from django.utils.html import mark_safe
from django.utils.translation import ugettext_lazy as _
from content_editor.admin import ContentEditorInline
__all__ = ("HTML", "HTMLInline", "render_html")
class HTML(models.Model):
"""
Raw HTML plugin
"""
html = models.TextField(
"HTML",
help_text=_(
"The content will be inserted directly into the page."
" It is VERY important that the HTML snippet is well-formed!"
),
)
class Meta:
abstract = True
verbose_name = "HTML"
verbose_name_plural = "HTML"
def __str__(self):
return ""
class HTMLInline(ContentEditorInline):
"""
Just available for consistency, absolutely no difference to a standard
``ContentEditorInline``.
"""
pass
def render_html(plugin, **kwargs):
"""
Return the HTML code as safe string so that it is not escaped. Of course
the contents are not guaranteed to be safe at all
"""
return mark_safe(plugin.html)
<commit_msg>Make the default HTML textarea smaller<commit_after>"""
Plugin providing a simple textarea where raw HTML, CSS and JS code can be
entered.
Most useful for people wanting to shoot themselves in the foot.
"""
from django import forms
from django.db import models
from django.utils.html import mark_safe
from django.utils.translation import ugettext_lazy as _
from content_editor.admin import ContentEditorInline
__all__ = ("HTML", "HTMLInline", "render_html")
class HTML(models.Model):
"""
Raw HTML plugin
"""
html = models.TextField(
"HTML",
help_text=_(
"The content will be inserted directly into the page."
" It is VERY important that the HTML snippet is well-formed!"
),
)
class Meta:
abstract = True
verbose_name = "HTML"
verbose_name_plural = "HTML"
def __str__(self):
return ""
class HTMLInline(ContentEditorInline):
"""
Just available for consistency, absolutely no difference to a standard
``ContentEditorInline``.
"""
formfield_overrides = {
models.TextField: {
"widget": forms.Textarea(
attrs={"rows": 3, "cols": 40, "class": "vLargeTextField"}
)
}
}
def render_html(plugin, **kwargs):
"""
Return the HTML code as safe string so that it is not escaped. Of course
the contents are not guaranteed to be safe at all
"""
return mark_safe(plugin.html)
|
720833e96e24ffe73822a3a1280e3dc901e52829
|
anchorhub/lib/filetolist.py
|
anchorhub/lib/filetolist.py
|
"""
Class for FileToList
"""
class FileToList(object):
"""
FileToList is a helper class used to import text files and turn them into
lists, with each index in the list representing a single line from the
text file.
"""
@staticmethod
def to_list(file_path):
"""
Static method. Takes in a file path, and outputs a list of stings.
Each element in the list corresponds to a line in the file.
:param file_path: string file path
:return: A list of strings, with elements in the list corresponding
to lines in the file pointed to in file_path
"""
l = []
f = open(file_path, 'rb')
for line in f:
l.append(line)
f.close()
return l
|
"""
Class for FileToList
"""
class FileToList(object):
"""
FileToList is a helper class used to import text files and turn them into
lists, with each index in the list representing a single line from the
text file.
"""
@staticmethod
def to_list(file_path):
"""
Static method. Takes in a file path, and outputs a list of stings.
Each element in the list corresponds to a line in the file.
:param file_path: string file path
:return: A list of strings, with elements in the list corresponding
to lines in the file pointed to in file_path
"""
l = []
f = open(file_path, 'r')
for line in f:
l.append(line)
f.close()
return l
|
Remove 'b' classifer on FileToList's read() usage
|
Remove 'b' classifer on FileToList's read() usage
|
Python
|
apache-2.0
|
samjabrahams/anchorhub
|
"""
Class for FileToList
"""
class FileToList(object):
"""
FileToList is a helper class used to import text files and turn them into
lists, with each index in the list representing a single line from the
text file.
"""
@staticmethod
def to_list(file_path):
"""
Static method. Takes in a file path, and outputs a list of stings.
Each element in the list corresponds to a line in the file.
:param file_path: string file path
:return: A list of strings, with elements in the list corresponding
to lines in the file pointed to in file_path
"""
l = []
f = open(file_path, 'rb')
for line in f:
l.append(line)
f.close()
return l
Remove 'b' classifer on FileToList's read() usage
|
"""
Class for FileToList
"""
class FileToList(object):
"""
FileToList is a helper class used to import text files and turn them into
lists, with each index in the list representing a single line from the
text file.
"""
@staticmethod
def to_list(file_path):
"""
Static method. Takes in a file path, and outputs a list of stings.
Each element in the list corresponds to a line in the file.
:param file_path: string file path
:return: A list of strings, with elements in the list corresponding
to lines in the file pointed to in file_path
"""
l = []
f = open(file_path, 'r')
for line in f:
l.append(line)
f.close()
return l
|
<commit_before>"""
Class for FileToList
"""
class FileToList(object):
"""
FileToList is a helper class used to import text files and turn them into
lists, with each index in the list representing a single line from the
text file.
"""
@staticmethod
def to_list(file_path):
"""
Static method. Takes in a file path, and outputs a list of stings.
Each element in the list corresponds to a line in the file.
:param file_path: string file path
:return: A list of strings, with elements in the list corresponding
to lines in the file pointed to in file_path
"""
l = []
f = open(file_path, 'rb')
for line in f:
l.append(line)
f.close()
return l
<commit_msg>Remove 'b' classifer on FileToList's read() usage<commit_after>
|
"""
Class for FileToList
"""
class FileToList(object):
"""
FileToList is a helper class used to import text files and turn them into
lists, with each index in the list representing a single line from the
text file.
"""
@staticmethod
def to_list(file_path):
"""
Static method. Takes in a file path, and outputs a list of stings.
Each element in the list corresponds to a line in the file.
:param file_path: string file path
:return: A list of strings, with elements in the list corresponding
to lines in the file pointed to in file_path
"""
l = []
f = open(file_path, 'r')
for line in f:
l.append(line)
f.close()
return l
|
"""
Class for FileToList
"""
class FileToList(object):
"""
FileToList is a helper class used to import text files and turn them into
lists, with each index in the list representing a single line from the
text file.
"""
@staticmethod
def to_list(file_path):
"""
Static method. Takes in a file path, and outputs a list of stings.
Each element in the list corresponds to a line in the file.
:param file_path: string file path
:return: A list of strings, with elements in the list corresponding
to lines in the file pointed to in file_path
"""
l = []
f = open(file_path, 'rb')
for line in f:
l.append(line)
f.close()
return l
Remove 'b' classifer on FileToList's read() usage"""
Class for FileToList
"""
class FileToList(object):
"""
FileToList is a helper class used to import text files and turn them into
lists, with each index in the list representing a single line from the
text file.
"""
@staticmethod
def to_list(file_path):
"""
Static method. Takes in a file path, and outputs a list of stings.
Each element in the list corresponds to a line in the file.
:param file_path: string file path
:return: A list of strings, with elements in the list corresponding
to lines in the file pointed to in file_path
"""
l = []
f = open(file_path, 'r')
for line in f:
l.append(line)
f.close()
return l
|
<commit_before>"""
Class for FileToList
"""
class FileToList(object):
"""
FileToList is a helper class used to import text files and turn them into
lists, with each index in the list representing a single line from the
text file.
"""
@staticmethod
def to_list(file_path):
"""
Static method. Takes in a file path, and outputs a list of stings.
Each element in the list corresponds to a line in the file.
:param file_path: string file path
:return: A list of strings, with elements in the list corresponding
to lines in the file pointed to in file_path
"""
l = []
f = open(file_path, 'rb')
for line in f:
l.append(line)
f.close()
return l
<commit_msg>Remove 'b' classifer on FileToList's read() usage<commit_after>"""
Class for FileToList
"""
class FileToList(object):
"""
FileToList is a helper class used to import text files and turn them into
lists, with each index in the list representing a single line from the
text file.
"""
@staticmethod
def to_list(file_path):
"""
Static method. Takes in a file path, and outputs a list of stings.
Each element in the list corresponds to a line in the file.
:param file_path: string file path
:return: A list of strings, with elements in the list corresponding
to lines in the file pointed to in file_path
"""
l = []
f = open(file_path, 'r')
for line in f:
l.append(line)
f.close()
return l
|
614f83d826c51a51ebb4feb01371a441473af423
|
featureflow/__init__.py
|
featureflow/__init__.py
|
__version__ = '1.16.14'
from model import BaseModel
from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \
PickleFeature
from extractor import Node, Graph, Aggregator, NotEnoughData
from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip
from data import \
IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \
KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \
InMemoryDatabase
from datawriter import DataWriter
from database_iterator import DatabaseIterator
from encoder import IdentityEncoder
from decoder import Decoder, PickleDecoder
from lmdbstore import LmdbDatabase
from objectstore import ObjectStoreDatabase
from persistence import PersistenceSettings
from iteratornode import IteratorNode
from eventlog import EventLog, RedisChannel
try:
from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \
BaseNumpyDecoder, NumpyMetaData, NumpyFeature
except ImportError:
pass
|
__version__ = '1.16.14'
from model import BaseModel
from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \
PickleFeature
from extractor import Node, Graph, Aggregator, NotEnoughData
from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip
from data import \
IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \
KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \
InMemoryDatabase
from datawriter import DataWriter
from database_iterator import DatabaseIterator
from encoder import IdentityEncoder, PickleEncoder
from decoder import Decoder, PickleDecoder
from lmdbstore import LmdbDatabase
from objectstore import ObjectStoreDatabase
from persistence import PersistenceSettings
from iteratornode import IteratorNode
from eventlog import EventLog, RedisChannel
try:
from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \
BaseNumpyDecoder, NumpyMetaData, NumpyFeature
except ImportError:
pass
|
Add PickleEncoder to the public API
|
Add PickleEncoder to the public API
|
Python
|
mit
|
JohnVinyard/featureflow,JohnVinyard/featureflow
|
__version__ = '1.16.14'
from model import BaseModel
from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \
PickleFeature
from extractor import Node, Graph, Aggregator, NotEnoughData
from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip
from data import \
IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \
KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \
InMemoryDatabase
from datawriter import DataWriter
from database_iterator import DatabaseIterator
from encoder import IdentityEncoder
from decoder import Decoder, PickleDecoder
from lmdbstore import LmdbDatabase
from objectstore import ObjectStoreDatabase
from persistence import PersistenceSettings
from iteratornode import IteratorNode
from eventlog import EventLog, RedisChannel
try:
from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \
BaseNumpyDecoder, NumpyMetaData, NumpyFeature
except ImportError:
pass
Add PickleEncoder to the public API
|
__version__ = '1.16.14'
from model import BaseModel
from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \
PickleFeature
from extractor import Node, Graph, Aggregator, NotEnoughData
from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip
from data import \
IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \
KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \
InMemoryDatabase
from datawriter import DataWriter
from database_iterator import DatabaseIterator
from encoder import IdentityEncoder, PickleEncoder
from decoder import Decoder, PickleDecoder
from lmdbstore import LmdbDatabase
from objectstore import ObjectStoreDatabase
from persistence import PersistenceSettings
from iteratornode import IteratorNode
from eventlog import EventLog, RedisChannel
try:
from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \
BaseNumpyDecoder, NumpyMetaData, NumpyFeature
except ImportError:
pass
|
<commit_before>__version__ = '1.16.14'
from model import BaseModel
from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \
PickleFeature
from extractor import Node, Graph, Aggregator, NotEnoughData
from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip
from data import \
IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \
KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \
InMemoryDatabase
from datawriter import DataWriter
from database_iterator import DatabaseIterator
from encoder import IdentityEncoder
from decoder import Decoder, PickleDecoder
from lmdbstore import LmdbDatabase
from objectstore import ObjectStoreDatabase
from persistence import PersistenceSettings
from iteratornode import IteratorNode
from eventlog import EventLog, RedisChannel
try:
from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \
BaseNumpyDecoder, NumpyMetaData, NumpyFeature
except ImportError:
pass
<commit_msg>Add PickleEncoder to the public API<commit_after>
|
__version__ = '1.16.14'
from model import BaseModel
from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \
PickleFeature
from extractor import Node, Graph, Aggregator, NotEnoughData
from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip
from data import \
IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \
KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \
InMemoryDatabase
from datawriter import DataWriter
from database_iterator import DatabaseIterator
from encoder import IdentityEncoder, PickleEncoder
from decoder import Decoder, PickleDecoder
from lmdbstore import LmdbDatabase
from objectstore import ObjectStoreDatabase
from persistence import PersistenceSettings
from iteratornode import IteratorNode
from eventlog import EventLog, RedisChannel
try:
from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \
BaseNumpyDecoder, NumpyMetaData, NumpyFeature
except ImportError:
pass
|
__version__ = '1.16.14'
from model import BaseModel
from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \
PickleFeature
from extractor import Node, Graph, Aggregator, NotEnoughData
from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip
from data import \
IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \
KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \
InMemoryDatabase
from datawriter import DataWriter
from database_iterator import DatabaseIterator
from encoder import IdentityEncoder
from decoder import Decoder, PickleDecoder
from lmdbstore import LmdbDatabase
from objectstore import ObjectStoreDatabase
from persistence import PersistenceSettings
from iteratornode import IteratorNode
from eventlog import EventLog, RedisChannel
try:
from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \
BaseNumpyDecoder, NumpyMetaData, NumpyFeature
except ImportError:
pass
Add PickleEncoder to the public API__version__ = '1.16.14'
from model import BaseModel
from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \
PickleFeature
from extractor import Node, Graph, Aggregator, NotEnoughData
from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip
from data import \
IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \
KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \
InMemoryDatabase
from datawriter import DataWriter
from database_iterator import DatabaseIterator
from encoder import IdentityEncoder, PickleEncoder
from decoder import Decoder, PickleDecoder
from lmdbstore import LmdbDatabase
from objectstore import ObjectStoreDatabase
from persistence import PersistenceSettings
from iteratornode import IteratorNode
from eventlog import EventLog, RedisChannel
try:
from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \
BaseNumpyDecoder, NumpyMetaData, NumpyFeature
except ImportError:
pass
|
<commit_before>__version__ = '1.16.14'
from model import BaseModel
from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \
PickleFeature
from extractor import Node, Graph, Aggregator, NotEnoughData
from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip
from data import \
IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \
KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \
InMemoryDatabase
from datawriter import DataWriter
from database_iterator import DatabaseIterator
from encoder import IdentityEncoder
from decoder import Decoder, PickleDecoder
from lmdbstore import LmdbDatabase
from objectstore import ObjectStoreDatabase
from persistence import PersistenceSettings
from iteratornode import IteratorNode
from eventlog import EventLog, RedisChannel
try:
from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \
BaseNumpyDecoder, NumpyMetaData, NumpyFeature
except ImportError:
pass
<commit_msg>Add PickleEncoder to the public API<commit_after>__version__ = '1.16.14'
from model import BaseModel
from feature import Feature, JSONFeature, TextFeature, CompressedFeature, \
PickleFeature
from extractor import Node, Graph, Aggregator, NotEnoughData
from bytestream import ByteStream, ByteStreamFeature, ZipWrapper, iter_zip
from data import \
IdProvider, UuidProvider, UserSpecifiedIdProvider, StaticIdProvider, \
KeyBuilder, StringDelimitedKeyBuilder, Database, FileSystemDatabase, \
InMemoryDatabase
from datawriter import DataWriter
from database_iterator import DatabaseIterator
from encoder import IdentityEncoder, PickleEncoder
from decoder import Decoder, PickleDecoder
from lmdbstore import LmdbDatabase
from objectstore import ObjectStoreDatabase
from persistence import PersistenceSettings
from iteratornode import IteratorNode
from eventlog import EventLog, RedisChannel
try:
from nmpy import NumpyEncoder, PackedNumpyEncoder, StreamingNumpyDecoder, \
BaseNumpyDecoder, NumpyMetaData, NumpyFeature
except ImportError:
pass
|
d087e0cc47697e6b7f222de90a4143e3bb612a66
|
radar/models/forms.py
|
radar/models/forms.py
|
from sqlalchemy import Column, Integer, ForeignKey, String
from sqlalchemy.orm import relationship
from sqlalchemy.dialects import postgresql
from radar.database import db
from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship
from radar.models.logs import log_changes
class Form(db.Model):
__tablename__ = 'forms'
id = Column(Integer, primary_key=True)
name = Column(String, nullable=False)
data = Column(postgresql.JSONB, nullable=False)
@log_changes
class Entry(db.Model, MetaModelMixin):
__tablename__ = 'entries'
id = uuid_pk_column()
patient_id = patient_id_column()
patient = patient_relationship('entries')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
data = Column(postgresql.JSONB, nullable=False)
class GroupForm(db.Model):
__tablename__ = 'group_forms'
id = Column(Integer, primary_key=True)
group_id = Column(Integer, ForeignKey('groups.id'), nullable=False)
group = relationship('Group')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
|
from sqlalchemy import Column, Integer, ForeignKey, String
from sqlalchemy.orm import relationship
from sqlalchemy.dialects import postgresql
from radar.database import db
from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship
from radar.models.logs import log_changes
class Form(db.Model):
__tablename__ = 'forms'
id = Column(Integer, primary_key=True)
name = Column(String, nullable=False)
data = Column(postgresql.JSONB, nullable=False)
@log_changes
class Entry(db.Model, MetaModelMixin):
__tablename__ = 'entries'
id = uuid_pk_column()
patient_id = patient_id_column()
patient = patient_relationship('entries')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
data = Column(postgresql.JSONB, nullable=False)
Index('entries_patient_idx', Entry.patient_id)
class GroupForm(db.Model):
__tablename__ = 'group_forms'
id = Column(Integer, primary_key=True)
group_id = Column(Integer, ForeignKey('groups.id'), nullable=False)
group = relationship('Group')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
|
Add index on patient id
|
Add index on patient id
|
Python
|
agpl-3.0
|
renalreg/radar,renalreg/radar,renalreg/radar,renalreg/radar
|
from sqlalchemy import Column, Integer, ForeignKey, String
from sqlalchemy.orm import relationship
from sqlalchemy.dialects import postgresql
from radar.database import db
from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship
from radar.models.logs import log_changes
class Form(db.Model):
__tablename__ = 'forms'
id = Column(Integer, primary_key=True)
name = Column(String, nullable=False)
data = Column(postgresql.JSONB, nullable=False)
@log_changes
class Entry(db.Model, MetaModelMixin):
__tablename__ = 'entries'
id = uuid_pk_column()
patient_id = patient_id_column()
patient = patient_relationship('entries')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
data = Column(postgresql.JSONB, nullable=False)
class GroupForm(db.Model):
__tablename__ = 'group_forms'
id = Column(Integer, primary_key=True)
group_id = Column(Integer, ForeignKey('groups.id'), nullable=False)
group = relationship('Group')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')Add index on patient id
|
from sqlalchemy import Column, Integer, ForeignKey, String
from sqlalchemy.orm import relationship
from sqlalchemy.dialects import postgresql
from radar.database import db
from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship
from radar.models.logs import log_changes
class Form(db.Model):
__tablename__ = 'forms'
id = Column(Integer, primary_key=True)
name = Column(String, nullable=False)
data = Column(postgresql.JSONB, nullable=False)
@log_changes
class Entry(db.Model, MetaModelMixin):
__tablename__ = 'entries'
id = uuid_pk_column()
patient_id = patient_id_column()
patient = patient_relationship('entries')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
data = Column(postgresql.JSONB, nullable=False)
Index('entries_patient_idx', Entry.patient_id)
class GroupForm(db.Model):
__tablename__ = 'group_forms'
id = Column(Integer, primary_key=True)
group_id = Column(Integer, ForeignKey('groups.id'), nullable=False)
group = relationship('Group')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
|
<commit_before>from sqlalchemy import Column, Integer, ForeignKey, String
from sqlalchemy.orm import relationship
from sqlalchemy.dialects import postgresql
from radar.database import db
from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship
from radar.models.logs import log_changes
class Form(db.Model):
__tablename__ = 'forms'
id = Column(Integer, primary_key=True)
name = Column(String, nullable=False)
data = Column(postgresql.JSONB, nullable=False)
@log_changes
class Entry(db.Model, MetaModelMixin):
__tablename__ = 'entries'
id = uuid_pk_column()
patient_id = patient_id_column()
patient = patient_relationship('entries')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
data = Column(postgresql.JSONB, nullable=False)
class GroupForm(db.Model):
__tablename__ = 'group_forms'
id = Column(Integer, primary_key=True)
group_id = Column(Integer, ForeignKey('groups.id'), nullable=False)
group = relationship('Group')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')<commit_msg>Add index on patient id<commit_after>
|
from sqlalchemy import Column, Integer, ForeignKey, String
from sqlalchemy.orm import relationship
from sqlalchemy.dialects import postgresql
from radar.database import db
from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship
from radar.models.logs import log_changes
class Form(db.Model):
__tablename__ = 'forms'
id = Column(Integer, primary_key=True)
name = Column(String, nullable=False)
data = Column(postgresql.JSONB, nullable=False)
@log_changes
class Entry(db.Model, MetaModelMixin):
__tablename__ = 'entries'
id = uuid_pk_column()
patient_id = patient_id_column()
patient = patient_relationship('entries')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
data = Column(postgresql.JSONB, nullable=False)
Index('entries_patient_idx', Entry.patient_id)
class GroupForm(db.Model):
__tablename__ = 'group_forms'
id = Column(Integer, primary_key=True)
group_id = Column(Integer, ForeignKey('groups.id'), nullable=False)
group = relationship('Group')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
|
from sqlalchemy import Column, Integer, ForeignKey, String
from sqlalchemy.orm import relationship
from sqlalchemy.dialects import postgresql
from radar.database import db
from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship
from radar.models.logs import log_changes
class Form(db.Model):
__tablename__ = 'forms'
id = Column(Integer, primary_key=True)
name = Column(String, nullable=False)
data = Column(postgresql.JSONB, nullable=False)
@log_changes
class Entry(db.Model, MetaModelMixin):
__tablename__ = 'entries'
id = uuid_pk_column()
patient_id = patient_id_column()
patient = patient_relationship('entries')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
data = Column(postgresql.JSONB, nullable=False)
class GroupForm(db.Model):
__tablename__ = 'group_forms'
id = Column(Integer, primary_key=True)
group_id = Column(Integer, ForeignKey('groups.id'), nullable=False)
group = relationship('Group')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')Add index on patient idfrom sqlalchemy import Column, Integer, ForeignKey, String
from sqlalchemy.orm import relationship
from sqlalchemy.dialects import postgresql
from radar.database import db
from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship
from radar.models.logs import log_changes
class Form(db.Model):
__tablename__ = 'forms'
id = Column(Integer, primary_key=True)
name = Column(String, nullable=False)
data = Column(postgresql.JSONB, nullable=False)
@log_changes
class Entry(db.Model, MetaModelMixin):
__tablename__ = 'entries'
id = uuid_pk_column()
patient_id = patient_id_column()
patient = patient_relationship('entries')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
data = Column(postgresql.JSONB, nullable=False)
Index('entries_patient_idx', Entry.patient_id)
class GroupForm(db.Model):
__tablename__ = 'group_forms'
id = Column(Integer, primary_key=True)
group_id = Column(Integer, ForeignKey('groups.id'), nullable=False)
group = relationship('Group')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
|
<commit_before>from sqlalchemy import Column, Integer, ForeignKey, String
from sqlalchemy.orm import relationship
from sqlalchemy.dialects import postgresql
from radar.database import db
from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship
from radar.models.logs import log_changes
class Form(db.Model):
__tablename__ = 'forms'
id = Column(Integer, primary_key=True)
name = Column(String, nullable=False)
data = Column(postgresql.JSONB, nullable=False)
@log_changes
class Entry(db.Model, MetaModelMixin):
__tablename__ = 'entries'
id = uuid_pk_column()
patient_id = patient_id_column()
patient = patient_relationship('entries')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
data = Column(postgresql.JSONB, nullable=False)
class GroupForm(db.Model):
__tablename__ = 'group_forms'
id = Column(Integer, primary_key=True)
group_id = Column(Integer, ForeignKey('groups.id'), nullable=False)
group = relationship('Group')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')<commit_msg>Add index on patient id<commit_after>from sqlalchemy import Column, Integer, ForeignKey, String
from sqlalchemy.orm import relationship
from sqlalchemy.dialects import postgresql
from radar.database import db
from radar.models.common import uuid_pk_column, MetaModelMixin, patient_id_column, patient_relationship
from radar.models.logs import log_changes
class Form(db.Model):
__tablename__ = 'forms'
id = Column(Integer, primary_key=True)
name = Column(String, nullable=False)
data = Column(postgresql.JSONB, nullable=False)
@log_changes
class Entry(db.Model, MetaModelMixin):
__tablename__ = 'entries'
id = uuid_pk_column()
patient_id = patient_id_column()
patient = patient_relationship('entries')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
data = Column(postgresql.JSONB, nullable=False)
Index('entries_patient_idx', Entry.patient_id)
class GroupForm(db.Model):
__tablename__ = 'group_forms'
id = Column(Integer, primary_key=True)
group_id = Column(Integer, ForeignKey('groups.id'), nullable=False)
group = relationship('Group')
form_id = Column(Integer, ForeignKey('forms.id'), nullable=False)
form = relationship('Form')
|
e00fb0d87b60a982c2d932864a67a70e7d5b4312
|
src/apps/rDSN.monitor/rDSN.Monitor.py
|
src/apps/rDSN.monitor/rDSN.Monitor.py
|
import sys
import os
import threading
import time
sys.path.append(os.getcwd() + '/app_package')
from MonitorApp import *
def start_dsn():
service_app = ServiceApp()
app_dict['monitor'] = MonitorService
service_app.register_app('monitor')
if len(sys.argv) < 2:
#rDSN.Monitor run as an embedded service
print "rDSN.Monitor runs in embedded mode"
Native.dsn_app_loader_signal()
time.sleep(1)
elif sys.argv[1] == 'standalone':
#rDSN.Monitor run as a caller calling the monitored program
print "rDSN.Monitor runs in standalone mode"
argv = (c_char_p*2)()
argv[0] = b'rDSN.Monitor.exe'
argv[1] = b'config.ini'
Native.dsn_run(2, argv, c_bool(1))
if __name__ == '__main__':
start_dsn()
|
import sys
import os
import threading
sys.path.append(os.getcwd() + '/app_package')
from MonitorApp import *
def start_dsn():
service_app = ServiceApp()
app_dict['monitor'] = MonitorService
service_app.register_app('monitor')
if len(sys.argv) < 2:
#rDSN.Monitor run as an embedded service
print "rDSN.Monitor runs in embedded mode"
Native.dsn_app_loader_signal()
#to be fix, hangs forever now to keep python interpreter alive
dummy_event = threading.Event()
dummy_event.wait()
elif sys.argv[1] == 'standalone':
#rDSN.Monitor run as a caller calling the monitored program
print "rDSN.Monitor runs in standalone mode"
argv = (c_char_p*2)()
argv[0] = b'rDSN.Monitor.exe'
argv[1] = b'config.ini'
Native.dsn_run(2, argv, c_bool(1))
if __name__ == '__main__':
start_dsn()
|
Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app starts
|
Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app starts
|
Python
|
mit
|
mcfatealan/rDSN.Python,rDSN-Projects/rDSN.Python,mcfatealan/rDSN.Python,mcfatealan/rDSN.Python,rDSN-Projects/rDSN.Python,mcfatealan/rDSN.Python,rDSN-Projects/rDSN.Python,rDSN-Projects/rDSN.Python,rDSN-Projects/rDSN.Python,mcfatealan/rDSN.Python
|
import sys
import os
import threading
import time
sys.path.append(os.getcwd() + '/app_package')
from MonitorApp import *
def start_dsn():
service_app = ServiceApp()
app_dict['monitor'] = MonitorService
service_app.register_app('monitor')
if len(sys.argv) < 2:
#rDSN.Monitor run as an embedded service
print "rDSN.Monitor runs in embedded mode"
Native.dsn_app_loader_signal()
time.sleep(1)
elif sys.argv[1] == 'standalone':
#rDSN.Monitor run as a caller calling the monitored program
print "rDSN.Monitor runs in standalone mode"
argv = (c_char_p*2)()
argv[0] = b'rDSN.Monitor.exe'
argv[1] = b'config.ini'
Native.dsn_run(2, argv, c_bool(1))
if __name__ == '__main__':
start_dsn()
Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app starts
|
import sys
import os
import threading
sys.path.append(os.getcwd() + '/app_package')
from MonitorApp import *
def start_dsn():
service_app = ServiceApp()
app_dict['monitor'] = MonitorService
service_app.register_app('monitor')
if len(sys.argv) < 2:
#rDSN.Monitor run as an embedded service
print "rDSN.Monitor runs in embedded mode"
Native.dsn_app_loader_signal()
#to be fix, hangs forever now to keep python interpreter alive
dummy_event = threading.Event()
dummy_event.wait()
elif sys.argv[1] == 'standalone':
#rDSN.Monitor run as a caller calling the monitored program
print "rDSN.Monitor runs in standalone mode"
argv = (c_char_p*2)()
argv[0] = b'rDSN.Monitor.exe'
argv[1] = b'config.ini'
Native.dsn_run(2, argv, c_bool(1))
if __name__ == '__main__':
start_dsn()
|
<commit_before>import sys
import os
import threading
import time
sys.path.append(os.getcwd() + '/app_package')
from MonitorApp import *
def start_dsn():
service_app = ServiceApp()
app_dict['monitor'] = MonitorService
service_app.register_app('monitor')
if len(sys.argv) < 2:
#rDSN.Monitor run as an embedded service
print "rDSN.Monitor runs in embedded mode"
Native.dsn_app_loader_signal()
time.sleep(1)
elif sys.argv[1] == 'standalone':
#rDSN.Monitor run as a caller calling the monitored program
print "rDSN.Monitor runs in standalone mode"
argv = (c_char_p*2)()
argv[0] = b'rDSN.Monitor.exe'
argv[1] = b'config.ini'
Native.dsn_run(2, argv, c_bool(1))
if __name__ == '__main__':
start_dsn()
<commit_msg>Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app starts<commit_after>
|
import sys
import os
import threading
sys.path.append(os.getcwd() + '/app_package')
from MonitorApp import *
def start_dsn():
service_app = ServiceApp()
app_dict['monitor'] = MonitorService
service_app.register_app('monitor')
if len(sys.argv) < 2:
#rDSN.Monitor run as an embedded service
print "rDSN.Monitor runs in embedded mode"
Native.dsn_app_loader_signal()
#to be fix, hangs forever now to keep python interpreter alive
dummy_event = threading.Event()
dummy_event.wait()
elif sys.argv[1] == 'standalone':
#rDSN.Monitor run as a caller calling the monitored program
print "rDSN.Monitor runs in standalone mode"
argv = (c_char_p*2)()
argv[0] = b'rDSN.Monitor.exe'
argv[1] = b'config.ini'
Native.dsn_run(2, argv, c_bool(1))
if __name__ == '__main__':
start_dsn()
|
import sys
import os
import threading
import time
sys.path.append(os.getcwd() + '/app_package')
from MonitorApp import *
def start_dsn():
service_app = ServiceApp()
app_dict['monitor'] = MonitorService
service_app.register_app('monitor')
if len(sys.argv) < 2:
#rDSN.Monitor run as an embedded service
print "rDSN.Monitor runs in embedded mode"
Native.dsn_app_loader_signal()
time.sleep(1)
elif sys.argv[1] == 'standalone':
#rDSN.Monitor run as a caller calling the monitored program
print "rDSN.Monitor runs in standalone mode"
argv = (c_char_p*2)()
argv[0] = b'rDSN.Monitor.exe'
argv[1] = b'config.ini'
Native.dsn_run(2, argv, c_bool(1))
if __name__ == '__main__':
start_dsn()
Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app startsimport sys
import os
import threading
sys.path.append(os.getcwd() + '/app_package')
from MonitorApp import *
def start_dsn():
service_app = ServiceApp()
app_dict['monitor'] = MonitorService
service_app.register_app('monitor')
if len(sys.argv) < 2:
#rDSN.Monitor run as an embedded service
print "rDSN.Monitor runs in embedded mode"
Native.dsn_app_loader_signal()
#to be fix, hangs forever now to keep python interpreter alive
dummy_event = threading.Event()
dummy_event.wait()
elif sys.argv[1] == 'standalone':
#rDSN.Monitor run as a caller calling the monitored program
print "rDSN.Monitor runs in standalone mode"
argv = (c_char_p*2)()
argv[0] = b'rDSN.Monitor.exe'
argv[1] = b'config.ini'
Native.dsn_run(2, argv, c_bool(1))
if __name__ == '__main__':
start_dsn()
|
<commit_before>import sys
import os
import threading
import time
sys.path.append(os.getcwd() + '/app_package')
from MonitorApp import *
def start_dsn():
service_app = ServiceApp()
app_dict['monitor'] = MonitorService
service_app.register_app('monitor')
if len(sys.argv) < 2:
#rDSN.Monitor run as an embedded service
print "rDSN.Monitor runs in embedded mode"
Native.dsn_app_loader_signal()
time.sleep(1)
elif sys.argv[1] == 'standalone':
#rDSN.Monitor run as a caller calling the monitored program
print "rDSN.Monitor runs in standalone mode"
argv = (c_char_p*2)()
argv[0] = b'rDSN.Monitor.exe'
argv[1] = b'config.ini'
Native.dsn_run(2, argv, c_bool(1))
if __name__ == '__main__':
start_dsn()
<commit_msg>Replace sleep() with wait() forever after monitor registers, this ensures the python interpreter alive before app starts<commit_after>import sys
import os
import threading
sys.path.append(os.getcwd() + '/app_package')
from MonitorApp import *
def start_dsn():
service_app = ServiceApp()
app_dict['monitor'] = MonitorService
service_app.register_app('monitor')
if len(sys.argv) < 2:
#rDSN.Monitor run as an embedded service
print "rDSN.Monitor runs in embedded mode"
Native.dsn_app_loader_signal()
#to be fix, hangs forever now to keep python interpreter alive
dummy_event = threading.Event()
dummy_event.wait()
elif sys.argv[1] == 'standalone':
#rDSN.Monitor run as a caller calling the monitored program
print "rDSN.Monitor runs in standalone mode"
argv = (c_char_p*2)()
argv[0] = b'rDSN.Monitor.exe'
argv[1] = b'config.ini'
Native.dsn_run(2, argv, c_bool(1))
if __name__ == '__main__':
start_dsn()
|
65daee8f169e8bb6e721ce016c7bcf6cb9893016
|
froide/problem/utils.py
|
froide/problem/utils.py
|
from django.core.mail import mail_managers
from django.conf import settings
from django.urls import reverse
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
def inform_managers(report):
admin_url = settings.SITE_URL + reverse(
'admin:problem_problemreport_change', args=(report.id,))
mail_managers(
_('New problem: {label} [#{reqid}]').format(
label=report.get_kind_display(),
reqid=report.message.request_id
),
'{}\n\n---\n\n{}\n'.format(
report.description,
report.get_absolute_domain_url(),
admin_url
)
)
def inform_user_problem_resolved(report):
if report.auto_submitted or not report.user:
return False
foirequest = report.message.request
subject = _('Problem resolved on your request')
body = render_to_string("problem/email_problem_resolved.txt", {
"user": report.user,
"title": foirequest.title,
"report": report,
"url": report.user.get_autologin_url(
report.message.get_absolute_short_url()
),
"site_name": settings.SITE_NAME
})
report.user.send_mail(subject, body)
return True
|
from django.core.mail import mail_managers
from django.conf import settings
from django.urls import reverse
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
def inform_managers(report):
admin_url = settings.SITE_URL + reverse(
'admin:problem_problemreport_change', args=(report.id,))
mail_managers(
_('New problem: {label} [#{reqid}]').format(
label=report.get_kind_display(),
reqid=report.message.request_id
),
'{}\n\n---\n\n{}\n{}'.format(
report.description,
report.get_absolute_domain_url(),
admin_url
)
)
def inform_user_problem_resolved(report):
if report.auto_submitted or not report.user:
return False
foirequest = report.message.request
subject = _('Problem resolved on your request')
body = render_to_string("problem/email_problem_resolved.txt", {
"user": report.user,
"title": foirequest.title,
"report": report,
"url": report.user.get_autologin_url(
report.message.get_absolute_short_url()
),
"site_name": settings.SITE_NAME
})
report.user.send_mail(subject, body)
return True
|
Add problem admin URL to problem report email
|
Add problem admin URL to problem report email
|
Python
|
mit
|
stefanw/froide,stefanw/froide,stefanw/froide,fin/froide,fin/froide,stefanw/froide,fin/froide,stefanw/froide,fin/froide
|
from django.core.mail import mail_managers
from django.conf import settings
from django.urls import reverse
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
def inform_managers(report):
admin_url = settings.SITE_URL + reverse(
'admin:problem_problemreport_change', args=(report.id,))
mail_managers(
_('New problem: {label} [#{reqid}]').format(
label=report.get_kind_display(),
reqid=report.message.request_id
),
'{}\n\n---\n\n{}\n'.format(
report.description,
report.get_absolute_domain_url(),
admin_url
)
)
def inform_user_problem_resolved(report):
if report.auto_submitted or not report.user:
return False
foirequest = report.message.request
subject = _('Problem resolved on your request')
body = render_to_string("problem/email_problem_resolved.txt", {
"user": report.user,
"title": foirequest.title,
"report": report,
"url": report.user.get_autologin_url(
report.message.get_absolute_short_url()
),
"site_name": settings.SITE_NAME
})
report.user.send_mail(subject, body)
return True
Add problem admin URL to problem report email
|
from django.core.mail import mail_managers
from django.conf import settings
from django.urls import reverse
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
def inform_managers(report):
admin_url = settings.SITE_URL + reverse(
'admin:problem_problemreport_change', args=(report.id,))
mail_managers(
_('New problem: {label} [#{reqid}]').format(
label=report.get_kind_display(),
reqid=report.message.request_id
),
'{}\n\n---\n\n{}\n{}'.format(
report.description,
report.get_absolute_domain_url(),
admin_url
)
)
def inform_user_problem_resolved(report):
if report.auto_submitted or not report.user:
return False
foirequest = report.message.request
subject = _('Problem resolved on your request')
body = render_to_string("problem/email_problem_resolved.txt", {
"user": report.user,
"title": foirequest.title,
"report": report,
"url": report.user.get_autologin_url(
report.message.get_absolute_short_url()
),
"site_name": settings.SITE_NAME
})
report.user.send_mail(subject, body)
return True
|
<commit_before>from django.core.mail import mail_managers
from django.conf import settings
from django.urls import reverse
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
def inform_managers(report):
admin_url = settings.SITE_URL + reverse(
'admin:problem_problemreport_change', args=(report.id,))
mail_managers(
_('New problem: {label} [#{reqid}]').format(
label=report.get_kind_display(),
reqid=report.message.request_id
),
'{}\n\n---\n\n{}\n'.format(
report.description,
report.get_absolute_domain_url(),
admin_url
)
)
def inform_user_problem_resolved(report):
if report.auto_submitted or not report.user:
return False
foirequest = report.message.request
subject = _('Problem resolved on your request')
body = render_to_string("problem/email_problem_resolved.txt", {
"user": report.user,
"title": foirequest.title,
"report": report,
"url": report.user.get_autologin_url(
report.message.get_absolute_short_url()
),
"site_name": settings.SITE_NAME
})
report.user.send_mail(subject, body)
return True
<commit_msg>Add problem admin URL to problem report email<commit_after>
|
from django.core.mail import mail_managers
from django.conf import settings
from django.urls import reverse
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
def inform_managers(report):
admin_url = settings.SITE_URL + reverse(
'admin:problem_problemreport_change', args=(report.id,))
mail_managers(
_('New problem: {label} [#{reqid}]').format(
label=report.get_kind_display(),
reqid=report.message.request_id
),
'{}\n\n---\n\n{}\n{}'.format(
report.description,
report.get_absolute_domain_url(),
admin_url
)
)
def inform_user_problem_resolved(report):
if report.auto_submitted or not report.user:
return False
foirequest = report.message.request
subject = _('Problem resolved on your request')
body = render_to_string("problem/email_problem_resolved.txt", {
"user": report.user,
"title": foirequest.title,
"report": report,
"url": report.user.get_autologin_url(
report.message.get_absolute_short_url()
),
"site_name": settings.SITE_NAME
})
report.user.send_mail(subject, body)
return True
|
from django.core.mail import mail_managers
from django.conf import settings
from django.urls import reverse
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
def inform_managers(report):
admin_url = settings.SITE_URL + reverse(
'admin:problem_problemreport_change', args=(report.id,))
mail_managers(
_('New problem: {label} [#{reqid}]').format(
label=report.get_kind_display(),
reqid=report.message.request_id
),
'{}\n\n---\n\n{}\n'.format(
report.description,
report.get_absolute_domain_url(),
admin_url
)
)
def inform_user_problem_resolved(report):
if report.auto_submitted or not report.user:
return False
foirequest = report.message.request
subject = _('Problem resolved on your request')
body = render_to_string("problem/email_problem_resolved.txt", {
"user": report.user,
"title": foirequest.title,
"report": report,
"url": report.user.get_autologin_url(
report.message.get_absolute_short_url()
),
"site_name": settings.SITE_NAME
})
report.user.send_mail(subject, body)
return True
Add problem admin URL to problem report emailfrom django.core.mail import mail_managers
from django.conf import settings
from django.urls import reverse
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
def inform_managers(report):
admin_url = settings.SITE_URL + reverse(
'admin:problem_problemreport_change', args=(report.id,))
mail_managers(
_('New problem: {label} [#{reqid}]').format(
label=report.get_kind_display(),
reqid=report.message.request_id
),
'{}\n\n---\n\n{}\n{}'.format(
report.description,
report.get_absolute_domain_url(),
admin_url
)
)
def inform_user_problem_resolved(report):
if report.auto_submitted or not report.user:
return False
foirequest = report.message.request
subject = _('Problem resolved on your request')
body = render_to_string("problem/email_problem_resolved.txt", {
"user": report.user,
"title": foirequest.title,
"report": report,
"url": report.user.get_autologin_url(
report.message.get_absolute_short_url()
),
"site_name": settings.SITE_NAME
})
report.user.send_mail(subject, body)
return True
|
<commit_before>from django.core.mail import mail_managers
from django.conf import settings
from django.urls import reverse
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
def inform_managers(report):
admin_url = settings.SITE_URL + reverse(
'admin:problem_problemreport_change', args=(report.id,))
mail_managers(
_('New problem: {label} [#{reqid}]').format(
label=report.get_kind_display(),
reqid=report.message.request_id
),
'{}\n\n---\n\n{}\n'.format(
report.description,
report.get_absolute_domain_url(),
admin_url
)
)
def inform_user_problem_resolved(report):
if report.auto_submitted or not report.user:
return False
foirequest = report.message.request
subject = _('Problem resolved on your request')
body = render_to_string("problem/email_problem_resolved.txt", {
"user": report.user,
"title": foirequest.title,
"report": report,
"url": report.user.get_autologin_url(
report.message.get_absolute_short_url()
),
"site_name": settings.SITE_NAME
})
report.user.send_mail(subject, body)
return True
<commit_msg>Add problem admin URL to problem report email<commit_after>from django.core.mail import mail_managers
from django.conf import settings
from django.urls import reverse
from django.template.loader import render_to_string
from django.utils.translation import ugettext_lazy as _
def inform_managers(report):
admin_url = settings.SITE_URL + reverse(
'admin:problem_problemreport_change', args=(report.id,))
mail_managers(
_('New problem: {label} [#{reqid}]').format(
label=report.get_kind_display(),
reqid=report.message.request_id
),
'{}\n\n---\n\n{}\n{}'.format(
report.description,
report.get_absolute_domain_url(),
admin_url
)
)
def inform_user_problem_resolved(report):
if report.auto_submitted or not report.user:
return False
foirequest = report.message.request
subject = _('Problem resolved on your request')
body = render_to_string("problem/email_problem_resolved.txt", {
"user": report.user,
"title": foirequest.title,
"report": report,
"url": report.user.get_autologin_url(
report.message.get_absolute_short_url()
),
"site_name": settings.SITE_NAME
})
report.user.send_mail(subject, body)
return True
|
6a4e16f9afa373233c03cc8f1ede7076e9a44058
|
basics/utils.py
|
basics/utils.py
|
import numpy as np
def arctan_transform(array, thresh):
return np.arctan(array/thresh)
|
import numpy as np
from functools import partial
def arctan_transform(array, thresh):
return np.arctan(array/thresh)
def dist_uppertri(cond_arr, shape):
dist_arr = np.zeros((shape, ) * 2, dtype=cond_arr.dtype)
def unrav_ind(i, j, n):
return n*j - j*(j+1)/2 + i - 1 - j
arr_ind = partial(unrav_ind, n=shape)
for i in xrange(shape):
for j in xrange(i):
dist_arr[i, j] = cond_arr[arr_ind(i, j)]
return dist_arr
|
Convert a condensed distance matrix (pdist) into an upper triangular matrix
|
Convert a condensed distance matrix (pdist) into an upper triangular matrix
|
Python
|
mit
|
e-koch/BaSiCs
|
import numpy as np
def arctan_transform(array, thresh):
return np.arctan(array/thresh)
Convert a condensed distance matrix (pdist) into an upper triangular matrix
|
import numpy as np
from functools import partial
def arctan_transform(array, thresh):
return np.arctan(array/thresh)
def dist_uppertri(cond_arr, shape):
dist_arr = np.zeros((shape, ) * 2, dtype=cond_arr.dtype)
def unrav_ind(i, j, n):
return n*j - j*(j+1)/2 + i - 1 - j
arr_ind = partial(unrav_ind, n=shape)
for i in xrange(shape):
for j in xrange(i):
dist_arr[i, j] = cond_arr[arr_ind(i, j)]
return dist_arr
|
<commit_before>
import numpy as np
def arctan_transform(array, thresh):
return np.arctan(array/thresh)
<commit_msg>Convert a condensed distance matrix (pdist) into an upper triangular matrix<commit_after>
|
import numpy as np
from functools import partial
def arctan_transform(array, thresh):
return np.arctan(array/thresh)
def dist_uppertri(cond_arr, shape):
dist_arr = np.zeros((shape, ) * 2, dtype=cond_arr.dtype)
def unrav_ind(i, j, n):
return n*j - j*(j+1)/2 + i - 1 - j
arr_ind = partial(unrav_ind, n=shape)
for i in xrange(shape):
for j in xrange(i):
dist_arr[i, j] = cond_arr[arr_ind(i, j)]
return dist_arr
|
import numpy as np
def arctan_transform(array, thresh):
return np.arctan(array/thresh)
Convert a condensed distance matrix (pdist) into an upper triangular matrix
import numpy as np
from functools import partial
def arctan_transform(array, thresh):
return np.arctan(array/thresh)
def dist_uppertri(cond_arr, shape):
dist_arr = np.zeros((shape, ) * 2, dtype=cond_arr.dtype)
def unrav_ind(i, j, n):
return n*j - j*(j+1)/2 + i - 1 - j
arr_ind = partial(unrav_ind, n=shape)
for i in xrange(shape):
for j in xrange(i):
dist_arr[i, j] = cond_arr[arr_ind(i, j)]
return dist_arr
|
<commit_before>
import numpy as np
def arctan_transform(array, thresh):
return np.arctan(array/thresh)
<commit_msg>Convert a condensed distance matrix (pdist) into an upper triangular matrix<commit_after>
import numpy as np
from functools import partial
def arctan_transform(array, thresh):
return np.arctan(array/thresh)
def dist_uppertri(cond_arr, shape):
dist_arr = np.zeros((shape, ) * 2, dtype=cond_arr.dtype)
def unrav_ind(i, j, n):
return n*j - j*(j+1)/2 + i - 1 - j
arr_ind = partial(unrav_ind, n=shape)
for i in xrange(shape):
for j in xrange(i):
dist_arr[i, j] = cond_arr[arr_ind(i, j)]
return dist_arr
|
e97dee6ec7c49cf3d33803504c7269a41c4d0a0f
|
authentication_app/views.py
|
authentication_app/views.py
|
from django.shortcuts import render
from django.http import HttpResponse
from .models import Greeting
# Create your views here.
def index(request):
return HttpResponse('Hello from Python!')
def db(request):
greeting = Greeting()
greeting.save()
greetings = Greeting.objects.all()
return render(request, 'db.html', {'greetings': greetings})
|
from rest_framework import permissions, viewsets
from authentication_app.models import Account
from authentication_app.permissions import IsAccountOwner
from authentication_app.serializers import AccountSerializer
'''
@name : AccountViewSerializer
@desc : Defines the serializer for the account view.
'''
class AccountViewSerializer(viewsets.ModelViewSet):
lookup_field = 'username'
queryset = Account.objects.all()
serializer_class = AccountSerializer
def get_permissions(self):
if self.reqiest.method in permissions.SAFE_METHODS:
return (permissions.AllowAny(),)
if self.request.method == 'POST':
return (permissions.AllowAny(),)
return (permissions.IsAuthenticated(), IsAccountOwner(),)
def create(self, request):
serializer = self.serializer_class(data=reqiest.data)
if serializer.is_valid():
Account.objects.create_user(**serializer.validated_data)
return Response(serializer.validated_data, status=status.HTTP_201_CREATED)
return Response({
'status' : 'Bad Request',
'message' : 'Account could not be created with the received data.'
}, status=status.HTTP_400_BAD_REQUEST)
|
Add the view serializer for the account model.
|
Add the view serializer for the account model.
|
Python
|
mit
|
mvpgomes/shopit-app,mvpgomes/shopit-app,mvpgomes/shopit-app,mvpgomes/shopit-app
|
from django.shortcuts import render
from django.http import HttpResponse
from .models import Greeting
# Create your views here.
def index(request):
return HttpResponse('Hello from Python!')
def db(request):
greeting = Greeting()
greeting.save()
greetings = Greeting.objects.all()
return render(request, 'db.html', {'greetings': greetings})
Add the view serializer for the account model.
|
from rest_framework import permissions, viewsets
from authentication_app.models import Account
from authentication_app.permissions import IsAccountOwner
from authentication_app.serializers import AccountSerializer
'''
@name : AccountViewSerializer
@desc : Defines the serializer for the account view.
'''
class AccountViewSerializer(viewsets.ModelViewSet):
lookup_field = 'username'
queryset = Account.objects.all()
serializer_class = AccountSerializer
def get_permissions(self):
if self.reqiest.method in permissions.SAFE_METHODS:
return (permissions.AllowAny(),)
if self.request.method == 'POST':
return (permissions.AllowAny(),)
return (permissions.IsAuthenticated(), IsAccountOwner(),)
def create(self, request):
serializer = self.serializer_class(data=reqiest.data)
if serializer.is_valid():
Account.objects.create_user(**serializer.validated_data)
return Response(serializer.validated_data, status=status.HTTP_201_CREATED)
return Response({
'status' : 'Bad Request',
'message' : 'Account could not be created with the received data.'
}, status=status.HTTP_400_BAD_REQUEST)
|
<commit_before>from django.shortcuts import render
from django.http import HttpResponse
from .models import Greeting
# Create your views here.
def index(request):
return HttpResponse('Hello from Python!')
def db(request):
greeting = Greeting()
greeting.save()
greetings = Greeting.objects.all()
return render(request, 'db.html', {'greetings': greetings})
<commit_msg>Add the view serializer for the account model.<commit_after>
|
from rest_framework import permissions, viewsets
from authentication_app.models import Account
from authentication_app.permissions import IsAccountOwner
from authentication_app.serializers import AccountSerializer
'''
@name : AccountViewSerializer
@desc : Defines the serializer for the account view.
'''
class AccountViewSerializer(viewsets.ModelViewSet):
lookup_field = 'username'
queryset = Account.objects.all()
serializer_class = AccountSerializer
def get_permissions(self):
if self.reqiest.method in permissions.SAFE_METHODS:
return (permissions.AllowAny(),)
if self.request.method == 'POST':
return (permissions.AllowAny(),)
return (permissions.IsAuthenticated(), IsAccountOwner(),)
def create(self, request):
serializer = self.serializer_class(data=reqiest.data)
if serializer.is_valid():
Account.objects.create_user(**serializer.validated_data)
return Response(serializer.validated_data, status=status.HTTP_201_CREATED)
return Response({
'status' : 'Bad Request',
'message' : 'Account could not be created with the received data.'
}, status=status.HTTP_400_BAD_REQUEST)
|
from django.shortcuts import render
from django.http import HttpResponse
from .models import Greeting
# Create your views here.
def index(request):
return HttpResponse('Hello from Python!')
def db(request):
greeting = Greeting()
greeting.save()
greetings = Greeting.objects.all()
return render(request, 'db.html', {'greetings': greetings})
Add the view serializer for the account model.from rest_framework import permissions, viewsets
from authentication_app.models import Account
from authentication_app.permissions import IsAccountOwner
from authentication_app.serializers import AccountSerializer
'''
@name : AccountViewSerializer
@desc : Defines the serializer for the account view.
'''
class AccountViewSerializer(viewsets.ModelViewSet):
lookup_field = 'username'
queryset = Account.objects.all()
serializer_class = AccountSerializer
def get_permissions(self):
if self.reqiest.method in permissions.SAFE_METHODS:
return (permissions.AllowAny(),)
if self.request.method == 'POST':
return (permissions.AllowAny(),)
return (permissions.IsAuthenticated(), IsAccountOwner(),)
def create(self, request):
serializer = self.serializer_class(data=reqiest.data)
if serializer.is_valid():
Account.objects.create_user(**serializer.validated_data)
return Response(serializer.validated_data, status=status.HTTP_201_CREATED)
return Response({
'status' : 'Bad Request',
'message' : 'Account could not be created with the received data.'
}, status=status.HTTP_400_BAD_REQUEST)
|
<commit_before>from django.shortcuts import render
from django.http import HttpResponse
from .models import Greeting
# Create your views here.
def index(request):
return HttpResponse('Hello from Python!')
def db(request):
greeting = Greeting()
greeting.save()
greetings = Greeting.objects.all()
return render(request, 'db.html', {'greetings': greetings})
<commit_msg>Add the view serializer for the account model.<commit_after>from rest_framework import permissions, viewsets
from authentication_app.models import Account
from authentication_app.permissions import IsAccountOwner
from authentication_app.serializers import AccountSerializer
'''
@name : AccountViewSerializer
@desc : Defines the serializer for the account view.
'''
class AccountViewSerializer(viewsets.ModelViewSet):
lookup_field = 'username'
queryset = Account.objects.all()
serializer_class = AccountSerializer
def get_permissions(self):
if self.reqiest.method in permissions.SAFE_METHODS:
return (permissions.AllowAny(),)
if self.request.method == 'POST':
return (permissions.AllowAny(),)
return (permissions.IsAuthenticated(), IsAccountOwner(),)
def create(self, request):
serializer = self.serializer_class(data=reqiest.data)
if serializer.is_valid():
Account.objects.create_user(**serializer.validated_data)
return Response(serializer.validated_data, status=status.HTTP_201_CREATED)
return Response({
'status' : 'Bad Request',
'message' : 'Account could not be created with the received data.'
}, status=status.HTTP_400_BAD_REQUEST)
|
f8304bb26151fdb999a77da9afbea8ff653a37f8
|
artists/views.py
|
artists/views.py
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
# TODO re-update old cumulative similarity if artist name changed
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
Add note to update old similarities
|
Add note to update old similarities
|
Python
|
bsd-3-clause
|
FreeMusicNinja/api.freemusic.ninja
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
Add note to update old similarities
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
# TODO re-update old cumulative similarity if artist name changed
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
<commit_before>from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
<commit_msg>Add note to update old similarities<commit_after>
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
# TODO re-update old cumulative similarity if artist name changed
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
Add note to update old similaritiesfrom django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
# TODO re-update old cumulative similarity if artist name changed
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
<commit_before>from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
<commit_msg>Add note to update old similarities<commit_after>from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
# TODO re-update old cumulative similarity if artist name changed
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
261fb861015ee96771e4c387bcd2b2c7d5c369db
|
hellopython/__init__.py
|
hellopython/__init__.py
|
__version__ = '1.0.0'
from story.story import BaseStory
from . import print_method
class Story(BaseStory):
name = 'hellopython'
adventures = [
print_method
]
|
__version__ = '1.0.0'
from story.story import BaseStory
from . import print_method
class Story(BaseStory):
name = 'hellopython'
title = 'Introuction to python'
adventures = [
print_method
]
|
Add a title to the story the story
|
Add a title to the story the story
|
Python
|
mit
|
pyschool/hipyschool
|
__version__ = '1.0.0'
from story.story import BaseStory
from . import print_method
class Story(BaseStory):
name = 'hellopython'
adventures = [
print_method
]
Add a title to the story the story
|
__version__ = '1.0.0'
from story.story import BaseStory
from . import print_method
class Story(BaseStory):
name = 'hellopython'
title = 'Introuction to python'
adventures = [
print_method
]
|
<commit_before>__version__ = '1.0.0'
from story.story import BaseStory
from . import print_method
class Story(BaseStory):
name = 'hellopython'
adventures = [
print_method
]
<commit_msg>Add a title to the story the story<commit_after>
|
__version__ = '1.0.0'
from story.story import BaseStory
from . import print_method
class Story(BaseStory):
name = 'hellopython'
title = 'Introuction to python'
adventures = [
print_method
]
|
__version__ = '1.0.0'
from story.story import BaseStory
from . import print_method
class Story(BaseStory):
name = 'hellopython'
adventures = [
print_method
]
Add a title to the story the story__version__ = '1.0.0'
from story.story import BaseStory
from . import print_method
class Story(BaseStory):
name = 'hellopython'
title = 'Introuction to python'
adventures = [
print_method
]
|
<commit_before>__version__ = '1.0.0'
from story.story import BaseStory
from . import print_method
class Story(BaseStory):
name = 'hellopython'
adventures = [
print_method
]
<commit_msg>Add a title to the story the story<commit_after>__version__ = '1.0.0'
from story.story import BaseStory
from . import print_method
class Story(BaseStory):
name = 'hellopython'
title = 'Introuction to python'
adventures = [
print_method
]
|
8d11c6854e9c2309abb74a2e4b960a5206a27a0c
|
funbox/iterators_ordered.py
|
funbox/iterators_ordered.py
|
#! /usr/bin/env python
"""Functions on iterators, optimised for case when iterators are sorted.
Note sift_o is hidden as _sift_o at the moment because it is broken.
Please don't use it.
Once fixed, I'll remove the leading underscore again.
"""
import itertools
import iterators
def partition_o(left_function, items):
"""Return a pair of iterators: left and right
Items for which left_function returns a true value go into left.
Items for which left_function returns a false value go into right.
Items must be sorted such that left_function may be true for an
initial set of items, but once an item is found such that
left_function(item) is false, it will remain false for the rest of the
items.
In other words the following must hold:
for all N where 0 <= N < (len(items) - 1) :
not(func(item[n])) => not(func(item[n+1]))
For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but
not for [1,3,4,2,5].
>>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5])
>>> list(left)
[-1, 0, 2]
>>> list(right)
[3, 4, 5]
"""
left = itertools.takewhile(left_function, items)
right = itertools.dropwhile(left_function, items)
return left, right
if __name__ == "__main__":
import doctest
doctest.testmod()
|
#! /usr/bin/env python
"""Functions on iterators, optimised for case when iterators are sorted.
"""
import itertools
import iterators
def partition_o(left_function, items):
"""Return a pair of iterators: left and right
Items for which left_function returns a true value go into left.
Items for which left_function returns a false value go into right.
Items must be sorted such that left_function may be true for an
initial set of items, but once an item is found such that
left_function(item) is false, it will remain false for the rest of the
items.
In other words the following must hold:
for all N where 0 <= N < (len(items) - 1) :
not(func(item[n])) => not(func(item[n+1]))
For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but
not for [1,3,4,2,5].
>>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5])
>>> list(left)
[-1, 0, 2]
>>> list(right)
[3, 4, 5]
"""
left = itertools.takewhile(left_function, items)
right = itertools.dropwhile(left_function, items)
return left, right
if __name__ == "__main__":
import doctest
doctest.testmod()
|
Remove reference in docs to removed function.
|
Remove reference in docs to removed function.
|
Python
|
mit
|
nmbooker/python-funbox,nmbooker/python-funbox
|
#! /usr/bin/env python
"""Functions on iterators, optimised for case when iterators are sorted.
Note sift_o is hidden as _sift_o at the moment because it is broken.
Please don't use it.
Once fixed, I'll remove the leading underscore again.
"""
import itertools
import iterators
def partition_o(left_function, items):
"""Return a pair of iterators: left and right
Items for which left_function returns a true value go into left.
Items for which left_function returns a false value go into right.
Items must be sorted such that left_function may be true for an
initial set of items, but once an item is found such that
left_function(item) is false, it will remain false for the rest of the
items.
In other words the following must hold:
for all N where 0 <= N < (len(items) - 1) :
not(func(item[n])) => not(func(item[n+1]))
For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but
not for [1,3,4,2,5].
>>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5])
>>> list(left)
[-1, 0, 2]
>>> list(right)
[3, 4, 5]
"""
left = itertools.takewhile(left_function, items)
right = itertools.dropwhile(left_function, items)
return left, right
if __name__ == "__main__":
import doctest
doctest.testmod()
Remove reference in docs to removed function.
|
#! /usr/bin/env python
"""Functions on iterators, optimised for case when iterators are sorted.
"""
import itertools
import iterators
def partition_o(left_function, items):
"""Return a pair of iterators: left and right
Items for which left_function returns a true value go into left.
Items for which left_function returns a false value go into right.
Items must be sorted such that left_function may be true for an
initial set of items, but once an item is found such that
left_function(item) is false, it will remain false for the rest of the
items.
In other words the following must hold:
for all N where 0 <= N < (len(items) - 1) :
not(func(item[n])) => not(func(item[n+1]))
For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but
not for [1,3,4,2,5].
>>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5])
>>> list(left)
[-1, 0, 2]
>>> list(right)
[3, 4, 5]
"""
left = itertools.takewhile(left_function, items)
right = itertools.dropwhile(left_function, items)
return left, right
if __name__ == "__main__":
import doctest
doctest.testmod()
|
<commit_before>#! /usr/bin/env python
"""Functions on iterators, optimised for case when iterators are sorted.
Note sift_o is hidden as _sift_o at the moment because it is broken.
Please don't use it.
Once fixed, I'll remove the leading underscore again.
"""
import itertools
import iterators
def partition_o(left_function, items):
"""Return a pair of iterators: left and right
Items for which left_function returns a true value go into left.
Items for which left_function returns a false value go into right.
Items must be sorted such that left_function may be true for an
initial set of items, but once an item is found such that
left_function(item) is false, it will remain false for the rest of the
items.
In other words the following must hold:
for all N where 0 <= N < (len(items) - 1) :
not(func(item[n])) => not(func(item[n+1]))
For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but
not for [1,3,4,2,5].
>>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5])
>>> list(left)
[-1, 0, 2]
>>> list(right)
[3, 4, 5]
"""
left = itertools.takewhile(left_function, items)
right = itertools.dropwhile(left_function, items)
return left, right
if __name__ == "__main__":
import doctest
doctest.testmod()
<commit_msg>Remove reference in docs to removed function.<commit_after>
|
#! /usr/bin/env python
"""Functions on iterators, optimised for case when iterators are sorted.
"""
import itertools
import iterators
def partition_o(left_function, items):
"""Return a pair of iterators: left and right
Items for which left_function returns a true value go into left.
Items for which left_function returns a false value go into right.
Items must be sorted such that left_function may be true for an
initial set of items, but once an item is found such that
left_function(item) is false, it will remain false for the rest of the
items.
In other words the following must hold:
for all N where 0 <= N < (len(items) - 1) :
not(func(item[n])) => not(func(item[n+1]))
For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but
not for [1,3,4,2,5].
>>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5])
>>> list(left)
[-1, 0, 2]
>>> list(right)
[3, 4, 5]
"""
left = itertools.takewhile(left_function, items)
right = itertools.dropwhile(left_function, items)
return left, right
if __name__ == "__main__":
import doctest
doctest.testmod()
|
#! /usr/bin/env python
"""Functions on iterators, optimised for case when iterators are sorted.
Note sift_o is hidden as _sift_o at the moment because it is broken.
Please don't use it.
Once fixed, I'll remove the leading underscore again.
"""
import itertools
import iterators
def partition_o(left_function, items):
"""Return a pair of iterators: left and right
Items for which left_function returns a true value go into left.
Items for which left_function returns a false value go into right.
Items must be sorted such that left_function may be true for an
initial set of items, but once an item is found such that
left_function(item) is false, it will remain false for the rest of the
items.
In other words the following must hold:
for all N where 0 <= N < (len(items) - 1) :
not(func(item[n])) => not(func(item[n+1]))
For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but
not for [1,3,4,2,5].
>>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5])
>>> list(left)
[-1, 0, 2]
>>> list(right)
[3, 4, 5]
"""
left = itertools.takewhile(left_function, items)
right = itertools.dropwhile(left_function, items)
return left, right
if __name__ == "__main__":
import doctest
doctest.testmod()
Remove reference in docs to removed function.#! /usr/bin/env python
"""Functions on iterators, optimised for case when iterators are sorted.
"""
import itertools
import iterators
def partition_o(left_function, items):
"""Return a pair of iterators: left and right
Items for which left_function returns a true value go into left.
Items for which left_function returns a false value go into right.
Items must be sorted such that left_function may be true for an
initial set of items, but once an item is found such that
left_function(item) is false, it will remain false for the rest of the
items.
In other words the following must hold:
for all N where 0 <= N < (len(items) - 1) :
not(func(item[n])) => not(func(item[n+1]))
For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but
not for [1,3,4,2,5].
>>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5])
>>> list(left)
[-1, 0, 2]
>>> list(right)
[3, 4, 5]
"""
left = itertools.takewhile(left_function, items)
right = itertools.dropwhile(left_function, items)
return left, right
if __name__ == "__main__":
import doctest
doctest.testmod()
|
<commit_before>#! /usr/bin/env python
"""Functions on iterators, optimised for case when iterators are sorted.
Note sift_o is hidden as _sift_o at the moment because it is broken.
Please don't use it.
Once fixed, I'll remove the leading underscore again.
"""
import itertools
import iterators
def partition_o(left_function, items):
"""Return a pair of iterators: left and right
Items for which left_function returns a true value go into left.
Items for which left_function returns a false value go into right.
Items must be sorted such that left_function may be true for an
initial set of items, but once an item is found such that
left_function(item) is false, it will remain false for the rest of the
items.
In other words the following must hold:
for all N where 0 <= N < (len(items) - 1) :
not(func(item[n])) => not(func(item[n+1]))
For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but
not for [1,3,4,2,5].
>>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5])
>>> list(left)
[-1, 0, 2]
>>> list(right)
[3, 4, 5]
"""
left = itertools.takewhile(left_function, items)
right = itertools.dropwhile(left_function, items)
return left, right
if __name__ == "__main__":
import doctest
doctest.testmod()
<commit_msg>Remove reference in docs to removed function.<commit_after>#! /usr/bin/env python
"""Functions on iterators, optimised for case when iterators are sorted.
"""
import itertools
import iterators
def partition_o(left_function, items):
"""Return a pair of iterators: left and right
Items for which left_function returns a true value go into left.
Items for which left_function returns a false value go into right.
Items must be sorted such that left_function may be true for an
initial set of items, but once an item is found such that
left_function(item) is false, it will remain false for the rest of the
items.
In other words the following must hold:
for all N where 0 <= N < (len(items) - 1) :
not(func(item[n])) => not(func(item[n+1]))
For example lambda x: x < 3 is a valid function for [1,2,3,4,5], but
not for [1,3,4,2,5].
>>> (left, right) = partition_o(lambda x: x < 3, [-1, 0, 2, 3, 4, 5])
>>> list(left)
[-1, 0, 2]
>>> list(right)
[3, 4, 5]
"""
left = itertools.takewhile(left_function, items)
right = itertools.dropwhile(left_function, items)
return left, right
if __name__ == "__main__":
import doctest
doctest.testmod()
|
276df9f8fbb5ad15fd768db6a13040a37037e7d6
|
service/urls.py
|
service/urls.py
|
from django.conf.urls import url, include
from rest_framework import routers
import service.authors.views
import service.friendrequest.views
import service.users.views
import service.posts.views
router = routers.DefaultRouter()
router.register(r'users', service.users.views.UserViewSet)
router.register(r'nodes', service.nodes.views.NodeViewSet)
router.register(r'author', service.authors.views.AuthorViewSet, base_name="author")
router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post")
# Wire up our API using automatic URL routing.
# Additionally, we include login URLs for the browseable API.
urlpatterns = [
url(r'^', include(router.urls)),
url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'),
]
|
from django.conf.urls import url, include
from rest_framework import routers
import service.authors.views
import service.friendrequest.views
import service.users.views
import service.nodes.views
import service.posts.views
router = routers.DefaultRouter()
router.register(r'users', service.users.views.UserViewSet)
router.register(r'nodes', service.nodes.views.NodeViewSet)
router.register(r'author', service.authors.views.AuthorViewSet, base_name="author")
router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post")
# Wire up our API using automatic URL routing.
# Additionally, we include login URLs for the browseable API.
urlpatterns = [
url(r'^', include(router.urls)),
url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'),
]
|
Add missing Node view import
|
Add missing Node view import
|
Python
|
apache-2.0
|
TeamAADGT/CMPUT404-project-socialdistribution,TeamAADGT/CMPUT404-project-socialdistribution,TeamAADGT/CMPUT404-project-socialdistribution
|
from django.conf.urls import url, include
from rest_framework import routers
import service.authors.views
import service.friendrequest.views
import service.users.views
import service.posts.views
router = routers.DefaultRouter()
router.register(r'users', service.users.views.UserViewSet)
router.register(r'nodes', service.nodes.views.NodeViewSet)
router.register(r'author', service.authors.views.AuthorViewSet, base_name="author")
router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post")
# Wire up our API using automatic URL routing.
# Additionally, we include login URLs for the browseable API.
urlpatterns = [
url(r'^', include(router.urls)),
url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'),
]
Add missing Node view import
|
from django.conf.urls import url, include
from rest_framework import routers
import service.authors.views
import service.friendrequest.views
import service.users.views
import service.nodes.views
import service.posts.views
router = routers.DefaultRouter()
router.register(r'users', service.users.views.UserViewSet)
router.register(r'nodes', service.nodes.views.NodeViewSet)
router.register(r'author', service.authors.views.AuthorViewSet, base_name="author")
router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post")
# Wire up our API using automatic URL routing.
# Additionally, we include login URLs for the browseable API.
urlpatterns = [
url(r'^', include(router.urls)),
url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'),
]
|
<commit_before>from django.conf.urls import url, include
from rest_framework import routers
import service.authors.views
import service.friendrequest.views
import service.users.views
import service.posts.views
router = routers.DefaultRouter()
router.register(r'users', service.users.views.UserViewSet)
router.register(r'nodes', service.nodes.views.NodeViewSet)
router.register(r'author', service.authors.views.AuthorViewSet, base_name="author")
router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post")
# Wire up our API using automatic URL routing.
# Additionally, we include login URLs for the browseable API.
urlpatterns = [
url(r'^', include(router.urls)),
url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'),
]
<commit_msg>Add missing Node view import<commit_after>
|
from django.conf.urls import url, include
from rest_framework import routers
import service.authors.views
import service.friendrequest.views
import service.users.views
import service.nodes.views
import service.posts.views
router = routers.DefaultRouter()
router.register(r'users', service.users.views.UserViewSet)
router.register(r'nodes', service.nodes.views.NodeViewSet)
router.register(r'author', service.authors.views.AuthorViewSet, base_name="author")
router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post")
# Wire up our API using automatic URL routing.
# Additionally, we include login URLs for the browseable API.
urlpatterns = [
url(r'^', include(router.urls)),
url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'),
]
|
from django.conf.urls import url, include
from rest_framework import routers
import service.authors.views
import service.friendrequest.views
import service.users.views
import service.posts.views
router = routers.DefaultRouter()
router.register(r'users', service.users.views.UserViewSet)
router.register(r'nodes', service.nodes.views.NodeViewSet)
router.register(r'author', service.authors.views.AuthorViewSet, base_name="author")
router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post")
# Wire up our API using automatic URL routing.
# Additionally, we include login URLs for the browseable API.
urlpatterns = [
url(r'^', include(router.urls)),
url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'),
]
Add missing Node view importfrom django.conf.urls import url, include
from rest_framework import routers
import service.authors.views
import service.friendrequest.views
import service.users.views
import service.nodes.views
import service.posts.views
router = routers.DefaultRouter()
router.register(r'users', service.users.views.UserViewSet)
router.register(r'nodes', service.nodes.views.NodeViewSet)
router.register(r'author', service.authors.views.AuthorViewSet, base_name="author")
router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post")
# Wire up our API using automatic URL routing.
# Additionally, we include login URLs for the browseable API.
urlpatterns = [
url(r'^', include(router.urls)),
url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'),
]
|
<commit_before>from django.conf.urls import url, include
from rest_framework import routers
import service.authors.views
import service.friendrequest.views
import service.users.views
import service.posts.views
router = routers.DefaultRouter()
router.register(r'users', service.users.views.UserViewSet)
router.register(r'nodes', service.nodes.views.NodeViewSet)
router.register(r'author', service.authors.views.AuthorViewSet, base_name="author")
router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post")
# Wire up our API using automatic URL routing.
# Additionally, we include login URLs for the browseable API.
urlpatterns = [
url(r'^', include(router.urls)),
url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'),
]
<commit_msg>Add missing Node view import<commit_after>from django.conf.urls import url, include
from rest_framework import routers
import service.authors.views
import service.friendrequest.views
import service.users.views
import service.nodes.views
import service.posts.views
router = routers.DefaultRouter()
router.register(r'users', service.users.views.UserViewSet)
router.register(r'nodes', service.nodes.views.NodeViewSet)
router.register(r'author', service.authors.views.AuthorViewSet, base_name="author")
router.register(r'posts', service.posts.views.PublicPostsViewSet, base_name="post")
# Wire up our API using automatic URL routing.
# Additionally, we include login URLs for the browseable API.
urlpatterns = [
url(r'^', include(router.urls)),
url(r'^auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^friendrequest/', service.friendrequest.views.friendrequest, name='friend-request'),
]
|
73d22cc63a2a37bd3c99774bf098ca12c81d54ae
|
funnels.py
|
funnels.py
|
import pyglet
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
window = pyglet.window.Window()#fullscreen=True)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
|
import pyglet
import argparse
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
def main(fullscreen):
window = pyglet.window.Window(fullscreen=fullscreen)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Arithemetic practice game.")
parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True')
parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False')
parser.set_defaults(fullscreen=True)
results = parser.parse_args()
main(results.fullscreen)
|
Add argparse to turn on/off fullscreen behavior
|
Add argparse to turn on/off fullscreen behavior
|
Python
|
mit
|
simeonf/claire
|
import pyglet
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
window = pyglet.window.Window()#fullscreen=True)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
Add argparse to turn on/off fullscreen behavior
|
import pyglet
import argparse
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
def main(fullscreen):
window = pyglet.window.Window(fullscreen=fullscreen)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Arithemetic practice game.")
parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True')
parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False')
parser.set_defaults(fullscreen=True)
results = parser.parse_args()
main(results.fullscreen)
|
<commit_before>import pyglet
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
window = pyglet.window.Window()#fullscreen=True)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
<commit_msg>Add argparse to turn on/off fullscreen behavior<commit_after>
|
import pyglet
import argparse
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
def main(fullscreen):
window = pyglet.window.Window(fullscreen=fullscreen)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Arithemetic practice game.")
parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True')
parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False')
parser.set_defaults(fullscreen=True)
results = parser.parse_args()
main(results.fullscreen)
|
import pyglet
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
window = pyglet.window.Window()#fullscreen=True)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
Add argparse to turn on/off fullscreen behaviorimport pyglet
import argparse
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
def main(fullscreen):
window = pyglet.window.Window(fullscreen=fullscreen)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Arithemetic practice game.")
parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True')
parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False')
parser.set_defaults(fullscreen=True)
results = parser.parse_args()
main(results.fullscreen)
|
<commit_before>import pyglet
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
window = pyglet.window.Window()#fullscreen=True)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
<commit_msg>Add argparse to turn on/off fullscreen behavior<commit_after>import pyglet
import argparse
from levels import GameOver, IntroScreen, TheGame
from levels.levels import Levels
def main(fullscreen):
window = pyglet.window.Window(fullscreen=fullscreen)
levels = Levels([IntroScreen(window), TheGame(window), GameOver(window)])
pyglet.clock.schedule(levels.clock)
@window.event
def on_key_press(symbol, modifiers):
levels.key(symbol, modifiers)
@window.event
def on_draw():
levels.draw()
pyglet.app.run()
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Arithemetic practice game.")
parser.add_argument('--fullscreen', action="store_true", help='Turn on fullscreen. Defaults to True')
parser.add_argument('--no-fullscreen', dest="fullscreen", action="store_false", help='Turn off fullscreen. Defaults to False')
parser.set_defaults(fullscreen=True)
results = parser.parse_args()
main(results.fullscreen)
|
86edd9a5d060d88b011d280b72e208716e001c3a
|
phy/__init__.py
|
phy/__init__.py
|
# -*- coding: utf-8 -*-
# flake8: noqa
"""
phy is an open source electrophysiological data analysis package in Python
for neuronal recordings made with high-density multielectrode arrays
containing up to thousands of channels.
"""
#------------------------------------------------------------------------------
# Imports
#------------------------------------------------------------------------------
import os.path as op
from pkg_resources import get_distribution, DistributionNotFound
from .utils.logging import _default_logger, set_level
from .utils.datasets import download_test_data
from .utils.dock import enable_qt, qt_app
#------------------------------------------------------------------------------
# Global variables and functions
#------------------------------------------------------------------------------
__author__ = 'Kwik team'
__email__ = 'cyrille.rossant at gmail.com'
__version__ = '0.1.0.dev0'
__all__ = ['debug', 'set_level']
# Set up the default logger.
_default_logger()
def debug(enable=True):
"""Enable debug logging mode."""
if enable:
set_level('debug')
else:
set_level('info')
|
# -*- coding: utf-8 -*-
# flake8: noqa
"""
phy is an open source electrophysiological data analysis package in Python
for neuronal recordings made with high-density multielectrode arrays
containing up to thousands of channels.
"""
#------------------------------------------------------------------------------
# Imports
#------------------------------------------------------------------------------
# Define a mock @profile decorator to avoid NameErrors when profiling.
import sys
if sys.version_info[0] == 3:
import builtins
else:
import __builtin__ as builtins
builtins.__dict__['profile'] = lambda func: func
import os.path as op
from pkg_resources import get_distribution, DistributionNotFound
from .utils.logging import _default_logger, set_level
from .utils.datasets import download_test_data
from .utils.dock import enable_qt, qt_app
#------------------------------------------------------------------------------
# Global variables and functions
#------------------------------------------------------------------------------
__author__ = 'Kwik team'
__email__ = 'cyrille.rossant at gmail.com'
__version__ = '0.1.0.dev0'
__all__ = ['debug', 'set_level']
# Set up the default logger.
_default_logger()
def debug(enable=True):
"""Enable debug logging mode."""
if enable:
set_level('debug')
else:
set_level('info')
|
Define mock @profile decorator in builtins.
|
Define mock @profile decorator in builtins.
|
Python
|
bsd-3-clause
|
nippoo/phy,kwikteam/phy,nsteinme/phy,rossant/phy,kwikteam/phy,rossant/phy,nsteinme/phy,rossant/phy,nippoo/phy,kwikteam/phy
|
# -*- coding: utf-8 -*-
# flake8: noqa
"""
phy is an open source electrophysiological data analysis package in Python
for neuronal recordings made with high-density multielectrode arrays
containing up to thousands of channels.
"""
#------------------------------------------------------------------------------
# Imports
#------------------------------------------------------------------------------
import os.path as op
from pkg_resources import get_distribution, DistributionNotFound
from .utils.logging import _default_logger, set_level
from .utils.datasets import download_test_data
from .utils.dock import enable_qt, qt_app
#------------------------------------------------------------------------------
# Global variables and functions
#------------------------------------------------------------------------------
__author__ = 'Kwik team'
__email__ = 'cyrille.rossant at gmail.com'
__version__ = '0.1.0.dev0'
__all__ = ['debug', 'set_level']
# Set up the default logger.
_default_logger()
def debug(enable=True):
"""Enable debug logging mode."""
if enable:
set_level('debug')
else:
set_level('info')
Define mock @profile decorator in builtins.
|
# -*- coding: utf-8 -*-
# flake8: noqa
"""
phy is an open source electrophysiological data analysis package in Python
for neuronal recordings made with high-density multielectrode arrays
containing up to thousands of channels.
"""
#------------------------------------------------------------------------------
# Imports
#------------------------------------------------------------------------------
# Define a mock @profile decorator to avoid NameErrors when profiling.
import sys
if sys.version_info[0] == 3:
import builtins
else:
import __builtin__ as builtins
builtins.__dict__['profile'] = lambda func: func
import os.path as op
from pkg_resources import get_distribution, DistributionNotFound
from .utils.logging import _default_logger, set_level
from .utils.datasets import download_test_data
from .utils.dock import enable_qt, qt_app
#------------------------------------------------------------------------------
# Global variables and functions
#------------------------------------------------------------------------------
__author__ = 'Kwik team'
__email__ = 'cyrille.rossant at gmail.com'
__version__ = '0.1.0.dev0'
__all__ = ['debug', 'set_level']
# Set up the default logger.
_default_logger()
def debug(enable=True):
"""Enable debug logging mode."""
if enable:
set_level('debug')
else:
set_level('info')
|
<commit_before># -*- coding: utf-8 -*-
# flake8: noqa
"""
phy is an open source electrophysiological data analysis package in Python
for neuronal recordings made with high-density multielectrode arrays
containing up to thousands of channels.
"""
#------------------------------------------------------------------------------
# Imports
#------------------------------------------------------------------------------
import os.path as op
from pkg_resources import get_distribution, DistributionNotFound
from .utils.logging import _default_logger, set_level
from .utils.datasets import download_test_data
from .utils.dock import enable_qt, qt_app
#------------------------------------------------------------------------------
# Global variables and functions
#------------------------------------------------------------------------------
__author__ = 'Kwik team'
__email__ = 'cyrille.rossant at gmail.com'
__version__ = '0.1.0.dev0'
__all__ = ['debug', 'set_level']
# Set up the default logger.
_default_logger()
def debug(enable=True):
"""Enable debug logging mode."""
if enable:
set_level('debug')
else:
set_level('info')
<commit_msg>Define mock @profile decorator in builtins.<commit_after>
|
# -*- coding: utf-8 -*-
# flake8: noqa
"""
phy is an open source electrophysiological data analysis package in Python
for neuronal recordings made with high-density multielectrode arrays
containing up to thousands of channels.
"""
#------------------------------------------------------------------------------
# Imports
#------------------------------------------------------------------------------
# Define a mock @profile decorator to avoid NameErrors when profiling.
import sys
if sys.version_info[0] == 3:
import builtins
else:
import __builtin__ as builtins
builtins.__dict__['profile'] = lambda func: func
import os.path as op
from pkg_resources import get_distribution, DistributionNotFound
from .utils.logging import _default_logger, set_level
from .utils.datasets import download_test_data
from .utils.dock import enable_qt, qt_app
#------------------------------------------------------------------------------
# Global variables and functions
#------------------------------------------------------------------------------
__author__ = 'Kwik team'
__email__ = 'cyrille.rossant at gmail.com'
__version__ = '0.1.0.dev0'
__all__ = ['debug', 'set_level']
# Set up the default logger.
_default_logger()
def debug(enable=True):
"""Enable debug logging mode."""
if enable:
set_level('debug')
else:
set_level('info')
|
# -*- coding: utf-8 -*-
# flake8: noqa
"""
phy is an open source electrophysiological data analysis package in Python
for neuronal recordings made with high-density multielectrode arrays
containing up to thousands of channels.
"""
#------------------------------------------------------------------------------
# Imports
#------------------------------------------------------------------------------
import os.path as op
from pkg_resources import get_distribution, DistributionNotFound
from .utils.logging import _default_logger, set_level
from .utils.datasets import download_test_data
from .utils.dock import enable_qt, qt_app
#------------------------------------------------------------------------------
# Global variables and functions
#------------------------------------------------------------------------------
__author__ = 'Kwik team'
__email__ = 'cyrille.rossant at gmail.com'
__version__ = '0.1.0.dev0'
__all__ = ['debug', 'set_level']
# Set up the default logger.
_default_logger()
def debug(enable=True):
"""Enable debug logging mode."""
if enable:
set_level('debug')
else:
set_level('info')
Define mock @profile decorator in builtins.# -*- coding: utf-8 -*-
# flake8: noqa
"""
phy is an open source electrophysiological data analysis package in Python
for neuronal recordings made with high-density multielectrode arrays
containing up to thousands of channels.
"""
#------------------------------------------------------------------------------
# Imports
#------------------------------------------------------------------------------
# Define a mock @profile decorator to avoid NameErrors when profiling.
import sys
if sys.version_info[0] == 3:
import builtins
else:
import __builtin__ as builtins
builtins.__dict__['profile'] = lambda func: func
import os.path as op
from pkg_resources import get_distribution, DistributionNotFound
from .utils.logging import _default_logger, set_level
from .utils.datasets import download_test_data
from .utils.dock import enable_qt, qt_app
#------------------------------------------------------------------------------
# Global variables and functions
#------------------------------------------------------------------------------
__author__ = 'Kwik team'
__email__ = 'cyrille.rossant at gmail.com'
__version__ = '0.1.0.dev0'
__all__ = ['debug', 'set_level']
# Set up the default logger.
_default_logger()
def debug(enable=True):
"""Enable debug logging mode."""
if enable:
set_level('debug')
else:
set_level('info')
|
<commit_before># -*- coding: utf-8 -*-
# flake8: noqa
"""
phy is an open source electrophysiological data analysis package in Python
for neuronal recordings made with high-density multielectrode arrays
containing up to thousands of channels.
"""
#------------------------------------------------------------------------------
# Imports
#------------------------------------------------------------------------------
import os.path as op
from pkg_resources import get_distribution, DistributionNotFound
from .utils.logging import _default_logger, set_level
from .utils.datasets import download_test_data
from .utils.dock import enable_qt, qt_app
#------------------------------------------------------------------------------
# Global variables and functions
#------------------------------------------------------------------------------
__author__ = 'Kwik team'
__email__ = 'cyrille.rossant at gmail.com'
__version__ = '0.1.0.dev0'
__all__ = ['debug', 'set_level']
# Set up the default logger.
_default_logger()
def debug(enable=True):
"""Enable debug logging mode."""
if enable:
set_level('debug')
else:
set_level('info')
<commit_msg>Define mock @profile decorator in builtins.<commit_after># -*- coding: utf-8 -*-
# flake8: noqa
"""
phy is an open source electrophysiological data analysis package in Python
for neuronal recordings made with high-density multielectrode arrays
containing up to thousands of channels.
"""
#------------------------------------------------------------------------------
# Imports
#------------------------------------------------------------------------------
# Define a mock @profile decorator to avoid NameErrors when profiling.
import sys
if sys.version_info[0] == 3:
import builtins
else:
import __builtin__ as builtins
builtins.__dict__['profile'] = lambda func: func
import os.path as op
from pkg_resources import get_distribution, DistributionNotFound
from .utils.logging import _default_logger, set_level
from .utils.datasets import download_test_data
from .utils.dock import enable_qt, qt_app
#------------------------------------------------------------------------------
# Global variables and functions
#------------------------------------------------------------------------------
__author__ = 'Kwik team'
__email__ = 'cyrille.rossant at gmail.com'
__version__ = '0.1.0.dev0'
__all__ = ['debug', 'set_level']
# Set up the default logger.
_default_logger()
def debug(enable=True):
"""Enable debug logging mode."""
if enable:
set_level('debug')
else:
set_level('info')
|
1b23e939a40652f8ef870e3ee7146f62fd131933
|
getlost.py
|
getlost.py
|
from os import environ
from urllib2 import urlopen
from math import log
from flask import Flask, json, jsonify
app = Flask(__name__)
from hip import get_ranking_array
from utils import jsonp
url = 'http://open.mapquestapi.com/directions/v2/route'
params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian'
rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}'
@app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>")
@jsonp
def route(from_lat, from_lng, to_lat, to_lng):
resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'],
flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng))
route = json.loads(resp.read().decode("utf-8"))
coords = [(man['startPoint']['lat'], man['startPoint']['lng'])
for leg in route['route']['legs']
for man in leg['maneuvers']]
hip_rank, total_rank = get_ranking_array(coords)
return jsonify(route=route,
hip_rank=list(hip_rank),
total_rank=log(total_rank))
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
|
from os import environ
from urllib2 import urlopen
from math import log, sqrt
from flask import Flask, json, jsonify
app = Flask(__name__)
from hip import get_ranking_array
from utils import jsonp
url = 'http://open.mapquestapi.com/directions/v2/route'
params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian'
rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}'
@app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>")
@jsonp
def route(from_lat, from_lng, to_lat, to_lng):
resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'],
flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng))
route = json.loads(resp.read().decode("utf-8"))
coords = [(man['startPoint']['lat'], man['startPoint']['lng'])
for leg in route['route']['legs']
for man in leg['maneuvers']]
hip_rank, total_rank = get_ranking_array(coords)
total_rank /= sqrt((float(to_lng) - float(from_lng))**2 +
(float(to_lat) - float(from_lat))**2)
return jsonify(route=route,
hip_rank=list(hip_rank),
total_rank=log(total_rank))
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
|
Normalize total rank by distance between start and end
|
Normalize total rank by distance between start and end
|
Python
|
apache-2.0
|
kynan/GetLost
|
from os import environ
from urllib2 import urlopen
from math import log
from flask import Flask, json, jsonify
app = Flask(__name__)
from hip import get_ranking_array
from utils import jsonp
url = 'http://open.mapquestapi.com/directions/v2/route'
params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian'
rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}'
@app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>")
@jsonp
def route(from_lat, from_lng, to_lat, to_lng):
resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'],
flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng))
route = json.loads(resp.read().decode("utf-8"))
coords = [(man['startPoint']['lat'], man['startPoint']['lng'])
for leg in route['route']['legs']
for man in leg['maneuvers']]
hip_rank, total_rank = get_ranking_array(coords)
return jsonify(route=route,
hip_rank=list(hip_rank),
total_rank=log(total_rank))
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
Normalize total rank by distance between start and end
|
from os import environ
from urllib2 import urlopen
from math import log, sqrt
from flask import Flask, json, jsonify
app = Flask(__name__)
from hip import get_ranking_array
from utils import jsonp
url = 'http://open.mapquestapi.com/directions/v2/route'
params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian'
rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}'
@app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>")
@jsonp
def route(from_lat, from_lng, to_lat, to_lng):
resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'],
flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng))
route = json.loads(resp.read().decode("utf-8"))
coords = [(man['startPoint']['lat'], man['startPoint']['lng'])
for leg in route['route']['legs']
for man in leg['maneuvers']]
hip_rank, total_rank = get_ranking_array(coords)
total_rank /= sqrt((float(to_lng) - float(from_lng))**2 +
(float(to_lat) - float(from_lat))**2)
return jsonify(route=route,
hip_rank=list(hip_rank),
total_rank=log(total_rank))
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
|
<commit_before>from os import environ
from urllib2 import urlopen
from math import log
from flask import Flask, json, jsonify
app = Flask(__name__)
from hip import get_ranking_array
from utils import jsonp
url = 'http://open.mapquestapi.com/directions/v2/route'
params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian'
rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}'
@app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>")
@jsonp
def route(from_lat, from_lng, to_lat, to_lng):
resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'],
flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng))
route = json.loads(resp.read().decode("utf-8"))
coords = [(man['startPoint']['lat'], man['startPoint']['lng'])
for leg in route['route']['legs']
for man in leg['maneuvers']]
hip_rank, total_rank = get_ranking_array(coords)
return jsonify(route=route,
hip_rank=list(hip_rank),
total_rank=log(total_rank))
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
<commit_msg>Normalize total rank by distance between start and end<commit_after>
|
from os import environ
from urllib2 import urlopen
from math import log, sqrt
from flask import Flask, json, jsonify
app = Flask(__name__)
from hip import get_ranking_array
from utils import jsonp
url = 'http://open.mapquestapi.com/directions/v2/route'
params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian'
rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}'
@app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>")
@jsonp
def route(from_lat, from_lng, to_lat, to_lng):
resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'],
flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng))
route = json.loads(resp.read().decode("utf-8"))
coords = [(man['startPoint']['lat'], man['startPoint']['lng'])
for leg in route['route']['legs']
for man in leg['maneuvers']]
hip_rank, total_rank = get_ranking_array(coords)
total_rank /= sqrt((float(to_lng) - float(from_lng))**2 +
(float(to_lat) - float(from_lat))**2)
return jsonify(route=route,
hip_rank=list(hip_rank),
total_rank=log(total_rank))
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
|
from os import environ
from urllib2 import urlopen
from math import log
from flask import Flask, json, jsonify
app = Flask(__name__)
from hip import get_ranking_array
from utils import jsonp
url = 'http://open.mapquestapi.com/directions/v2/route'
params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian'
rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}'
@app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>")
@jsonp
def route(from_lat, from_lng, to_lat, to_lng):
resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'],
flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng))
route = json.loads(resp.read().decode("utf-8"))
coords = [(man['startPoint']['lat'], man['startPoint']['lng'])
for leg in route['route']['legs']
for man in leg['maneuvers']]
hip_rank, total_rank = get_ranking_array(coords)
return jsonify(route=route,
hip_rank=list(hip_rank),
total_rank=log(total_rank))
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
Normalize total rank by distance between start and endfrom os import environ
from urllib2 import urlopen
from math import log, sqrt
from flask import Flask, json, jsonify
app = Flask(__name__)
from hip import get_ranking_array
from utils import jsonp
url = 'http://open.mapquestapi.com/directions/v2/route'
params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian'
rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}'
@app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>")
@jsonp
def route(from_lat, from_lng, to_lat, to_lng):
resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'],
flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng))
route = json.loads(resp.read().decode("utf-8"))
coords = [(man['startPoint']['lat'], man['startPoint']['lng'])
for leg in route['route']['legs']
for man in leg['maneuvers']]
hip_rank, total_rank = get_ranking_array(coords)
total_rank /= sqrt((float(to_lng) - float(from_lng))**2 +
(float(to_lat) - float(from_lat))**2)
return jsonify(route=route,
hip_rank=list(hip_rank),
total_rank=log(total_rank))
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
|
<commit_before>from os import environ
from urllib2 import urlopen
from math import log
from flask import Flask, json, jsonify
app = Flask(__name__)
from hip import get_ranking_array
from utils import jsonp
url = 'http://open.mapquestapi.com/directions/v2/route'
params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian'
rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}'
@app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>")
@jsonp
def route(from_lat, from_lng, to_lat, to_lng):
resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'],
flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng))
route = json.loads(resp.read().decode("utf-8"))
coords = [(man['startPoint']['lat'], man['startPoint']['lng'])
for leg in route['route']['legs']
for man in leg['maneuvers']]
hip_rank, total_rank = get_ranking_array(coords)
return jsonify(route=route,
hip_rank=list(hip_rank),
total_rank=log(total_rank))
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
<commit_msg>Normalize total rank by distance between start and end<commit_after>from os import environ
from urllib2 import urlopen
from math import log, sqrt
from flask import Flask, json, jsonify
app = Flask(__name__)
from hip import get_ranking_array
from utils import jsonp
url = 'http://open.mapquestapi.com/directions/v2/route'
params = '?key={apikey}&ambiguities=ignore&routeType=pedestrian'
rel = url + params + '&from={flat},{flng}&to={tlat},{tlng}'
@app.route("/route/<from_lat>,<from_lng>/<to_lat>,<to_lng>")
@jsonp
def route(from_lat, from_lng, to_lat, to_lng):
resp = urlopen(rel.format(apikey=environ['MAPQUEST_API_KEY'],
flat=from_lat, flng=from_lng, tlat=to_lat, tlng=to_lng))
route = json.loads(resp.read().decode("utf-8"))
coords = [(man['startPoint']['lat'], man['startPoint']['lng'])
for leg in route['route']['legs']
for man in leg['maneuvers']]
hip_rank, total_rank = get_ranking_array(coords)
total_rank /= sqrt((float(to_lng) - float(from_lng))**2 +
(float(to_lat) - float(from_lat))**2)
return jsonify(route=route,
hip_rank=list(hip_rank),
total_rank=log(total_rank))
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=environ.get('FLASK_DEBUG', False))
|
96bcf7f55a50895dead660add9fc949af197f550
|
networking_sfc/tests/functional/services/sfc/agent/extensions/test_ovs_agent_sfc_extension.py
|
networking_sfc/tests/functional/services/sfc/agent/extensions/test_ovs_agent_sfc_extension.py
|
# Copyright (c) 2016 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron.tests.functional.agent.l2 import base
class TestOVSAgentSfcExtension(base.OVSAgentTestFramework):
def setUp(self):
super(TestOVSAgentSfcExtension, self).setUp()
self.config.set_override('extensions', ['sfc'], 'agent')
def test_run(self):
agent = self.create_agent()
self.start_agent(agent)
agent_state = agent.state_rpc.report_state.call_args[0][1]
self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
|
# Copyright (c) 2016 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron.tests.functional.agent.l2 import base
class TestOVSAgentSfcExtension(base.OVSAgentTestFramework):
def setUp(self):
super(TestOVSAgentSfcExtension, self).setUp()
self.config.set_override('extensions', ['sfc'], 'agent')
self.agent = self.create_agent()
def test_run(self):
self.agent._report_state()
agent_state = self.agent.state_rpc.report_state.call_args[0][1]
self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
|
Fix extension loading functional test
|
Fix extension loading functional test
Call the agent _report_state() before checking the report state itself
Change-Id: Idbf552d5ca5968bc95b0a3c395499c3f2d215729
Closes-Bug: 1658089
|
Python
|
apache-2.0
|
openstack/networking-sfc,openstack/networking-sfc
|
# Copyright (c) 2016 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron.tests.functional.agent.l2 import base
class TestOVSAgentSfcExtension(base.OVSAgentTestFramework):
def setUp(self):
super(TestOVSAgentSfcExtension, self).setUp()
self.config.set_override('extensions', ['sfc'], 'agent')
def test_run(self):
agent = self.create_agent()
self.start_agent(agent)
agent_state = agent.state_rpc.report_state.call_args[0][1]
self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
Fix extension loading functional test
Call the agent _report_state() before checking the report state itself
Change-Id: Idbf552d5ca5968bc95b0a3c395499c3f2d215729
Closes-Bug: 1658089
|
# Copyright (c) 2016 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron.tests.functional.agent.l2 import base
class TestOVSAgentSfcExtension(base.OVSAgentTestFramework):
def setUp(self):
super(TestOVSAgentSfcExtension, self).setUp()
self.config.set_override('extensions', ['sfc'], 'agent')
self.agent = self.create_agent()
def test_run(self):
self.agent._report_state()
agent_state = self.agent.state_rpc.report_state.call_args[0][1]
self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
|
<commit_before># Copyright (c) 2016 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron.tests.functional.agent.l2 import base
class TestOVSAgentSfcExtension(base.OVSAgentTestFramework):
def setUp(self):
super(TestOVSAgentSfcExtension, self).setUp()
self.config.set_override('extensions', ['sfc'], 'agent')
def test_run(self):
agent = self.create_agent()
self.start_agent(agent)
agent_state = agent.state_rpc.report_state.call_args[0][1]
self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
<commit_msg>Fix extension loading functional test
Call the agent _report_state() before checking the report state itself
Change-Id: Idbf552d5ca5968bc95b0a3c395499c3f2d215729
Closes-Bug: 1658089<commit_after>
|
# Copyright (c) 2016 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron.tests.functional.agent.l2 import base
class TestOVSAgentSfcExtension(base.OVSAgentTestFramework):
def setUp(self):
super(TestOVSAgentSfcExtension, self).setUp()
self.config.set_override('extensions', ['sfc'], 'agent')
self.agent = self.create_agent()
def test_run(self):
self.agent._report_state()
agent_state = self.agent.state_rpc.report_state.call_args[0][1]
self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
|
# Copyright (c) 2016 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron.tests.functional.agent.l2 import base
class TestOVSAgentSfcExtension(base.OVSAgentTestFramework):
def setUp(self):
super(TestOVSAgentSfcExtension, self).setUp()
self.config.set_override('extensions', ['sfc'], 'agent')
def test_run(self):
agent = self.create_agent()
self.start_agent(agent)
agent_state = agent.state_rpc.report_state.call_args[0][1]
self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
Fix extension loading functional test
Call the agent _report_state() before checking the report state itself
Change-Id: Idbf552d5ca5968bc95b0a3c395499c3f2d215729
Closes-Bug: 1658089# Copyright (c) 2016 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron.tests.functional.agent.l2 import base
class TestOVSAgentSfcExtension(base.OVSAgentTestFramework):
def setUp(self):
super(TestOVSAgentSfcExtension, self).setUp()
self.config.set_override('extensions', ['sfc'], 'agent')
self.agent = self.create_agent()
def test_run(self):
self.agent._report_state()
agent_state = self.agent.state_rpc.report_state.call_args[0][1]
self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
|
<commit_before># Copyright (c) 2016 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron.tests.functional.agent.l2 import base
class TestOVSAgentSfcExtension(base.OVSAgentTestFramework):
def setUp(self):
super(TestOVSAgentSfcExtension, self).setUp()
self.config.set_override('extensions', ['sfc'], 'agent')
def test_run(self):
agent = self.create_agent()
self.start_agent(agent)
agent_state = agent.state_rpc.report_state.call_args[0][1]
self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
<commit_msg>Fix extension loading functional test
Call the agent _report_state() before checking the report state itself
Change-Id: Idbf552d5ca5968bc95b0a3c395499c3f2d215729
Closes-Bug: 1658089<commit_after># Copyright (c) 2016 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron.tests.functional.agent.l2 import base
class TestOVSAgentSfcExtension(base.OVSAgentTestFramework):
def setUp(self):
super(TestOVSAgentSfcExtension, self).setUp()
self.config.set_override('extensions', ['sfc'], 'agent')
self.agent = self.create_agent()
def test_run(self):
self.agent._report_state()
agent_state = self.agent.state_rpc.report_state.call_args[0][1]
self.assertEqual(['sfc'], agent_state['configurations']['extensions'])
|
907298a325e966f6e03c766c90f22e1b03c25c1e
|
data/propaganda2mongo.py
|
data/propaganda2mongo.py
|
import bson.json_util
from bson.objectid import ObjectId
import json
import sys
def main():
node_table = {}
while True:
line = sys.stdin.readline()
if not line:
break
record = json.loads(line)
ident = str(record["twitter_id"])
aoid = node_table.get(ident)
if aoid is None:
node_table[ident] = aoid = ObjectId()
print bson.json_util.dumps({"_id": aoid,
"type": "node",
"data": {"twitter_id": ident,
"type": "audience",
"propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"],
"geos": record["geos"],
"timestamps_of_propaganda": record["timestamps_of_propaganda"]}})
for p in record["propagandists_followed"]:
oid = node_table.get(p)
if oid is None:
node_table[ident] = oid = ObjectId()
print bson.json_util.dumps({"_id": oid,
"type": "node",
"data": {"twitter_id": ident,
"type": "propagandist"}})
print bson.json_util.dumps({"_id": ObjectId(),
"type": "link",
"source": aoid,
"target": oid,
"data": {}})
if __name__ == "__main__":
sys.exit(main())
|
import bson.json_util
from bson.objectid import ObjectId
import json
import sys
def main():
node_table = {}
while True:
line = sys.stdin.readline()
if not line:
break
record = json.loads(line)
ident = str(record["twitter_id"])
aoid = node_table.get(ident)
if aoid is None:
node_table[ident] = aoid = ObjectId()
print bson.json_util.dumps({"_id": aoid,
"type": "node",
"data": {"twitter_id": ident,
"type": "audience",
"propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"],
"geos": record["geos"],
"timestamps_of_propaganda": record["timestamps_of_propaganda"]}})
for p in record["propagandists_followed"]:
oid = node_table.get(p)
if oid is None:
node_table[ident] = oid = ObjectId()
print bson.json_util.dumps({"_id": oid,
"type": "node",
"data": {"twitter_id": p,
"type": "propagandist"}})
print bson.json_util.dumps({"_id": ObjectId(),
"type": "link",
"source": aoid,
"target": oid,
"data": {}})
if __name__ == "__main__":
sys.exit(main())
|
Fix bug in data collection
|
Fix bug in data collection
|
Python
|
apache-2.0
|
XDATA-Year-3/clique-propaganda,XDATA-Year-3/clique-propaganda,XDATA-Year-3/clique-propaganda
|
import bson.json_util
from bson.objectid import ObjectId
import json
import sys
def main():
node_table = {}
while True:
line = sys.stdin.readline()
if not line:
break
record = json.loads(line)
ident = str(record["twitter_id"])
aoid = node_table.get(ident)
if aoid is None:
node_table[ident] = aoid = ObjectId()
print bson.json_util.dumps({"_id": aoid,
"type": "node",
"data": {"twitter_id": ident,
"type": "audience",
"propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"],
"geos": record["geos"],
"timestamps_of_propaganda": record["timestamps_of_propaganda"]}})
for p in record["propagandists_followed"]:
oid = node_table.get(p)
if oid is None:
node_table[ident] = oid = ObjectId()
print bson.json_util.dumps({"_id": oid,
"type": "node",
"data": {"twitter_id": ident,
"type": "propagandist"}})
print bson.json_util.dumps({"_id": ObjectId(),
"type": "link",
"source": aoid,
"target": oid,
"data": {}})
if __name__ == "__main__":
sys.exit(main())
Fix bug in data collection
|
import bson.json_util
from bson.objectid import ObjectId
import json
import sys
def main():
node_table = {}
while True:
line = sys.stdin.readline()
if not line:
break
record = json.loads(line)
ident = str(record["twitter_id"])
aoid = node_table.get(ident)
if aoid is None:
node_table[ident] = aoid = ObjectId()
print bson.json_util.dumps({"_id": aoid,
"type": "node",
"data": {"twitter_id": ident,
"type": "audience",
"propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"],
"geos": record["geos"],
"timestamps_of_propaganda": record["timestamps_of_propaganda"]}})
for p in record["propagandists_followed"]:
oid = node_table.get(p)
if oid is None:
node_table[ident] = oid = ObjectId()
print bson.json_util.dumps({"_id": oid,
"type": "node",
"data": {"twitter_id": p,
"type": "propagandist"}})
print bson.json_util.dumps({"_id": ObjectId(),
"type": "link",
"source": aoid,
"target": oid,
"data": {}})
if __name__ == "__main__":
sys.exit(main())
|
<commit_before>import bson.json_util
from bson.objectid import ObjectId
import json
import sys
def main():
node_table = {}
while True:
line = sys.stdin.readline()
if not line:
break
record = json.loads(line)
ident = str(record["twitter_id"])
aoid = node_table.get(ident)
if aoid is None:
node_table[ident] = aoid = ObjectId()
print bson.json_util.dumps({"_id": aoid,
"type": "node",
"data": {"twitter_id": ident,
"type": "audience",
"propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"],
"geos": record["geos"],
"timestamps_of_propaganda": record["timestamps_of_propaganda"]}})
for p in record["propagandists_followed"]:
oid = node_table.get(p)
if oid is None:
node_table[ident] = oid = ObjectId()
print bson.json_util.dumps({"_id": oid,
"type": "node",
"data": {"twitter_id": ident,
"type": "propagandist"}})
print bson.json_util.dumps({"_id": ObjectId(),
"type": "link",
"source": aoid,
"target": oid,
"data": {}})
if __name__ == "__main__":
sys.exit(main())
<commit_msg>Fix bug in data collection<commit_after>
|
import bson.json_util
from bson.objectid import ObjectId
import json
import sys
def main():
node_table = {}
while True:
line = sys.stdin.readline()
if not line:
break
record = json.loads(line)
ident = str(record["twitter_id"])
aoid = node_table.get(ident)
if aoid is None:
node_table[ident] = aoid = ObjectId()
print bson.json_util.dumps({"_id": aoid,
"type": "node",
"data": {"twitter_id": ident,
"type": "audience",
"propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"],
"geos": record["geos"],
"timestamps_of_propaganda": record["timestamps_of_propaganda"]}})
for p in record["propagandists_followed"]:
oid = node_table.get(p)
if oid is None:
node_table[ident] = oid = ObjectId()
print bson.json_util.dumps({"_id": oid,
"type": "node",
"data": {"twitter_id": p,
"type": "propagandist"}})
print bson.json_util.dumps({"_id": ObjectId(),
"type": "link",
"source": aoid,
"target": oid,
"data": {}})
if __name__ == "__main__":
sys.exit(main())
|
import bson.json_util
from bson.objectid import ObjectId
import json
import sys
def main():
node_table = {}
while True:
line = sys.stdin.readline()
if not line:
break
record = json.loads(line)
ident = str(record["twitter_id"])
aoid = node_table.get(ident)
if aoid is None:
node_table[ident] = aoid = ObjectId()
print bson.json_util.dumps({"_id": aoid,
"type": "node",
"data": {"twitter_id": ident,
"type": "audience",
"propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"],
"geos": record["geos"],
"timestamps_of_propaganda": record["timestamps_of_propaganda"]}})
for p in record["propagandists_followed"]:
oid = node_table.get(p)
if oid is None:
node_table[ident] = oid = ObjectId()
print bson.json_util.dumps({"_id": oid,
"type": "node",
"data": {"twitter_id": ident,
"type": "propagandist"}})
print bson.json_util.dumps({"_id": ObjectId(),
"type": "link",
"source": aoid,
"target": oid,
"data": {}})
if __name__ == "__main__":
sys.exit(main())
Fix bug in data collectionimport bson.json_util
from bson.objectid import ObjectId
import json
import sys
def main():
node_table = {}
while True:
line = sys.stdin.readline()
if not line:
break
record = json.loads(line)
ident = str(record["twitter_id"])
aoid = node_table.get(ident)
if aoid is None:
node_table[ident] = aoid = ObjectId()
print bson.json_util.dumps({"_id": aoid,
"type": "node",
"data": {"twitter_id": ident,
"type": "audience",
"propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"],
"geos": record["geos"],
"timestamps_of_propaganda": record["timestamps_of_propaganda"]}})
for p in record["propagandists_followed"]:
oid = node_table.get(p)
if oid is None:
node_table[ident] = oid = ObjectId()
print bson.json_util.dumps({"_id": oid,
"type": "node",
"data": {"twitter_id": p,
"type": "propagandist"}})
print bson.json_util.dumps({"_id": ObjectId(),
"type": "link",
"source": aoid,
"target": oid,
"data": {}})
if __name__ == "__main__":
sys.exit(main())
|
<commit_before>import bson.json_util
from bson.objectid import ObjectId
import json
import sys
def main():
node_table = {}
while True:
line = sys.stdin.readline()
if not line:
break
record = json.loads(line)
ident = str(record["twitter_id"])
aoid = node_table.get(ident)
if aoid is None:
node_table[ident] = aoid = ObjectId()
print bson.json_util.dumps({"_id": aoid,
"type": "node",
"data": {"twitter_id": ident,
"type": "audience",
"propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"],
"geos": record["geos"],
"timestamps_of_propaganda": record["timestamps_of_propaganda"]}})
for p in record["propagandists_followed"]:
oid = node_table.get(p)
if oid is None:
node_table[ident] = oid = ObjectId()
print bson.json_util.dumps({"_id": oid,
"type": "node",
"data": {"twitter_id": ident,
"type": "propagandist"}})
print bson.json_util.dumps({"_id": ObjectId(),
"type": "link",
"source": aoid,
"target": oid,
"data": {}})
if __name__ == "__main__":
sys.exit(main())
<commit_msg>Fix bug in data collection<commit_after>import bson.json_util
from bson.objectid import ObjectId
import json
import sys
def main():
node_table = {}
while True:
line = sys.stdin.readline()
if not line:
break
record = json.loads(line)
ident = str(record["twitter_id"])
aoid = node_table.get(ident)
if aoid is None:
node_table[ident] = aoid = ObjectId()
print bson.json_util.dumps({"_id": aoid,
"type": "node",
"data": {"twitter_id": ident,
"type": "audience",
"propaganda_urls_exposed_to": record["propaganda_urls_exposed_to"],
"geos": record["geos"],
"timestamps_of_propaganda": record["timestamps_of_propaganda"]}})
for p in record["propagandists_followed"]:
oid = node_table.get(p)
if oid is None:
node_table[ident] = oid = ObjectId()
print bson.json_util.dumps({"_id": oid,
"type": "node",
"data": {"twitter_id": p,
"type": "propagandist"}})
print bson.json_util.dumps({"_id": ObjectId(),
"type": "link",
"source": aoid,
"target": oid,
"data": {}})
if __name__ == "__main__":
sys.exit(main())
|
2e5a8adb47491be58d3cdc48a4984812538f55a6
|
golang/main.py
|
golang/main.py
|
from evolution_master.runners import pkg, download
# Install for Arch
with pkg.pacman() as pkg_man:
pkg_man.install('go')
# Install for Debian & Ubuntu
with pkg.apt() as pkg_man:
pkg_man.install('golang')
# TODO: make this a runner and require a switch to enable this
pkg_man.install('golang-go-darwin-amd64',
'golang-go-freebsd-amd64',
'golang-go-netbsd-amd64',
'golang-go-windows-amd64')
# Install for OSX
with pkg.brew() as pkg_man:
pkg_man.install('go')
# Install for Windows
with download.https() as downloader, pkg.msiexec() as installer:
downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi')
downloader.checksum('sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba')
installer.install_flags('/qn' '/norestart')
installer.await(downloader.finished())
|
from genes import apt, brew, pacman, http_downloader, checksum, msiexec
import platform
opsys = platform.system()
dist = platform.linux_distribution()
if platform == 'Linux' and dist == 'Arch':
pacman.update()
pacman.sync('go')
if platform == 'Linux' and (dist == 'Debian' or dist == 'Ubuntu'):
apt.update()
apt.install('golang')
# TODO: make this a runner and require a switch to enable this
apt.install('golang-go-darwin-amd64',
'golang-go-freebsd-amd64',
'golang-go-netbsd-amd64',
'golang-go-windows-amd64')
if platform == 'Darwin':
brew.update()
brew.install('go')
if platform == 'Windows':
installer = http_downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi')
checksum.check(installer, 'sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba')
install_flags = ('/qn' '/norestart')
msiexec.run(installer, install_flags)
|
Format go to the new design
|
Format go to the new design
|
Python
|
mit
|
hatchery/Genepool2,hatchery/genepool
|
from evolution_master.runners import pkg, download
# Install for Arch
with pkg.pacman() as pkg_man:
pkg_man.install('go')
# Install for Debian & Ubuntu
with pkg.apt() as pkg_man:
pkg_man.install('golang')
# TODO: make this a runner and require a switch to enable this
pkg_man.install('golang-go-darwin-amd64',
'golang-go-freebsd-amd64',
'golang-go-netbsd-amd64',
'golang-go-windows-amd64')
# Install for OSX
with pkg.brew() as pkg_man:
pkg_man.install('go')
# Install for Windows
with download.https() as downloader, pkg.msiexec() as installer:
downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi')
downloader.checksum('sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba')
installer.install_flags('/qn' '/norestart')
installer.await(downloader.finished())
Format go to the new design
|
from genes import apt, brew, pacman, http_downloader, checksum, msiexec
import platform
opsys = platform.system()
dist = platform.linux_distribution()
if platform == 'Linux' and dist == 'Arch':
pacman.update()
pacman.sync('go')
if platform == 'Linux' and (dist == 'Debian' or dist == 'Ubuntu'):
apt.update()
apt.install('golang')
# TODO: make this a runner and require a switch to enable this
apt.install('golang-go-darwin-amd64',
'golang-go-freebsd-amd64',
'golang-go-netbsd-amd64',
'golang-go-windows-amd64')
if platform == 'Darwin':
brew.update()
brew.install('go')
if platform == 'Windows':
installer = http_downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi')
checksum.check(installer, 'sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba')
install_flags = ('/qn' '/norestart')
msiexec.run(installer, install_flags)
|
<commit_before>from evolution_master.runners import pkg, download
# Install for Arch
with pkg.pacman() as pkg_man:
pkg_man.install('go')
# Install for Debian & Ubuntu
with pkg.apt() as pkg_man:
pkg_man.install('golang')
# TODO: make this a runner and require a switch to enable this
pkg_man.install('golang-go-darwin-amd64',
'golang-go-freebsd-amd64',
'golang-go-netbsd-amd64',
'golang-go-windows-amd64')
# Install for OSX
with pkg.brew() as pkg_man:
pkg_man.install('go')
# Install for Windows
with download.https() as downloader, pkg.msiexec() as installer:
downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi')
downloader.checksum('sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba')
installer.install_flags('/qn' '/norestart')
installer.await(downloader.finished())
<commit_msg>Format go to the new design<commit_after>
|
from genes import apt, brew, pacman, http_downloader, checksum, msiexec
import platform
opsys = platform.system()
dist = platform.linux_distribution()
if platform == 'Linux' and dist == 'Arch':
pacman.update()
pacman.sync('go')
if platform == 'Linux' and (dist == 'Debian' or dist == 'Ubuntu'):
apt.update()
apt.install('golang')
# TODO: make this a runner and require a switch to enable this
apt.install('golang-go-darwin-amd64',
'golang-go-freebsd-amd64',
'golang-go-netbsd-amd64',
'golang-go-windows-amd64')
if platform == 'Darwin':
brew.update()
brew.install('go')
if platform == 'Windows':
installer = http_downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi')
checksum.check(installer, 'sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba')
install_flags = ('/qn' '/norestart')
msiexec.run(installer, install_flags)
|
from evolution_master.runners import pkg, download
# Install for Arch
with pkg.pacman() as pkg_man:
pkg_man.install('go')
# Install for Debian & Ubuntu
with pkg.apt() as pkg_man:
pkg_man.install('golang')
# TODO: make this a runner and require a switch to enable this
pkg_man.install('golang-go-darwin-amd64',
'golang-go-freebsd-amd64',
'golang-go-netbsd-amd64',
'golang-go-windows-amd64')
# Install for OSX
with pkg.brew() as pkg_man:
pkg_man.install('go')
# Install for Windows
with download.https() as downloader, pkg.msiexec() as installer:
downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi')
downloader.checksum('sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba')
installer.install_flags('/qn' '/norestart')
installer.await(downloader.finished())
Format go to the new designfrom genes import apt, brew, pacman, http_downloader, checksum, msiexec
import platform
opsys = platform.system()
dist = platform.linux_distribution()
if platform == 'Linux' and dist == 'Arch':
pacman.update()
pacman.sync('go')
if platform == 'Linux' and (dist == 'Debian' or dist == 'Ubuntu'):
apt.update()
apt.install('golang')
# TODO: make this a runner and require a switch to enable this
apt.install('golang-go-darwin-amd64',
'golang-go-freebsd-amd64',
'golang-go-netbsd-amd64',
'golang-go-windows-amd64')
if platform == 'Darwin':
brew.update()
brew.install('go')
if platform == 'Windows':
installer = http_downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi')
checksum.check(installer, 'sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba')
install_flags = ('/qn' '/norestart')
msiexec.run(installer, install_flags)
|
<commit_before>from evolution_master.runners import pkg, download
# Install for Arch
with pkg.pacman() as pkg_man:
pkg_man.install('go')
# Install for Debian & Ubuntu
with pkg.apt() as pkg_man:
pkg_man.install('golang')
# TODO: make this a runner and require a switch to enable this
pkg_man.install('golang-go-darwin-amd64',
'golang-go-freebsd-amd64',
'golang-go-netbsd-amd64',
'golang-go-windows-amd64')
# Install for OSX
with pkg.brew() as pkg_man:
pkg_man.install('go')
# Install for Windows
with download.https() as downloader, pkg.msiexec() as installer:
downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi')
downloader.checksum('sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba')
installer.install_flags('/qn' '/norestart')
installer.await(downloader.finished())
<commit_msg>Format go to the new design<commit_after>from genes import apt, brew, pacman, http_downloader, checksum, msiexec
import platform
opsys = platform.system()
dist = platform.linux_distribution()
if platform == 'Linux' and dist == 'Arch':
pacman.update()
pacman.sync('go')
if platform == 'Linux' and (dist == 'Debian' or dist == 'Ubuntu'):
apt.update()
apt.install('golang')
# TODO: make this a runner and require a switch to enable this
apt.install('golang-go-darwin-amd64',
'golang-go-freebsd-amd64',
'golang-go-netbsd-amd64',
'golang-go-windows-amd64')
if platform == 'Darwin':
brew.update()
brew.install('go')
if platform == 'Windows':
installer = http_downloader.get('https://storage.googleapis.com/golang/go1.5.1.windows-amd64.msi')
checksum.check(installer, 'sha1', '0a439f49b546b82f85adf84a79bbf40de2b3d5ba')
install_flags = ('/qn' '/norestart')
msiexec.run(installer, install_flags)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.