commit stringlengths 40 40 | old_file stringlengths 4 118 | new_file stringlengths 4 118 | old_contents stringlengths 0 2.94k | new_contents stringlengths 1 4.43k | subject stringlengths 15 444 | message stringlengths 16 3.45k | lang stringclasses 1 value | license stringclasses 13 values | repos stringlengths 5 43.2k | prompt stringlengths 17 4.58k | response stringlengths 1 4.43k | prompt_tagged stringlengths 58 4.62k | response_tagged stringlengths 1 4.43k | text stringlengths 132 7.29k | text_tagged stringlengths 173 7.33k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
e721f5fc7481e7970ba5e281e37b426123b415c3 | ruledxml/tests/test_order.py | ruledxml/tests/test_order.py | #!/usr/bin/env python3
import io
import unittest
import ruledxml
from . import utils
class TestRuledXmlForeach(unittest.TestCase):
def test_030(self):
result = io.BytesIO()
with open(utils.data('030_source.xml')) as src:
ruledxml.run(src, utils.data('030_rules.py'), result)
with open(utils.data('030_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def test_031(self):
result = io.BytesIO()
with open(utils.data('031_source.xml')) as src:
ruledxml.run(src, utils.data('031_rules.py'), result)
with open(utils.data('031_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def run():
unittest.main()
if __name__ == '__main__':
run()
| #!/usr/bin/env python3
import io
import unittest
import ruledxml
from . import utils
class TestRuledXmlOrder(unittest.TestCase):
def test_030(self):
result = io.BytesIO()
with open(utils.data('030_source.xml')) as src:
ruledxml.run(src, utils.data('030_rules.py'), result)
with open(utils.data('030_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def test_031(self):
result = io.BytesIO()
with open(utils.data('031_source.xml')) as src:
ruledxml.run(src, utils.data('031_rules.py'), result)
with open(utils.data('031_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def run():
unittest.main()
if __name__ == '__main__':
run()
| Fix name: should be TestRuledXmlOrder, not TestRuledXmlForeach. | Fix name: should be TestRuledXmlOrder, not TestRuledXmlForeach.
| Python | bsd-3-clause | meisterluk/ruledxml | #!/usr/bin/env python3
import io
import unittest
import ruledxml
from . import utils
class TestRuledXmlForeach(unittest.TestCase):
def test_030(self):
result = io.BytesIO()
with open(utils.data('030_source.xml')) as src:
ruledxml.run(src, utils.data('030_rules.py'), result)
with open(utils.data('030_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def test_031(self):
result = io.BytesIO()
with open(utils.data('031_source.xml')) as src:
ruledxml.run(src, utils.data('031_rules.py'), result)
with open(utils.data('031_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def run():
unittest.main()
if __name__ == '__main__':
run()
Fix name: should be TestRuledXmlOrder, not TestRuledXmlForeach. | #!/usr/bin/env python3
import io
import unittest
import ruledxml
from . import utils
class TestRuledXmlOrder(unittest.TestCase):
def test_030(self):
result = io.BytesIO()
with open(utils.data('030_source.xml')) as src:
ruledxml.run(src, utils.data('030_rules.py'), result)
with open(utils.data('030_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def test_031(self):
result = io.BytesIO()
with open(utils.data('031_source.xml')) as src:
ruledxml.run(src, utils.data('031_rules.py'), result)
with open(utils.data('031_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def run():
unittest.main()
if __name__ == '__main__':
run()
| <commit_before>#!/usr/bin/env python3
import io
import unittest
import ruledxml
from . import utils
class TestRuledXmlForeach(unittest.TestCase):
def test_030(self):
result = io.BytesIO()
with open(utils.data('030_source.xml')) as src:
ruledxml.run(src, utils.data('030_rules.py'), result)
with open(utils.data('030_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def test_031(self):
result = io.BytesIO()
with open(utils.data('031_source.xml')) as src:
ruledxml.run(src, utils.data('031_rules.py'), result)
with open(utils.data('031_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def run():
unittest.main()
if __name__ == '__main__':
run()
<commit_msg>Fix name: should be TestRuledXmlOrder, not TestRuledXmlForeach.<commit_after> | #!/usr/bin/env python3
import io
import unittest
import ruledxml
from . import utils
class TestRuledXmlOrder(unittest.TestCase):
def test_030(self):
result = io.BytesIO()
with open(utils.data('030_source.xml')) as src:
ruledxml.run(src, utils.data('030_rules.py'), result)
with open(utils.data('030_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def test_031(self):
result = io.BytesIO()
with open(utils.data('031_source.xml')) as src:
ruledxml.run(src, utils.data('031_rules.py'), result)
with open(utils.data('031_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def run():
unittest.main()
if __name__ == '__main__':
run()
| #!/usr/bin/env python3
import io
import unittest
import ruledxml
from . import utils
class TestRuledXmlForeach(unittest.TestCase):
def test_030(self):
result = io.BytesIO()
with open(utils.data('030_source.xml')) as src:
ruledxml.run(src, utils.data('030_rules.py'), result)
with open(utils.data('030_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def test_031(self):
result = io.BytesIO()
with open(utils.data('031_source.xml')) as src:
ruledxml.run(src, utils.data('031_rules.py'), result)
with open(utils.data('031_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def run():
unittest.main()
if __name__ == '__main__':
run()
Fix name: should be TestRuledXmlOrder, not TestRuledXmlForeach.#!/usr/bin/env python3
import io
import unittest
import ruledxml
from . import utils
class TestRuledXmlOrder(unittest.TestCase):
def test_030(self):
result = io.BytesIO()
with open(utils.data('030_source.xml')) as src:
ruledxml.run(src, utils.data('030_rules.py'), result)
with open(utils.data('030_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def test_031(self):
result = io.BytesIO()
with open(utils.data('031_source.xml')) as src:
ruledxml.run(src, utils.data('031_rules.py'), result)
with open(utils.data('031_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def run():
unittest.main()
if __name__ == '__main__':
run()
| <commit_before>#!/usr/bin/env python3
import io
import unittest
import ruledxml
from . import utils
class TestRuledXmlForeach(unittest.TestCase):
def test_030(self):
result = io.BytesIO()
with open(utils.data('030_source.xml')) as src:
ruledxml.run(src, utils.data('030_rules.py'), result)
with open(utils.data('030_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def test_031(self):
result = io.BytesIO()
with open(utils.data('031_source.xml')) as src:
ruledxml.run(src, utils.data('031_rules.py'), result)
with open(utils.data('031_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def run():
unittest.main()
if __name__ == '__main__':
run()
<commit_msg>Fix name: should be TestRuledXmlOrder, not TestRuledXmlForeach.<commit_after>#!/usr/bin/env python3
import io
import unittest
import ruledxml
from . import utils
class TestRuledXmlOrder(unittest.TestCase):
def test_030(self):
result = io.BytesIO()
with open(utils.data('030_source.xml')) as src:
ruledxml.run(src, utils.data('030_rules.py'), result)
with open(utils.data('030_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def test_031(self):
result = io.BytesIO()
with open(utils.data('031_source.xml')) as src:
ruledxml.run(src, utils.data('031_rules.py'), result)
with open(utils.data('031_target.xml'), 'rb') as target:
utils.xmlEquals(self, result.getvalue(), target.read())
def run():
unittest.main()
if __name__ == '__main__':
run()
|
5a8f107f987198740a0f0b9f1ee1f79d90662109 | txircd/modules/rfc/cmode_n.py | txircd/modules/rfc/cmode_n.py | from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import IMode, IModuleData, Mode, ModuleData
from txircd.utils import ModeType
from zope.interface import implements
class NoExtMsgMode(ModuleData, Mode):
implements(IPlugin, IModuleData, IMode)
name = "NoExtMsgMode"
core = True
affectedActions = [ "commandmodify-PRIVMSG", "commandmodify-NOTICE" ]
def hookIRCd(self, ircd):
self.ircd = ircd
def channelModes(self):
return [ ("n", ModeType.NoParam, self) ]
def actions(self):
return [ ("modeactioncheck-channel-n-commandpermission-PRIVMSG", 1, self.channelHasMode),
("modeactioncheck-channel-n-commandpermission-NOTICE", 1, self.channelHasMode) ]
def apply(self, actionType, channel, param, user, command, data):
if user not in channel.users and channel in data["targetchans"]:
del data["targetchans"][channel]
user.sendMessage(irc.ERR_CANNOTSENDTOCHAN, channel.name, ":Cannot send to channel (no external messages)")
def channelHasMode(self, channel, user, command, data):
if "n" in channel.modes:
return ""
return None
noExtMsgMode = NoExtMsgMode() | from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import IMode, IModuleData, Mode, ModuleData
from txircd.utils import ModeType
from zope.interface import implements
class NoExtMsgMode(ModuleData, Mode):
implements(IPlugin, IModuleData, IMode)
name = "NoExtMsgMode"
core = True
affectedActions = [ "commandmodify-PRIVMSG", "commandmodify-NOTICE" ]
def hookIRCd(self, ircd):
self.ircd = ircd
def channelModes(self):
return [ ("n", ModeType.NoParam, self) ]
def actions(self):
return [ ("modeactioncheck-channel-n-commandmodify-PRIVMSG", 1, self.channelHasMode),
("modeactioncheck-channel-n-commandmodify-NOTICE", 1, self.channelHasMode) ]
def apply(self, actionType, channel, param, user, command, data):
if user not in channel.users and channel in data["targetchans"]:
del data["targetchans"][channel]
user.sendMessage(irc.ERR_CANNOTSENDTOCHAN, channel.name, ":Cannot send to channel (no external messages)")
def channelHasMode(self, channel, user, command, data):
if "n" in channel.modes:
return ""
return None
noExtMsgMode = NoExtMsgMode() | Fix mode +n specification so that it actually fires ever | Fix mode +n specification so that it actually fires ever
| Python | bsd-3-clause | Heufneutje/txircd,ElementalAlchemist/txircd | from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import IMode, IModuleData, Mode, ModuleData
from txircd.utils import ModeType
from zope.interface import implements
class NoExtMsgMode(ModuleData, Mode):
implements(IPlugin, IModuleData, IMode)
name = "NoExtMsgMode"
core = True
affectedActions = [ "commandmodify-PRIVMSG", "commandmodify-NOTICE" ]
def hookIRCd(self, ircd):
self.ircd = ircd
def channelModes(self):
return [ ("n", ModeType.NoParam, self) ]
def actions(self):
return [ ("modeactioncheck-channel-n-commandpermission-PRIVMSG", 1, self.channelHasMode),
("modeactioncheck-channel-n-commandpermission-NOTICE", 1, self.channelHasMode) ]
def apply(self, actionType, channel, param, user, command, data):
if user not in channel.users and channel in data["targetchans"]:
del data["targetchans"][channel]
user.sendMessage(irc.ERR_CANNOTSENDTOCHAN, channel.name, ":Cannot send to channel (no external messages)")
def channelHasMode(self, channel, user, command, data):
if "n" in channel.modes:
return ""
return None
noExtMsgMode = NoExtMsgMode()Fix mode +n specification so that it actually fires ever | from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import IMode, IModuleData, Mode, ModuleData
from txircd.utils import ModeType
from zope.interface import implements
class NoExtMsgMode(ModuleData, Mode):
implements(IPlugin, IModuleData, IMode)
name = "NoExtMsgMode"
core = True
affectedActions = [ "commandmodify-PRIVMSG", "commandmodify-NOTICE" ]
def hookIRCd(self, ircd):
self.ircd = ircd
def channelModes(self):
return [ ("n", ModeType.NoParam, self) ]
def actions(self):
return [ ("modeactioncheck-channel-n-commandmodify-PRIVMSG", 1, self.channelHasMode),
("modeactioncheck-channel-n-commandmodify-NOTICE", 1, self.channelHasMode) ]
def apply(self, actionType, channel, param, user, command, data):
if user not in channel.users and channel in data["targetchans"]:
del data["targetchans"][channel]
user.sendMessage(irc.ERR_CANNOTSENDTOCHAN, channel.name, ":Cannot send to channel (no external messages)")
def channelHasMode(self, channel, user, command, data):
if "n" in channel.modes:
return ""
return None
noExtMsgMode = NoExtMsgMode() | <commit_before>from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import IMode, IModuleData, Mode, ModuleData
from txircd.utils import ModeType
from zope.interface import implements
class NoExtMsgMode(ModuleData, Mode):
implements(IPlugin, IModuleData, IMode)
name = "NoExtMsgMode"
core = True
affectedActions = [ "commandmodify-PRIVMSG", "commandmodify-NOTICE" ]
def hookIRCd(self, ircd):
self.ircd = ircd
def channelModes(self):
return [ ("n", ModeType.NoParam, self) ]
def actions(self):
return [ ("modeactioncheck-channel-n-commandpermission-PRIVMSG", 1, self.channelHasMode),
("modeactioncheck-channel-n-commandpermission-NOTICE", 1, self.channelHasMode) ]
def apply(self, actionType, channel, param, user, command, data):
if user not in channel.users and channel in data["targetchans"]:
del data["targetchans"][channel]
user.sendMessage(irc.ERR_CANNOTSENDTOCHAN, channel.name, ":Cannot send to channel (no external messages)")
def channelHasMode(self, channel, user, command, data):
if "n" in channel.modes:
return ""
return None
noExtMsgMode = NoExtMsgMode()<commit_msg>Fix mode +n specification so that it actually fires ever<commit_after> | from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import IMode, IModuleData, Mode, ModuleData
from txircd.utils import ModeType
from zope.interface import implements
class NoExtMsgMode(ModuleData, Mode):
implements(IPlugin, IModuleData, IMode)
name = "NoExtMsgMode"
core = True
affectedActions = [ "commandmodify-PRIVMSG", "commandmodify-NOTICE" ]
def hookIRCd(self, ircd):
self.ircd = ircd
def channelModes(self):
return [ ("n", ModeType.NoParam, self) ]
def actions(self):
return [ ("modeactioncheck-channel-n-commandmodify-PRIVMSG", 1, self.channelHasMode),
("modeactioncheck-channel-n-commandmodify-NOTICE", 1, self.channelHasMode) ]
def apply(self, actionType, channel, param, user, command, data):
if user not in channel.users and channel in data["targetchans"]:
del data["targetchans"][channel]
user.sendMessage(irc.ERR_CANNOTSENDTOCHAN, channel.name, ":Cannot send to channel (no external messages)")
def channelHasMode(self, channel, user, command, data):
if "n" in channel.modes:
return ""
return None
noExtMsgMode = NoExtMsgMode() | from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import IMode, IModuleData, Mode, ModuleData
from txircd.utils import ModeType
from zope.interface import implements
class NoExtMsgMode(ModuleData, Mode):
implements(IPlugin, IModuleData, IMode)
name = "NoExtMsgMode"
core = True
affectedActions = [ "commandmodify-PRIVMSG", "commandmodify-NOTICE" ]
def hookIRCd(self, ircd):
self.ircd = ircd
def channelModes(self):
return [ ("n", ModeType.NoParam, self) ]
def actions(self):
return [ ("modeactioncheck-channel-n-commandpermission-PRIVMSG", 1, self.channelHasMode),
("modeactioncheck-channel-n-commandpermission-NOTICE", 1, self.channelHasMode) ]
def apply(self, actionType, channel, param, user, command, data):
if user not in channel.users and channel in data["targetchans"]:
del data["targetchans"][channel]
user.sendMessage(irc.ERR_CANNOTSENDTOCHAN, channel.name, ":Cannot send to channel (no external messages)")
def channelHasMode(self, channel, user, command, data):
if "n" in channel.modes:
return ""
return None
noExtMsgMode = NoExtMsgMode()Fix mode +n specification so that it actually fires everfrom twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import IMode, IModuleData, Mode, ModuleData
from txircd.utils import ModeType
from zope.interface import implements
class NoExtMsgMode(ModuleData, Mode):
implements(IPlugin, IModuleData, IMode)
name = "NoExtMsgMode"
core = True
affectedActions = [ "commandmodify-PRIVMSG", "commandmodify-NOTICE" ]
def hookIRCd(self, ircd):
self.ircd = ircd
def channelModes(self):
return [ ("n", ModeType.NoParam, self) ]
def actions(self):
return [ ("modeactioncheck-channel-n-commandmodify-PRIVMSG", 1, self.channelHasMode),
("modeactioncheck-channel-n-commandmodify-NOTICE", 1, self.channelHasMode) ]
def apply(self, actionType, channel, param, user, command, data):
if user not in channel.users and channel in data["targetchans"]:
del data["targetchans"][channel]
user.sendMessage(irc.ERR_CANNOTSENDTOCHAN, channel.name, ":Cannot send to channel (no external messages)")
def channelHasMode(self, channel, user, command, data):
if "n" in channel.modes:
return ""
return None
noExtMsgMode = NoExtMsgMode() | <commit_before>from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import IMode, IModuleData, Mode, ModuleData
from txircd.utils import ModeType
from zope.interface import implements
class NoExtMsgMode(ModuleData, Mode):
implements(IPlugin, IModuleData, IMode)
name = "NoExtMsgMode"
core = True
affectedActions = [ "commandmodify-PRIVMSG", "commandmodify-NOTICE" ]
def hookIRCd(self, ircd):
self.ircd = ircd
def channelModes(self):
return [ ("n", ModeType.NoParam, self) ]
def actions(self):
return [ ("modeactioncheck-channel-n-commandpermission-PRIVMSG", 1, self.channelHasMode),
("modeactioncheck-channel-n-commandpermission-NOTICE", 1, self.channelHasMode) ]
def apply(self, actionType, channel, param, user, command, data):
if user not in channel.users and channel in data["targetchans"]:
del data["targetchans"][channel]
user.sendMessage(irc.ERR_CANNOTSENDTOCHAN, channel.name, ":Cannot send to channel (no external messages)")
def channelHasMode(self, channel, user, command, data):
if "n" in channel.modes:
return ""
return None
noExtMsgMode = NoExtMsgMode()<commit_msg>Fix mode +n specification so that it actually fires ever<commit_after>from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import IMode, IModuleData, Mode, ModuleData
from txircd.utils import ModeType
from zope.interface import implements
class NoExtMsgMode(ModuleData, Mode):
implements(IPlugin, IModuleData, IMode)
name = "NoExtMsgMode"
core = True
affectedActions = [ "commandmodify-PRIVMSG", "commandmodify-NOTICE" ]
def hookIRCd(self, ircd):
self.ircd = ircd
def channelModes(self):
return [ ("n", ModeType.NoParam, self) ]
def actions(self):
return [ ("modeactioncheck-channel-n-commandmodify-PRIVMSG", 1, self.channelHasMode),
("modeactioncheck-channel-n-commandmodify-NOTICE", 1, self.channelHasMode) ]
def apply(self, actionType, channel, param, user, command, data):
if user not in channel.users and channel in data["targetchans"]:
del data["targetchans"][channel]
user.sendMessage(irc.ERR_CANNOTSENDTOCHAN, channel.name, ":Cannot send to channel (no external messages)")
def channelHasMode(self, channel, user, command, data):
if "n" in channel.modes:
return ""
return None
noExtMsgMode = NoExtMsgMode() |
3ff9f60e857c9ffbd7c72c53403ae7bf3afecab8 | test/features/steps/system.py | test/features/steps/system.py | from __future__ import print_function
import sys
import subprocess
import os
@given('a system executable {exe}')
def step_impl(context, exe):
binary = None
if sys.platform.startswith('win'):
try:
binary = subprocess.check_output(["where", exe]).decode('utf8').strip()
except:
pass
else:
try:
binary = subprocess.check_output(["which", exe]).decode('utf8').strip()
except:
pass
if binary is None:
print(
"Skipping scenario", context.scenario,
"(executable %s not found)" % exe,
file = sys.stderr
)
context.scenario.skip("The executable '%s' is not present" % exe)
else:
print(
"Found executable '%s' at '%s'" % (exe, binary),
file = sys.stderr
)
@then('{exe} is a static executable')
def step_impl(ctx, exe):
if sys.platform.lower().startswith('darwin'):
context.scenario.skip("Static runtime linking is not supported on OS X")
if sys.platform.startswith('win'):
lines = subprocess.check_output(["dumpbin.exe", "/DEPENDENTS", exe]).decode('utf8').split('\r\n')
for line in lines:
if 'msvcrt' in line.lower():
assert False, 'Found MSVCRT: %s' % line
else:
out = subprocess.check_output(["file", exe]).decode('utf8')
assert 'statically linked' in out, "Not a static executable: %s" % out
| from __future__ import print_function
import sys
import subprocess
import os
@given('a system executable {exe}')
def step_impl(context, exe):
binary = None
if sys.platform.startswith('win'):
try:
binary = subprocess.check_output(["where", exe]).decode('utf8').strip()
except:
pass
else:
try:
binary = subprocess.check_output(["which", exe]).decode('utf8').strip()
except:
pass
if binary is None:
print(
"Skipping scenario", context.scenario,
"(executable %s not found)" % exe,
file = sys.stderr
)
context.scenario.skip("The executable '%s' is not present" % exe)
else:
print(
"Found executable '%s' at '%s'" % (exe, binary),
file = sys.stderr
)
@then('{exe} is a static executable')
def step_impl(ctx, exe):
if sys.platform.lower().startswith('darwin'):
ctx.scenario.skip("Static runtime linking is not supported on OS X")
return
if sys.platform.startswith('win'):
lines = subprocess.check_output(["dumpbin.exe", "/DEPENDENTS", exe]).decode('utf8').split('\r\n')
for line in lines:
if 'msvcrt' in line.lower():
assert False, 'Found MSVCRT: %s' % line
else:
out = subprocess.check_output(["file", exe]).decode('utf8')
assert 'statically linked' in out, "Not a static executable: %s" % out
| Fix OS X test skip. | tests.features: Fix OS X test skip.
| Python | bsd-3-clause | hotgloupi/configure,hotgloupi/configure,hotgloupi/configure,hotgloupi/configure,hotgloupi/configure | from __future__ import print_function
import sys
import subprocess
import os
@given('a system executable {exe}')
def step_impl(context, exe):
binary = None
if sys.platform.startswith('win'):
try:
binary = subprocess.check_output(["where", exe]).decode('utf8').strip()
except:
pass
else:
try:
binary = subprocess.check_output(["which", exe]).decode('utf8').strip()
except:
pass
if binary is None:
print(
"Skipping scenario", context.scenario,
"(executable %s not found)" % exe,
file = sys.stderr
)
context.scenario.skip("The executable '%s' is not present" % exe)
else:
print(
"Found executable '%s' at '%s'" % (exe, binary),
file = sys.stderr
)
@then('{exe} is a static executable')
def step_impl(ctx, exe):
if sys.platform.lower().startswith('darwin'):
context.scenario.skip("Static runtime linking is not supported on OS X")
if sys.platform.startswith('win'):
lines = subprocess.check_output(["dumpbin.exe", "/DEPENDENTS", exe]).decode('utf8').split('\r\n')
for line in lines:
if 'msvcrt' in line.lower():
assert False, 'Found MSVCRT: %s' % line
else:
out = subprocess.check_output(["file", exe]).decode('utf8')
assert 'statically linked' in out, "Not a static executable: %s" % out
tests.features: Fix OS X test skip. | from __future__ import print_function
import sys
import subprocess
import os
@given('a system executable {exe}')
def step_impl(context, exe):
binary = None
if sys.platform.startswith('win'):
try:
binary = subprocess.check_output(["where", exe]).decode('utf8').strip()
except:
pass
else:
try:
binary = subprocess.check_output(["which", exe]).decode('utf8').strip()
except:
pass
if binary is None:
print(
"Skipping scenario", context.scenario,
"(executable %s not found)" % exe,
file = sys.stderr
)
context.scenario.skip("The executable '%s' is not present" % exe)
else:
print(
"Found executable '%s' at '%s'" % (exe, binary),
file = sys.stderr
)
@then('{exe} is a static executable')
def step_impl(ctx, exe):
if sys.platform.lower().startswith('darwin'):
ctx.scenario.skip("Static runtime linking is not supported on OS X")
return
if sys.platform.startswith('win'):
lines = subprocess.check_output(["dumpbin.exe", "/DEPENDENTS", exe]).decode('utf8').split('\r\n')
for line in lines:
if 'msvcrt' in line.lower():
assert False, 'Found MSVCRT: %s' % line
else:
out = subprocess.check_output(["file", exe]).decode('utf8')
assert 'statically linked' in out, "Not a static executable: %s" % out
| <commit_before>from __future__ import print_function
import sys
import subprocess
import os
@given('a system executable {exe}')
def step_impl(context, exe):
binary = None
if sys.platform.startswith('win'):
try:
binary = subprocess.check_output(["where", exe]).decode('utf8').strip()
except:
pass
else:
try:
binary = subprocess.check_output(["which", exe]).decode('utf8').strip()
except:
pass
if binary is None:
print(
"Skipping scenario", context.scenario,
"(executable %s not found)" % exe,
file = sys.stderr
)
context.scenario.skip("The executable '%s' is not present" % exe)
else:
print(
"Found executable '%s' at '%s'" % (exe, binary),
file = sys.stderr
)
@then('{exe} is a static executable')
def step_impl(ctx, exe):
if sys.platform.lower().startswith('darwin'):
context.scenario.skip("Static runtime linking is not supported on OS X")
if sys.platform.startswith('win'):
lines = subprocess.check_output(["dumpbin.exe", "/DEPENDENTS", exe]).decode('utf8').split('\r\n')
for line in lines:
if 'msvcrt' in line.lower():
assert False, 'Found MSVCRT: %s' % line
else:
out = subprocess.check_output(["file", exe]).decode('utf8')
assert 'statically linked' in out, "Not a static executable: %s" % out
<commit_msg>tests.features: Fix OS X test skip.<commit_after> | from __future__ import print_function
import sys
import subprocess
import os
@given('a system executable {exe}')
def step_impl(context, exe):
binary = None
if sys.platform.startswith('win'):
try:
binary = subprocess.check_output(["where", exe]).decode('utf8').strip()
except:
pass
else:
try:
binary = subprocess.check_output(["which", exe]).decode('utf8').strip()
except:
pass
if binary is None:
print(
"Skipping scenario", context.scenario,
"(executable %s not found)" % exe,
file = sys.stderr
)
context.scenario.skip("The executable '%s' is not present" % exe)
else:
print(
"Found executable '%s' at '%s'" % (exe, binary),
file = sys.stderr
)
@then('{exe} is a static executable')
def step_impl(ctx, exe):
if sys.platform.lower().startswith('darwin'):
ctx.scenario.skip("Static runtime linking is not supported on OS X")
return
if sys.platform.startswith('win'):
lines = subprocess.check_output(["dumpbin.exe", "/DEPENDENTS", exe]).decode('utf8').split('\r\n')
for line in lines:
if 'msvcrt' in line.lower():
assert False, 'Found MSVCRT: %s' % line
else:
out = subprocess.check_output(["file", exe]).decode('utf8')
assert 'statically linked' in out, "Not a static executable: %s" % out
| from __future__ import print_function
import sys
import subprocess
import os
@given('a system executable {exe}')
def step_impl(context, exe):
binary = None
if sys.platform.startswith('win'):
try:
binary = subprocess.check_output(["where", exe]).decode('utf8').strip()
except:
pass
else:
try:
binary = subprocess.check_output(["which", exe]).decode('utf8').strip()
except:
pass
if binary is None:
print(
"Skipping scenario", context.scenario,
"(executable %s not found)" % exe,
file = sys.stderr
)
context.scenario.skip("The executable '%s' is not present" % exe)
else:
print(
"Found executable '%s' at '%s'" % (exe, binary),
file = sys.stderr
)
@then('{exe} is a static executable')
def step_impl(ctx, exe):
if sys.platform.lower().startswith('darwin'):
context.scenario.skip("Static runtime linking is not supported on OS X")
if sys.platform.startswith('win'):
lines = subprocess.check_output(["dumpbin.exe", "/DEPENDENTS", exe]).decode('utf8').split('\r\n')
for line in lines:
if 'msvcrt' in line.lower():
assert False, 'Found MSVCRT: %s' % line
else:
out = subprocess.check_output(["file", exe]).decode('utf8')
assert 'statically linked' in out, "Not a static executable: %s" % out
tests.features: Fix OS X test skip.from __future__ import print_function
import sys
import subprocess
import os
@given('a system executable {exe}')
def step_impl(context, exe):
binary = None
if sys.platform.startswith('win'):
try:
binary = subprocess.check_output(["where", exe]).decode('utf8').strip()
except:
pass
else:
try:
binary = subprocess.check_output(["which", exe]).decode('utf8').strip()
except:
pass
if binary is None:
print(
"Skipping scenario", context.scenario,
"(executable %s not found)" % exe,
file = sys.stderr
)
context.scenario.skip("The executable '%s' is not present" % exe)
else:
print(
"Found executable '%s' at '%s'" % (exe, binary),
file = sys.stderr
)
@then('{exe} is a static executable')
def step_impl(ctx, exe):
if sys.platform.lower().startswith('darwin'):
ctx.scenario.skip("Static runtime linking is not supported on OS X")
return
if sys.platform.startswith('win'):
lines = subprocess.check_output(["dumpbin.exe", "/DEPENDENTS", exe]).decode('utf8').split('\r\n')
for line in lines:
if 'msvcrt' in line.lower():
assert False, 'Found MSVCRT: %s' % line
else:
out = subprocess.check_output(["file", exe]).decode('utf8')
assert 'statically linked' in out, "Not a static executable: %s" % out
| <commit_before>from __future__ import print_function
import sys
import subprocess
import os
@given('a system executable {exe}')
def step_impl(context, exe):
binary = None
if sys.platform.startswith('win'):
try:
binary = subprocess.check_output(["where", exe]).decode('utf8').strip()
except:
pass
else:
try:
binary = subprocess.check_output(["which", exe]).decode('utf8').strip()
except:
pass
if binary is None:
print(
"Skipping scenario", context.scenario,
"(executable %s not found)" % exe,
file = sys.stderr
)
context.scenario.skip("The executable '%s' is not present" % exe)
else:
print(
"Found executable '%s' at '%s'" % (exe, binary),
file = sys.stderr
)
@then('{exe} is a static executable')
def step_impl(ctx, exe):
if sys.platform.lower().startswith('darwin'):
context.scenario.skip("Static runtime linking is not supported on OS X")
if sys.platform.startswith('win'):
lines = subprocess.check_output(["dumpbin.exe", "/DEPENDENTS", exe]).decode('utf8').split('\r\n')
for line in lines:
if 'msvcrt' in line.lower():
assert False, 'Found MSVCRT: %s' % line
else:
out = subprocess.check_output(["file", exe]).decode('utf8')
assert 'statically linked' in out, "Not a static executable: %s" % out
<commit_msg>tests.features: Fix OS X test skip.<commit_after>from __future__ import print_function
import sys
import subprocess
import os
@given('a system executable {exe}')
def step_impl(context, exe):
binary = None
if sys.platform.startswith('win'):
try:
binary = subprocess.check_output(["where", exe]).decode('utf8').strip()
except:
pass
else:
try:
binary = subprocess.check_output(["which", exe]).decode('utf8').strip()
except:
pass
if binary is None:
print(
"Skipping scenario", context.scenario,
"(executable %s not found)" % exe,
file = sys.stderr
)
context.scenario.skip("The executable '%s' is not present" % exe)
else:
print(
"Found executable '%s' at '%s'" % (exe, binary),
file = sys.stderr
)
@then('{exe} is a static executable')
def step_impl(ctx, exe):
if sys.platform.lower().startswith('darwin'):
ctx.scenario.skip("Static runtime linking is not supported on OS X")
return
if sys.platform.startswith('win'):
lines = subprocess.check_output(["dumpbin.exe", "/DEPENDENTS", exe]).decode('utf8').split('\r\n')
for line in lines:
if 'msvcrt' in line.lower():
assert False, 'Found MSVCRT: %s' % line
else:
out = subprocess.check_output(["file", exe]).decode('utf8')
assert 'statically linked' in out, "Not a static executable: %s" % out
|
1dfe45d9ce6c81e5ae2396f97cc979192251c906 | selectable/apps.py | selectable/apps.py | try:
from django.apps import AppConfig
except ImportError:
AppConfig = object
class SelectableConfig(AppConfig):
"""App configuration for django-selectable."""
name = 'selectable'
def ready(self):
self.module.registry.autodiscover()
| try:
from django.apps import AppConfig
except ImportError:
AppConfig = object
class SelectableConfig(AppConfig):
"""App configuration for django-selectable."""
name = 'selectable'
def ready(self):
from . import registry
registry.autodiscover()
| Update auto-registration to work while running the tests. | Update auto-registration to work while running the tests.
| Python | bsd-2-clause | affan2/django-selectable,affan2/django-selectable,mlavin/django-selectable,affan2/django-selectable,mlavin/django-selectable,mlavin/django-selectable | try:
from django.apps import AppConfig
except ImportError:
AppConfig = object
class SelectableConfig(AppConfig):
"""App configuration for django-selectable."""
name = 'selectable'
def ready(self):
self.module.registry.autodiscover()
Update auto-registration to work while running the tests. | try:
from django.apps import AppConfig
except ImportError:
AppConfig = object
class SelectableConfig(AppConfig):
"""App configuration for django-selectable."""
name = 'selectable'
def ready(self):
from . import registry
registry.autodiscover()
| <commit_before>try:
from django.apps import AppConfig
except ImportError:
AppConfig = object
class SelectableConfig(AppConfig):
"""App configuration for django-selectable."""
name = 'selectable'
def ready(self):
self.module.registry.autodiscover()
<commit_msg>Update auto-registration to work while running the tests.<commit_after> | try:
from django.apps import AppConfig
except ImportError:
AppConfig = object
class SelectableConfig(AppConfig):
"""App configuration for django-selectable."""
name = 'selectable'
def ready(self):
from . import registry
registry.autodiscover()
| try:
from django.apps import AppConfig
except ImportError:
AppConfig = object
class SelectableConfig(AppConfig):
"""App configuration for django-selectable."""
name = 'selectable'
def ready(self):
self.module.registry.autodiscover()
Update auto-registration to work while running the tests.try:
from django.apps import AppConfig
except ImportError:
AppConfig = object
class SelectableConfig(AppConfig):
"""App configuration for django-selectable."""
name = 'selectable'
def ready(self):
from . import registry
registry.autodiscover()
| <commit_before>try:
from django.apps import AppConfig
except ImportError:
AppConfig = object
class SelectableConfig(AppConfig):
"""App configuration for django-selectable."""
name = 'selectable'
def ready(self):
self.module.registry.autodiscover()
<commit_msg>Update auto-registration to work while running the tests.<commit_after>try:
from django.apps import AppConfig
except ImportError:
AppConfig = object
class SelectableConfig(AppConfig):
"""App configuration for django-selectable."""
name = 'selectable'
def ready(self):
from . import registry
registry.autodiscover()
|
96ace17d9cd800a5649ad32a8cb496a55d73ca9f | wapps/templatetags/wagtail.py | wapps/templatetags/wagtail.py | import jinja2
from django.conf import settings
from django_jinja import library
from jinja2.ext import Extension
from wagtail.wagtailcore.models import Page
from wagtail.contrib.wagtailroutablepage.templatetags.wagtailroutablepage_tags import (
routablepageurl as dj_routablepageurl
)
from wapps.utils import get_image_url
@library.global_function
def menu():
return Page.objects.live().in_menu().filter(depth__lte=3)
@library.global_function
@jinja2.contextfunction
def is_site_root(context, page):
if 'request' not in context or not context['request'].site:
return False
site = context['request'].site
return site.root_page.pk == page.pk
@library.global_function
def image_url(image, specs):
return get_image_url(image, specs)
@library.global_function
@jinja2.contextfunction
def routablepageurl(context, page, name, *args, **kwargs):
return dj_routablepageurl(context, page, name, *args, **kwargs)
@library.extension
class WagtailSettings(Extension):
def __init__(self, environment):
super(WagtailSettings, self).__init__(environment)
environment.globals['WAGTAIL_SITE_NAME'] = getattr(settings, 'WAGTAIL_SITE_NAME', None)
| import jinja2
from django.conf import settings
from django_jinja import library
from jinja2.ext import Extension
from wagtail.wagtailcore.models import Page
from wagtail.contrib.wagtailroutablepage.templatetags.wagtailroutablepage_tags import (
routablepageurl as dj_routablepageurl
)
from wapps.utils import get_image_url
@library.global_function
def menu():
return Page.objects.live().in_menu().filter(depth__lte=3)
@library.global_function
@jinja2.contextfunction
def is_site_root(context, page):
if 'request' not in context or not context['request'].site or not page:
return False
site = context['request'].site
return site.root_page.pk == page.pk
@library.global_function
def image_url(image, specs):
return get_image_url(image, specs)
@library.global_function
@jinja2.contextfunction
def routablepageurl(context, page, name, *args, **kwargs):
return dj_routablepageurl(context, page, name, *args, **kwargs)
@library.extension
class WagtailSettings(Extension):
def __init__(self, environment):
super(WagtailSettings, self).__init__(environment)
environment.globals['WAGTAIL_SITE_NAME'] = getattr(settings, 'WAGTAIL_SITE_NAME', None)
| Fix is_site_root when no page | Fix is_site_root when no page
| Python | mit | apihackers/wapps,apihackers/wapps,apihackers/wapps,apihackers/wapps | import jinja2
from django.conf import settings
from django_jinja import library
from jinja2.ext import Extension
from wagtail.wagtailcore.models import Page
from wagtail.contrib.wagtailroutablepage.templatetags.wagtailroutablepage_tags import (
routablepageurl as dj_routablepageurl
)
from wapps.utils import get_image_url
@library.global_function
def menu():
return Page.objects.live().in_menu().filter(depth__lte=3)
@library.global_function
@jinja2.contextfunction
def is_site_root(context, page):
if 'request' not in context or not context['request'].site:
return False
site = context['request'].site
return site.root_page.pk == page.pk
@library.global_function
def image_url(image, specs):
return get_image_url(image, specs)
@library.global_function
@jinja2.contextfunction
def routablepageurl(context, page, name, *args, **kwargs):
return dj_routablepageurl(context, page, name, *args, **kwargs)
@library.extension
class WagtailSettings(Extension):
def __init__(self, environment):
super(WagtailSettings, self).__init__(environment)
environment.globals['WAGTAIL_SITE_NAME'] = getattr(settings, 'WAGTAIL_SITE_NAME', None)
Fix is_site_root when no page | import jinja2
from django.conf import settings
from django_jinja import library
from jinja2.ext import Extension
from wagtail.wagtailcore.models import Page
from wagtail.contrib.wagtailroutablepage.templatetags.wagtailroutablepage_tags import (
routablepageurl as dj_routablepageurl
)
from wapps.utils import get_image_url
@library.global_function
def menu():
return Page.objects.live().in_menu().filter(depth__lte=3)
@library.global_function
@jinja2.contextfunction
def is_site_root(context, page):
if 'request' not in context or not context['request'].site or not page:
return False
site = context['request'].site
return site.root_page.pk == page.pk
@library.global_function
def image_url(image, specs):
return get_image_url(image, specs)
@library.global_function
@jinja2.contextfunction
def routablepageurl(context, page, name, *args, **kwargs):
return dj_routablepageurl(context, page, name, *args, **kwargs)
@library.extension
class WagtailSettings(Extension):
def __init__(self, environment):
super(WagtailSettings, self).__init__(environment)
environment.globals['WAGTAIL_SITE_NAME'] = getattr(settings, 'WAGTAIL_SITE_NAME', None)
| <commit_before>import jinja2
from django.conf import settings
from django_jinja import library
from jinja2.ext import Extension
from wagtail.wagtailcore.models import Page
from wagtail.contrib.wagtailroutablepage.templatetags.wagtailroutablepage_tags import (
routablepageurl as dj_routablepageurl
)
from wapps.utils import get_image_url
@library.global_function
def menu():
return Page.objects.live().in_menu().filter(depth__lte=3)
@library.global_function
@jinja2.contextfunction
def is_site_root(context, page):
if 'request' not in context or not context['request'].site:
return False
site = context['request'].site
return site.root_page.pk == page.pk
@library.global_function
def image_url(image, specs):
return get_image_url(image, specs)
@library.global_function
@jinja2.contextfunction
def routablepageurl(context, page, name, *args, **kwargs):
return dj_routablepageurl(context, page, name, *args, **kwargs)
@library.extension
class WagtailSettings(Extension):
def __init__(self, environment):
super(WagtailSettings, self).__init__(environment)
environment.globals['WAGTAIL_SITE_NAME'] = getattr(settings, 'WAGTAIL_SITE_NAME', None)
<commit_msg>Fix is_site_root when no page<commit_after> | import jinja2
from django.conf import settings
from django_jinja import library
from jinja2.ext import Extension
from wagtail.wagtailcore.models import Page
from wagtail.contrib.wagtailroutablepage.templatetags.wagtailroutablepage_tags import (
routablepageurl as dj_routablepageurl
)
from wapps.utils import get_image_url
@library.global_function
def menu():
return Page.objects.live().in_menu().filter(depth__lte=3)
@library.global_function
@jinja2.contextfunction
def is_site_root(context, page):
if 'request' not in context or not context['request'].site or not page:
return False
site = context['request'].site
return site.root_page.pk == page.pk
@library.global_function
def image_url(image, specs):
return get_image_url(image, specs)
@library.global_function
@jinja2.contextfunction
def routablepageurl(context, page, name, *args, **kwargs):
return dj_routablepageurl(context, page, name, *args, **kwargs)
@library.extension
class WagtailSettings(Extension):
def __init__(self, environment):
super(WagtailSettings, self).__init__(environment)
environment.globals['WAGTAIL_SITE_NAME'] = getattr(settings, 'WAGTAIL_SITE_NAME', None)
| import jinja2
from django.conf import settings
from django_jinja import library
from jinja2.ext import Extension
from wagtail.wagtailcore.models import Page
from wagtail.contrib.wagtailroutablepage.templatetags.wagtailroutablepage_tags import (
routablepageurl as dj_routablepageurl
)
from wapps.utils import get_image_url
@library.global_function
def menu():
return Page.objects.live().in_menu().filter(depth__lte=3)
@library.global_function
@jinja2.contextfunction
def is_site_root(context, page):
if 'request' not in context or not context['request'].site:
return False
site = context['request'].site
return site.root_page.pk == page.pk
@library.global_function
def image_url(image, specs):
return get_image_url(image, specs)
@library.global_function
@jinja2.contextfunction
def routablepageurl(context, page, name, *args, **kwargs):
return dj_routablepageurl(context, page, name, *args, **kwargs)
@library.extension
class WagtailSettings(Extension):
def __init__(self, environment):
super(WagtailSettings, self).__init__(environment)
environment.globals['WAGTAIL_SITE_NAME'] = getattr(settings, 'WAGTAIL_SITE_NAME', None)
Fix is_site_root when no pageimport jinja2
from django.conf import settings
from django_jinja import library
from jinja2.ext import Extension
from wagtail.wagtailcore.models import Page
from wagtail.contrib.wagtailroutablepage.templatetags.wagtailroutablepage_tags import (
routablepageurl as dj_routablepageurl
)
from wapps.utils import get_image_url
@library.global_function
def menu():
return Page.objects.live().in_menu().filter(depth__lte=3)
@library.global_function
@jinja2.contextfunction
def is_site_root(context, page):
if 'request' not in context or not context['request'].site or not page:
return False
site = context['request'].site
return site.root_page.pk == page.pk
@library.global_function
def image_url(image, specs):
return get_image_url(image, specs)
@library.global_function
@jinja2.contextfunction
def routablepageurl(context, page, name, *args, **kwargs):
return dj_routablepageurl(context, page, name, *args, **kwargs)
@library.extension
class WagtailSettings(Extension):
def __init__(self, environment):
super(WagtailSettings, self).__init__(environment)
environment.globals['WAGTAIL_SITE_NAME'] = getattr(settings, 'WAGTAIL_SITE_NAME', None)
| <commit_before>import jinja2
from django.conf import settings
from django_jinja import library
from jinja2.ext import Extension
from wagtail.wagtailcore.models import Page
from wagtail.contrib.wagtailroutablepage.templatetags.wagtailroutablepage_tags import (
routablepageurl as dj_routablepageurl
)
from wapps.utils import get_image_url
@library.global_function
def menu():
return Page.objects.live().in_menu().filter(depth__lte=3)
@library.global_function
@jinja2.contextfunction
def is_site_root(context, page):
if 'request' not in context or not context['request'].site:
return False
site = context['request'].site
return site.root_page.pk == page.pk
@library.global_function
def image_url(image, specs):
return get_image_url(image, specs)
@library.global_function
@jinja2.contextfunction
def routablepageurl(context, page, name, *args, **kwargs):
return dj_routablepageurl(context, page, name, *args, **kwargs)
@library.extension
class WagtailSettings(Extension):
def __init__(self, environment):
super(WagtailSettings, self).__init__(environment)
environment.globals['WAGTAIL_SITE_NAME'] = getattr(settings, 'WAGTAIL_SITE_NAME', None)
<commit_msg>Fix is_site_root when no page<commit_after>import jinja2
from django.conf import settings
from django_jinja import library
from jinja2.ext import Extension
from wagtail.wagtailcore.models import Page
from wagtail.contrib.wagtailroutablepage.templatetags.wagtailroutablepage_tags import (
routablepageurl as dj_routablepageurl
)
from wapps.utils import get_image_url
@library.global_function
def menu():
return Page.objects.live().in_menu().filter(depth__lte=3)
@library.global_function
@jinja2.contextfunction
def is_site_root(context, page):
if 'request' not in context or not context['request'].site or not page:
return False
site = context['request'].site
return site.root_page.pk == page.pk
@library.global_function
def image_url(image, specs):
return get_image_url(image, specs)
@library.global_function
@jinja2.contextfunction
def routablepageurl(context, page, name, *args, **kwargs):
return dj_routablepageurl(context, page, name, *args, **kwargs)
@library.extension
class WagtailSettings(Extension):
def __init__(self, environment):
super(WagtailSettings, self).__init__(environment)
environment.globals['WAGTAIL_SITE_NAME'] = getattr(settings, 'WAGTAIL_SITE_NAME', None)
|
020ffbe8436da2f7ee654fa6a12d50f9915db17f | examples/collection/views.py | examples/collection/views.py | from cruditor.contrib.collection import CollectionViewMixin
from cruditor.views import CruditorAddView, CruditorChangeView, CruditorDeleteView, CruditorListView
from django.urls import reverse, reverse_lazy
from examples.mixins import ExamplesMixin
from store.models import Person
from .filters import PersonFilter
from .forms import PersonForm
from .tables import PersonTable
class PersonViewMixin(ExamplesMixin, CollectionViewMixin):
model = Person
collection_list_title = 'Persons'
collection_list_urlname = 'collection:list'
collection_detail_urlname = 'collection:change'
class PersonListView(PersonViewMixin, CruditorListView):
title = 'Persons'
table_class = PersonTable
def get_titlebuttons(self):
return [{'url': reverse('collection:add'), 'label': 'Add person'}]
class PersonFilterView(PersonListView):
filter_class = PersonFilter
class PersonAddView(PersonViewMixin, CruditorAddView):
success_url = reverse_lazy('collection:lits')
form_class = PersonForm
class PersonChangeView(PersonViewMixin, CruditorChangeView):
form_class = PersonForm
def get_delete_url(self):
return reverse('collection:delete', args=(self.object.pk,))
class PersonDeleteView(PersonViewMixin, CruditorDeleteView):
pass
| from cruditor.contrib.collection import CollectionViewMixin
from cruditor.views import CruditorAddView, CruditorChangeView, CruditorDeleteView, CruditorListView
from django.urls import reverse, reverse_lazy
from examples.mixins import ExamplesMixin
from store.models import Person
from .filters import PersonFilter
from .forms import PersonForm
from .tables import PersonTable
class PersonViewMixin(ExamplesMixin, CollectionViewMixin):
model = Person
collection_list_title = 'Persons'
collection_list_urlname = 'collection:list'
collection_detail_urlname = 'collection:change'
class PersonListView(PersonViewMixin, CruditorListView):
title = 'Persons'
def get_titlebuttons(self):
return [{'url': reverse('collection:add'), 'label': 'Add person'}]
class PersonFilterView(PersonListView):
filter_class = PersonFilter
table_class = PersonTable
class PersonAddView(PersonViewMixin, CruditorAddView):
success_url = reverse_lazy('collection:lits')
form_class = PersonForm
class PersonChangeView(PersonViewMixin, CruditorChangeView):
form_class = PersonForm
def get_delete_url(self):
return reverse('collection:delete', args=(self.object.pk,))
class PersonDeleteView(PersonViewMixin, CruditorDeleteView):
pass
| Make use of auto generated table classes. | Make use of auto generated table classes.
| Python | mit | moccu/django-cruditor,moccu/django-cruditor,moccu/django-cruditor | from cruditor.contrib.collection import CollectionViewMixin
from cruditor.views import CruditorAddView, CruditorChangeView, CruditorDeleteView, CruditorListView
from django.urls import reverse, reverse_lazy
from examples.mixins import ExamplesMixin
from store.models import Person
from .filters import PersonFilter
from .forms import PersonForm
from .tables import PersonTable
class PersonViewMixin(ExamplesMixin, CollectionViewMixin):
model = Person
collection_list_title = 'Persons'
collection_list_urlname = 'collection:list'
collection_detail_urlname = 'collection:change'
class PersonListView(PersonViewMixin, CruditorListView):
title = 'Persons'
table_class = PersonTable
def get_titlebuttons(self):
return [{'url': reverse('collection:add'), 'label': 'Add person'}]
class PersonFilterView(PersonListView):
filter_class = PersonFilter
class PersonAddView(PersonViewMixin, CruditorAddView):
success_url = reverse_lazy('collection:lits')
form_class = PersonForm
class PersonChangeView(PersonViewMixin, CruditorChangeView):
form_class = PersonForm
def get_delete_url(self):
return reverse('collection:delete', args=(self.object.pk,))
class PersonDeleteView(PersonViewMixin, CruditorDeleteView):
pass
Make use of auto generated table classes. | from cruditor.contrib.collection import CollectionViewMixin
from cruditor.views import CruditorAddView, CruditorChangeView, CruditorDeleteView, CruditorListView
from django.urls import reverse, reverse_lazy
from examples.mixins import ExamplesMixin
from store.models import Person
from .filters import PersonFilter
from .forms import PersonForm
from .tables import PersonTable
class PersonViewMixin(ExamplesMixin, CollectionViewMixin):
model = Person
collection_list_title = 'Persons'
collection_list_urlname = 'collection:list'
collection_detail_urlname = 'collection:change'
class PersonListView(PersonViewMixin, CruditorListView):
title = 'Persons'
def get_titlebuttons(self):
return [{'url': reverse('collection:add'), 'label': 'Add person'}]
class PersonFilterView(PersonListView):
filter_class = PersonFilter
table_class = PersonTable
class PersonAddView(PersonViewMixin, CruditorAddView):
success_url = reverse_lazy('collection:lits')
form_class = PersonForm
class PersonChangeView(PersonViewMixin, CruditorChangeView):
form_class = PersonForm
def get_delete_url(self):
return reverse('collection:delete', args=(self.object.pk,))
class PersonDeleteView(PersonViewMixin, CruditorDeleteView):
pass
| <commit_before>from cruditor.contrib.collection import CollectionViewMixin
from cruditor.views import CruditorAddView, CruditorChangeView, CruditorDeleteView, CruditorListView
from django.urls import reverse, reverse_lazy
from examples.mixins import ExamplesMixin
from store.models import Person
from .filters import PersonFilter
from .forms import PersonForm
from .tables import PersonTable
class PersonViewMixin(ExamplesMixin, CollectionViewMixin):
model = Person
collection_list_title = 'Persons'
collection_list_urlname = 'collection:list'
collection_detail_urlname = 'collection:change'
class PersonListView(PersonViewMixin, CruditorListView):
title = 'Persons'
table_class = PersonTable
def get_titlebuttons(self):
return [{'url': reverse('collection:add'), 'label': 'Add person'}]
class PersonFilterView(PersonListView):
filter_class = PersonFilter
class PersonAddView(PersonViewMixin, CruditorAddView):
success_url = reverse_lazy('collection:lits')
form_class = PersonForm
class PersonChangeView(PersonViewMixin, CruditorChangeView):
form_class = PersonForm
def get_delete_url(self):
return reverse('collection:delete', args=(self.object.pk,))
class PersonDeleteView(PersonViewMixin, CruditorDeleteView):
pass
<commit_msg>Make use of auto generated table classes.<commit_after> | from cruditor.contrib.collection import CollectionViewMixin
from cruditor.views import CruditorAddView, CruditorChangeView, CruditorDeleteView, CruditorListView
from django.urls import reverse, reverse_lazy
from examples.mixins import ExamplesMixin
from store.models import Person
from .filters import PersonFilter
from .forms import PersonForm
from .tables import PersonTable
class PersonViewMixin(ExamplesMixin, CollectionViewMixin):
model = Person
collection_list_title = 'Persons'
collection_list_urlname = 'collection:list'
collection_detail_urlname = 'collection:change'
class PersonListView(PersonViewMixin, CruditorListView):
title = 'Persons'
def get_titlebuttons(self):
return [{'url': reverse('collection:add'), 'label': 'Add person'}]
class PersonFilterView(PersonListView):
filter_class = PersonFilter
table_class = PersonTable
class PersonAddView(PersonViewMixin, CruditorAddView):
success_url = reverse_lazy('collection:lits')
form_class = PersonForm
class PersonChangeView(PersonViewMixin, CruditorChangeView):
form_class = PersonForm
def get_delete_url(self):
return reverse('collection:delete', args=(self.object.pk,))
class PersonDeleteView(PersonViewMixin, CruditorDeleteView):
pass
| from cruditor.contrib.collection import CollectionViewMixin
from cruditor.views import CruditorAddView, CruditorChangeView, CruditorDeleteView, CruditorListView
from django.urls import reverse, reverse_lazy
from examples.mixins import ExamplesMixin
from store.models import Person
from .filters import PersonFilter
from .forms import PersonForm
from .tables import PersonTable
class PersonViewMixin(ExamplesMixin, CollectionViewMixin):
model = Person
collection_list_title = 'Persons'
collection_list_urlname = 'collection:list'
collection_detail_urlname = 'collection:change'
class PersonListView(PersonViewMixin, CruditorListView):
title = 'Persons'
table_class = PersonTable
def get_titlebuttons(self):
return [{'url': reverse('collection:add'), 'label': 'Add person'}]
class PersonFilterView(PersonListView):
filter_class = PersonFilter
class PersonAddView(PersonViewMixin, CruditorAddView):
success_url = reverse_lazy('collection:lits')
form_class = PersonForm
class PersonChangeView(PersonViewMixin, CruditorChangeView):
form_class = PersonForm
def get_delete_url(self):
return reverse('collection:delete', args=(self.object.pk,))
class PersonDeleteView(PersonViewMixin, CruditorDeleteView):
pass
Make use of auto generated table classes.from cruditor.contrib.collection import CollectionViewMixin
from cruditor.views import CruditorAddView, CruditorChangeView, CruditorDeleteView, CruditorListView
from django.urls import reverse, reverse_lazy
from examples.mixins import ExamplesMixin
from store.models import Person
from .filters import PersonFilter
from .forms import PersonForm
from .tables import PersonTable
class PersonViewMixin(ExamplesMixin, CollectionViewMixin):
model = Person
collection_list_title = 'Persons'
collection_list_urlname = 'collection:list'
collection_detail_urlname = 'collection:change'
class PersonListView(PersonViewMixin, CruditorListView):
title = 'Persons'
def get_titlebuttons(self):
return [{'url': reverse('collection:add'), 'label': 'Add person'}]
class PersonFilterView(PersonListView):
filter_class = PersonFilter
table_class = PersonTable
class PersonAddView(PersonViewMixin, CruditorAddView):
success_url = reverse_lazy('collection:lits')
form_class = PersonForm
class PersonChangeView(PersonViewMixin, CruditorChangeView):
form_class = PersonForm
def get_delete_url(self):
return reverse('collection:delete', args=(self.object.pk,))
class PersonDeleteView(PersonViewMixin, CruditorDeleteView):
pass
| <commit_before>from cruditor.contrib.collection import CollectionViewMixin
from cruditor.views import CruditorAddView, CruditorChangeView, CruditorDeleteView, CruditorListView
from django.urls import reverse, reverse_lazy
from examples.mixins import ExamplesMixin
from store.models import Person
from .filters import PersonFilter
from .forms import PersonForm
from .tables import PersonTable
class PersonViewMixin(ExamplesMixin, CollectionViewMixin):
model = Person
collection_list_title = 'Persons'
collection_list_urlname = 'collection:list'
collection_detail_urlname = 'collection:change'
class PersonListView(PersonViewMixin, CruditorListView):
title = 'Persons'
table_class = PersonTable
def get_titlebuttons(self):
return [{'url': reverse('collection:add'), 'label': 'Add person'}]
class PersonFilterView(PersonListView):
filter_class = PersonFilter
class PersonAddView(PersonViewMixin, CruditorAddView):
success_url = reverse_lazy('collection:lits')
form_class = PersonForm
class PersonChangeView(PersonViewMixin, CruditorChangeView):
form_class = PersonForm
def get_delete_url(self):
return reverse('collection:delete', args=(self.object.pk,))
class PersonDeleteView(PersonViewMixin, CruditorDeleteView):
pass
<commit_msg>Make use of auto generated table classes.<commit_after>from cruditor.contrib.collection import CollectionViewMixin
from cruditor.views import CruditorAddView, CruditorChangeView, CruditorDeleteView, CruditorListView
from django.urls import reverse, reverse_lazy
from examples.mixins import ExamplesMixin
from store.models import Person
from .filters import PersonFilter
from .forms import PersonForm
from .tables import PersonTable
class PersonViewMixin(ExamplesMixin, CollectionViewMixin):
model = Person
collection_list_title = 'Persons'
collection_list_urlname = 'collection:list'
collection_detail_urlname = 'collection:change'
class PersonListView(PersonViewMixin, CruditorListView):
title = 'Persons'
def get_titlebuttons(self):
return [{'url': reverse('collection:add'), 'label': 'Add person'}]
class PersonFilterView(PersonListView):
filter_class = PersonFilter
table_class = PersonTable
class PersonAddView(PersonViewMixin, CruditorAddView):
success_url = reverse_lazy('collection:lits')
form_class = PersonForm
class PersonChangeView(PersonViewMixin, CruditorChangeView):
form_class = PersonForm
def get_delete_url(self):
return reverse('collection:delete', args=(self.object.pk,))
class PersonDeleteView(PersonViewMixin, CruditorDeleteView):
pass
|
4ce7f8ce338c84b44e7ad16475ff68bc0fad970e | dddp/accounts/tests.py | dddp/accounts/tests.py | """Django DDP Accounts test suite."""
from __future__ import unicode_literals
import sys
from dddp import tests
class AccountsTestCase(tests.DDPServerTestCase):
# gevent-websocket doesn't work with Python 3 yet
@tests.expected_failure_if(sys.version_info.major == 3)
def test_login_no_accounts(self):
sockjs = self.server.sockjs('/sockjs/1/a/websocket')
resp = sockjs.websocket.recv()
self.assertEqual(resp, 'o')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'server_id': '0'},
],
)
sockjs.connect('1', 'pre2', 'pre1')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'msg': 'connected', 'session': msgs[0]['session']},
],
)
id_ = sockjs.call(
'login', {'user': 'invalid@example.com', 'password': 'foo'},
)
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{
'msg': 'result', 'id': id_,
'error': {
'error': 403, 'reason': 'Authentication failed.',
},
},
],
)
sockjs.close()
| """Django DDP Accounts test suite."""
from __future__ import unicode_literals
import sys
from dddp import tests
# gevent-websocket doesn't work with Python 3 yet
@tests.expected_failure_if(sys.version_info.major == 3)
class AccountsTestCase(tests.DDPServerTestCase):
def test_login_no_accounts(self):
sockjs = self.server.sockjs('/sockjs/1/a/websocket')
resp = sockjs.websocket.recv()
self.assertEqual(resp, 'o')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'server_id': '0'},
],
)
sockjs.connect('1', 'pre2', 'pre1')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'msg': 'connected', 'session': msgs[0]['session']},
],
)
id_ = sockjs.call(
'login', {'user': 'invalid@example.com', 'password': 'foo'},
)
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{
'msg': 'result', 'id': id_,
'error': {
'error': 403, 'reason': 'Authentication failed.',
},
},
],
)
sockjs.close()
| Move expected test failure to TestCase class. | Move expected test failure to TestCase class.
| Python | mit | commoncode/django-ddp,django-ddp/django-ddp,commoncode/django-ddp,django-ddp/django-ddp,django-ddp/django-ddp,commoncode/django-ddp,django-ddp/django-ddp,commoncode/django-ddp | """Django DDP Accounts test suite."""
from __future__ import unicode_literals
import sys
from dddp import tests
class AccountsTestCase(tests.DDPServerTestCase):
# gevent-websocket doesn't work with Python 3 yet
@tests.expected_failure_if(sys.version_info.major == 3)
def test_login_no_accounts(self):
sockjs = self.server.sockjs('/sockjs/1/a/websocket')
resp = sockjs.websocket.recv()
self.assertEqual(resp, 'o')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'server_id': '0'},
],
)
sockjs.connect('1', 'pre2', 'pre1')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'msg': 'connected', 'session': msgs[0]['session']},
],
)
id_ = sockjs.call(
'login', {'user': 'invalid@example.com', 'password': 'foo'},
)
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{
'msg': 'result', 'id': id_,
'error': {
'error': 403, 'reason': 'Authentication failed.',
},
},
],
)
sockjs.close()
Move expected test failure to TestCase class. | """Django DDP Accounts test suite."""
from __future__ import unicode_literals
import sys
from dddp import tests
# gevent-websocket doesn't work with Python 3 yet
@tests.expected_failure_if(sys.version_info.major == 3)
class AccountsTestCase(tests.DDPServerTestCase):
def test_login_no_accounts(self):
sockjs = self.server.sockjs('/sockjs/1/a/websocket')
resp = sockjs.websocket.recv()
self.assertEqual(resp, 'o')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'server_id': '0'},
],
)
sockjs.connect('1', 'pre2', 'pre1')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'msg': 'connected', 'session': msgs[0]['session']},
],
)
id_ = sockjs.call(
'login', {'user': 'invalid@example.com', 'password': 'foo'},
)
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{
'msg': 'result', 'id': id_,
'error': {
'error': 403, 'reason': 'Authentication failed.',
},
},
],
)
sockjs.close()
| <commit_before>"""Django DDP Accounts test suite."""
from __future__ import unicode_literals
import sys
from dddp import tests
class AccountsTestCase(tests.DDPServerTestCase):
# gevent-websocket doesn't work with Python 3 yet
@tests.expected_failure_if(sys.version_info.major == 3)
def test_login_no_accounts(self):
sockjs = self.server.sockjs('/sockjs/1/a/websocket')
resp = sockjs.websocket.recv()
self.assertEqual(resp, 'o')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'server_id': '0'},
],
)
sockjs.connect('1', 'pre2', 'pre1')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'msg': 'connected', 'session': msgs[0]['session']},
],
)
id_ = sockjs.call(
'login', {'user': 'invalid@example.com', 'password': 'foo'},
)
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{
'msg': 'result', 'id': id_,
'error': {
'error': 403, 'reason': 'Authentication failed.',
},
},
],
)
sockjs.close()
<commit_msg>Move expected test failure to TestCase class.<commit_after> | """Django DDP Accounts test suite."""
from __future__ import unicode_literals
import sys
from dddp import tests
# gevent-websocket doesn't work with Python 3 yet
@tests.expected_failure_if(sys.version_info.major == 3)
class AccountsTestCase(tests.DDPServerTestCase):
def test_login_no_accounts(self):
sockjs = self.server.sockjs('/sockjs/1/a/websocket')
resp = sockjs.websocket.recv()
self.assertEqual(resp, 'o')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'server_id': '0'},
],
)
sockjs.connect('1', 'pre2', 'pre1')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'msg': 'connected', 'session': msgs[0]['session']},
],
)
id_ = sockjs.call(
'login', {'user': 'invalid@example.com', 'password': 'foo'},
)
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{
'msg': 'result', 'id': id_,
'error': {
'error': 403, 'reason': 'Authentication failed.',
},
},
],
)
sockjs.close()
| """Django DDP Accounts test suite."""
from __future__ import unicode_literals
import sys
from dddp import tests
class AccountsTestCase(tests.DDPServerTestCase):
# gevent-websocket doesn't work with Python 3 yet
@tests.expected_failure_if(sys.version_info.major == 3)
def test_login_no_accounts(self):
sockjs = self.server.sockjs('/sockjs/1/a/websocket')
resp = sockjs.websocket.recv()
self.assertEqual(resp, 'o')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'server_id': '0'},
],
)
sockjs.connect('1', 'pre2', 'pre1')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'msg': 'connected', 'session': msgs[0]['session']},
],
)
id_ = sockjs.call(
'login', {'user': 'invalid@example.com', 'password': 'foo'},
)
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{
'msg': 'result', 'id': id_,
'error': {
'error': 403, 'reason': 'Authentication failed.',
},
},
],
)
sockjs.close()
Move expected test failure to TestCase class."""Django DDP Accounts test suite."""
from __future__ import unicode_literals
import sys
from dddp import tests
# gevent-websocket doesn't work with Python 3 yet
@tests.expected_failure_if(sys.version_info.major == 3)
class AccountsTestCase(tests.DDPServerTestCase):
def test_login_no_accounts(self):
sockjs = self.server.sockjs('/sockjs/1/a/websocket')
resp = sockjs.websocket.recv()
self.assertEqual(resp, 'o')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'server_id': '0'},
],
)
sockjs.connect('1', 'pre2', 'pre1')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'msg': 'connected', 'session': msgs[0]['session']},
],
)
id_ = sockjs.call(
'login', {'user': 'invalid@example.com', 'password': 'foo'},
)
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{
'msg': 'result', 'id': id_,
'error': {
'error': 403, 'reason': 'Authentication failed.',
},
},
],
)
sockjs.close()
| <commit_before>"""Django DDP Accounts test suite."""
from __future__ import unicode_literals
import sys
from dddp import tests
class AccountsTestCase(tests.DDPServerTestCase):
# gevent-websocket doesn't work with Python 3 yet
@tests.expected_failure_if(sys.version_info.major == 3)
def test_login_no_accounts(self):
sockjs = self.server.sockjs('/sockjs/1/a/websocket')
resp = sockjs.websocket.recv()
self.assertEqual(resp, 'o')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'server_id': '0'},
],
)
sockjs.connect('1', 'pre2', 'pre1')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'msg': 'connected', 'session': msgs[0]['session']},
],
)
id_ = sockjs.call(
'login', {'user': 'invalid@example.com', 'password': 'foo'},
)
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{
'msg': 'result', 'id': id_,
'error': {
'error': 403, 'reason': 'Authentication failed.',
},
},
],
)
sockjs.close()
<commit_msg>Move expected test failure to TestCase class.<commit_after>"""Django DDP Accounts test suite."""
from __future__ import unicode_literals
import sys
from dddp import tests
# gevent-websocket doesn't work with Python 3 yet
@tests.expected_failure_if(sys.version_info.major == 3)
class AccountsTestCase(tests.DDPServerTestCase):
def test_login_no_accounts(self):
sockjs = self.server.sockjs('/sockjs/1/a/websocket')
resp = sockjs.websocket.recv()
self.assertEqual(resp, 'o')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'server_id': '0'},
],
)
sockjs.connect('1', 'pre2', 'pre1')
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{'msg': 'connected', 'session': msgs[0]['session']},
],
)
id_ = sockjs.call(
'login', {'user': 'invalid@example.com', 'password': 'foo'},
)
msgs = sockjs.recv()
self.assertEqual(
msgs, [
{
'msg': 'result', 'id': id_,
'error': {
'error': 403, 'reason': 'Authentication failed.',
},
},
],
)
sockjs.close()
|
02f7a546cda7b8b3ce31616a74f3aa3518632885 | djangocms_spa_vue_js/templatetags/router_tags.py | djangocms_spa_vue_js/templatetags/router_tags.py | import json
from django import template
from django.utils.safestring import mark_safe
from ..menu_helpers import get_vue_js_router
register = template.Library()
@register.simple_tag(takes_context=True)
def vue_js_router(context):
if context.has_key('vue_js_router'):
router = context['vue_js_router']
else:
router = get_vue_js_router(context=context)
router_json = json.dumps(router)
escaped_router_json = router_json.replace("'", "'") # Escape apostrophes to prevent JS errors.
return mark_safe(escaped_router_json)
| import json
from django import template
from django.utils.safestring import mark_safe
from ..menu_helpers import get_vue_js_router
register = template.Library()
@register.simple_tag(takes_context=True)
def vue_js_router(context):
if 'vue_js_router' in context:
router = context['vue_js_router']
else:
router = get_vue_js_router(context=context)
router_json = json.dumps(router)
escaped_router_json = router_json.replace("'", "'") # Escape apostrophes to prevent JS errors.
return mark_safe(escaped_router_json)
| Use `in` rather than `has_key` | Use `in` rather than `has_key`
| Python | mit | dreipol/djangocms-spa-vue-js | import json
from django import template
from django.utils.safestring import mark_safe
from ..menu_helpers import get_vue_js_router
register = template.Library()
@register.simple_tag(takes_context=True)
def vue_js_router(context):
if context.has_key('vue_js_router'):
router = context['vue_js_router']
else:
router = get_vue_js_router(context=context)
router_json = json.dumps(router)
escaped_router_json = router_json.replace("'", "'") # Escape apostrophes to prevent JS errors.
return mark_safe(escaped_router_json)
Use `in` rather than `has_key` | import json
from django import template
from django.utils.safestring import mark_safe
from ..menu_helpers import get_vue_js_router
register = template.Library()
@register.simple_tag(takes_context=True)
def vue_js_router(context):
if 'vue_js_router' in context:
router = context['vue_js_router']
else:
router = get_vue_js_router(context=context)
router_json = json.dumps(router)
escaped_router_json = router_json.replace("'", "'") # Escape apostrophes to prevent JS errors.
return mark_safe(escaped_router_json)
| <commit_before>import json
from django import template
from django.utils.safestring import mark_safe
from ..menu_helpers import get_vue_js_router
register = template.Library()
@register.simple_tag(takes_context=True)
def vue_js_router(context):
if context.has_key('vue_js_router'):
router = context['vue_js_router']
else:
router = get_vue_js_router(context=context)
router_json = json.dumps(router)
escaped_router_json = router_json.replace("'", "'") # Escape apostrophes to prevent JS errors.
return mark_safe(escaped_router_json)
<commit_msg>Use `in` rather than `has_key`<commit_after> | import json
from django import template
from django.utils.safestring import mark_safe
from ..menu_helpers import get_vue_js_router
register = template.Library()
@register.simple_tag(takes_context=True)
def vue_js_router(context):
if 'vue_js_router' in context:
router = context['vue_js_router']
else:
router = get_vue_js_router(context=context)
router_json = json.dumps(router)
escaped_router_json = router_json.replace("'", "'") # Escape apostrophes to prevent JS errors.
return mark_safe(escaped_router_json)
| import json
from django import template
from django.utils.safestring import mark_safe
from ..menu_helpers import get_vue_js_router
register = template.Library()
@register.simple_tag(takes_context=True)
def vue_js_router(context):
if context.has_key('vue_js_router'):
router = context['vue_js_router']
else:
router = get_vue_js_router(context=context)
router_json = json.dumps(router)
escaped_router_json = router_json.replace("'", "'") # Escape apostrophes to prevent JS errors.
return mark_safe(escaped_router_json)
Use `in` rather than `has_key`import json
from django import template
from django.utils.safestring import mark_safe
from ..menu_helpers import get_vue_js_router
register = template.Library()
@register.simple_tag(takes_context=True)
def vue_js_router(context):
if 'vue_js_router' in context:
router = context['vue_js_router']
else:
router = get_vue_js_router(context=context)
router_json = json.dumps(router)
escaped_router_json = router_json.replace("'", "'") # Escape apostrophes to prevent JS errors.
return mark_safe(escaped_router_json)
| <commit_before>import json
from django import template
from django.utils.safestring import mark_safe
from ..menu_helpers import get_vue_js_router
register = template.Library()
@register.simple_tag(takes_context=True)
def vue_js_router(context):
if context.has_key('vue_js_router'):
router = context['vue_js_router']
else:
router = get_vue_js_router(context=context)
router_json = json.dumps(router)
escaped_router_json = router_json.replace("'", "'") # Escape apostrophes to prevent JS errors.
return mark_safe(escaped_router_json)
<commit_msg>Use `in` rather than `has_key`<commit_after>import json
from django import template
from django.utils.safestring import mark_safe
from ..menu_helpers import get_vue_js_router
register = template.Library()
@register.simple_tag(takes_context=True)
def vue_js_router(context):
if 'vue_js_router' in context:
router = context['vue_js_router']
else:
router = get_vue_js_router(context=context)
router_json = json.dumps(router)
escaped_router_json = router_json.replace("'", "'") # Escape apostrophes to prevent JS errors.
return mark_safe(escaped_router_json)
|
15bbe3aaaa017513ac652bf246b906139a71be00 | doc/tutorials/examples/general/client/headers.py | doc/tutorials/examples/general/client/headers.py | import base64
from pyamf.remoting.client import RemotingService
gw = RemotingService('http://demo.pyamf.org/gateway/recordset')
gw.addHTTPHeader('Set-Cookie', 'sessionid=QT3cUmACNeKQo5oPeM0')
gw.removeHTTPHeader('Set-Cookie')
username = 'admin'
password = 'admin'
auth = base64.encodestring('%s:%s' % (username, password))[:-1]
gw.addHTTPHeader("Authorization", "Basic %s" % auth) | from pyamf.remoting.client import RemotingService
gw = RemotingService('http://demo.pyamf.org/gateway/recordset')
gw.addHTTPHeader('Set-Cookie', 'sessionid=QT3cUmACNeKQo5oPeM0')
gw.removeHTTPHeader('Set-Cookie')
username = 'admin'
password = 'admin'
auth = ('%s:%s' % (username, password)).encode('base64')[:-1]
gw.addHTTPHeader("Authorization", "Basic %s" % auth)
service = gw.getService('service')
print service.getLanguages()
| Apply client authorization fix from wiki | Apply client authorization fix from wiki
git-svn-id: f3978d5834b2aa37aa734927aace4f0b92cf88c5@2985 2dde4cc4-cf3c-0410-b1a3-a9b8ff274da5
| Python | mit | cardmagic/PyAMF,cardmagic/PyAMF,cardmagic/PyAMF | import base64
from pyamf.remoting.client import RemotingService
gw = RemotingService('http://demo.pyamf.org/gateway/recordset')
gw.addHTTPHeader('Set-Cookie', 'sessionid=QT3cUmACNeKQo5oPeM0')
gw.removeHTTPHeader('Set-Cookie')
username = 'admin'
password = 'admin'
auth = base64.encodestring('%s:%s' % (username, password))[:-1]
gw.addHTTPHeader("Authorization", "Basic %s" % auth)Apply client authorization fix from wiki
git-svn-id: f3978d5834b2aa37aa734927aace4f0b92cf88c5@2985 2dde4cc4-cf3c-0410-b1a3-a9b8ff274da5 | from pyamf.remoting.client import RemotingService
gw = RemotingService('http://demo.pyamf.org/gateway/recordset')
gw.addHTTPHeader('Set-Cookie', 'sessionid=QT3cUmACNeKQo5oPeM0')
gw.removeHTTPHeader('Set-Cookie')
username = 'admin'
password = 'admin'
auth = ('%s:%s' % (username, password)).encode('base64')[:-1]
gw.addHTTPHeader("Authorization", "Basic %s" % auth)
service = gw.getService('service')
print service.getLanguages()
| <commit_before>import base64
from pyamf.remoting.client import RemotingService
gw = RemotingService('http://demo.pyamf.org/gateway/recordset')
gw.addHTTPHeader('Set-Cookie', 'sessionid=QT3cUmACNeKQo5oPeM0')
gw.removeHTTPHeader('Set-Cookie')
username = 'admin'
password = 'admin'
auth = base64.encodestring('%s:%s' % (username, password))[:-1]
gw.addHTTPHeader("Authorization", "Basic %s" % auth)<commit_msg>Apply client authorization fix from wiki
git-svn-id: f3978d5834b2aa37aa734927aace4f0b92cf88c5@2985 2dde4cc4-cf3c-0410-b1a3-a9b8ff274da5<commit_after> | from pyamf.remoting.client import RemotingService
gw = RemotingService('http://demo.pyamf.org/gateway/recordset')
gw.addHTTPHeader('Set-Cookie', 'sessionid=QT3cUmACNeKQo5oPeM0')
gw.removeHTTPHeader('Set-Cookie')
username = 'admin'
password = 'admin'
auth = ('%s:%s' % (username, password)).encode('base64')[:-1]
gw.addHTTPHeader("Authorization", "Basic %s" % auth)
service = gw.getService('service')
print service.getLanguages()
| import base64
from pyamf.remoting.client import RemotingService
gw = RemotingService('http://demo.pyamf.org/gateway/recordset')
gw.addHTTPHeader('Set-Cookie', 'sessionid=QT3cUmACNeKQo5oPeM0')
gw.removeHTTPHeader('Set-Cookie')
username = 'admin'
password = 'admin'
auth = base64.encodestring('%s:%s' % (username, password))[:-1]
gw.addHTTPHeader("Authorization", "Basic %s" % auth)Apply client authorization fix from wiki
git-svn-id: f3978d5834b2aa37aa734927aace4f0b92cf88c5@2985 2dde4cc4-cf3c-0410-b1a3-a9b8ff274da5from pyamf.remoting.client import RemotingService
gw = RemotingService('http://demo.pyamf.org/gateway/recordset')
gw.addHTTPHeader('Set-Cookie', 'sessionid=QT3cUmACNeKQo5oPeM0')
gw.removeHTTPHeader('Set-Cookie')
username = 'admin'
password = 'admin'
auth = ('%s:%s' % (username, password)).encode('base64')[:-1]
gw.addHTTPHeader("Authorization", "Basic %s" % auth)
service = gw.getService('service')
print service.getLanguages()
| <commit_before>import base64
from pyamf.remoting.client import RemotingService
gw = RemotingService('http://demo.pyamf.org/gateway/recordset')
gw.addHTTPHeader('Set-Cookie', 'sessionid=QT3cUmACNeKQo5oPeM0')
gw.removeHTTPHeader('Set-Cookie')
username = 'admin'
password = 'admin'
auth = base64.encodestring('%s:%s' % (username, password))[:-1]
gw.addHTTPHeader("Authorization", "Basic %s" % auth)<commit_msg>Apply client authorization fix from wiki
git-svn-id: f3978d5834b2aa37aa734927aace4f0b92cf88c5@2985 2dde4cc4-cf3c-0410-b1a3-a9b8ff274da5<commit_after>from pyamf.remoting.client import RemotingService
gw = RemotingService('http://demo.pyamf.org/gateway/recordset')
gw.addHTTPHeader('Set-Cookie', 'sessionid=QT3cUmACNeKQo5oPeM0')
gw.removeHTTPHeader('Set-Cookie')
username = 'admin'
password = 'admin'
auth = ('%s:%s' % (username, password)).encode('base64')[:-1]
gw.addHTTPHeader("Authorization", "Basic %s" % auth)
service = gw.getService('service')
print service.getLanguages()
|
a3c3a6ed4d01f1857fc4728b10505e330af9e6ae | code/helper/easierlife.py | code/helper/easierlife.py | #! /usr/bin/env python3
""" Helper functions to make our life easier.
Originally obtained from the 'pharm' repository, but modified.
"""
import fileinput
import json
import os.path
from dstruct import Sentence
## BASE_DIR denotes the application directory
BASE_DIR, throwaway = os.path.split(os.path.realpath(__file__))
BASE_DIR = os.path.realpath(BASE_DIR + "/../..")
## Return the start and end indexes of all subsets of words in the sentence
## sent, with size at most max_phrase_length
def get_all_phrases_in_sentence(sent, max_phrase_length):
for start in range(len(sent.words)):
for end in reversed(range(start + 1, min(len(sent.words), start + 1 + max_phrase_length))):
yield (start, end)
## Return Sentence objects from input lines
def get_input_sentences(input_files=[]):
for line in fileinput.input(input_files):
sent_dict = json.loads(line)
yield Sentence(sent_dict["doc_id"], sent_dict["sent_id"],
sent_dict["wordidxs"], sent_dict["words"],
sent_dict["poses"], sent_dict["ners"], sent_dict["lemmas"],
sent_dict["dep_paths"], sent_dict["dep_parents"],
sent_dict["bounding_boxes"])
| #! /usr/bin/env python3
""" Helper functions to make our life easier.
Originally obtained from the 'pharm' repository, but modified.
"""
import fileinput
import json
import os.path
import sys
from dstruct.Sentence import Sentence
## BASE_DIR denotes the application directory
BASE_DIR, throwaway = os.path.split(os.path.realpath(__file__))
BASE_DIR = os.path.realpath(BASE_DIR + "/../..")
## Return the start and end indexes of all subsets of words in the sentence
## sent, with size at most max_phrase_length
def get_all_phrases_in_sentence(sent, max_phrase_length):
for start in range(len(sent.words)):
for end in reversed(range(start + 1, min(len(sent.words), start + 1 + max_phrase_length))):
yield (start, end)
## Return Sentence objects from input lines
def get_input_sentences(input_files=sys.argv[1:]):
with fileinput.input(files=input_files) as f:
for line in f:
sent_dict = json.loads(line)
yield Sentence(sent_dict["doc_id"], sent_dict["sent_id"],
sent_dict["wordidxs"], sent_dict["words"],
sent_dict["poses"], sent_dict["ners"], sent_dict["lemmas"],
sent_dict["dep_paths"], sent_dict["dep_parents"],
sent_dict["bounding_boxes"])
| Fix import, use fileinput.iput as context, and fix its argument | Fix import, use fileinput.iput as context, and fix its argument
| Python | apache-2.0 | amwenger/dd-genomics,rionda/dd-genomics,HazyResearch/dd-genomics,amwenger/dd-genomics,HazyResearch/dd-genomics,HazyResearch/dd-genomics,HazyResearch/dd-genomics,rionda/dd-genomics,amwenger/dd-genomics,HazyResearch/dd-genomics | #! /usr/bin/env python3
""" Helper functions to make our life easier.
Originally obtained from the 'pharm' repository, but modified.
"""
import fileinput
import json
import os.path
from dstruct import Sentence
## BASE_DIR denotes the application directory
BASE_DIR, throwaway = os.path.split(os.path.realpath(__file__))
BASE_DIR = os.path.realpath(BASE_DIR + "/../..")
## Return the start and end indexes of all subsets of words in the sentence
## sent, with size at most max_phrase_length
def get_all_phrases_in_sentence(sent, max_phrase_length):
for start in range(len(sent.words)):
for end in reversed(range(start + 1, min(len(sent.words), start + 1 + max_phrase_length))):
yield (start, end)
## Return Sentence objects from input lines
def get_input_sentences(input_files=[]):
for line in fileinput.input(input_files):
sent_dict = json.loads(line)
yield Sentence(sent_dict["doc_id"], sent_dict["sent_id"],
sent_dict["wordidxs"], sent_dict["words"],
sent_dict["poses"], sent_dict["ners"], sent_dict["lemmas"],
sent_dict["dep_paths"], sent_dict["dep_parents"],
sent_dict["bounding_boxes"])
Fix import, use fileinput.iput as context, and fix its argument | #! /usr/bin/env python3
""" Helper functions to make our life easier.
Originally obtained from the 'pharm' repository, but modified.
"""
import fileinput
import json
import os.path
import sys
from dstruct.Sentence import Sentence
## BASE_DIR denotes the application directory
BASE_DIR, throwaway = os.path.split(os.path.realpath(__file__))
BASE_DIR = os.path.realpath(BASE_DIR + "/../..")
## Return the start and end indexes of all subsets of words in the sentence
## sent, with size at most max_phrase_length
def get_all_phrases_in_sentence(sent, max_phrase_length):
for start in range(len(sent.words)):
for end in reversed(range(start + 1, min(len(sent.words), start + 1 + max_phrase_length))):
yield (start, end)
## Return Sentence objects from input lines
def get_input_sentences(input_files=sys.argv[1:]):
with fileinput.input(files=input_files) as f:
for line in f:
sent_dict = json.loads(line)
yield Sentence(sent_dict["doc_id"], sent_dict["sent_id"],
sent_dict["wordidxs"], sent_dict["words"],
sent_dict["poses"], sent_dict["ners"], sent_dict["lemmas"],
sent_dict["dep_paths"], sent_dict["dep_parents"],
sent_dict["bounding_boxes"])
| <commit_before>#! /usr/bin/env python3
""" Helper functions to make our life easier.
Originally obtained from the 'pharm' repository, but modified.
"""
import fileinput
import json
import os.path
from dstruct import Sentence
## BASE_DIR denotes the application directory
BASE_DIR, throwaway = os.path.split(os.path.realpath(__file__))
BASE_DIR = os.path.realpath(BASE_DIR + "/../..")
## Return the start and end indexes of all subsets of words in the sentence
## sent, with size at most max_phrase_length
def get_all_phrases_in_sentence(sent, max_phrase_length):
for start in range(len(sent.words)):
for end in reversed(range(start + 1, min(len(sent.words), start + 1 + max_phrase_length))):
yield (start, end)
## Return Sentence objects from input lines
def get_input_sentences(input_files=[]):
for line in fileinput.input(input_files):
sent_dict = json.loads(line)
yield Sentence(sent_dict["doc_id"], sent_dict["sent_id"],
sent_dict["wordidxs"], sent_dict["words"],
sent_dict["poses"], sent_dict["ners"], sent_dict["lemmas"],
sent_dict["dep_paths"], sent_dict["dep_parents"],
sent_dict["bounding_boxes"])
<commit_msg>Fix import, use fileinput.iput as context, and fix its argument<commit_after> | #! /usr/bin/env python3
""" Helper functions to make our life easier.
Originally obtained from the 'pharm' repository, but modified.
"""
import fileinput
import json
import os.path
import sys
from dstruct.Sentence import Sentence
## BASE_DIR denotes the application directory
BASE_DIR, throwaway = os.path.split(os.path.realpath(__file__))
BASE_DIR = os.path.realpath(BASE_DIR + "/../..")
## Return the start and end indexes of all subsets of words in the sentence
## sent, with size at most max_phrase_length
def get_all_phrases_in_sentence(sent, max_phrase_length):
for start in range(len(sent.words)):
for end in reversed(range(start + 1, min(len(sent.words), start + 1 + max_phrase_length))):
yield (start, end)
## Return Sentence objects from input lines
def get_input_sentences(input_files=sys.argv[1:]):
with fileinput.input(files=input_files) as f:
for line in f:
sent_dict = json.loads(line)
yield Sentence(sent_dict["doc_id"], sent_dict["sent_id"],
sent_dict["wordidxs"], sent_dict["words"],
sent_dict["poses"], sent_dict["ners"], sent_dict["lemmas"],
sent_dict["dep_paths"], sent_dict["dep_parents"],
sent_dict["bounding_boxes"])
| #! /usr/bin/env python3
""" Helper functions to make our life easier.
Originally obtained from the 'pharm' repository, but modified.
"""
import fileinput
import json
import os.path
from dstruct import Sentence
## BASE_DIR denotes the application directory
BASE_DIR, throwaway = os.path.split(os.path.realpath(__file__))
BASE_DIR = os.path.realpath(BASE_DIR + "/../..")
## Return the start and end indexes of all subsets of words in the sentence
## sent, with size at most max_phrase_length
def get_all_phrases_in_sentence(sent, max_phrase_length):
for start in range(len(sent.words)):
for end in reversed(range(start + 1, min(len(sent.words), start + 1 + max_phrase_length))):
yield (start, end)
## Return Sentence objects from input lines
def get_input_sentences(input_files=[]):
for line in fileinput.input(input_files):
sent_dict = json.loads(line)
yield Sentence(sent_dict["doc_id"], sent_dict["sent_id"],
sent_dict["wordidxs"], sent_dict["words"],
sent_dict["poses"], sent_dict["ners"], sent_dict["lemmas"],
sent_dict["dep_paths"], sent_dict["dep_parents"],
sent_dict["bounding_boxes"])
Fix import, use fileinput.iput as context, and fix its argument#! /usr/bin/env python3
""" Helper functions to make our life easier.
Originally obtained from the 'pharm' repository, but modified.
"""
import fileinput
import json
import os.path
import sys
from dstruct.Sentence import Sentence
## BASE_DIR denotes the application directory
BASE_DIR, throwaway = os.path.split(os.path.realpath(__file__))
BASE_DIR = os.path.realpath(BASE_DIR + "/../..")
## Return the start and end indexes of all subsets of words in the sentence
## sent, with size at most max_phrase_length
def get_all_phrases_in_sentence(sent, max_phrase_length):
for start in range(len(sent.words)):
for end in reversed(range(start + 1, min(len(sent.words), start + 1 + max_phrase_length))):
yield (start, end)
## Return Sentence objects from input lines
def get_input_sentences(input_files=sys.argv[1:]):
with fileinput.input(files=input_files) as f:
for line in f:
sent_dict = json.loads(line)
yield Sentence(sent_dict["doc_id"], sent_dict["sent_id"],
sent_dict["wordidxs"], sent_dict["words"],
sent_dict["poses"], sent_dict["ners"], sent_dict["lemmas"],
sent_dict["dep_paths"], sent_dict["dep_parents"],
sent_dict["bounding_boxes"])
| <commit_before>#! /usr/bin/env python3
""" Helper functions to make our life easier.
Originally obtained from the 'pharm' repository, but modified.
"""
import fileinput
import json
import os.path
from dstruct import Sentence
## BASE_DIR denotes the application directory
BASE_DIR, throwaway = os.path.split(os.path.realpath(__file__))
BASE_DIR = os.path.realpath(BASE_DIR + "/../..")
## Return the start and end indexes of all subsets of words in the sentence
## sent, with size at most max_phrase_length
def get_all_phrases_in_sentence(sent, max_phrase_length):
for start in range(len(sent.words)):
for end in reversed(range(start + 1, min(len(sent.words), start + 1 + max_phrase_length))):
yield (start, end)
## Return Sentence objects from input lines
def get_input_sentences(input_files=[]):
for line in fileinput.input(input_files):
sent_dict = json.loads(line)
yield Sentence(sent_dict["doc_id"], sent_dict["sent_id"],
sent_dict["wordidxs"], sent_dict["words"],
sent_dict["poses"], sent_dict["ners"], sent_dict["lemmas"],
sent_dict["dep_paths"], sent_dict["dep_parents"],
sent_dict["bounding_boxes"])
<commit_msg>Fix import, use fileinput.iput as context, and fix its argument<commit_after>#! /usr/bin/env python3
""" Helper functions to make our life easier.
Originally obtained from the 'pharm' repository, but modified.
"""
import fileinput
import json
import os.path
import sys
from dstruct.Sentence import Sentence
## BASE_DIR denotes the application directory
BASE_DIR, throwaway = os.path.split(os.path.realpath(__file__))
BASE_DIR = os.path.realpath(BASE_DIR + "/../..")
## Return the start and end indexes of all subsets of words in the sentence
## sent, with size at most max_phrase_length
def get_all_phrases_in_sentence(sent, max_phrase_length):
for start in range(len(sent.words)):
for end in reversed(range(start + 1, min(len(sent.words), start + 1 + max_phrase_length))):
yield (start, end)
## Return Sentence objects from input lines
def get_input_sentences(input_files=sys.argv[1:]):
with fileinput.input(files=input_files) as f:
for line in f:
sent_dict = json.loads(line)
yield Sentence(sent_dict["doc_id"], sent_dict["sent_id"],
sent_dict["wordidxs"], sent_dict["words"],
sent_dict["poses"], sent_dict["ners"], sent_dict["lemmas"],
sent_dict["dep_paths"], sent_dict["dep_parents"],
sent_dict["bounding_boxes"])
|
91aa7ed06d168700692a33fd3c51add585d60ac0 | backend/uclapi/roombookings/migrations/0007_auto_20170327_1323.py | backend/uclapi/roombookings/migrations/0007_auto_20170327_1323.py | # -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-27 13:23
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('roombookings', '0006_bookinga_bookingb_lock'),
]
operations = [
migrations.AddField(
model_name='bookinga',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AddField(
model_name='bookingb',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AlterField(
model_name='bookinga',
name='slotid',
field=models.BigIntegerField(),
),
migrations.AlterField(
model_name='bookingb',
name='slotid',
field=models.BigIntegerField(),
),
]
| # -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-27 13:23
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('roombookings', '0006_bookinga_bookingb_lock'),
]
operations = [
migrations.AlterField(
model_name='bookinga',
name='slotid',
field=models.BigIntegerField(null=True, primary_key=False),
),
migrations.AlterField(
model_name='bookingb',
name='slotid',
field=models.BigIntegerField(null=True, primary_key=False),
),
migrations.AddField(
model_name='bookinga',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AddField(
model_name='bookingb',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
]
| Fix up migration to have only one PK | Fix up migration to have only one PK
| Python | mit | uclapi/uclapi,uclapi/uclapi,uclapi/uclapi,uclapi/uclapi | # -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-27 13:23
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('roombookings', '0006_bookinga_bookingb_lock'),
]
operations = [
migrations.AddField(
model_name='bookinga',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AddField(
model_name='bookingb',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AlterField(
model_name='bookinga',
name='slotid',
field=models.BigIntegerField(),
),
migrations.AlterField(
model_name='bookingb',
name='slotid',
field=models.BigIntegerField(),
),
]
Fix up migration to have only one PK | # -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-27 13:23
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('roombookings', '0006_bookinga_bookingb_lock'),
]
operations = [
migrations.AlterField(
model_name='bookinga',
name='slotid',
field=models.BigIntegerField(null=True, primary_key=False),
),
migrations.AlterField(
model_name='bookingb',
name='slotid',
field=models.BigIntegerField(null=True, primary_key=False),
),
migrations.AddField(
model_name='bookinga',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AddField(
model_name='bookingb',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
]
| <commit_before># -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-27 13:23
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('roombookings', '0006_bookinga_bookingb_lock'),
]
operations = [
migrations.AddField(
model_name='bookinga',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AddField(
model_name='bookingb',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AlterField(
model_name='bookinga',
name='slotid',
field=models.BigIntegerField(),
),
migrations.AlterField(
model_name='bookingb',
name='slotid',
field=models.BigIntegerField(),
),
]
<commit_msg>Fix up migration to have only one PK<commit_after> | # -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-27 13:23
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('roombookings', '0006_bookinga_bookingb_lock'),
]
operations = [
migrations.AlterField(
model_name='bookinga',
name='slotid',
field=models.BigIntegerField(null=True, primary_key=False),
),
migrations.AlterField(
model_name='bookingb',
name='slotid',
field=models.BigIntegerField(null=True, primary_key=False),
),
migrations.AddField(
model_name='bookinga',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AddField(
model_name='bookingb',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
]
| # -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-27 13:23
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('roombookings', '0006_bookinga_bookingb_lock'),
]
operations = [
migrations.AddField(
model_name='bookinga',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AddField(
model_name='bookingb',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AlterField(
model_name='bookinga',
name='slotid',
field=models.BigIntegerField(),
),
migrations.AlterField(
model_name='bookingb',
name='slotid',
field=models.BigIntegerField(),
),
]
Fix up migration to have only one PK# -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-27 13:23
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('roombookings', '0006_bookinga_bookingb_lock'),
]
operations = [
migrations.AlterField(
model_name='bookinga',
name='slotid',
field=models.BigIntegerField(null=True, primary_key=False),
),
migrations.AlterField(
model_name='bookingb',
name='slotid',
field=models.BigIntegerField(null=True, primary_key=False),
),
migrations.AddField(
model_name='bookinga',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AddField(
model_name='bookingb',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
]
| <commit_before># -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-27 13:23
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('roombookings', '0006_bookinga_bookingb_lock'),
]
operations = [
migrations.AddField(
model_name='bookinga',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AddField(
model_name='bookingb',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AlterField(
model_name='bookinga',
name='slotid',
field=models.BigIntegerField(),
),
migrations.AlterField(
model_name='bookingb',
name='slotid',
field=models.BigIntegerField(),
),
]
<commit_msg>Fix up migration to have only one PK<commit_after># -*- coding: utf-8 -*-
# Generated by Django 1.10.5 on 2017-03-27 13:23
from __future__ import unicode_literals
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('roombookings', '0006_bookinga_bookingb_lock'),
]
operations = [
migrations.AlterField(
model_name='bookinga',
name='slotid',
field=models.BigIntegerField(null=True, primary_key=False),
),
migrations.AlterField(
model_name='bookingb',
name='slotid',
field=models.BigIntegerField(null=True, primary_key=False),
),
migrations.AddField(
model_name='bookinga',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
migrations.AddField(
model_name='bookingb',
name='id',
field=models.AutoField(primary_key=True, serialize=False),
preserve_default=False,
),
]
|
61e30e91ffc87a7a8f575d32fba43e61a65b477a | bot/storage/data_source/data_sources/sqlite/sqlite.py | bot/storage/data_source/data_sources/sqlite/sqlite.py | from sqlite_framework.log.logger import SqliteLogger
from sqlite_framework.session.session import SqliteSession
from bot.storage.data_source.data_source import StorageDataSource
class SqliteStorageDataSource(StorageDataSource):
def __init__(self, database_filename: str, debug: bool, logger: SqliteLogger):
super().__init__()
self.session = SqliteSession(database_filename, debug)
self.logger = logger
def init(self):
self.session.init()
def context_manager(self):
return self.session.context_manager()
| from sqlite_framework.log.logger import SqliteLogger
from sqlite_framework.session.session import SqliteSession
from bot.storage.data_source.data_source import StorageDataSource
class SqliteStorageDataSource(StorageDataSource):
def __init__(self, session: SqliteSession, logger: SqliteLogger):
super().__init__()
self.session = session
self.logger = logger
def init(self):
self.session.init()
def context_manager(self):
return self.session.context_manager()
| Update SqliteStorageDataSource to receive the SqliteSession already built, so that clients can have more control over its construction | Update SqliteStorageDataSource to receive the SqliteSession already built, so that clients can have more control over its construction
| Python | agpl-3.0 | alvarogzp/telegram-bot,alvarogzp/telegram-bot | from sqlite_framework.log.logger import SqliteLogger
from sqlite_framework.session.session import SqliteSession
from bot.storage.data_source.data_source import StorageDataSource
class SqliteStorageDataSource(StorageDataSource):
def __init__(self, database_filename: str, debug: bool, logger: SqliteLogger):
super().__init__()
self.session = SqliteSession(database_filename, debug)
self.logger = logger
def init(self):
self.session.init()
def context_manager(self):
return self.session.context_manager()
Update SqliteStorageDataSource to receive the SqliteSession already built, so that clients can have more control over its construction | from sqlite_framework.log.logger import SqliteLogger
from sqlite_framework.session.session import SqliteSession
from bot.storage.data_source.data_source import StorageDataSource
class SqliteStorageDataSource(StorageDataSource):
def __init__(self, session: SqliteSession, logger: SqliteLogger):
super().__init__()
self.session = session
self.logger = logger
def init(self):
self.session.init()
def context_manager(self):
return self.session.context_manager()
| <commit_before>from sqlite_framework.log.logger import SqliteLogger
from sqlite_framework.session.session import SqliteSession
from bot.storage.data_source.data_source import StorageDataSource
class SqliteStorageDataSource(StorageDataSource):
def __init__(self, database_filename: str, debug: bool, logger: SqliteLogger):
super().__init__()
self.session = SqliteSession(database_filename, debug)
self.logger = logger
def init(self):
self.session.init()
def context_manager(self):
return self.session.context_manager()
<commit_msg>Update SqliteStorageDataSource to receive the SqliteSession already built, so that clients can have more control over its construction<commit_after> | from sqlite_framework.log.logger import SqliteLogger
from sqlite_framework.session.session import SqliteSession
from bot.storage.data_source.data_source import StorageDataSource
class SqliteStorageDataSource(StorageDataSource):
def __init__(self, session: SqliteSession, logger: SqliteLogger):
super().__init__()
self.session = session
self.logger = logger
def init(self):
self.session.init()
def context_manager(self):
return self.session.context_manager()
| from sqlite_framework.log.logger import SqliteLogger
from sqlite_framework.session.session import SqliteSession
from bot.storage.data_source.data_source import StorageDataSource
class SqliteStorageDataSource(StorageDataSource):
def __init__(self, database_filename: str, debug: bool, logger: SqliteLogger):
super().__init__()
self.session = SqliteSession(database_filename, debug)
self.logger = logger
def init(self):
self.session.init()
def context_manager(self):
return self.session.context_manager()
Update SqliteStorageDataSource to receive the SqliteSession already built, so that clients can have more control over its constructionfrom sqlite_framework.log.logger import SqliteLogger
from sqlite_framework.session.session import SqliteSession
from bot.storage.data_source.data_source import StorageDataSource
class SqliteStorageDataSource(StorageDataSource):
def __init__(self, session: SqliteSession, logger: SqliteLogger):
super().__init__()
self.session = session
self.logger = logger
def init(self):
self.session.init()
def context_manager(self):
return self.session.context_manager()
| <commit_before>from sqlite_framework.log.logger import SqliteLogger
from sqlite_framework.session.session import SqliteSession
from bot.storage.data_source.data_source import StorageDataSource
class SqliteStorageDataSource(StorageDataSource):
def __init__(self, database_filename: str, debug: bool, logger: SqliteLogger):
super().__init__()
self.session = SqliteSession(database_filename, debug)
self.logger = logger
def init(self):
self.session.init()
def context_manager(self):
return self.session.context_manager()
<commit_msg>Update SqliteStorageDataSource to receive the SqliteSession already built, so that clients can have more control over its construction<commit_after>from sqlite_framework.log.logger import SqliteLogger
from sqlite_framework.session.session import SqliteSession
from bot.storage.data_source.data_source import StorageDataSource
class SqliteStorageDataSource(StorageDataSource):
def __init__(self, session: SqliteSession, logger: SqliteLogger):
super().__init__()
self.session = session
self.logger = logger
def init(self):
self.session.init()
def context_manager(self):
return self.session.context_manager()
|
9207009ae26324650f904b010a065d98c2f41300 | server/server/debug.py | server/server/debug.py | from django.http import HttpResponse
import json
class NonHtmlDebugToolbarMiddleware(object):
"""
The Django Debug Toolbar usually only works for views that return HTML.
This middleware wraps any non-HTML response in HTML if the request has a
'debug' query parameter (e.g. http://localhost/foo?debug) Special handling
for json (pretty printing) and binary data (only show data length).
Based on http://stackoverflow.com/a/19249559/10817
"""
@staticmethod
def process_response(request, response):
if request.GET.get('debug') == '':
if response['Content-Type'] == 'application/octet-stream':
new_content = '<html><body>Binary Data, ' \
'Length: {}</body></html>'.format(len(response.content))
response = HttpResponse(new_content)
elif response['Content-Type'] != 'text/html':
content = response.content
try:
json_ = json.loads(content)
content = json.dumps(json_, sort_keys=True, indent=2)
except ValueError:
pass
response = HttpResponse('<html><body><pre>{}'
'</pre></body></html>'.format(content))
return response
# Middleware classes for debug toolbar.
middleware = ('debug_toolbar.middleware.DebugToolbarMiddleware',
'server.debug.NonHtmlDebugToolbarMiddleware')
| from django.http import HttpResponse
import json
class NonHtmlDebugToolbarMiddleware(object):
"""
The Django Debug Toolbar usually only works for views that return HTML.
This middleware wraps any non-HTML response in HTML if the request has a
'debug' query parameter (e.g. http://localhost/foo?debug) Special handling
for json (pretty printing) and binary data (only show data length).
Based on http://stackoverflow.com/a/19249559/10817
"""
@staticmethod
def process_response(request, response):
if request.GET.get('debug') == '':
if response['Content-Type'] == 'application/octet-stream':
new_content = '<html><body>Binary Data, ' \
'Length: {}</body></html>'.format(len(response.content))
response = HttpResponse(new_content)
elif not response['Content-Type'].startswith('text/html'):
content = response.content
try:
json_ = json.loads(content)
content = json.dumps(json_, sort_keys=True, indent=2)
except ValueError:
pass
response = HttpResponse('<html><body><pre>{}'
'</pre></body></html>'.format(content))
return response
# Middleware classes for debug toolbar.
middleware = ('debug_toolbar.middleware.DebugToolbarMiddleware',
'server.debug.NonHtmlDebugToolbarMiddleware')
| Drop wrap content starting with text/html | Drop wrap content starting with text/html
New versions of Django give HTML responses with the content type
'text/html; charset: utf-8'. We don't want to wrap that, so only check
for a start of text/html.
| Python | apache-2.0 | auvsi-suas/interop,auvsi-suas/interop,auvsi-suas/interop,justineaster/interop,justineaster/interop,justineaster/interop,auvsi-suas/interop,justineaster/interop,justineaster/interop | from django.http import HttpResponse
import json
class NonHtmlDebugToolbarMiddleware(object):
"""
The Django Debug Toolbar usually only works for views that return HTML.
This middleware wraps any non-HTML response in HTML if the request has a
'debug' query parameter (e.g. http://localhost/foo?debug) Special handling
for json (pretty printing) and binary data (only show data length).
Based on http://stackoverflow.com/a/19249559/10817
"""
@staticmethod
def process_response(request, response):
if request.GET.get('debug') == '':
if response['Content-Type'] == 'application/octet-stream':
new_content = '<html><body>Binary Data, ' \
'Length: {}</body></html>'.format(len(response.content))
response = HttpResponse(new_content)
elif response['Content-Type'] != 'text/html':
content = response.content
try:
json_ = json.loads(content)
content = json.dumps(json_, sort_keys=True, indent=2)
except ValueError:
pass
response = HttpResponse('<html><body><pre>{}'
'</pre></body></html>'.format(content))
return response
# Middleware classes for debug toolbar.
middleware = ('debug_toolbar.middleware.DebugToolbarMiddleware',
'server.debug.NonHtmlDebugToolbarMiddleware')
Drop wrap content starting with text/html
New versions of Django give HTML responses with the content type
'text/html; charset: utf-8'. We don't want to wrap that, so only check
for a start of text/html. | from django.http import HttpResponse
import json
class NonHtmlDebugToolbarMiddleware(object):
"""
The Django Debug Toolbar usually only works for views that return HTML.
This middleware wraps any non-HTML response in HTML if the request has a
'debug' query parameter (e.g. http://localhost/foo?debug) Special handling
for json (pretty printing) and binary data (only show data length).
Based on http://stackoverflow.com/a/19249559/10817
"""
@staticmethod
def process_response(request, response):
if request.GET.get('debug') == '':
if response['Content-Type'] == 'application/octet-stream':
new_content = '<html><body>Binary Data, ' \
'Length: {}</body></html>'.format(len(response.content))
response = HttpResponse(new_content)
elif not response['Content-Type'].startswith('text/html'):
content = response.content
try:
json_ = json.loads(content)
content = json.dumps(json_, sort_keys=True, indent=2)
except ValueError:
pass
response = HttpResponse('<html><body><pre>{}'
'</pre></body></html>'.format(content))
return response
# Middleware classes for debug toolbar.
middleware = ('debug_toolbar.middleware.DebugToolbarMiddleware',
'server.debug.NonHtmlDebugToolbarMiddleware')
| <commit_before>from django.http import HttpResponse
import json
class NonHtmlDebugToolbarMiddleware(object):
"""
The Django Debug Toolbar usually only works for views that return HTML.
This middleware wraps any non-HTML response in HTML if the request has a
'debug' query parameter (e.g. http://localhost/foo?debug) Special handling
for json (pretty printing) and binary data (only show data length).
Based on http://stackoverflow.com/a/19249559/10817
"""
@staticmethod
def process_response(request, response):
if request.GET.get('debug') == '':
if response['Content-Type'] == 'application/octet-stream':
new_content = '<html><body>Binary Data, ' \
'Length: {}</body></html>'.format(len(response.content))
response = HttpResponse(new_content)
elif response['Content-Type'] != 'text/html':
content = response.content
try:
json_ = json.loads(content)
content = json.dumps(json_, sort_keys=True, indent=2)
except ValueError:
pass
response = HttpResponse('<html><body><pre>{}'
'</pre></body></html>'.format(content))
return response
# Middleware classes for debug toolbar.
middleware = ('debug_toolbar.middleware.DebugToolbarMiddleware',
'server.debug.NonHtmlDebugToolbarMiddleware')
<commit_msg>Drop wrap content starting with text/html
New versions of Django give HTML responses with the content type
'text/html; charset: utf-8'. We don't want to wrap that, so only check
for a start of text/html.<commit_after> | from django.http import HttpResponse
import json
class NonHtmlDebugToolbarMiddleware(object):
"""
The Django Debug Toolbar usually only works for views that return HTML.
This middleware wraps any non-HTML response in HTML if the request has a
'debug' query parameter (e.g. http://localhost/foo?debug) Special handling
for json (pretty printing) and binary data (only show data length).
Based on http://stackoverflow.com/a/19249559/10817
"""
@staticmethod
def process_response(request, response):
if request.GET.get('debug') == '':
if response['Content-Type'] == 'application/octet-stream':
new_content = '<html><body>Binary Data, ' \
'Length: {}</body></html>'.format(len(response.content))
response = HttpResponse(new_content)
elif not response['Content-Type'].startswith('text/html'):
content = response.content
try:
json_ = json.loads(content)
content = json.dumps(json_, sort_keys=True, indent=2)
except ValueError:
pass
response = HttpResponse('<html><body><pre>{}'
'</pre></body></html>'.format(content))
return response
# Middleware classes for debug toolbar.
middleware = ('debug_toolbar.middleware.DebugToolbarMiddleware',
'server.debug.NonHtmlDebugToolbarMiddleware')
| from django.http import HttpResponse
import json
class NonHtmlDebugToolbarMiddleware(object):
"""
The Django Debug Toolbar usually only works for views that return HTML.
This middleware wraps any non-HTML response in HTML if the request has a
'debug' query parameter (e.g. http://localhost/foo?debug) Special handling
for json (pretty printing) and binary data (only show data length).
Based on http://stackoverflow.com/a/19249559/10817
"""
@staticmethod
def process_response(request, response):
if request.GET.get('debug') == '':
if response['Content-Type'] == 'application/octet-stream':
new_content = '<html><body>Binary Data, ' \
'Length: {}</body></html>'.format(len(response.content))
response = HttpResponse(new_content)
elif response['Content-Type'] != 'text/html':
content = response.content
try:
json_ = json.loads(content)
content = json.dumps(json_, sort_keys=True, indent=2)
except ValueError:
pass
response = HttpResponse('<html><body><pre>{}'
'</pre></body></html>'.format(content))
return response
# Middleware classes for debug toolbar.
middleware = ('debug_toolbar.middleware.DebugToolbarMiddleware',
'server.debug.NonHtmlDebugToolbarMiddleware')
Drop wrap content starting with text/html
New versions of Django give HTML responses with the content type
'text/html; charset: utf-8'. We don't want to wrap that, so only check
for a start of text/html.from django.http import HttpResponse
import json
class NonHtmlDebugToolbarMiddleware(object):
"""
The Django Debug Toolbar usually only works for views that return HTML.
This middleware wraps any non-HTML response in HTML if the request has a
'debug' query parameter (e.g. http://localhost/foo?debug) Special handling
for json (pretty printing) and binary data (only show data length).
Based on http://stackoverflow.com/a/19249559/10817
"""
@staticmethod
def process_response(request, response):
if request.GET.get('debug') == '':
if response['Content-Type'] == 'application/octet-stream':
new_content = '<html><body>Binary Data, ' \
'Length: {}</body></html>'.format(len(response.content))
response = HttpResponse(new_content)
elif not response['Content-Type'].startswith('text/html'):
content = response.content
try:
json_ = json.loads(content)
content = json.dumps(json_, sort_keys=True, indent=2)
except ValueError:
pass
response = HttpResponse('<html><body><pre>{}'
'</pre></body></html>'.format(content))
return response
# Middleware classes for debug toolbar.
middleware = ('debug_toolbar.middleware.DebugToolbarMiddleware',
'server.debug.NonHtmlDebugToolbarMiddleware')
| <commit_before>from django.http import HttpResponse
import json
class NonHtmlDebugToolbarMiddleware(object):
"""
The Django Debug Toolbar usually only works for views that return HTML.
This middleware wraps any non-HTML response in HTML if the request has a
'debug' query parameter (e.g. http://localhost/foo?debug) Special handling
for json (pretty printing) and binary data (only show data length).
Based on http://stackoverflow.com/a/19249559/10817
"""
@staticmethod
def process_response(request, response):
if request.GET.get('debug') == '':
if response['Content-Type'] == 'application/octet-stream':
new_content = '<html><body>Binary Data, ' \
'Length: {}</body></html>'.format(len(response.content))
response = HttpResponse(new_content)
elif response['Content-Type'] != 'text/html':
content = response.content
try:
json_ = json.loads(content)
content = json.dumps(json_, sort_keys=True, indent=2)
except ValueError:
pass
response = HttpResponse('<html><body><pre>{}'
'</pre></body></html>'.format(content))
return response
# Middleware classes for debug toolbar.
middleware = ('debug_toolbar.middleware.DebugToolbarMiddleware',
'server.debug.NonHtmlDebugToolbarMiddleware')
<commit_msg>Drop wrap content starting with text/html
New versions of Django give HTML responses with the content type
'text/html; charset: utf-8'. We don't want to wrap that, so only check
for a start of text/html.<commit_after>from django.http import HttpResponse
import json
class NonHtmlDebugToolbarMiddleware(object):
"""
The Django Debug Toolbar usually only works for views that return HTML.
This middleware wraps any non-HTML response in HTML if the request has a
'debug' query parameter (e.g. http://localhost/foo?debug) Special handling
for json (pretty printing) and binary data (only show data length).
Based on http://stackoverflow.com/a/19249559/10817
"""
@staticmethod
def process_response(request, response):
if request.GET.get('debug') == '':
if response['Content-Type'] == 'application/octet-stream':
new_content = '<html><body>Binary Data, ' \
'Length: {}</body></html>'.format(len(response.content))
response = HttpResponse(new_content)
elif not response['Content-Type'].startswith('text/html'):
content = response.content
try:
json_ = json.loads(content)
content = json.dumps(json_, sort_keys=True, indent=2)
except ValueError:
pass
response = HttpResponse('<html><body><pre>{}'
'</pre></body></html>'.format(content))
return response
# Middleware classes for debug toolbar.
middleware = ('debug_toolbar.middleware.DebugToolbarMiddleware',
'server.debug.NonHtmlDebugToolbarMiddleware')
|
023e814e6661c11bfe58a4e3e4ce4167ae63cd7f | rdio_dl/cli.py | rdio_dl/cli.py | import click
import youtube_dl
from .config import storage_load
from .extractor import RdioIE
@click.command()
@click.option(u'-u', u'--user', help=u'A Rdio user')
@click.option(u'-p', u'--password', help=u'The password')
@click.argument(u'urls', required=True, nargs=-1)
def main(user, password, urls):
storage = storage_load()
with youtube_dl.YoutubeDL() as ydl:
ydl.add_info_extractor(RdioIE(storage, user, password))
ydl.download(urls)
| # -*- coding: utf-8 -*-
import click
import youtube_dl
from .config import storage_load
from .extractor import RdioIE
def add_info_extractor_above_generic(ydl, ie):
generic = ydl._ies.pop()
ydl.add_info_extractor(ie)
ydl.add_info_extractor(generic)
@click.command()
@click.option(u'-u', u'--user', help=u'A Rdio user')
@click.option(u'-p', u'--password', help=u'The password')
@click.argument(u'urls', required=True, nargs=-1)
def main(user, password, urls):
storage = storage_load()
with youtube_dl.YoutubeDL() as ydl:
add_info_extractor_above_generic(ydl, RdioIE(storage, user, password))
ydl.download(urls)
| Fix generic extractor being always selected | Fix generic extractor being always selected
Turns out our extractor was being inserted *after* the GenericIE.
Now we are inserting our RdioIE right above GenericIE.
| Python | mit | ravishi/rdio-dl | import click
import youtube_dl
from .config import storage_load
from .extractor import RdioIE
@click.command()
@click.option(u'-u', u'--user', help=u'A Rdio user')
@click.option(u'-p', u'--password', help=u'The password')
@click.argument(u'urls', required=True, nargs=-1)
def main(user, password, urls):
storage = storage_load()
with youtube_dl.YoutubeDL() as ydl:
ydl.add_info_extractor(RdioIE(storage, user, password))
ydl.download(urls)
Fix generic extractor being always selected
Turns out our extractor was being inserted *after* the GenericIE.
Now we are inserting our RdioIE right above GenericIE. | # -*- coding: utf-8 -*-
import click
import youtube_dl
from .config import storage_load
from .extractor import RdioIE
def add_info_extractor_above_generic(ydl, ie):
generic = ydl._ies.pop()
ydl.add_info_extractor(ie)
ydl.add_info_extractor(generic)
@click.command()
@click.option(u'-u', u'--user', help=u'A Rdio user')
@click.option(u'-p', u'--password', help=u'The password')
@click.argument(u'urls', required=True, nargs=-1)
def main(user, password, urls):
storage = storage_load()
with youtube_dl.YoutubeDL() as ydl:
add_info_extractor_above_generic(ydl, RdioIE(storage, user, password))
ydl.download(urls)
| <commit_before>import click
import youtube_dl
from .config import storage_load
from .extractor import RdioIE
@click.command()
@click.option(u'-u', u'--user', help=u'A Rdio user')
@click.option(u'-p', u'--password', help=u'The password')
@click.argument(u'urls', required=True, nargs=-1)
def main(user, password, urls):
storage = storage_load()
with youtube_dl.YoutubeDL() as ydl:
ydl.add_info_extractor(RdioIE(storage, user, password))
ydl.download(urls)
<commit_msg>Fix generic extractor being always selected
Turns out our extractor was being inserted *after* the GenericIE.
Now we are inserting our RdioIE right above GenericIE.<commit_after> | # -*- coding: utf-8 -*-
import click
import youtube_dl
from .config import storage_load
from .extractor import RdioIE
def add_info_extractor_above_generic(ydl, ie):
generic = ydl._ies.pop()
ydl.add_info_extractor(ie)
ydl.add_info_extractor(generic)
@click.command()
@click.option(u'-u', u'--user', help=u'A Rdio user')
@click.option(u'-p', u'--password', help=u'The password')
@click.argument(u'urls', required=True, nargs=-1)
def main(user, password, urls):
storage = storage_load()
with youtube_dl.YoutubeDL() as ydl:
add_info_extractor_above_generic(ydl, RdioIE(storage, user, password))
ydl.download(urls)
| import click
import youtube_dl
from .config import storage_load
from .extractor import RdioIE
@click.command()
@click.option(u'-u', u'--user', help=u'A Rdio user')
@click.option(u'-p', u'--password', help=u'The password')
@click.argument(u'urls', required=True, nargs=-1)
def main(user, password, urls):
storage = storage_load()
with youtube_dl.YoutubeDL() as ydl:
ydl.add_info_extractor(RdioIE(storage, user, password))
ydl.download(urls)
Fix generic extractor being always selected
Turns out our extractor was being inserted *after* the GenericIE.
Now we are inserting our RdioIE right above GenericIE.# -*- coding: utf-8 -*-
import click
import youtube_dl
from .config import storage_load
from .extractor import RdioIE
def add_info_extractor_above_generic(ydl, ie):
generic = ydl._ies.pop()
ydl.add_info_extractor(ie)
ydl.add_info_extractor(generic)
@click.command()
@click.option(u'-u', u'--user', help=u'A Rdio user')
@click.option(u'-p', u'--password', help=u'The password')
@click.argument(u'urls', required=True, nargs=-1)
def main(user, password, urls):
storage = storage_load()
with youtube_dl.YoutubeDL() as ydl:
add_info_extractor_above_generic(ydl, RdioIE(storage, user, password))
ydl.download(urls)
| <commit_before>import click
import youtube_dl
from .config import storage_load
from .extractor import RdioIE
@click.command()
@click.option(u'-u', u'--user', help=u'A Rdio user')
@click.option(u'-p', u'--password', help=u'The password')
@click.argument(u'urls', required=True, nargs=-1)
def main(user, password, urls):
storage = storage_load()
with youtube_dl.YoutubeDL() as ydl:
ydl.add_info_extractor(RdioIE(storage, user, password))
ydl.download(urls)
<commit_msg>Fix generic extractor being always selected
Turns out our extractor was being inserted *after* the GenericIE.
Now we are inserting our RdioIE right above GenericIE.<commit_after># -*- coding: utf-8 -*-
import click
import youtube_dl
from .config import storage_load
from .extractor import RdioIE
def add_info_extractor_above_generic(ydl, ie):
generic = ydl._ies.pop()
ydl.add_info_extractor(ie)
ydl.add_info_extractor(generic)
@click.command()
@click.option(u'-u', u'--user', help=u'A Rdio user')
@click.option(u'-p', u'--password', help=u'The password')
@click.argument(u'urls', required=True, nargs=-1)
def main(user, password, urls):
storage = storage_load()
with youtube_dl.YoutubeDL() as ydl:
add_info_extractor_above_generic(ydl, RdioIE(storage, user, password))
ydl.download(urls)
|
fb3f1023faedda37e5ca16b87d2b9ddc38a2196c | deployer/tasks/util.py | deployer/tasks/util.py | from celery.result import ResultBase, AsyncResult, GroupResult
import deployer
from deployer.tasks.exceptions import TaskExecutionException
__author__ = 'sukrit'
def check_or_raise_task_exception(result):
if isinstance(result, AsyncResult) and result.failed():
if isinstance(result.result, TaskExecutionException):
raise result.result
else:
raise TaskExecutionException(result.result, result.traceback)
def _check_error(result):
if not result or not isinstance(result, AsyncResult):
return
check_or_raise_task_exception(result)
_check_error(result.parent)
def simple_result(result):
# DO not remove line below
# Explanation: https://github.com/celery/celery/issues/2315
deployer.celery.app.set_current()
if isinstance(result, GroupResult):
return simple_result(result.results)
elif hasattr(result, '__iter__') and not isinstance(result, dict):
return [simple_result(each_result)
for each_result in result]
elif isinstance(result, ResultBase):
_check_error(result)
if result.ready():
check_or_raise_task_exception(result)
return simple_result(result.result)
else:
raise TaskNotReadyException()
return result
class TaskNotReadyException(Exception):
pass
| import socket
from celery.result import ResultBase, AsyncResult, GroupResult
import deployer
from deployer.tasks.exceptions import TaskExecutionException
from deployer.util import retry
__author__ = 'sukrit'
def check_or_raise_task_exception(result):
if isinstance(result, AsyncResult) and result.failed():
if isinstance(result.result, TaskExecutionException):
raise result.result
else:
raise TaskExecutionException(result.result, result.traceback)
def _check_error(result):
if not result or not isinstance(result, AsyncResult):
return
check_or_raise_task_exception(result)
_check_error(result.parent)
@retry(10, delay=5, backoff=1, except_on=(IOError, socket.error))
def simple_result(result):
# DO not remove line below
# Explanation: https://github.com/celery/celery/issues/2315
deployer.celery.app.set_current()
if isinstance(result, GroupResult):
return simple_result(result.results)
elif hasattr(result, '__iter__') and not isinstance(result, dict):
return [simple_result(each_result)
for each_result in result]
elif isinstance(result, ResultBase):
_check_error(result)
if result.ready():
check_or_raise_task_exception(result)
return simple_result(result.result)
else:
raise TaskNotReadyException()
return result
class TaskNotReadyException(Exception):
pass
| Add retry for socket error | Add retry for socket error
| Python | mit | totem/cluster-deployer,totem/cluster-deployer,totem/cluster-deployer | from celery.result import ResultBase, AsyncResult, GroupResult
import deployer
from deployer.tasks.exceptions import TaskExecutionException
__author__ = 'sukrit'
def check_or_raise_task_exception(result):
if isinstance(result, AsyncResult) and result.failed():
if isinstance(result.result, TaskExecutionException):
raise result.result
else:
raise TaskExecutionException(result.result, result.traceback)
def _check_error(result):
if not result or not isinstance(result, AsyncResult):
return
check_or_raise_task_exception(result)
_check_error(result.parent)
def simple_result(result):
# DO not remove line below
# Explanation: https://github.com/celery/celery/issues/2315
deployer.celery.app.set_current()
if isinstance(result, GroupResult):
return simple_result(result.results)
elif hasattr(result, '__iter__') and not isinstance(result, dict):
return [simple_result(each_result)
for each_result in result]
elif isinstance(result, ResultBase):
_check_error(result)
if result.ready():
check_or_raise_task_exception(result)
return simple_result(result.result)
else:
raise TaskNotReadyException()
return result
class TaskNotReadyException(Exception):
pass
Add retry for socket error | import socket
from celery.result import ResultBase, AsyncResult, GroupResult
import deployer
from deployer.tasks.exceptions import TaskExecutionException
from deployer.util import retry
__author__ = 'sukrit'
def check_or_raise_task_exception(result):
if isinstance(result, AsyncResult) and result.failed():
if isinstance(result.result, TaskExecutionException):
raise result.result
else:
raise TaskExecutionException(result.result, result.traceback)
def _check_error(result):
if not result or not isinstance(result, AsyncResult):
return
check_or_raise_task_exception(result)
_check_error(result.parent)
@retry(10, delay=5, backoff=1, except_on=(IOError, socket.error))
def simple_result(result):
# DO not remove line below
# Explanation: https://github.com/celery/celery/issues/2315
deployer.celery.app.set_current()
if isinstance(result, GroupResult):
return simple_result(result.results)
elif hasattr(result, '__iter__') and not isinstance(result, dict):
return [simple_result(each_result)
for each_result in result]
elif isinstance(result, ResultBase):
_check_error(result)
if result.ready():
check_or_raise_task_exception(result)
return simple_result(result.result)
else:
raise TaskNotReadyException()
return result
class TaskNotReadyException(Exception):
pass
| <commit_before>from celery.result import ResultBase, AsyncResult, GroupResult
import deployer
from deployer.tasks.exceptions import TaskExecutionException
__author__ = 'sukrit'
def check_or_raise_task_exception(result):
if isinstance(result, AsyncResult) and result.failed():
if isinstance(result.result, TaskExecutionException):
raise result.result
else:
raise TaskExecutionException(result.result, result.traceback)
def _check_error(result):
if not result or not isinstance(result, AsyncResult):
return
check_or_raise_task_exception(result)
_check_error(result.parent)
def simple_result(result):
# DO not remove line below
# Explanation: https://github.com/celery/celery/issues/2315
deployer.celery.app.set_current()
if isinstance(result, GroupResult):
return simple_result(result.results)
elif hasattr(result, '__iter__') and not isinstance(result, dict):
return [simple_result(each_result)
for each_result in result]
elif isinstance(result, ResultBase):
_check_error(result)
if result.ready():
check_or_raise_task_exception(result)
return simple_result(result.result)
else:
raise TaskNotReadyException()
return result
class TaskNotReadyException(Exception):
pass
<commit_msg>Add retry for socket error<commit_after> | import socket
from celery.result import ResultBase, AsyncResult, GroupResult
import deployer
from deployer.tasks.exceptions import TaskExecutionException
from deployer.util import retry
__author__ = 'sukrit'
def check_or_raise_task_exception(result):
if isinstance(result, AsyncResult) and result.failed():
if isinstance(result.result, TaskExecutionException):
raise result.result
else:
raise TaskExecutionException(result.result, result.traceback)
def _check_error(result):
if not result or not isinstance(result, AsyncResult):
return
check_or_raise_task_exception(result)
_check_error(result.parent)
@retry(10, delay=5, backoff=1, except_on=(IOError, socket.error))
def simple_result(result):
# DO not remove line below
# Explanation: https://github.com/celery/celery/issues/2315
deployer.celery.app.set_current()
if isinstance(result, GroupResult):
return simple_result(result.results)
elif hasattr(result, '__iter__') and not isinstance(result, dict):
return [simple_result(each_result)
for each_result in result]
elif isinstance(result, ResultBase):
_check_error(result)
if result.ready():
check_or_raise_task_exception(result)
return simple_result(result.result)
else:
raise TaskNotReadyException()
return result
class TaskNotReadyException(Exception):
pass
| from celery.result import ResultBase, AsyncResult, GroupResult
import deployer
from deployer.tasks.exceptions import TaskExecutionException
__author__ = 'sukrit'
def check_or_raise_task_exception(result):
if isinstance(result, AsyncResult) and result.failed():
if isinstance(result.result, TaskExecutionException):
raise result.result
else:
raise TaskExecutionException(result.result, result.traceback)
def _check_error(result):
if not result or not isinstance(result, AsyncResult):
return
check_or_raise_task_exception(result)
_check_error(result.parent)
def simple_result(result):
# DO not remove line below
# Explanation: https://github.com/celery/celery/issues/2315
deployer.celery.app.set_current()
if isinstance(result, GroupResult):
return simple_result(result.results)
elif hasattr(result, '__iter__') and not isinstance(result, dict):
return [simple_result(each_result)
for each_result in result]
elif isinstance(result, ResultBase):
_check_error(result)
if result.ready():
check_or_raise_task_exception(result)
return simple_result(result.result)
else:
raise TaskNotReadyException()
return result
class TaskNotReadyException(Exception):
pass
Add retry for socket errorimport socket
from celery.result import ResultBase, AsyncResult, GroupResult
import deployer
from deployer.tasks.exceptions import TaskExecutionException
from deployer.util import retry
__author__ = 'sukrit'
def check_or_raise_task_exception(result):
if isinstance(result, AsyncResult) and result.failed():
if isinstance(result.result, TaskExecutionException):
raise result.result
else:
raise TaskExecutionException(result.result, result.traceback)
def _check_error(result):
if not result or not isinstance(result, AsyncResult):
return
check_or_raise_task_exception(result)
_check_error(result.parent)
@retry(10, delay=5, backoff=1, except_on=(IOError, socket.error))
def simple_result(result):
# DO not remove line below
# Explanation: https://github.com/celery/celery/issues/2315
deployer.celery.app.set_current()
if isinstance(result, GroupResult):
return simple_result(result.results)
elif hasattr(result, '__iter__') and not isinstance(result, dict):
return [simple_result(each_result)
for each_result in result]
elif isinstance(result, ResultBase):
_check_error(result)
if result.ready():
check_or_raise_task_exception(result)
return simple_result(result.result)
else:
raise TaskNotReadyException()
return result
class TaskNotReadyException(Exception):
pass
| <commit_before>from celery.result import ResultBase, AsyncResult, GroupResult
import deployer
from deployer.tasks.exceptions import TaskExecutionException
__author__ = 'sukrit'
def check_or_raise_task_exception(result):
if isinstance(result, AsyncResult) and result.failed():
if isinstance(result.result, TaskExecutionException):
raise result.result
else:
raise TaskExecutionException(result.result, result.traceback)
def _check_error(result):
if not result or not isinstance(result, AsyncResult):
return
check_or_raise_task_exception(result)
_check_error(result.parent)
def simple_result(result):
# DO not remove line below
# Explanation: https://github.com/celery/celery/issues/2315
deployer.celery.app.set_current()
if isinstance(result, GroupResult):
return simple_result(result.results)
elif hasattr(result, '__iter__') and not isinstance(result, dict):
return [simple_result(each_result)
for each_result in result]
elif isinstance(result, ResultBase):
_check_error(result)
if result.ready():
check_or_raise_task_exception(result)
return simple_result(result.result)
else:
raise TaskNotReadyException()
return result
class TaskNotReadyException(Exception):
pass
<commit_msg>Add retry for socket error<commit_after>import socket
from celery.result import ResultBase, AsyncResult, GroupResult
import deployer
from deployer.tasks.exceptions import TaskExecutionException
from deployer.util import retry
__author__ = 'sukrit'
def check_or_raise_task_exception(result):
if isinstance(result, AsyncResult) and result.failed():
if isinstance(result.result, TaskExecutionException):
raise result.result
else:
raise TaskExecutionException(result.result, result.traceback)
def _check_error(result):
if not result or not isinstance(result, AsyncResult):
return
check_or_raise_task_exception(result)
_check_error(result.parent)
@retry(10, delay=5, backoff=1, except_on=(IOError, socket.error))
def simple_result(result):
# DO not remove line below
# Explanation: https://github.com/celery/celery/issues/2315
deployer.celery.app.set_current()
if isinstance(result, GroupResult):
return simple_result(result.results)
elif hasattr(result, '__iter__') and not isinstance(result, dict):
return [simple_result(each_result)
for each_result in result]
elif isinstance(result, ResultBase):
_check_error(result)
if result.ready():
check_or_raise_task_exception(result)
return simple_result(result.result)
else:
raise TaskNotReadyException()
return result
class TaskNotReadyException(Exception):
pass
|
d0ea27a56013af944ef9e7fef9ebe1c8f44e3aab | community_blog/__openerp__.py | community_blog/__openerp__.py | # -*- coding: utf-8 -*-
##############################################################################
#
# Author: Yannick Buron
# Copyright 2013 Yannick Buron
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
{'name': 'OpenERP CommunityTools - Website Blog',
'version': '1.0',
'category': 'Community',
'depends': ['community',
'website_blog',
],
'author': 'Yannick Buron',
'license': 'AGPL-3',
'website': 'https://launchpad.net/openerp-communitytools',
'description': """
OpenERP for Communities - Blog
=================
""",
'data': ['security/community_blog_security.xml'],
'demo': [],
'installable': True,
'application': True,
}
| # -*- coding: utf-8 -*-
##############################################################################
#
# Author: Yannick Buron
# Copyright 2013 Yannick Buron
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
{'name': 'Odoo for Communities - Website Blog',
'version': '1.0',
'category': 'Community',
'depends': ['community',
'website_blog',
],
'author': 'Yannick Buron',
'license': 'AGPL-3',
'website': 'https://launchpad.net/openerp-communitytools',
'description': """
Odoo for Communities - Blog
=================
""",
'data': ['security/community_blog_security.xml'],
'demo': [],
'installable': True,
'application': True,
}
| Change name of module community_blog | Change name of module community_blog
Changement de nom | Python | agpl-3.0 | YannickB/vertical-community,Valeureux/wezer-exchange,Valeureux/wezer-exchange,codoo/vertical-community,open-synergy/vertical-community,Valeureux/wezer-exchange,Valeureux/wezer-exchange | # -*- coding: utf-8 -*-
##############################################################################
#
# Author: Yannick Buron
# Copyright 2013 Yannick Buron
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
{'name': 'OpenERP CommunityTools - Website Blog',
'version': '1.0',
'category': 'Community',
'depends': ['community',
'website_blog',
],
'author': 'Yannick Buron',
'license': 'AGPL-3',
'website': 'https://launchpad.net/openerp-communitytools',
'description': """
OpenERP for Communities - Blog
=================
""",
'data': ['security/community_blog_security.xml'],
'demo': [],
'installable': True,
'application': True,
}
Change name of module community_blog
Changement de nom | # -*- coding: utf-8 -*-
##############################################################################
#
# Author: Yannick Buron
# Copyright 2013 Yannick Buron
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
{'name': 'Odoo for Communities - Website Blog',
'version': '1.0',
'category': 'Community',
'depends': ['community',
'website_blog',
],
'author': 'Yannick Buron',
'license': 'AGPL-3',
'website': 'https://launchpad.net/openerp-communitytools',
'description': """
Odoo for Communities - Blog
=================
""",
'data': ['security/community_blog_security.xml'],
'demo': [],
'installable': True,
'application': True,
}
| <commit_before># -*- coding: utf-8 -*-
##############################################################################
#
# Author: Yannick Buron
# Copyright 2013 Yannick Buron
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
{'name': 'OpenERP CommunityTools - Website Blog',
'version': '1.0',
'category': 'Community',
'depends': ['community',
'website_blog',
],
'author': 'Yannick Buron',
'license': 'AGPL-3',
'website': 'https://launchpad.net/openerp-communitytools',
'description': """
OpenERP for Communities - Blog
=================
""",
'data': ['security/community_blog_security.xml'],
'demo': [],
'installable': True,
'application': True,
}
<commit_msg>Change name of module community_blog
Changement de nom<commit_after> | # -*- coding: utf-8 -*-
##############################################################################
#
# Author: Yannick Buron
# Copyright 2013 Yannick Buron
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
{'name': 'Odoo for Communities - Website Blog',
'version': '1.0',
'category': 'Community',
'depends': ['community',
'website_blog',
],
'author': 'Yannick Buron',
'license': 'AGPL-3',
'website': 'https://launchpad.net/openerp-communitytools',
'description': """
Odoo for Communities - Blog
=================
""",
'data': ['security/community_blog_security.xml'],
'demo': [],
'installable': True,
'application': True,
}
| # -*- coding: utf-8 -*-
##############################################################################
#
# Author: Yannick Buron
# Copyright 2013 Yannick Buron
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
{'name': 'OpenERP CommunityTools - Website Blog',
'version': '1.0',
'category': 'Community',
'depends': ['community',
'website_blog',
],
'author': 'Yannick Buron',
'license': 'AGPL-3',
'website': 'https://launchpad.net/openerp-communitytools',
'description': """
OpenERP for Communities - Blog
=================
""",
'data': ['security/community_blog_security.xml'],
'demo': [],
'installable': True,
'application': True,
}
Change name of module community_blog
Changement de nom# -*- coding: utf-8 -*-
##############################################################################
#
# Author: Yannick Buron
# Copyright 2013 Yannick Buron
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
{'name': 'Odoo for Communities - Website Blog',
'version': '1.0',
'category': 'Community',
'depends': ['community',
'website_blog',
],
'author': 'Yannick Buron',
'license': 'AGPL-3',
'website': 'https://launchpad.net/openerp-communitytools',
'description': """
Odoo for Communities - Blog
=================
""",
'data': ['security/community_blog_security.xml'],
'demo': [],
'installable': True,
'application': True,
}
| <commit_before># -*- coding: utf-8 -*-
##############################################################################
#
# Author: Yannick Buron
# Copyright 2013 Yannick Buron
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
{'name': 'OpenERP CommunityTools - Website Blog',
'version': '1.0',
'category': 'Community',
'depends': ['community',
'website_blog',
],
'author': 'Yannick Buron',
'license': 'AGPL-3',
'website': 'https://launchpad.net/openerp-communitytools',
'description': """
OpenERP for Communities - Blog
=================
""",
'data': ['security/community_blog_security.xml'],
'demo': [],
'installable': True,
'application': True,
}
<commit_msg>Change name of module community_blog
Changement de nom<commit_after># -*- coding: utf-8 -*-
##############################################################################
#
# Author: Yannick Buron
# Copyright 2013 Yannick Buron
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
{'name': 'Odoo for Communities - Website Blog',
'version': '1.0',
'category': 'Community',
'depends': ['community',
'website_blog',
],
'author': 'Yannick Buron',
'license': 'AGPL-3',
'website': 'https://launchpad.net/openerp-communitytools',
'description': """
Odoo for Communities - Blog
=================
""",
'data': ['security/community_blog_security.xml'],
'demo': [],
'installable': True,
'application': True,
}
|
b7bad7823384ec5261271e3f54ed272775a7562f | sparqllib/formatter.py | sparqllib/formatter.py | import abc
import re
class Formatter:
@abc.abstractmethod
def format(self, query):
''' Should return a human-readable version of the query string
'''
pass
class BasicFormatter(Formatter):
''' Provides a basic default formatting for query strings
This formatter provides only indentation levels and newlines at
open braces.
'''
def __init__(self):
self.indent_str = " "
def format(self, query):
#TODO handle braces inside literals correctly
formatted_query = ""
indent_level = 0
for letter in query:
# newline and reindent on open brace
if letter == "{":
indent_level += 1
formatted_query += "{\n" + self.indent_str*indent_level
# newline and reindent on close brace
elif letter == "}":
indent_level -= 1
formatted_query += "\n" + self.indent_str*indent_level + "}"
# reindent after any newline
elif len(formatted_query) and formatted_query[-1] == '\n':
formatted_query += self.indent_str*indent_level + letter
# otherwise just add the letter
else:
formatted_query += letter
# remove duplicate newlines
formatted_query = re.sub(r'(\n+)', '\n', formatted_query, flags=re.MULTILINE)
return formatted_query
| import abc
import re
class Formatter:
@abc.abstractmethod
def format(self, query):
''' Should return a human-readable version of the query string
'''
pass
class BasicFormatter(Formatter):
''' Provides a basic default formatting for query strings
This formatter provides only indentation levels and newlines at
open braces.
'''
def __init__(self):
self.indent_str = " "
def format(self, query):
if not isinstance(query, str):
query = query.serialize()
#TODO handle braces inside literals correctly
formatted_query = ""
indent_level = 0
for letter in query:
# newline and reindent on open brace
if letter == "{":
indent_level += 1
formatted_query += "{\n" + self.indent_str*indent_level
# newline and reindent on close brace
elif letter == "}":
indent_level -= 1
formatted_query += "\n" + self.indent_str*indent_level + "}"
# reindent after any newline
elif len(formatted_query) and formatted_query[-1] == '\n':
formatted_query += self.indent_str*indent_level + letter
# otherwise just add the letter
else:
formatted_query += letter
# trim whitespace
formatted_query = re.sub(r'(.)\s+\n', '\g<1>\n', formatted_query, flags=re.MULTILINE)
# remove duplicate newlines
formatted_query = re.sub(r'(\n+)', '\n', formatted_query, flags=re.MULTILINE)
return formatted_query
| Trim trailing whitespace with BasicFormatter | Trim trailing whitespace with BasicFormatter
| Python | mit | ALSchwalm/sparqllib | import abc
import re
class Formatter:
@abc.abstractmethod
def format(self, query):
''' Should return a human-readable version of the query string
'''
pass
class BasicFormatter(Formatter):
''' Provides a basic default formatting for query strings
This formatter provides only indentation levels and newlines at
open braces.
'''
def __init__(self):
self.indent_str = " "
def format(self, query):
#TODO handle braces inside literals correctly
formatted_query = ""
indent_level = 0
for letter in query:
# newline and reindent on open brace
if letter == "{":
indent_level += 1
formatted_query += "{\n" + self.indent_str*indent_level
# newline and reindent on close brace
elif letter == "}":
indent_level -= 1
formatted_query += "\n" + self.indent_str*indent_level + "}"
# reindent after any newline
elif len(formatted_query) and formatted_query[-1] == '\n':
formatted_query += self.indent_str*indent_level + letter
# otherwise just add the letter
else:
formatted_query += letter
# remove duplicate newlines
formatted_query = re.sub(r'(\n+)', '\n', formatted_query, flags=re.MULTILINE)
return formatted_query
Trim trailing whitespace with BasicFormatter | import abc
import re
class Formatter:
@abc.abstractmethod
def format(self, query):
''' Should return a human-readable version of the query string
'''
pass
class BasicFormatter(Formatter):
''' Provides a basic default formatting for query strings
This formatter provides only indentation levels and newlines at
open braces.
'''
def __init__(self):
self.indent_str = " "
def format(self, query):
if not isinstance(query, str):
query = query.serialize()
#TODO handle braces inside literals correctly
formatted_query = ""
indent_level = 0
for letter in query:
# newline and reindent on open brace
if letter == "{":
indent_level += 1
formatted_query += "{\n" + self.indent_str*indent_level
# newline and reindent on close brace
elif letter == "}":
indent_level -= 1
formatted_query += "\n" + self.indent_str*indent_level + "}"
# reindent after any newline
elif len(formatted_query) and formatted_query[-1] == '\n':
formatted_query += self.indent_str*indent_level + letter
# otherwise just add the letter
else:
formatted_query += letter
# trim whitespace
formatted_query = re.sub(r'(.)\s+\n', '\g<1>\n', formatted_query, flags=re.MULTILINE)
# remove duplicate newlines
formatted_query = re.sub(r'(\n+)', '\n', formatted_query, flags=re.MULTILINE)
return formatted_query
| <commit_before>import abc
import re
class Formatter:
@abc.abstractmethod
def format(self, query):
''' Should return a human-readable version of the query string
'''
pass
class BasicFormatter(Formatter):
''' Provides a basic default formatting for query strings
This formatter provides only indentation levels and newlines at
open braces.
'''
def __init__(self):
self.indent_str = " "
def format(self, query):
#TODO handle braces inside literals correctly
formatted_query = ""
indent_level = 0
for letter in query:
# newline and reindent on open brace
if letter == "{":
indent_level += 1
formatted_query += "{\n" + self.indent_str*indent_level
# newline and reindent on close brace
elif letter == "}":
indent_level -= 1
formatted_query += "\n" + self.indent_str*indent_level + "}"
# reindent after any newline
elif len(formatted_query) and formatted_query[-1] == '\n':
formatted_query += self.indent_str*indent_level + letter
# otherwise just add the letter
else:
formatted_query += letter
# remove duplicate newlines
formatted_query = re.sub(r'(\n+)', '\n', formatted_query, flags=re.MULTILINE)
return formatted_query
<commit_msg>Trim trailing whitespace with BasicFormatter<commit_after> | import abc
import re
class Formatter:
@abc.abstractmethod
def format(self, query):
''' Should return a human-readable version of the query string
'''
pass
class BasicFormatter(Formatter):
''' Provides a basic default formatting for query strings
This formatter provides only indentation levels and newlines at
open braces.
'''
def __init__(self):
self.indent_str = " "
def format(self, query):
if not isinstance(query, str):
query = query.serialize()
#TODO handle braces inside literals correctly
formatted_query = ""
indent_level = 0
for letter in query:
# newline and reindent on open brace
if letter == "{":
indent_level += 1
formatted_query += "{\n" + self.indent_str*indent_level
# newline and reindent on close brace
elif letter == "}":
indent_level -= 1
formatted_query += "\n" + self.indent_str*indent_level + "}"
# reindent after any newline
elif len(formatted_query) and formatted_query[-1] == '\n':
formatted_query += self.indent_str*indent_level + letter
# otherwise just add the letter
else:
formatted_query += letter
# trim whitespace
formatted_query = re.sub(r'(.)\s+\n', '\g<1>\n', formatted_query, flags=re.MULTILINE)
# remove duplicate newlines
formatted_query = re.sub(r'(\n+)', '\n', formatted_query, flags=re.MULTILINE)
return formatted_query
| import abc
import re
class Formatter:
@abc.abstractmethod
def format(self, query):
''' Should return a human-readable version of the query string
'''
pass
class BasicFormatter(Formatter):
''' Provides a basic default formatting for query strings
This formatter provides only indentation levels and newlines at
open braces.
'''
def __init__(self):
self.indent_str = " "
def format(self, query):
#TODO handle braces inside literals correctly
formatted_query = ""
indent_level = 0
for letter in query:
# newline and reindent on open brace
if letter == "{":
indent_level += 1
formatted_query += "{\n" + self.indent_str*indent_level
# newline and reindent on close brace
elif letter == "}":
indent_level -= 1
formatted_query += "\n" + self.indent_str*indent_level + "}"
# reindent after any newline
elif len(formatted_query) and formatted_query[-1] == '\n':
formatted_query += self.indent_str*indent_level + letter
# otherwise just add the letter
else:
formatted_query += letter
# remove duplicate newlines
formatted_query = re.sub(r'(\n+)', '\n', formatted_query, flags=re.MULTILINE)
return formatted_query
Trim trailing whitespace with BasicFormatterimport abc
import re
class Formatter:
@abc.abstractmethod
def format(self, query):
''' Should return a human-readable version of the query string
'''
pass
class BasicFormatter(Formatter):
''' Provides a basic default formatting for query strings
This formatter provides only indentation levels and newlines at
open braces.
'''
def __init__(self):
self.indent_str = " "
def format(self, query):
if not isinstance(query, str):
query = query.serialize()
#TODO handle braces inside literals correctly
formatted_query = ""
indent_level = 0
for letter in query:
# newline and reindent on open brace
if letter == "{":
indent_level += 1
formatted_query += "{\n" + self.indent_str*indent_level
# newline and reindent on close brace
elif letter == "}":
indent_level -= 1
formatted_query += "\n" + self.indent_str*indent_level + "}"
# reindent after any newline
elif len(formatted_query) and formatted_query[-1] == '\n':
formatted_query += self.indent_str*indent_level + letter
# otherwise just add the letter
else:
formatted_query += letter
# trim whitespace
formatted_query = re.sub(r'(.)\s+\n', '\g<1>\n', formatted_query, flags=re.MULTILINE)
# remove duplicate newlines
formatted_query = re.sub(r'(\n+)', '\n', formatted_query, flags=re.MULTILINE)
return formatted_query
| <commit_before>import abc
import re
class Formatter:
@abc.abstractmethod
def format(self, query):
''' Should return a human-readable version of the query string
'''
pass
class BasicFormatter(Formatter):
''' Provides a basic default formatting for query strings
This formatter provides only indentation levels and newlines at
open braces.
'''
def __init__(self):
self.indent_str = " "
def format(self, query):
#TODO handle braces inside literals correctly
formatted_query = ""
indent_level = 0
for letter in query:
# newline and reindent on open brace
if letter == "{":
indent_level += 1
formatted_query += "{\n" + self.indent_str*indent_level
# newline and reindent on close brace
elif letter == "}":
indent_level -= 1
formatted_query += "\n" + self.indent_str*indent_level + "}"
# reindent after any newline
elif len(formatted_query) and formatted_query[-1] == '\n':
formatted_query += self.indent_str*indent_level + letter
# otherwise just add the letter
else:
formatted_query += letter
# remove duplicate newlines
formatted_query = re.sub(r'(\n+)', '\n', formatted_query, flags=re.MULTILINE)
return formatted_query
<commit_msg>Trim trailing whitespace with BasicFormatter<commit_after>import abc
import re
class Formatter:
@abc.abstractmethod
def format(self, query):
''' Should return a human-readable version of the query string
'''
pass
class BasicFormatter(Formatter):
''' Provides a basic default formatting for query strings
This formatter provides only indentation levels and newlines at
open braces.
'''
def __init__(self):
self.indent_str = " "
def format(self, query):
if not isinstance(query, str):
query = query.serialize()
#TODO handle braces inside literals correctly
formatted_query = ""
indent_level = 0
for letter in query:
# newline and reindent on open brace
if letter == "{":
indent_level += 1
formatted_query += "{\n" + self.indent_str*indent_level
# newline and reindent on close brace
elif letter == "}":
indent_level -= 1
formatted_query += "\n" + self.indent_str*indent_level + "}"
# reindent after any newline
elif len(formatted_query) and formatted_query[-1] == '\n':
formatted_query += self.indent_str*indent_level + letter
# otherwise just add the letter
else:
formatted_query += letter
# trim whitespace
formatted_query = re.sub(r'(.)\s+\n', '\g<1>\n', formatted_query, flags=re.MULTILINE)
# remove duplicate newlines
formatted_query = re.sub(r'(\n+)', '\n', formatted_query, flags=re.MULTILINE)
return formatted_query
|
a7f467589c49020977328e45eed4eff5b607231f | checker/tests/downstream/test_check_files_menu_agreements.py | checker/tests/downstream/test_check_files_menu_agreements.py | import magic
import os.path as op
from checker.base import BakeryTestCase as TestCase, tags
from checker.metadata import Metadata
class CheckFontsMenuAgreements(TestCase):
path = '.'
name = __name__
targets = ['metadata']
tool = 'lint'
def menufile(self, font_metadata):
return '%s.menu' % font_metadata.post_script_name
@tags('required')
def test_menu_file_agreement(self):
""" Menu file have font-name-style.menu format """
contents = self.read_metadata_contents()
fm = Metadata.get_family_metadata(contents)
for font_metadata in fm.fonts:
menufile = self.menufile(font_metadata)
path = op.join(op.dirname(self.path), menufile)
if not op.exists(path):
self.fail('%s does not exist' % menufile)
if magic.from_file("%s.menu" % self.fname) != 'TrueType font data':
self.fail('%s is not actual TTF file' % menufile)
| import magic
import os.path as op
from checker.base import BakeryTestCase as TestCase, tags
from checker.metadata import Metadata
class CheckFontsMenuAgreements(TestCase):
path = '.'
name = __name__
targets = ['metadata']
tool = 'lint'
def read_metadata_contents(self):
return open(self.path).read()
def menufile(self, font_metadata):
return '%s.menu' % font_metadata.post_script_name
@tags('required')
def test_menu_file_agreement(self):
""" Menu file have font-name-style.menu format """
contents = self.read_metadata_contents()
fm = Metadata.get_family_metadata(contents)
for font_metadata in fm.fonts:
menufile = self.menufile(font_metadata)
path = op.join(op.dirname(self.path), menufile)
if not op.exists(path):
self.fail('%s does not exist' % menufile)
if magic.from_file("%s.menu" % self.fname) != 'TrueType font data':
self.fail('%s is not actual TTF file' % menufile)
| Fix check menu files agreement test | Fix check menu files agreement test
| Python | apache-2.0 | davelab6/fontbakery,googlefonts/fontbakery,moyogo/fontbakery,moyogo/fontbakery,graphicore/fontbakery,jessamynsmith/fontbakery,graphicore/fontbakery,googlefonts/fontbakery,graphicore/fontbakery,moyogo/fontbakery,googlefonts/fontbakery | import magic
import os.path as op
from checker.base import BakeryTestCase as TestCase, tags
from checker.metadata import Metadata
class CheckFontsMenuAgreements(TestCase):
path = '.'
name = __name__
targets = ['metadata']
tool = 'lint'
def menufile(self, font_metadata):
return '%s.menu' % font_metadata.post_script_name
@tags('required')
def test_menu_file_agreement(self):
""" Menu file have font-name-style.menu format """
contents = self.read_metadata_contents()
fm = Metadata.get_family_metadata(contents)
for font_metadata in fm.fonts:
menufile = self.menufile(font_metadata)
path = op.join(op.dirname(self.path), menufile)
if not op.exists(path):
self.fail('%s does not exist' % menufile)
if magic.from_file("%s.menu" % self.fname) != 'TrueType font data':
self.fail('%s is not actual TTF file' % menufile)
Fix check menu files agreement test | import magic
import os.path as op
from checker.base import BakeryTestCase as TestCase, tags
from checker.metadata import Metadata
class CheckFontsMenuAgreements(TestCase):
path = '.'
name = __name__
targets = ['metadata']
tool = 'lint'
def read_metadata_contents(self):
return open(self.path).read()
def menufile(self, font_metadata):
return '%s.menu' % font_metadata.post_script_name
@tags('required')
def test_menu_file_agreement(self):
""" Menu file have font-name-style.menu format """
contents = self.read_metadata_contents()
fm = Metadata.get_family_metadata(contents)
for font_metadata in fm.fonts:
menufile = self.menufile(font_metadata)
path = op.join(op.dirname(self.path), menufile)
if not op.exists(path):
self.fail('%s does not exist' % menufile)
if magic.from_file("%s.menu" % self.fname) != 'TrueType font data':
self.fail('%s is not actual TTF file' % menufile)
| <commit_before>import magic
import os.path as op
from checker.base import BakeryTestCase as TestCase, tags
from checker.metadata import Metadata
class CheckFontsMenuAgreements(TestCase):
path = '.'
name = __name__
targets = ['metadata']
tool = 'lint'
def menufile(self, font_metadata):
return '%s.menu' % font_metadata.post_script_name
@tags('required')
def test_menu_file_agreement(self):
""" Menu file have font-name-style.menu format """
contents = self.read_metadata_contents()
fm = Metadata.get_family_metadata(contents)
for font_metadata in fm.fonts:
menufile = self.menufile(font_metadata)
path = op.join(op.dirname(self.path), menufile)
if not op.exists(path):
self.fail('%s does not exist' % menufile)
if magic.from_file("%s.menu" % self.fname) != 'TrueType font data':
self.fail('%s is not actual TTF file' % menufile)
<commit_msg>Fix check menu files agreement test<commit_after> | import magic
import os.path as op
from checker.base import BakeryTestCase as TestCase, tags
from checker.metadata import Metadata
class CheckFontsMenuAgreements(TestCase):
path = '.'
name = __name__
targets = ['metadata']
tool = 'lint'
def read_metadata_contents(self):
return open(self.path).read()
def menufile(self, font_metadata):
return '%s.menu' % font_metadata.post_script_name
@tags('required')
def test_menu_file_agreement(self):
""" Menu file have font-name-style.menu format """
contents = self.read_metadata_contents()
fm = Metadata.get_family_metadata(contents)
for font_metadata in fm.fonts:
menufile = self.menufile(font_metadata)
path = op.join(op.dirname(self.path), menufile)
if not op.exists(path):
self.fail('%s does not exist' % menufile)
if magic.from_file("%s.menu" % self.fname) != 'TrueType font data':
self.fail('%s is not actual TTF file' % menufile)
| import magic
import os.path as op
from checker.base import BakeryTestCase as TestCase, tags
from checker.metadata import Metadata
class CheckFontsMenuAgreements(TestCase):
path = '.'
name = __name__
targets = ['metadata']
tool = 'lint'
def menufile(self, font_metadata):
return '%s.menu' % font_metadata.post_script_name
@tags('required')
def test_menu_file_agreement(self):
""" Menu file have font-name-style.menu format """
contents = self.read_metadata_contents()
fm = Metadata.get_family_metadata(contents)
for font_metadata in fm.fonts:
menufile = self.menufile(font_metadata)
path = op.join(op.dirname(self.path), menufile)
if not op.exists(path):
self.fail('%s does not exist' % menufile)
if magic.from_file("%s.menu" % self.fname) != 'TrueType font data':
self.fail('%s is not actual TTF file' % menufile)
Fix check menu files agreement testimport magic
import os.path as op
from checker.base import BakeryTestCase as TestCase, tags
from checker.metadata import Metadata
class CheckFontsMenuAgreements(TestCase):
path = '.'
name = __name__
targets = ['metadata']
tool = 'lint'
def read_metadata_contents(self):
return open(self.path).read()
def menufile(self, font_metadata):
return '%s.menu' % font_metadata.post_script_name
@tags('required')
def test_menu_file_agreement(self):
""" Menu file have font-name-style.menu format """
contents = self.read_metadata_contents()
fm = Metadata.get_family_metadata(contents)
for font_metadata in fm.fonts:
menufile = self.menufile(font_metadata)
path = op.join(op.dirname(self.path), menufile)
if not op.exists(path):
self.fail('%s does not exist' % menufile)
if magic.from_file("%s.menu" % self.fname) != 'TrueType font data':
self.fail('%s is not actual TTF file' % menufile)
| <commit_before>import magic
import os.path as op
from checker.base import BakeryTestCase as TestCase, tags
from checker.metadata import Metadata
class CheckFontsMenuAgreements(TestCase):
path = '.'
name = __name__
targets = ['metadata']
tool = 'lint'
def menufile(self, font_metadata):
return '%s.menu' % font_metadata.post_script_name
@tags('required')
def test_menu_file_agreement(self):
""" Menu file have font-name-style.menu format """
contents = self.read_metadata_contents()
fm = Metadata.get_family_metadata(contents)
for font_metadata in fm.fonts:
menufile = self.menufile(font_metadata)
path = op.join(op.dirname(self.path), menufile)
if not op.exists(path):
self.fail('%s does not exist' % menufile)
if magic.from_file("%s.menu" % self.fname) != 'TrueType font data':
self.fail('%s is not actual TTF file' % menufile)
<commit_msg>Fix check menu files agreement test<commit_after>import magic
import os.path as op
from checker.base import BakeryTestCase as TestCase, tags
from checker.metadata import Metadata
class CheckFontsMenuAgreements(TestCase):
path = '.'
name = __name__
targets = ['metadata']
tool = 'lint'
def read_metadata_contents(self):
return open(self.path).read()
def menufile(self, font_metadata):
return '%s.menu' % font_metadata.post_script_name
@tags('required')
def test_menu_file_agreement(self):
""" Menu file have font-name-style.menu format """
contents = self.read_metadata_contents()
fm = Metadata.get_family_metadata(contents)
for font_metadata in fm.fonts:
menufile = self.menufile(font_metadata)
path = op.join(op.dirname(self.path), menufile)
if not op.exists(path):
self.fail('%s does not exist' % menufile)
if magic.from_file("%s.menu" % self.fname) != 'TrueType font data':
self.fail('%s is not actual TTF file' % menufile)
|
4fc108a39476f92acf0d42b66466012cba868b1d | h2o-py/tests/testdir_algos/rf/pyunit_vi_toy_testRF.py | h2o-py/tests/testdir_algos/rf/pyunit_vi_toy_testRF.py | import sys
sys.path.insert(1, "../../../")
import h2o
def vi_toy_test(ip,port):
# Connect to h2o
h2o.init(ip,port)
toy_data = h2o.import_frame(path=h2o.locate("smalldata/gbm_test/toy_data_RF.csv"))
#toy_data.summary()
toy_data[6] = toy_data[6].asfactor()
toy_data.show()
rf = h2o.random_forest(x=toy_data[[0,1,2,3,4,5]], y=toy_data[6], ntrees=500, max_depth=20, nbins=100)
ranking = [rf._model_json['output']['variable_importances'].cell_values[v][0] for v in range(toy_data.ncol()-1)]
print(ranking)
assert tuple(ranking) == tuple(["V3","V2","V6","V1","V5","V4"]), "expected specific variable importance ranking"
if __name__ == "__main__":
h2o.run_test(sys.argv, vi_toy_test)
| import sys
sys.path.insert(1, "../../../")
import h2o
def vi_toy_test(ip,port):
# Connect to h2o
h2o.init(ip,port)
toy_data = h2o.import_frame(path=h2o.locate("smalldata/gbm_test/toy_data_RF.csv"))
#toy_data.summary()
toy_data[6] = toy_data[6].asfactor()
toy_data.show()
rf = h2o.random_forest(x=toy_data[[0,1,2,3,4,5]], y=toy_data[6], ntrees=500, max_depth=20, nbins=100, seed=0)
ranking = [rf._model_json['output']['variable_importances'].cell_values[v][0] for v in range(toy_data.ncol()-1)]
print(ranking)
assert tuple(ranking) == tuple(["V3","V2","V6","V1","V5","V4"]), "expected specific variable importance ranking"
if __name__ == "__main__":
h2o.run_test(sys.argv, vi_toy_test)
| Fix the seed for RF test. | Fix the seed for RF test.
| Python | apache-2.0 | ChristosChristofidis/h2o-3,YzPaul3/h2o-3,h2oai/h2o-3,bospetersen/h2o-3,spennihana/h2o-3,mathemage/h2o-3,weaver-viii/h2o-3,h2oai/h2o-dev,michalkurka/h2o-3,h2oai/h2o-3,mathemage/h2o-3,PawarPawan/h2o-v3,michalkurka/h2o-3,datachand/h2o-3,PawarPawan/h2o-v3,printedheart/h2o-3,kyoren/https-github.com-h2oai-h2o-3,mrgloom/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,printedheart/h2o-3,h2oai/h2o-dev,YzPaul3/h2o-3,tarasane/h2o-3,pchmieli/h2o-3,YzPaul3/h2o-3,h2oai/h2o-3,tarasane/h2o-3,PawarPawan/h2o-v3,printedheart/h2o-3,jangorecki/h2o-3,mrgloom/h2o-3,h2oai/h2o-dev,kyoren/https-github.com-h2oai-h2o-3,nilbody/h2o-3,pchmieli/h2o-3,h2oai/h2o-3,PawarPawan/h2o-v3,kyoren/https-github.com-h2oai-h2o-3,mathemage/h2o-3,tarasane/h2o-3,ChristosChristofidis/h2o-3,michalkurka/h2o-3,spennihana/h2o-3,nilbody/h2o-3,nilbody/h2o-3,weaver-viii/h2o-3,bospetersen/h2o-3,pchmieli/h2o-3,junwucs/h2o-3,YzPaul3/h2o-3,junwucs/h2o-3,spennihana/h2o-3,spennihana/h2o-3,spennihana/h2o-3,YzPaul3/h2o-3,weaver-viii/h2o-3,kyoren/https-github.com-h2oai-h2o-3,mathemage/h2o-3,tarasane/h2o-3,pchmieli/h2o-3,nilbody/h2o-3,brightchen/h2o-3,tarasane/h2o-3,mrgloom/h2o-3,mrgloom/h2o-3,jangorecki/h2o-3,datachand/h2o-3,nilbody/h2o-3,madmax983/h2o-3,h2oai/h2o-3,h2oai/h2o-dev,ChristosChristofidis/h2o-3,ChristosChristofidis/h2o-3,h2oai/h2o-dev,bospetersen/h2o-3,ChristosChristofidis/h2o-3,brightchen/h2o-3,brightchen/h2o-3,madmax983/h2o-3,datachand/h2o-3,PawarPawan/h2o-v3,bospetersen/h2o-3,datachand/h2o-3,PawarPawan/h2o-v3,kyoren/https-github.com-h2oai-h2o-3,nilbody/h2o-3,michalkurka/h2o-3,mathemage/h2o-3,YzPaul3/h2o-3,brightchen/h2o-3,printedheart/h2o-3,bospetersen/h2o-3,weaver-viii/h2o-3,junwucs/h2o-3,jangorecki/h2o-3,printedheart/h2o-3,mathemage/h2o-3,pchmieli/h2o-3,jangorecki/h2o-3,nilbody/h2o-3,kyoren/https-github.com-h2oai-h2o-3,weaver-viii/h2o-3,madmax983/h2o-3,mrgloom/h2o-3,h2oai/h2o-dev,YzPaul3/h2o-3,brightchen/h2o-3,tarasane/h2o-3,jangorecki/h2o-3,pchmieli/h2o-3,h2oai/h2o-3,mathemage/h2o-3,printedheart/h2o-3,datachand/h2o-3,PawarPawan/h2o-v3,pchmieli/h2o-3,junwucs/h2o-3,h2oai/h2o-dev,printedheart/h2o-3,madmax983/h2o-3,mrgloom/h2o-3,jangorecki/h2o-3,datachand/h2o-3,weaver-viii/h2o-3,junwucs/h2o-3,datachand/h2o-3,jangorecki/h2o-3,spennihana/h2o-3,ChristosChristofidis/h2o-3,junwucs/h2o-3,ChristosChristofidis/h2o-3,bospetersen/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,madmax983/h2o-3,junwucs/h2o-3,michalkurka/h2o-3,bospetersen/h2o-3,mrgloom/h2o-3,spennihana/h2o-3,brightchen/h2o-3,madmax983/h2o-3,madmax983/h2o-3,kyoren/https-github.com-h2oai-h2o-3,weaver-viii/h2o-3,tarasane/h2o-3,brightchen/h2o-3 | import sys
sys.path.insert(1, "../../../")
import h2o
def vi_toy_test(ip,port):
# Connect to h2o
h2o.init(ip,port)
toy_data = h2o.import_frame(path=h2o.locate("smalldata/gbm_test/toy_data_RF.csv"))
#toy_data.summary()
toy_data[6] = toy_data[6].asfactor()
toy_data.show()
rf = h2o.random_forest(x=toy_data[[0,1,2,3,4,5]], y=toy_data[6], ntrees=500, max_depth=20, nbins=100)
ranking = [rf._model_json['output']['variable_importances'].cell_values[v][0] for v in range(toy_data.ncol()-1)]
print(ranking)
assert tuple(ranking) == tuple(["V3","V2","V6","V1","V5","V4"]), "expected specific variable importance ranking"
if __name__ == "__main__":
h2o.run_test(sys.argv, vi_toy_test)
Fix the seed for RF test. | import sys
sys.path.insert(1, "../../../")
import h2o
def vi_toy_test(ip,port):
# Connect to h2o
h2o.init(ip,port)
toy_data = h2o.import_frame(path=h2o.locate("smalldata/gbm_test/toy_data_RF.csv"))
#toy_data.summary()
toy_data[6] = toy_data[6].asfactor()
toy_data.show()
rf = h2o.random_forest(x=toy_data[[0,1,2,3,4,5]], y=toy_data[6], ntrees=500, max_depth=20, nbins=100, seed=0)
ranking = [rf._model_json['output']['variable_importances'].cell_values[v][0] for v in range(toy_data.ncol()-1)]
print(ranking)
assert tuple(ranking) == tuple(["V3","V2","V6","V1","V5","V4"]), "expected specific variable importance ranking"
if __name__ == "__main__":
h2o.run_test(sys.argv, vi_toy_test)
| <commit_before>import sys
sys.path.insert(1, "../../../")
import h2o
def vi_toy_test(ip,port):
# Connect to h2o
h2o.init(ip,port)
toy_data = h2o.import_frame(path=h2o.locate("smalldata/gbm_test/toy_data_RF.csv"))
#toy_data.summary()
toy_data[6] = toy_data[6].asfactor()
toy_data.show()
rf = h2o.random_forest(x=toy_data[[0,1,2,3,4,5]], y=toy_data[6], ntrees=500, max_depth=20, nbins=100)
ranking = [rf._model_json['output']['variable_importances'].cell_values[v][0] for v in range(toy_data.ncol()-1)]
print(ranking)
assert tuple(ranking) == tuple(["V3","V2","V6","V1","V5","V4"]), "expected specific variable importance ranking"
if __name__ == "__main__":
h2o.run_test(sys.argv, vi_toy_test)
<commit_msg>Fix the seed for RF test.<commit_after> | import sys
sys.path.insert(1, "../../../")
import h2o
def vi_toy_test(ip,port):
# Connect to h2o
h2o.init(ip,port)
toy_data = h2o.import_frame(path=h2o.locate("smalldata/gbm_test/toy_data_RF.csv"))
#toy_data.summary()
toy_data[6] = toy_data[6].asfactor()
toy_data.show()
rf = h2o.random_forest(x=toy_data[[0,1,2,3,4,5]], y=toy_data[6], ntrees=500, max_depth=20, nbins=100, seed=0)
ranking = [rf._model_json['output']['variable_importances'].cell_values[v][0] for v in range(toy_data.ncol()-1)]
print(ranking)
assert tuple(ranking) == tuple(["V3","V2","V6","V1","V5","V4"]), "expected specific variable importance ranking"
if __name__ == "__main__":
h2o.run_test(sys.argv, vi_toy_test)
| import sys
sys.path.insert(1, "../../../")
import h2o
def vi_toy_test(ip,port):
# Connect to h2o
h2o.init(ip,port)
toy_data = h2o.import_frame(path=h2o.locate("smalldata/gbm_test/toy_data_RF.csv"))
#toy_data.summary()
toy_data[6] = toy_data[6].asfactor()
toy_data.show()
rf = h2o.random_forest(x=toy_data[[0,1,2,3,4,5]], y=toy_data[6], ntrees=500, max_depth=20, nbins=100)
ranking = [rf._model_json['output']['variable_importances'].cell_values[v][0] for v in range(toy_data.ncol()-1)]
print(ranking)
assert tuple(ranking) == tuple(["V3","V2","V6","V1","V5","V4"]), "expected specific variable importance ranking"
if __name__ == "__main__":
h2o.run_test(sys.argv, vi_toy_test)
Fix the seed for RF test.import sys
sys.path.insert(1, "../../../")
import h2o
def vi_toy_test(ip,port):
# Connect to h2o
h2o.init(ip,port)
toy_data = h2o.import_frame(path=h2o.locate("smalldata/gbm_test/toy_data_RF.csv"))
#toy_data.summary()
toy_data[6] = toy_data[6].asfactor()
toy_data.show()
rf = h2o.random_forest(x=toy_data[[0,1,2,3,4,5]], y=toy_data[6], ntrees=500, max_depth=20, nbins=100, seed=0)
ranking = [rf._model_json['output']['variable_importances'].cell_values[v][0] for v in range(toy_data.ncol()-1)]
print(ranking)
assert tuple(ranking) == tuple(["V3","V2","V6","V1","V5","V4"]), "expected specific variable importance ranking"
if __name__ == "__main__":
h2o.run_test(sys.argv, vi_toy_test)
| <commit_before>import sys
sys.path.insert(1, "../../../")
import h2o
def vi_toy_test(ip,port):
# Connect to h2o
h2o.init(ip,port)
toy_data = h2o.import_frame(path=h2o.locate("smalldata/gbm_test/toy_data_RF.csv"))
#toy_data.summary()
toy_data[6] = toy_data[6].asfactor()
toy_data.show()
rf = h2o.random_forest(x=toy_data[[0,1,2,3,4,5]], y=toy_data[6], ntrees=500, max_depth=20, nbins=100)
ranking = [rf._model_json['output']['variable_importances'].cell_values[v][0] for v in range(toy_data.ncol()-1)]
print(ranking)
assert tuple(ranking) == tuple(["V3","V2","V6","V1","V5","V4"]), "expected specific variable importance ranking"
if __name__ == "__main__":
h2o.run_test(sys.argv, vi_toy_test)
<commit_msg>Fix the seed for RF test.<commit_after>import sys
sys.path.insert(1, "../../../")
import h2o
def vi_toy_test(ip,port):
# Connect to h2o
h2o.init(ip,port)
toy_data = h2o.import_frame(path=h2o.locate("smalldata/gbm_test/toy_data_RF.csv"))
#toy_data.summary()
toy_data[6] = toy_data[6].asfactor()
toy_data.show()
rf = h2o.random_forest(x=toy_data[[0,1,2,3,4,5]], y=toy_data[6], ntrees=500, max_depth=20, nbins=100, seed=0)
ranking = [rf._model_json['output']['variable_importances'].cell_values[v][0] for v in range(toy_data.ncol()-1)]
print(ranking)
assert tuple(ranking) == tuple(["V3","V2","V6","V1","V5","V4"]), "expected specific variable importance ranking"
if __name__ == "__main__":
h2o.run_test(sys.argv, vi_toy_test)
|
ead5d7aa7a4a6fe4557c0e792ebc11e25359722f | rx/concurrency/scheduleditem.py | rx/concurrency/scheduleditem.py | from rx.disposables import SingleAssignmentDisposable
def default_sub_comparer(x, y):
return 0 if x == y else 1 if x > y else -1
class ScheduledItem(object):
def __init__(self, scheduler, state, action, duetime, comparer=None):
self.scheduler = scheduler
self.state = state
self.action = action
self.duetime = duetime
self.comparer = comparer or default_sub_comparer
self.disposable = SingleAssignmentDisposable()
def invoke(self):
self.disposable.disposable = self.invoke_core()
def compare_to(self, other):
return self.comparer(self.duetime, other.duetime)
def cancel(self):
"""Cancels the work item by disposing the resource returned by
invoke_core as soon as possible."""
self.disposable.dispose()
def is_cancelled(self):
return self.disposable.is_disposed
def invoke_core(self):
return self.action(self.scheduler, self.state)
def __lt__(self, other):
return self.compare_to(other) < 0
def __gt__(self, other):
return self.compare_to(other) > 0
def __eq__(self, other):
return self.compare_to(other) == 0
| from rx.core import Disposable
from rx.disposables import SingleAssignmentDisposable
def default_sub_comparer(x, y):
return 0 if x == y else 1 if x > y else -1
class ScheduledItem(object):
def __init__(self, scheduler, state, action, duetime, comparer=None):
self.scheduler = scheduler
self.state = state
self.action = action
self.duetime = duetime
self.comparer = comparer or default_sub_comparer
self.disposable = SingleAssignmentDisposable()
def invoke(self):
ret = self.action(self.scheduler, self.state)
if isinstance(ret, Disposable):
self.disposable.disposable = ret
def compare_to(self, other):
return self.comparer(self.duetime, other.duetime)
def cancel(self):
"""Cancels the work item by disposing the resource returned by
invoke_core as soon as possible."""
self.disposable.dispose()
def is_cancelled(self):
return self.disposable.is_disposed
def __lt__(self, other):
return self.compare_to(other) < 0
def __gt__(self, other):
return self.compare_to(other) > 0
def __eq__(self, other):
return self.compare_to(other) == 0
| Check if action returns disposable | Check if action returns disposable
| Python | mit | ReactiveX/RxPY,ReactiveX/RxPY | from rx.disposables import SingleAssignmentDisposable
def default_sub_comparer(x, y):
return 0 if x == y else 1 if x > y else -1
class ScheduledItem(object):
def __init__(self, scheduler, state, action, duetime, comparer=None):
self.scheduler = scheduler
self.state = state
self.action = action
self.duetime = duetime
self.comparer = comparer or default_sub_comparer
self.disposable = SingleAssignmentDisposable()
def invoke(self):
self.disposable.disposable = self.invoke_core()
def compare_to(self, other):
return self.comparer(self.duetime, other.duetime)
def cancel(self):
"""Cancels the work item by disposing the resource returned by
invoke_core as soon as possible."""
self.disposable.dispose()
def is_cancelled(self):
return self.disposable.is_disposed
def invoke_core(self):
return self.action(self.scheduler, self.state)
def __lt__(self, other):
return self.compare_to(other) < 0
def __gt__(self, other):
return self.compare_to(other) > 0
def __eq__(self, other):
return self.compare_to(other) == 0
Check if action returns disposable | from rx.core import Disposable
from rx.disposables import SingleAssignmentDisposable
def default_sub_comparer(x, y):
return 0 if x == y else 1 if x > y else -1
class ScheduledItem(object):
def __init__(self, scheduler, state, action, duetime, comparer=None):
self.scheduler = scheduler
self.state = state
self.action = action
self.duetime = duetime
self.comparer = comparer or default_sub_comparer
self.disposable = SingleAssignmentDisposable()
def invoke(self):
ret = self.action(self.scheduler, self.state)
if isinstance(ret, Disposable):
self.disposable.disposable = ret
def compare_to(self, other):
return self.comparer(self.duetime, other.duetime)
def cancel(self):
"""Cancels the work item by disposing the resource returned by
invoke_core as soon as possible."""
self.disposable.dispose()
def is_cancelled(self):
return self.disposable.is_disposed
def __lt__(self, other):
return self.compare_to(other) < 0
def __gt__(self, other):
return self.compare_to(other) > 0
def __eq__(self, other):
return self.compare_to(other) == 0
| <commit_before>from rx.disposables import SingleAssignmentDisposable
def default_sub_comparer(x, y):
return 0 if x == y else 1 if x > y else -1
class ScheduledItem(object):
def __init__(self, scheduler, state, action, duetime, comparer=None):
self.scheduler = scheduler
self.state = state
self.action = action
self.duetime = duetime
self.comparer = comparer or default_sub_comparer
self.disposable = SingleAssignmentDisposable()
def invoke(self):
self.disposable.disposable = self.invoke_core()
def compare_to(self, other):
return self.comparer(self.duetime, other.duetime)
def cancel(self):
"""Cancels the work item by disposing the resource returned by
invoke_core as soon as possible."""
self.disposable.dispose()
def is_cancelled(self):
return self.disposable.is_disposed
def invoke_core(self):
return self.action(self.scheduler, self.state)
def __lt__(self, other):
return self.compare_to(other) < 0
def __gt__(self, other):
return self.compare_to(other) > 0
def __eq__(self, other):
return self.compare_to(other) == 0
<commit_msg>Check if action returns disposable<commit_after> | from rx.core import Disposable
from rx.disposables import SingleAssignmentDisposable
def default_sub_comparer(x, y):
return 0 if x == y else 1 if x > y else -1
class ScheduledItem(object):
def __init__(self, scheduler, state, action, duetime, comparer=None):
self.scheduler = scheduler
self.state = state
self.action = action
self.duetime = duetime
self.comparer = comparer or default_sub_comparer
self.disposable = SingleAssignmentDisposable()
def invoke(self):
ret = self.action(self.scheduler, self.state)
if isinstance(ret, Disposable):
self.disposable.disposable = ret
def compare_to(self, other):
return self.comparer(self.duetime, other.duetime)
def cancel(self):
"""Cancels the work item by disposing the resource returned by
invoke_core as soon as possible."""
self.disposable.dispose()
def is_cancelled(self):
return self.disposable.is_disposed
def __lt__(self, other):
return self.compare_to(other) < 0
def __gt__(self, other):
return self.compare_to(other) > 0
def __eq__(self, other):
return self.compare_to(other) == 0
| from rx.disposables import SingleAssignmentDisposable
def default_sub_comparer(x, y):
return 0 if x == y else 1 if x > y else -1
class ScheduledItem(object):
def __init__(self, scheduler, state, action, duetime, comparer=None):
self.scheduler = scheduler
self.state = state
self.action = action
self.duetime = duetime
self.comparer = comparer or default_sub_comparer
self.disposable = SingleAssignmentDisposable()
def invoke(self):
self.disposable.disposable = self.invoke_core()
def compare_to(self, other):
return self.comparer(self.duetime, other.duetime)
def cancel(self):
"""Cancels the work item by disposing the resource returned by
invoke_core as soon as possible."""
self.disposable.dispose()
def is_cancelled(self):
return self.disposable.is_disposed
def invoke_core(self):
return self.action(self.scheduler, self.state)
def __lt__(self, other):
return self.compare_to(other) < 0
def __gt__(self, other):
return self.compare_to(other) > 0
def __eq__(self, other):
return self.compare_to(other) == 0
Check if action returns disposablefrom rx.core import Disposable
from rx.disposables import SingleAssignmentDisposable
def default_sub_comparer(x, y):
return 0 if x == y else 1 if x > y else -1
class ScheduledItem(object):
def __init__(self, scheduler, state, action, duetime, comparer=None):
self.scheduler = scheduler
self.state = state
self.action = action
self.duetime = duetime
self.comparer = comparer or default_sub_comparer
self.disposable = SingleAssignmentDisposable()
def invoke(self):
ret = self.action(self.scheduler, self.state)
if isinstance(ret, Disposable):
self.disposable.disposable = ret
def compare_to(self, other):
return self.comparer(self.duetime, other.duetime)
def cancel(self):
"""Cancels the work item by disposing the resource returned by
invoke_core as soon as possible."""
self.disposable.dispose()
def is_cancelled(self):
return self.disposable.is_disposed
def __lt__(self, other):
return self.compare_to(other) < 0
def __gt__(self, other):
return self.compare_to(other) > 0
def __eq__(self, other):
return self.compare_to(other) == 0
| <commit_before>from rx.disposables import SingleAssignmentDisposable
def default_sub_comparer(x, y):
return 0 if x == y else 1 if x > y else -1
class ScheduledItem(object):
def __init__(self, scheduler, state, action, duetime, comparer=None):
self.scheduler = scheduler
self.state = state
self.action = action
self.duetime = duetime
self.comparer = comparer or default_sub_comparer
self.disposable = SingleAssignmentDisposable()
def invoke(self):
self.disposable.disposable = self.invoke_core()
def compare_to(self, other):
return self.comparer(self.duetime, other.duetime)
def cancel(self):
"""Cancels the work item by disposing the resource returned by
invoke_core as soon as possible."""
self.disposable.dispose()
def is_cancelled(self):
return self.disposable.is_disposed
def invoke_core(self):
return self.action(self.scheduler, self.state)
def __lt__(self, other):
return self.compare_to(other) < 0
def __gt__(self, other):
return self.compare_to(other) > 0
def __eq__(self, other):
return self.compare_to(other) == 0
<commit_msg>Check if action returns disposable<commit_after>from rx.core import Disposable
from rx.disposables import SingleAssignmentDisposable
def default_sub_comparer(x, y):
return 0 if x == y else 1 if x > y else -1
class ScheduledItem(object):
def __init__(self, scheduler, state, action, duetime, comparer=None):
self.scheduler = scheduler
self.state = state
self.action = action
self.duetime = duetime
self.comparer = comparer or default_sub_comparer
self.disposable = SingleAssignmentDisposable()
def invoke(self):
ret = self.action(self.scheduler, self.state)
if isinstance(ret, Disposable):
self.disposable.disposable = ret
def compare_to(self, other):
return self.comparer(self.duetime, other.duetime)
def cancel(self):
"""Cancels the work item by disposing the resource returned by
invoke_core as soon as possible."""
self.disposable.dispose()
def is_cancelled(self):
return self.disposable.is_disposed
def __lt__(self, other):
return self.compare_to(other) < 0
def __gt__(self, other):
return self.compare_to(other) > 0
def __eq__(self, other):
return self.compare_to(other) == 0
|
840aef8fee59c9f1a9863177e060b05b09fcacd4 | tests/utils.py | tests/utils.py | # -*- coding: utf-8 -*-
def has_no_django():
try:
import django
return False
except ImportError:
return True
| # -*- coding: utf-8 -*-
def has_no_django():
try:
import django # noqa isort:skip
return False
except ImportError:
return True
| Add noqa to conditional import | Add noqa to conditional import
| Python | mit | python-thumbnails/python-thumbnails,relekang/python-thumbnails | # -*- coding: utf-8 -*-
def has_no_django():
try:
import django
return False
except ImportError:
return True
Add noqa to conditional import | # -*- coding: utf-8 -*-
def has_no_django():
try:
import django # noqa isort:skip
return False
except ImportError:
return True
| <commit_before># -*- coding: utf-8 -*-
def has_no_django():
try:
import django
return False
except ImportError:
return True
<commit_msg>Add noqa to conditional import<commit_after> | # -*- coding: utf-8 -*-
def has_no_django():
try:
import django # noqa isort:skip
return False
except ImportError:
return True
| # -*- coding: utf-8 -*-
def has_no_django():
try:
import django
return False
except ImportError:
return True
Add noqa to conditional import# -*- coding: utf-8 -*-
def has_no_django():
try:
import django # noqa isort:skip
return False
except ImportError:
return True
| <commit_before># -*- coding: utf-8 -*-
def has_no_django():
try:
import django
return False
except ImportError:
return True
<commit_msg>Add noqa to conditional import<commit_after># -*- coding: utf-8 -*-
def has_no_django():
try:
import django # noqa isort:skip
return False
except ImportError:
return True
|
5ddde4a43ede87770543984e96eb8ccaf1d829b2 | lib/methods/drupalconsole.py | lib/methods/drupalconsole.py | from base import BaseMethod
from fabric.api import *
from lib.utils import SSHTunnel, RemoteSSHTunnel
from fabric.colors import green, red
from lib import configuration
import copy
class DrupalConsoleMethod(BaseMethod):
@staticmethod
def supports(methodName):
return methodName == 'drupalconsole'
def install(self, config):
with cd(config['tmpFolder']):
run('curl https://drupalconsole.com/installer -L -o drupal.phar')
run('mv drupal.phar /usr/local/bin/drupal')
run('chmod +x /usr/local/bin/drupal')
run('drupal init')
print green('Drupal Console installed successfully.')
def run_drupalconsole(self, config, command):
with cd(config['rootFolder']):
run('drupal %s' % command)
def drupalconsole(self, config, **kwargs):
if kwargs['command'] == 'install':
self.install(config)
return
self.run_drupalconsole(config, kwargs['command'])
| from base import BaseMethod
from fabric.api import *
from lib.utils import SSHTunnel, RemoteSSHTunnel
from fabric.colors import green, red
from lib import configuration
import copy
class DrupalConsoleMethod(BaseMethod):
@staticmethod
def supports(methodName):
return methodName == 'drupalconsole'
def install(self, config, **kwargs):
with cd(config['tmpFolder']):
run('curl https://drupalconsole.com/installer -L -o drupal.phar')
run('mv drupal.phar /usr/local/bin/drupal')
run('chmod +x /usr/local/bin/drupal')
run('drupal init')
print green('Drupal Console installed successfully.')
def run_drupalconsole(self, config, command):
with cd(config['rootFolder']):
run('drupal %s' % command)
def drupalconsole(self, config, **kwargs):
if kwargs['command'] == 'install':
self.install(config)
return
self.run_drupalconsole(config, kwargs['command'])
| Fix exception when running install-task | Fix exception when running install-task
| Python | mit | factorial-io/fabalicious,factorial-io/fabalicious | from base import BaseMethod
from fabric.api import *
from lib.utils import SSHTunnel, RemoteSSHTunnel
from fabric.colors import green, red
from lib import configuration
import copy
class DrupalConsoleMethod(BaseMethod):
@staticmethod
def supports(methodName):
return methodName == 'drupalconsole'
def install(self, config):
with cd(config['tmpFolder']):
run('curl https://drupalconsole.com/installer -L -o drupal.phar')
run('mv drupal.phar /usr/local/bin/drupal')
run('chmod +x /usr/local/bin/drupal')
run('drupal init')
print green('Drupal Console installed successfully.')
def run_drupalconsole(self, config, command):
with cd(config['rootFolder']):
run('drupal %s' % command)
def drupalconsole(self, config, **kwargs):
if kwargs['command'] == 'install':
self.install(config)
return
self.run_drupalconsole(config, kwargs['command'])
Fix exception when running install-task | from base import BaseMethod
from fabric.api import *
from lib.utils import SSHTunnel, RemoteSSHTunnel
from fabric.colors import green, red
from lib import configuration
import copy
class DrupalConsoleMethod(BaseMethod):
@staticmethod
def supports(methodName):
return methodName == 'drupalconsole'
def install(self, config, **kwargs):
with cd(config['tmpFolder']):
run('curl https://drupalconsole.com/installer -L -o drupal.phar')
run('mv drupal.phar /usr/local/bin/drupal')
run('chmod +x /usr/local/bin/drupal')
run('drupal init')
print green('Drupal Console installed successfully.')
def run_drupalconsole(self, config, command):
with cd(config['rootFolder']):
run('drupal %s' % command)
def drupalconsole(self, config, **kwargs):
if kwargs['command'] == 'install':
self.install(config)
return
self.run_drupalconsole(config, kwargs['command'])
| <commit_before>from base import BaseMethod
from fabric.api import *
from lib.utils import SSHTunnel, RemoteSSHTunnel
from fabric.colors import green, red
from lib import configuration
import copy
class DrupalConsoleMethod(BaseMethod):
@staticmethod
def supports(methodName):
return methodName == 'drupalconsole'
def install(self, config):
with cd(config['tmpFolder']):
run('curl https://drupalconsole.com/installer -L -o drupal.phar')
run('mv drupal.phar /usr/local/bin/drupal')
run('chmod +x /usr/local/bin/drupal')
run('drupal init')
print green('Drupal Console installed successfully.')
def run_drupalconsole(self, config, command):
with cd(config['rootFolder']):
run('drupal %s' % command)
def drupalconsole(self, config, **kwargs):
if kwargs['command'] == 'install':
self.install(config)
return
self.run_drupalconsole(config, kwargs['command'])
<commit_msg>Fix exception when running install-task<commit_after> | from base import BaseMethod
from fabric.api import *
from lib.utils import SSHTunnel, RemoteSSHTunnel
from fabric.colors import green, red
from lib import configuration
import copy
class DrupalConsoleMethod(BaseMethod):
@staticmethod
def supports(methodName):
return methodName == 'drupalconsole'
def install(self, config, **kwargs):
with cd(config['tmpFolder']):
run('curl https://drupalconsole.com/installer -L -o drupal.phar')
run('mv drupal.phar /usr/local/bin/drupal')
run('chmod +x /usr/local/bin/drupal')
run('drupal init')
print green('Drupal Console installed successfully.')
def run_drupalconsole(self, config, command):
with cd(config['rootFolder']):
run('drupal %s' % command)
def drupalconsole(self, config, **kwargs):
if kwargs['command'] == 'install':
self.install(config)
return
self.run_drupalconsole(config, kwargs['command'])
| from base import BaseMethod
from fabric.api import *
from lib.utils import SSHTunnel, RemoteSSHTunnel
from fabric.colors import green, red
from lib import configuration
import copy
class DrupalConsoleMethod(BaseMethod):
@staticmethod
def supports(methodName):
return methodName == 'drupalconsole'
def install(self, config):
with cd(config['tmpFolder']):
run('curl https://drupalconsole.com/installer -L -o drupal.phar')
run('mv drupal.phar /usr/local/bin/drupal')
run('chmod +x /usr/local/bin/drupal')
run('drupal init')
print green('Drupal Console installed successfully.')
def run_drupalconsole(self, config, command):
with cd(config['rootFolder']):
run('drupal %s' % command)
def drupalconsole(self, config, **kwargs):
if kwargs['command'] == 'install':
self.install(config)
return
self.run_drupalconsole(config, kwargs['command'])
Fix exception when running install-taskfrom base import BaseMethod
from fabric.api import *
from lib.utils import SSHTunnel, RemoteSSHTunnel
from fabric.colors import green, red
from lib import configuration
import copy
class DrupalConsoleMethod(BaseMethod):
@staticmethod
def supports(methodName):
return methodName == 'drupalconsole'
def install(self, config, **kwargs):
with cd(config['tmpFolder']):
run('curl https://drupalconsole.com/installer -L -o drupal.phar')
run('mv drupal.phar /usr/local/bin/drupal')
run('chmod +x /usr/local/bin/drupal')
run('drupal init')
print green('Drupal Console installed successfully.')
def run_drupalconsole(self, config, command):
with cd(config['rootFolder']):
run('drupal %s' % command)
def drupalconsole(self, config, **kwargs):
if kwargs['command'] == 'install':
self.install(config)
return
self.run_drupalconsole(config, kwargs['command'])
| <commit_before>from base import BaseMethod
from fabric.api import *
from lib.utils import SSHTunnel, RemoteSSHTunnel
from fabric.colors import green, red
from lib import configuration
import copy
class DrupalConsoleMethod(BaseMethod):
@staticmethod
def supports(methodName):
return methodName == 'drupalconsole'
def install(self, config):
with cd(config['tmpFolder']):
run('curl https://drupalconsole.com/installer -L -o drupal.phar')
run('mv drupal.phar /usr/local/bin/drupal')
run('chmod +x /usr/local/bin/drupal')
run('drupal init')
print green('Drupal Console installed successfully.')
def run_drupalconsole(self, config, command):
with cd(config['rootFolder']):
run('drupal %s' % command)
def drupalconsole(self, config, **kwargs):
if kwargs['command'] == 'install':
self.install(config)
return
self.run_drupalconsole(config, kwargs['command'])
<commit_msg>Fix exception when running install-task<commit_after>from base import BaseMethod
from fabric.api import *
from lib.utils import SSHTunnel, RemoteSSHTunnel
from fabric.colors import green, red
from lib import configuration
import copy
class DrupalConsoleMethod(BaseMethod):
@staticmethod
def supports(methodName):
return methodName == 'drupalconsole'
def install(self, config, **kwargs):
with cd(config['tmpFolder']):
run('curl https://drupalconsole.com/installer -L -o drupal.phar')
run('mv drupal.phar /usr/local/bin/drupal')
run('chmod +x /usr/local/bin/drupal')
run('drupal init')
print green('Drupal Console installed successfully.')
def run_drupalconsole(self, config, command):
with cd(config['rootFolder']):
run('drupal %s' % command)
def drupalconsole(self, config, **kwargs):
if kwargs['command'] == 'install':
self.install(config)
return
self.run_drupalconsole(config, kwargs['command'])
|
a667b3503b0434f01459bae2d29df800d95ba1c4 | gapipy/resources/tour/departure.py | gapipy/resources/tour/departure.py | from __future__ import unicode_literals
from ...models import Address, AddOn, DepartureRoom, PP2aPrice
from ..base import Product
from .tour_dossier import TourDossier
from .departure_component import DepartureComponent
class Departure(Product):
_resource_name = 'departures'
_is_listable = True
_is_parent_resource = True
_as_is_fields = [
'id', 'href', 'availability', 'flags', 'nearest_start_airport',
'nearest_finish_airport', 'product_line', 'sku', 'requirements',
]
_date_fields = ['start_date', 'finish_date']
_date_time_fields_utc = ['date_created', 'date_last_modified']
_date_time_fields_local = ['latest_arrival_time', 'earliest_departure_time']
_resource_fields = [('tour', 'Tour'), ('tour_dossier', TourDossier)]
_resource_collection_fields = [
('components', DepartureComponent),
]
_model_fields = [('start_address', Address), ('finish_address', Address)]
_model_collection_fields = [
('addons', AddOn),
('rooms', DepartureRoom),
('lowest_pp2a_prices', PP2aPrice),
]
_deprecated_fields = ['add_ons']
| from __future__ import unicode_literals
from ...models import Address, AddOn, DepartureRoom, PP2aPrice
from ..base import Product
from .tour_dossier import TourDossier
from .departure_component import DepartureComponent
class Departure(Product):
_resource_name = 'departures'
_is_listable = True
_is_parent_resource = True
_as_is_fields = [
'id', 'href', 'name', 'availability', 'flags', 'nearest_start_airport',
'nearest_finish_airport', 'product_line', 'sku', 'requirements',
]
_date_fields = ['start_date', 'finish_date']
_date_time_fields_utc = ['date_created', 'date_last_modified']
_date_time_fields_local = ['latest_arrival_time', 'earliest_departure_time']
_resource_fields = [('tour', 'Tour'), ('tour_dossier', TourDossier)]
_resource_collection_fields = [
('components', DepartureComponent),
]
_model_fields = [('start_address', Address), ('finish_address', Address)]
_model_collection_fields = [
('addons', AddOn),
('rooms', DepartureRoom),
('lowest_pp2a_prices', PP2aPrice),
]
_deprecated_fields = ['add_ons']
| Add name to Departure resource | Add name to Departure resource
| Python | mit | gadventures/gapipy | from __future__ import unicode_literals
from ...models import Address, AddOn, DepartureRoom, PP2aPrice
from ..base import Product
from .tour_dossier import TourDossier
from .departure_component import DepartureComponent
class Departure(Product):
_resource_name = 'departures'
_is_listable = True
_is_parent_resource = True
_as_is_fields = [
'id', 'href', 'availability', 'flags', 'nearest_start_airport',
'nearest_finish_airport', 'product_line', 'sku', 'requirements',
]
_date_fields = ['start_date', 'finish_date']
_date_time_fields_utc = ['date_created', 'date_last_modified']
_date_time_fields_local = ['latest_arrival_time', 'earliest_departure_time']
_resource_fields = [('tour', 'Tour'), ('tour_dossier', TourDossier)]
_resource_collection_fields = [
('components', DepartureComponent),
]
_model_fields = [('start_address', Address), ('finish_address', Address)]
_model_collection_fields = [
('addons', AddOn),
('rooms', DepartureRoom),
('lowest_pp2a_prices', PP2aPrice),
]
_deprecated_fields = ['add_ons']
Add name to Departure resource | from __future__ import unicode_literals
from ...models import Address, AddOn, DepartureRoom, PP2aPrice
from ..base import Product
from .tour_dossier import TourDossier
from .departure_component import DepartureComponent
class Departure(Product):
_resource_name = 'departures'
_is_listable = True
_is_parent_resource = True
_as_is_fields = [
'id', 'href', 'name', 'availability', 'flags', 'nearest_start_airport',
'nearest_finish_airport', 'product_line', 'sku', 'requirements',
]
_date_fields = ['start_date', 'finish_date']
_date_time_fields_utc = ['date_created', 'date_last_modified']
_date_time_fields_local = ['latest_arrival_time', 'earliest_departure_time']
_resource_fields = [('tour', 'Tour'), ('tour_dossier', TourDossier)]
_resource_collection_fields = [
('components', DepartureComponent),
]
_model_fields = [('start_address', Address), ('finish_address', Address)]
_model_collection_fields = [
('addons', AddOn),
('rooms', DepartureRoom),
('lowest_pp2a_prices', PP2aPrice),
]
_deprecated_fields = ['add_ons']
| <commit_before>from __future__ import unicode_literals
from ...models import Address, AddOn, DepartureRoom, PP2aPrice
from ..base import Product
from .tour_dossier import TourDossier
from .departure_component import DepartureComponent
class Departure(Product):
_resource_name = 'departures'
_is_listable = True
_is_parent_resource = True
_as_is_fields = [
'id', 'href', 'availability', 'flags', 'nearest_start_airport',
'nearest_finish_airport', 'product_line', 'sku', 'requirements',
]
_date_fields = ['start_date', 'finish_date']
_date_time_fields_utc = ['date_created', 'date_last_modified']
_date_time_fields_local = ['latest_arrival_time', 'earliest_departure_time']
_resource_fields = [('tour', 'Tour'), ('tour_dossier', TourDossier)]
_resource_collection_fields = [
('components', DepartureComponent),
]
_model_fields = [('start_address', Address), ('finish_address', Address)]
_model_collection_fields = [
('addons', AddOn),
('rooms', DepartureRoom),
('lowest_pp2a_prices', PP2aPrice),
]
_deprecated_fields = ['add_ons']
<commit_msg>Add name to Departure resource<commit_after> | from __future__ import unicode_literals
from ...models import Address, AddOn, DepartureRoom, PP2aPrice
from ..base import Product
from .tour_dossier import TourDossier
from .departure_component import DepartureComponent
class Departure(Product):
_resource_name = 'departures'
_is_listable = True
_is_parent_resource = True
_as_is_fields = [
'id', 'href', 'name', 'availability', 'flags', 'nearest_start_airport',
'nearest_finish_airport', 'product_line', 'sku', 'requirements',
]
_date_fields = ['start_date', 'finish_date']
_date_time_fields_utc = ['date_created', 'date_last_modified']
_date_time_fields_local = ['latest_arrival_time', 'earliest_departure_time']
_resource_fields = [('tour', 'Tour'), ('tour_dossier', TourDossier)]
_resource_collection_fields = [
('components', DepartureComponent),
]
_model_fields = [('start_address', Address), ('finish_address', Address)]
_model_collection_fields = [
('addons', AddOn),
('rooms', DepartureRoom),
('lowest_pp2a_prices', PP2aPrice),
]
_deprecated_fields = ['add_ons']
| from __future__ import unicode_literals
from ...models import Address, AddOn, DepartureRoom, PP2aPrice
from ..base import Product
from .tour_dossier import TourDossier
from .departure_component import DepartureComponent
class Departure(Product):
_resource_name = 'departures'
_is_listable = True
_is_parent_resource = True
_as_is_fields = [
'id', 'href', 'availability', 'flags', 'nearest_start_airport',
'nearest_finish_airport', 'product_line', 'sku', 'requirements',
]
_date_fields = ['start_date', 'finish_date']
_date_time_fields_utc = ['date_created', 'date_last_modified']
_date_time_fields_local = ['latest_arrival_time', 'earliest_departure_time']
_resource_fields = [('tour', 'Tour'), ('tour_dossier', TourDossier)]
_resource_collection_fields = [
('components', DepartureComponent),
]
_model_fields = [('start_address', Address), ('finish_address', Address)]
_model_collection_fields = [
('addons', AddOn),
('rooms', DepartureRoom),
('lowest_pp2a_prices', PP2aPrice),
]
_deprecated_fields = ['add_ons']
Add name to Departure resourcefrom __future__ import unicode_literals
from ...models import Address, AddOn, DepartureRoom, PP2aPrice
from ..base import Product
from .tour_dossier import TourDossier
from .departure_component import DepartureComponent
class Departure(Product):
_resource_name = 'departures'
_is_listable = True
_is_parent_resource = True
_as_is_fields = [
'id', 'href', 'name', 'availability', 'flags', 'nearest_start_airport',
'nearest_finish_airport', 'product_line', 'sku', 'requirements',
]
_date_fields = ['start_date', 'finish_date']
_date_time_fields_utc = ['date_created', 'date_last_modified']
_date_time_fields_local = ['latest_arrival_time', 'earliest_departure_time']
_resource_fields = [('tour', 'Tour'), ('tour_dossier', TourDossier)]
_resource_collection_fields = [
('components', DepartureComponent),
]
_model_fields = [('start_address', Address), ('finish_address', Address)]
_model_collection_fields = [
('addons', AddOn),
('rooms', DepartureRoom),
('lowest_pp2a_prices', PP2aPrice),
]
_deprecated_fields = ['add_ons']
| <commit_before>from __future__ import unicode_literals
from ...models import Address, AddOn, DepartureRoom, PP2aPrice
from ..base import Product
from .tour_dossier import TourDossier
from .departure_component import DepartureComponent
class Departure(Product):
_resource_name = 'departures'
_is_listable = True
_is_parent_resource = True
_as_is_fields = [
'id', 'href', 'availability', 'flags', 'nearest_start_airport',
'nearest_finish_airport', 'product_line', 'sku', 'requirements',
]
_date_fields = ['start_date', 'finish_date']
_date_time_fields_utc = ['date_created', 'date_last_modified']
_date_time_fields_local = ['latest_arrival_time', 'earliest_departure_time']
_resource_fields = [('tour', 'Tour'), ('tour_dossier', TourDossier)]
_resource_collection_fields = [
('components', DepartureComponent),
]
_model_fields = [('start_address', Address), ('finish_address', Address)]
_model_collection_fields = [
('addons', AddOn),
('rooms', DepartureRoom),
('lowest_pp2a_prices', PP2aPrice),
]
_deprecated_fields = ['add_ons']
<commit_msg>Add name to Departure resource<commit_after>from __future__ import unicode_literals
from ...models import Address, AddOn, DepartureRoom, PP2aPrice
from ..base import Product
from .tour_dossier import TourDossier
from .departure_component import DepartureComponent
class Departure(Product):
_resource_name = 'departures'
_is_listable = True
_is_parent_resource = True
_as_is_fields = [
'id', 'href', 'name', 'availability', 'flags', 'nearest_start_airport',
'nearest_finish_airport', 'product_line', 'sku', 'requirements',
]
_date_fields = ['start_date', 'finish_date']
_date_time_fields_utc = ['date_created', 'date_last_modified']
_date_time_fields_local = ['latest_arrival_time', 'earliest_departure_time']
_resource_fields = [('tour', 'Tour'), ('tour_dossier', TourDossier)]
_resource_collection_fields = [
('components', DepartureComponent),
]
_model_fields = [('start_address', Address), ('finish_address', Address)]
_model_collection_fields = [
('addons', AddOn),
('rooms', DepartureRoom),
('lowest_pp2a_prices', PP2aPrice),
]
_deprecated_fields = ['add_ons']
|
ae1de4000a6e9f3fc70d14c6214038e83772a5f6 | Part2/main.py | Part2/main.py | import detectLang
import graph
# ====================================================================================================
# La detection est rapide car toute les perplexites sont stockées dans les fichiers binaires pp_EN etc
# Pour regénerer les fichiers :
# Executer detectLang.create_all_pp_and_save_to_disc()
# Les resultats des perplexités pour tous les fichiers seront serialisés dans les fichiers binaires.
# ces fichiers sont necessaire à la fontions detectLang.detect_language()
# Pour serialisé les dictionaires j'utilise la bibliotheque intégrer à python => pickle
# ====================================================================================================
test_file_number = 19
# detect_language(Numéro du fichier dans le repertoire de test, N du modele nGram)
# Print le code de la langue reconnue et la perplexité du modèle choisis.
detectLang.detect_language(test_file_number,1)
detectLang.detect_language(test_file_number,2)
detectLang.detect_language(test_file_number,3)
# ====================================================================================================
# Nécessite matplotlib
# Installé matplotlib avec => pip install matplotlib
# ====================================================================================================
# graphFile(Numéro du fichier dans le repertoire de test)
graph.graphFile(test_file_number)
| import detectLang
import graph
# ====================================================================================================
# La detection est rapide car toute les perplexites sont stockées dans les fichiers binaires pp_EN etc
# Pour regénerer les fichiers :
# Executer detectLang.create_all_pp_and_save_to_disc(), décommenter la ligne suivante
# detectLang.create_all_pp_and_save_to_disc()
# Les resultats des perplexités pour tous les fichiers seront serialisés dans les fichiers binaires.
# ces fichiers sont necessaire à la fontions detectLang.detect_language()
# Pour serialisé les dictionaires j'utilise la bibliotheque intégrer à python => pickle
# ====================================================================================================
test_file_number = 13
# detect_language(Numéro du fichier dans le repertoire de test, N du modele nGram)
# Print le code de la langue reconnue et la perplexité du modèle choisis.
# detectLang.detect_language(test_file_number,1)
# detectLang.detect_language(test_file_number,2)
# detectLang.detect_language(test_file_number,3)
# ====================================================================================================
# Nécessite matplotlib
# Installé matplotlib avec => pip install matplotlib
# ====================================================================================================
# Affiche sur un graphique les perplexité de tous les modeles sur un même fichier
# graph.graphFile(test_file_number)
# Pour donner le resultat sur tous les fichier test dans la console
detectLang.show_all_result()
| Update doc and add one call | Update doc and add one call
| Python | mit | Focom/NLPWork1,Focom/NLPWork1,Focom/NLPWork1 | import detectLang
import graph
# ====================================================================================================
# La detection est rapide car toute les perplexites sont stockées dans les fichiers binaires pp_EN etc
# Pour regénerer les fichiers :
# Executer detectLang.create_all_pp_and_save_to_disc()
# Les resultats des perplexités pour tous les fichiers seront serialisés dans les fichiers binaires.
# ces fichiers sont necessaire à la fontions detectLang.detect_language()
# Pour serialisé les dictionaires j'utilise la bibliotheque intégrer à python => pickle
# ====================================================================================================
test_file_number = 19
# detect_language(Numéro du fichier dans le repertoire de test, N du modele nGram)
# Print le code de la langue reconnue et la perplexité du modèle choisis.
detectLang.detect_language(test_file_number,1)
detectLang.detect_language(test_file_number,2)
detectLang.detect_language(test_file_number,3)
# ====================================================================================================
# Nécessite matplotlib
# Installé matplotlib avec => pip install matplotlib
# ====================================================================================================
# graphFile(Numéro du fichier dans le repertoire de test)
graph.graphFile(test_file_number)
Update doc and add one call | import detectLang
import graph
# ====================================================================================================
# La detection est rapide car toute les perplexites sont stockées dans les fichiers binaires pp_EN etc
# Pour regénerer les fichiers :
# Executer detectLang.create_all_pp_and_save_to_disc(), décommenter la ligne suivante
# detectLang.create_all_pp_and_save_to_disc()
# Les resultats des perplexités pour tous les fichiers seront serialisés dans les fichiers binaires.
# ces fichiers sont necessaire à la fontions detectLang.detect_language()
# Pour serialisé les dictionaires j'utilise la bibliotheque intégrer à python => pickle
# ====================================================================================================
test_file_number = 13
# detect_language(Numéro du fichier dans le repertoire de test, N du modele nGram)
# Print le code de la langue reconnue et la perplexité du modèle choisis.
# detectLang.detect_language(test_file_number,1)
# detectLang.detect_language(test_file_number,2)
# detectLang.detect_language(test_file_number,3)
# ====================================================================================================
# Nécessite matplotlib
# Installé matplotlib avec => pip install matplotlib
# ====================================================================================================
# Affiche sur un graphique les perplexité de tous les modeles sur un même fichier
# graph.graphFile(test_file_number)
# Pour donner le resultat sur tous les fichier test dans la console
detectLang.show_all_result()
| <commit_before>import detectLang
import graph
# ====================================================================================================
# La detection est rapide car toute les perplexites sont stockées dans les fichiers binaires pp_EN etc
# Pour regénerer les fichiers :
# Executer detectLang.create_all_pp_and_save_to_disc()
# Les resultats des perplexités pour tous les fichiers seront serialisés dans les fichiers binaires.
# ces fichiers sont necessaire à la fontions detectLang.detect_language()
# Pour serialisé les dictionaires j'utilise la bibliotheque intégrer à python => pickle
# ====================================================================================================
test_file_number = 19
# detect_language(Numéro du fichier dans le repertoire de test, N du modele nGram)
# Print le code de la langue reconnue et la perplexité du modèle choisis.
detectLang.detect_language(test_file_number,1)
detectLang.detect_language(test_file_number,2)
detectLang.detect_language(test_file_number,3)
# ====================================================================================================
# Nécessite matplotlib
# Installé matplotlib avec => pip install matplotlib
# ====================================================================================================
# graphFile(Numéro du fichier dans le repertoire de test)
graph.graphFile(test_file_number)
<commit_msg>Update doc and add one call<commit_after> | import detectLang
import graph
# ====================================================================================================
# La detection est rapide car toute les perplexites sont stockées dans les fichiers binaires pp_EN etc
# Pour regénerer les fichiers :
# Executer detectLang.create_all_pp_and_save_to_disc(), décommenter la ligne suivante
# detectLang.create_all_pp_and_save_to_disc()
# Les resultats des perplexités pour tous les fichiers seront serialisés dans les fichiers binaires.
# ces fichiers sont necessaire à la fontions detectLang.detect_language()
# Pour serialisé les dictionaires j'utilise la bibliotheque intégrer à python => pickle
# ====================================================================================================
test_file_number = 13
# detect_language(Numéro du fichier dans le repertoire de test, N du modele nGram)
# Print le code de la langue reconnue et la perplexité du modèle choisis.
# detectLang.detect_language(test_file_number,1)
# detectLang.detect_language(test_file_number,2)
# detectLang.detect_language(test_file_number,3)
# ====================================================================================================
# Nécessite matplotlib
# Installé matplotlib avec => pip install matplotlib
# ====================================================================================================
# Affiche sur un graphique les perplexité de tous les modeles sur un même fichier
# graph.graphFile(test_file_number)
# Pour donner le resultat sur tous les fichier test dans la console
detectLang.show_all_result()
| import detectLang
import graph
# ====================================================================================================
# La detection est rapide car toute les perplexites sont stockées dans les fichiers binaires pp_EN etc
# Pour regénerer les fichiers :
# Executer detectLang.create_all_pp_and_save_to_disc()
# Les resultats des perplexités pour tous les fichiers seront serialisés dans les fichiers binaires.
# ces fichiers sont necessaire à la fontions detectLang.detect_language()
# Pour serialisé les dictionaires j'utilise la bibliotheque intégrer à python => pickle
# ====================================================================================================
test_file_number = 19
# detect_language(Numéro du fichier dans le repertoire de test, N du modele nGram)
# Print le code de la langue reconnue et la perplexité du modèle choisis.
detectLang.detect_language(test_file_number,1)
detectLang.detect_language(test_file_number,2)
detectLang.detect_language(test_file_number,3)
# ====================================================================================================
# Nécessite matplotlib
# Installé matplotlib avec => pip install matplotlib
# ====================================================================================================
# graphFile(Numéro du fichier dans le repertoire de test)
graph.graphFile(test_file_number)
Update doc and add one callimport detectLang
import graph
# ====================================================================================================
# La detection est rapide car toute les perplexites sont stockées dans les fichiers binaires pp_EN etc
# Pour regénerer les fichiers :
# Executer detectLang.create_all_pp_and_save_to_disc(), décommenter la ligne suivante
# detectLang.create_all_pp_and_save_to_disc()
# Les resultats des perplexités pour tous les fichiers seront serialisés dans les fichiers binaires.
# ces fichiers sont necessaire à la fontions detectLang.detect_language()
# Pour serialisé les dictionaires j'utilise la bibliotheque intégrer à python => pickle
# ====================================================================================================
test_file_number = 13
# detect_language(Numéro du fichier dans le repertoire de test, N du modele nGram)
# Print le code de la langue reconnue et la perplexité du modèle choisis.
# detectLang.detect_language(test_file_number,1)
# detectLang.detect_language(test_file_number,2)
# detectLang.detect_language(test_file_number,3)
# ====================================================================================================
# Nécessite matplotlib
# Installé matplotlib avec => pip install matplotlib
# ====================================================================================================
# Affiche sur un graphique les perplexité de tous les modeles sur un même fichier
# graph.graphFile(test_file_number)
# Pour donner le resultat sur tous les fichier test dans la console
detectLang.show_all_result()
| <commit_before>import detectLang
import graph
# ====================================================================================================
# La detection est rapide car toute les perplexites sont stockées dans les fichiers binaires pp_EN etc
# Pour regénerer les fichiers :
# Executer detectLang.create_all_pp_and_save_to_disc()
# Les resultats des perplexités pour tous les fichiers seront serialisés dans les fichiers binaires.
# ces fichiers sont necessaire à la fontions detectLang.detect_language()
# Pour serialisé les dictionaires j'utilise la bibliotheque intégrer à python => pickle
# ====================================================================================================
test_file_number = 19
# detect_language(Numéro du fichier dans le repertoire de test, N du modele nGram)
# Print le code de la langue reconnue et la perplexité du modèle choisis.
detectLang.detect_language(test_file_number,1)
detectLang.detect_language(test_file_number,2)
detectLang.detect_language(test_file_number,3)
# ====================================================================================================
# Nécessite matplotlib
# Installé matplotlib avec => pip install matplotlib
# ====================================================================================================
# graphFile(Numéro du fichier dans le repertoire de test)
graph.graphFile(test_file_number)
<commit_msg>Update doc and add one call<commit_after>import detectLang
import graph
# ====================================================================================================
# La detection est rapide car toute les perplexites sont stockées dans les fichiers binaires pp_EN etc
# Pour regénerer les fichiers :
# Executer detectLang.create_all_pp_and_save_to_disc(), décommenter la ligne suivante
# detectLang.create_all_pp_and_save_to_disc()
# Les resultats des perplexités pour tous les fichiers seront serialisés dans les fichiers binaires.
# ces fichiers sont necessaire à la fontions detectLang.detect_language()
# Pour serialisé les dictionaires j'utilise la bibliotheque intégrer à python => pickle
# ====================================================================================================
test_file_number = 13
# detect_language(Numéro du fichier dans le repertoire de test, N du modele nGram)
# Print le code de la langue reconnue et la perplexité du modèle choisis.
# detectLang.detect_language(test_file_number,1)
# detectLang.detect_language(test_file_number,2)
# detectLang.detect_language(test_file_number,3)
# ====================================================================================================
# Nécessite matplotlib
# Installé matplotlib avec => pip install matplotlib
# ====================================================================================================
# Affiche sur un graphique les perplexité de tous les modeles sur un même fichier
# graph.graphFile(test_file_number)
# Pour donner le resultat sur tous les fichier test dans la console
detectLang.show_all_result()
|
19cb68209252615c66cee0a1c6df1069f81f6f77 | stock_request_picking_type/models/stock_request_order.py | stock_request_picking_type/models/stock_request_order.py | # Copyright 2019 Open Source Integrators
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockRequestOrder(models.Model):
_inherit = 'stock.request.order'
@api.model
def _get_default_picking_type(self):
return self.env['stock.picking.type'].search([
('code', '=', 'stock_request_order'),
('warehouse_id.company_id', 'in',
[self.env.context.get('company_id', self.env.user.company_id.id),
False])],
limit=1).id
picking_type_id = fields.Many2one(
'stock.picking.type', 'Operation Type',
default=_get_default_picking_type, required=True)
@api.onchange('warehouse_id')
def onchange_warehouse_picking_id(self):
if self.warehouse_id:
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', self.warehouse_id.id)], limit=1)
if picking_type_id:
self._origin.write({'picking_type_id': picking_type_id.id})
| # Copyright 2019 Open Source Integrators
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockRequestOrder(models.Model):
_inherit = 'stock.request.order'
@api.model
def _get_default_picking_type(self):
return self.env['stock.picking.type'].search([
('code', '=', 'stock_request_order'),
('warehouse_id.company_id', 'in',
[self.env.context.get('company_id', self.env.user.company_id.id),
False])],
limit=1).id
picking_type_id = fields.Many2one(
'stock.picking.type', 'Operation Type',
default=_get_default_picking_type, required=True)
@api.onchange('warehouse_id')
def onchange_warehouse_picking_id(self):
if self.warehouse_id:
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', self.warehouse_id.id)], limit=1)
if picking_type_id:
self._origin.write({'picking_type_id': picking_type_id.id})
@api.model
def create(self, vals):
if vals.get('warehouse_id', False):
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', vals['warehouse_id'])], limit=1)
if picking_type_id:
vals.update({'picking_type_id': picking_type_id.id})
return super().create(vals)
| Set Picking Type in Create | [IMP] Set Picking Type in Create
[IMP] Flake8
| Python | agpl-3.0 | Vauxoo/stock-logistics-warehouse,Vauxoo/stock-logistics-warehouse,Vauxoo/stock-logistics-warehouse | # Copyright 2019 Open Source Integrators
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockRequestOrder(models.Model):
_inherit = 'stock.request.order'
@api.model
def _get_default_picking_type(self):
return self.env['stock.picking.type'].search([
('code', '=', 'stock_request_order'),
('warehouse_id.company_id', 'in',
[self.env.context.get('company_id', self.env.user.company_id.id),
False])],
limit=1).id
picking_type_id = fields.Many2one(
'stock.picking.type', 'Operation Type',
default=_get_default_picking_type, required=True)
@api.onchange('warehouse_id')
def onchange_warehouse_picking_id(self):
if self.warehouse_id:
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', self.warehouse_id.id)], limit=1)
if picking_type_id:
self._origin.write({'picking_type_id': picking_type_id.id})
[IMP] Set Picking Type in Create
[IMP] Flake8 | # Copyright 2019 Open Source Integrators
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockRequestOrder(models.Model):
_inherit = 'stock.request.order'
@api.model
def _get_default_picking_type(self):
return self.env['stock.picking.type'].search([
('code', '=', 'stock_request_order'),
('warehouse_id.company_id', 'in',
[self.env.context.get('company_id', self.env.user.company_id.id),
False])],
limit=1).id
picking_type_id = fields.Many2one(
'stock.picking.type', 'Operation Type',
default=_get_default_picking_type, required=True)
@api.onchange('warehouse_id')
def onchange_warehouse_picking_id(self):
if self.warehouse_id:
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', self.warehouse_id.id)], limit=1)
if picking_type_id:
self._origin.write({'picking_type_id': picking_type_id.id})
@api.model
def create(self, vals):
if vals.get('warehouse_id', False):
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', vals['warehouse_id'])], limit=1)
if picking_type_id:
vals.update({'picking_type_id': picking_type_id.id})
return super().create(vals)
| <commit_before># Copyright 2019 Open Source Integrators
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockRequestOrder(models.Model):
_inherit = 'stock.request.order'
@api.model
def _get_default_picking_type(self):
return self.env['stock.picking.type'].search([
('code', '=', 'stock_request_order'),
('warehouse_id.company_id', 'in',
[self.env.context.get('company_id', self.env.user.company_id.id),
False])],
limit=1).id
picking_type_id = fields.Many2one(
'stock.picking.type', 'Operation Type',
default=_get_default_picking_type, required=True)
@api.onchange('warehouse_id')
def onchange_warehouse_picking_id(self):
if self.warehouse_id:
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', self.warehouse_id.id)], limit=1)
if picking_type_id:
self._origin.write({'picking_type_id': picking_type_id.id})
<commit_msg>[IMP] Set Picking Type in Create
[IMP] Flake8<commit_after> | # Copyright 2019 Open Source Integrators
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockRequestOrder(models.Model):
_inherit = 'stock.request.order'
@api.model
def _get_default_picking_type(self):
return self.env['stock.picking.type'].search([
('code', '=', 'stock_request_order'),
('warehouse_id.company_id', 'in',
[self.env.context.get('company_id', self.env.user.company_id.id),
False])],
limit=1).id
picking_type_id = fields.Many2one(
'stock.picking.type', 'Operation Type',
default=_get_default_picking_type, required=True)
@api.onchange('warehouse_id')
def onchange_warehouse_picking_id(self):
if self.warehouse_id:
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', self.warehouse_id.id)], limit=1)
if picking_type_id:
self._origin.write({'picking_type_id': picking_type_id.id})
@api.model
def create(self, vals):
if vals.get('warehouse_id', False):
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', vals['warehouse_id'])], limit=1)
if picking_type_id:
vals.update({'picking_type_id': picking_type_id.id})
return super().create(vals)
| # Copyright 2019 Open Source Integrators
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockRequestOrder(models.Model):
_inherit = 'stock.request.order'
@api.model
def _get_default_picking_type(self):
return self.env['stock.picking.type'].search([
('code', '=', 'stock_request_order'),
('warehouse_id.company_id', 'in',
[self.env.context.get('company_id', self.env.user.company_id.id),
False])],
limit=1).id
picking_type_id = fields.Many2one(
'stock.picking.type', 'Operation Type',
default=_get_default_picking_type, required=True)
@api.onchange('warehouse_id')
def onchange_warehouse_picking_id(self):
if self.warehouse_id:
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', self.warehouse_id.id)], limit=1)
if picking_type_id:
self._origin.write({'picking_type_id': picking_type_id.id})
[IMP] Set Picking Type in Create
[IMP] Flake8# Copyright 2019 Open Source Integrators
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockRequestOrder(models.Model):
_inherit = 'stock.request.order'
@api.model
def _get_default_picking_type(self):
return self.env['stock.picking.type'].search([
('code', '=', 'stock_request_order'),
('warehouse_id.company_id', 'in',
[self.env.context.get('company_id', self.env.user.company_id.id),
False])],
limit=1).id
picking_type_id = fields.Many2one(
'stock.picking.type', 'Operation Type',
default=_get_default_picking_type, required=True)
@api.onchange('warehouse_id')
def onchange_warehouse_picking_id(self):
if self.warehouse_id:
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', self.warehouse_id.id)], limit=1)
if picking_type_id:
self._origin.write({'picking_type_id': picking_type_id.id})
@api.model
def create(self, vals):
if vals.get('warehouse_id', False):
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', vals['warehouse_id'])], limit=1)
if picking_type_id:
vals.update({'picking_type_id': picking_type_id.id})
return super().create(vals)
| <commit_before># Copyright 2019 Open Source Integrators
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockRequestOrder(models.Model):
_inherit = 'stock.request.order'
@api.model
def _get_default_picking_type(self):
return self.env['stock.picking.type'].search([
('code', '=', 'stock_request_order'),
('warehouse_id.company_id', 'in',
[self.env.context.get('company_id', self.env.user.company_id.id),
False])],
limit=1).id
picking_type_id = fields.Many2one(
'stock.picking.type', 'Operation Type',
default=_get_default_picking_type, required=True)
@api.onchange('warehouse_id')
def onchange_warehouse_picking_id(self):
if self.warehouse_id:
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', self.warehouse_id.id)], limit=1)
if picking_type_id:
self._origin.write({'picking_type_id': picking_type_id.id})
<commit_msg>[IMP] Set Picking Type in Create
[IMP] Flake8<commit_after># Copyright 2019 Open Source Integrators
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockRequestOrder(models.Model):
_inherit = 'stock.request.order'
@api.model
def _get_default_picking_type(self):
return self.env['stock.picking.type'].search([
('code', '=', 'stock_request_order'),
('warehouse_id.company_id', 'in',
[self.env.context.get('company_id', self.env.user.company_id.id),
False])],
limit=1).id
picking_type_id = fields.Many2one(
'stock.picking.type', 'Operation Type',
default=_get_default_picking_type, required=True)
@api.onchange('warehouse_id')
def onchange_warehouse_picking_id(self):
if self.warehouse_id:
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', self.warehouse_id.id)], limit=1)
if picking_type_id:
self._origin.write({'picking_type_id': picking_type_id.id})
@api.model
def create(self, vals):
if vals.get('warehouse_id', False):
picking_type_id = self.env['stock.picking.type'].\
search([('code', '=', 'stock_request_order'),
('warehouse_id', '=', vals['warehouse_id'])], limit=1)
if picking_type_id:
vals.update({'picking_type_id': picking_type_id.id})
return super().create(vals)
|
b6139583bf5074c73c0de6626391b6f128ed6e34 | export_jars.py | export_jars.py | #!/usr/bin/env python
import os
import shutil
from glob import glob
from subprocess import call, check_output
OUTPUT_DIR_NAME = 'jars'
def call_unsafe(*args, **kwargs):
kwargs['shell'] = True
call(*args, **kwargs)
call_unsafe('./gradlew clean javadocRelease jarRelease')
try:
os.mkdir(OUTPUT_DIR_NAME)
except OSError:
pass
os.chdir(OUTPUT_DIR_NAME)
call_unsafe('cp ../beansdk/build/libs/*.jar .')
commit = check_output(['git', 'rev-parse', 'HEAD'])[:7]
for src in glob('*.jar'):
name, ext = os.path.splitext(src)
dest = name + '-' + commit + ext
shutil.move(src, dest)
call_unsafe('open .')
| #!/usr/bin/env python
import os
import shutil
from glob import glob
from subprocess import call, check_output
OUTPUT_DIR_NAME = 'jars'
def call_unsafe(*args, **kwargs):
kwargs['shell'] = True
call(*args, **kwargs)
call_unsafe('./gradlew clean javadocRelease jarRelease')
try:
os.mkdir(OUTPUT_DIR_NAME)
except OSError:
pass
os.chdir(OUTPUT_DIR_NAME)
call_unsafe('rm *.jar')
call_unsafe('cp ../beansdk/build/libs/*.jar .')
commit = check_output(['git', 'rev-parse', 'HEAD'])[:7]
for src in glob('*.jar'):
name, ext = os.path.splitext(src)
dest = name + '-' + commit + ext
shutil.move(src, dest)
call_unsafe('open .')
| Remove existing JARs before building new ones | Remove existing JARs before building new ones
| Python | mit | swstack/Bean-Android-SDK,PunchThrough/bean-sdk-android,colus001/Bean-Android-SDK,PunchThrough/Bean-Android-SDK,hongbinz/Bean-Android-SDK,androidgrl/Bean-Android-SDK,PunchThrough/Bean-Android-SDK,swstack/Bean-Android-SDK,PunchThrough/bean-sdk-android,androidgrl/Bean-Android-SDK,hongbinz/Bean-Android-SDK,colus001/Bean-Android-SDK | #!/usr/bin/env python
import os
import shutil
from glob import glob
from subprocess import call, check_output
OUTPUT_DIR_NAME = 'jars'
def call_unsafe(*args, **kwargs):
kwargs['shell'] = True
call(*args, **kwargs)
call_unsafe('./gradlew clean javadocRelease jarRelease')
try:
os.mkdir(OUTPUT_DIR_NAME)
except OSError:
pass
os.chdir(OUTPUT_DIR_NAME)
call_unsafe('cp ../beansdk/build/libs/*.jar .')
commit = check_output(['git', 'rev-parse', 'HEAD'])[:7]
for src in glob('*.jar'):
name, ext = os.path.splitext(src)
dest = name + '-' + commit + ext
shutil.move(src, dest)
call_unsafe('open .')
Remove existing JARs before building new ones | #!/usr/bin/env python
import os
import shutil
from glob import glob
from subprocess import call, check_output
OUTPUT_DIR_NAME = 'jars'
def call_unsafe(*args, **kwargs):
kwargs['shell'] = True
call(*args, **kwargs)
call_unsafe('./gradlew clean javadocRelease jarRelease')
try:
os.mkdir(OUTPUT_DIR_NAME)
except OSError:
pass
os.chdir(OUTPUT_DIR_NAME)
call_unsafe('rm *.jar')
call_unsafe('cp ../beansdk/build/libs/*.jar .')
commit = check_output(['git', 'rev-parse', 'HEAD'])[:7]
for src in glob('*.jar'):
name, ext = os.path.splitext(src)
dest = name + '-' + commit + ext
shutil.move(src, dest)
call_unsafe('open .')
| <commit_before>#!/usr/bin/env python
import os
import shutil
from glob import glob
from subprocess import call, check_output
OUTPUT_DIR_NAME = 'jars'
def call_unsafe(*args, **kwargs):
kwargs['shell'] = True
call(*args, **kwargs)
call_unsafe('./gradlew clean javadocRelease jarRelease')
try:
os.mkdir(OUTPUT_DIR_NAME)
except OSError:
pass
os.chdir(OUTPUT_DIR_NAME)
call_unsafe('cp ../beansdk/build/libs/*.jar .')
commit = check_output(['git', 'rev-parse', 'HEAD'])[:7]
for src in glob('*.jar'):
name, ext = os.path.splitext(src)
dest = name + '-' + commit + ext
shutil.move(src, dest)
call_unsafe('open .')
<commit_msg>Remove existing JARs before building new ones<commit_after> | #!/usr/bin/env python
import os
import shutil
from glob import glob
from subprocess import call, check_output
OUTPUT_DIR_NAME = 'jars'
def call_unsafe(*args, **kwargs):
kwargs['shell'] = True
call(*args, **kwargs)
call_unsafe('./gradlew clean javadocRelease jarRelease')
try:
os.mkdir(OUTPUT_DIR_NAME)
except OSError:
pass
os.chdir(OUTPUT_DIR_NAME)
call_unsafe('rm *.jar')
call_unsafe('cp ../beansdk/build/libs/*.jar .')
commit = check_output(['git', 'rev-parse', 'HEAD'])[:7]
for src in glob('*.jar'):
name, ext = os.path.splitext(src)
dest = name + '-' + commit + ext
shutil.move(src, dest)
call_unsafe('open .')
| #!/usr/bin/env python
import os
import shutil
from glob import glob
from subprocess import call, check_output
OUTPUT_DIR_NAME = 'jars'
def call_unsafe(*args, **kwargs):
kwargs['shell'] = True
call(*args, **kwargs)
call_unsafe('./gradlew clean javadocRelease jarRelease')
try:
os.mkdir(OUTPUT_DIR_NAME)
except OSError:
pass
os.chdir(OUTPUT_DIR_NAME)
call_unsafe('cp ../beansdk/build/libs/*.jar .')
commit = check_output(['git', 'rev-parse', 'HEAD'])[:7]
for src in glob('*.jar'):
name, ext = os.path.splitext(src)
dest = name + '-' + commit + ext
shutil.move(src, dest)
call_unsafe('open .')
Remove existing JARs before building new ones#!/usr/bin/env python
import os
import shutil
from glob import glob
from subprocess import call, check_output
OUTPUT_DIR_NAME = 'jars'
def call_unsafe(*args, **kwargs):
kwargs['shell'] = True
call(*args, **kwargs)
call_unsafe('./gradlew clean javadocRelease jarRelease')
try:
os.mkdir(OUTPUT_DIR_NAME)
except OSError:
pass
os.chdir(OUTPUT_DIR_NAME)
call_unsafe('rm *.jar')
call_unsafe('cp ../beansdk/build/libs/*.jar .')
commit = check_output(['git', 'rev-parse', 'HEAD'])[:7]
for src in glob('*.jar'):
name, ext = os.path.splitext(src)
dest = name + '-' + commit + ext
shutil.move(src, dest)
call_unsafe('open .')
| <commit_before>#!/usr/bin/env python
import os
import shutil
from glob import glob
from subprocess import call, check_output
OUTPUT_DIR_NAME = 'jars'
def call_unsafe(*args, **kwargs):
kwargs['shell'] = True
call(*args, **kwargs)
call_unsafe('./gradlew clean javadocRelease jarRelease')
try:
os.mkdir(OUTPUT_DIR_NAME)
except OSError:
pass
os.chdir(OUTPUT_DIR_NAME)
call_unsafe('cp ../beansdk/build/libs/*.jar .')
commit = check_output(['git', 'rev-parse', 'HEAD'])[:7]
for src in glob('*.jar'):
name, ext = os.path.splitext(src)
dest = name + '-' + commit + ext
shutil.move(src, dest)
call_unsafe('open .')
<commit_msg>Remove existing JARs before building new ones<commit_after>#!/usr/bin/env python
import os
import shutil
from glob import glob
from subprocess import call, check_output
OUTPUT_DIR_NAME = 'jars'
def call_unsafe(*args, **kwargs):
kwargs['shell'] = True
call(*args, **kwargs)
call_unsafe('./gradlew clean javadocRelease jarRelease')
try:
os.mkdir(OUTPUT_DIR_NAME)
except OSError:
pass
os.chdir(OUTPUT_DIR_NAME)
call_unsafe('rm *.jar')
call_unsafe('cp ../beansdk/build/libs/*.jar .')
commit = check_output(['git', 'rev-parse', 'HEAD'])[:7]
for src in glob('*.jar'):
name, ext = os.path.splitext(src)
dest = name + '-' + commit + ext
shutil.move(src, dest)
call_unsafe('open .')
|
1b7a3f045bf7a23ef993d136b481f22258c4a778 | wagtail/wagtailimages/rich_text.py | wagtail/wagtailimages/rich_text.py | from wagtail.wagtailimages.models import get_image_model
from wagtail.wagtailimages.formats import get_image_format
class ImageEmbedHandler(object):
"""
ImageEmbedHandler will be invoked whenever we encounter an element in HTML content
with an attribute of data-embedtype="image". The resulting element in the database
representation will be:
<embed embedtype="image" id="42" format="thumb" alt="some custom alt text">
"""
@staticmethod
def get_db_attributes(tag):
"""
Given a tag that we've identified as an image embed (because it has a
data-embedtype="image" attribute), return a dict of the attributes we should
have on the resulting <embed> element.
"""
return {
'id': tag['data-id'],
'format': tag['data-format'],
'alt': tag['data-alt'],
}
@staticmethod
def expand_db_attributes(attrs, for_editor):
"""
Given a dict of attributes from the <embed> tag, return the real HTML
representation.
"""
Image = get_image_model()
try:
image = Image.objects.get(id=attrs['id'])
image_format = get_image_format(attrs['format'])
if for_editor:
try:
return image_format.image_to_editor_html(image, attrs['alt'])
except:
return ''
else:
return image_format.image_to_html(image, attrs['alt'])
except Image.DoesNotExist:
return "<img>"
| from wagtail.wagtailimages.models import get_image_model
from wagtail.wagtailimages.formats import get_image_format
class ImageEmbedHandler(object):
"""
ImageEmbedHandler will be invoked whenever we encounter an element in HTML content
with an attribute of data-embedtype="image". The resulting element in the database
representation will be:
<embed embedtype="image" id="42" format="thumb" alt="some custom alt text">
"""
@staticmethod
def get_db_attributes(tag):
"""
Given a tag that we've identified as an image embed (because it has a
data-embedtype="image" attribute), return a dict of the attributes we should
have on the resulting <embed> element.
"""
return {
'id': tag['data-id'],
'format': tag['data-format'],
'alt': tag['data-alt'],
}
@staticmethod
def expand_db_attributes(attrs, for_editor):
"""
Given a dict of attributes from the <embed> tag, return the real HTML
representation.
"""
Image = get_image_model()
try:
image = Image.objects.get(id=attrs['id'])
except Image.DoesNotExist:
return "<img>"
image_format = get_image_format(attrs['format'])
if for_editor:
try:
return image_format.image_to_editor_html(image, attrs['alt'])
except:
return ''
else:
return image_format.image_to_html(image, attrs['alt'])
| Refactor try-catch block by limiting code in the try block | Refactor try-catch block by limiting code in the try block
Always good to know which line will raise an exception and limit the try block to that statement
| Python | bsd-3-clause | Toshakins/wagtail,timorieber/wagtail,nrsimha/wagtail,kurtrwall/wagtail,timorieber/wagtail,FlipperPA/wagtail,inonit/wagtail,davecranwell/wagtail,nealtodd/wagtail,iansprice/wagtail,thenewguy/wagtail,nutztherookie/wagtail,jnns/wagtail,kaedroho/wagtail,iansprice/wagtail,serzans/wagtail,inonit/wagtail,kurtw/wagtail,mixxorz/wagtail,takeflight/wagtail,jnns/wagtail,serzans/wagtail,kaedroho/wagtail,Toshakins/wagtail,wagtail/wagtail,mixxorz/wagtail,thenewguy/wagtail,chrxr/wagtail,thenewguy/wagtail,JoshBarr/wagtail,quru/wagtail,FlipperPA/wagtail,timorieber/wagtail,nimasmi/wagtail,Tivix/wagtail,nutztherookie/wagtail,wagtail/wagtail,mixxorz/wagtail,Tivix/wagtail,thenewguy/wagtail,nealtodd/wagtail,hamsterbacke23/wagtail,kaedroho/wagtail,gasman/wagtail,hamsterbacke23/wagtail,mikedingjan/wagtail,davecranwell/wagtail,nrsimha/wagtail,gogobook/wagtail,chrxr/wagtail,nimasmi/wagtail,zerolab/wagtail,timorieber/wagtail,mayapurmedia/wagtail,kurtw/wagtail,chrxr/wagtail,wagtail/wagtail,gasman/wagtail,nilnvoid/wagtail,wagtail/wagtail,nealtodd/wagtail,zerolab/wagtail,nimasmi/wagtail,jnns/wagtail,chrxr/wagtail,zerolab/wagtail,takeflight/wagtail,quru/wagtail,inonit/wagtail,nilnvoid/wagtail,rsalmaso/wagtail,mikedingjan/wagtail,nrsimha/wagtail,nutztherookie/wagtail,Tivix/wagtail,hanpama/wagtail,rsalmaso/wagtail,gogobook/wagtail,inonit/wagtail,mayapurmedia/wagtail,JoshBarr/wagtail,iansprice/wagtail,jnns/wagtail,Toshakins/wagtail,quru/wagtail,nilnvoid/wagtail,kurtrwall/wagtail,nutztherookie/wagtail,hamsterbacke23/wagtail,quru/wagtail,gasman/wagtail,FlipperPA/wagtail,kurtrwall/wagtail,hamsterbacke23/wagtail,serzans/wagtail,FlipperPA/wagtail,JoshBarr/wagtail,rsalmaso/wagtail,Toshakins/wagtail,mayapurmedia/wagtail,thenewguy/wagtail,mikedingjan/wagtail,hanpama/wagtail,mixxorz/wagtail,mikedingjan/wagtail,torchbox/wagtail,gasman/wagtail,zerolab/wagtail,mayapurmedia/wagtail,mixxorz/wagtail,gogobook/wagtail,rsalmaso/wagtail,hanpama/wagtail,serzans/wagtail,gogobook/wagtail,takeflight/wagtail,torchbox/wagtail,torchbox/wagtail,zerolab/wagtail,torchbox/wagtail,wagtail/wagtail,kaedroho/wagtail,JoshBarr/wagtail,rsalmaso/wagtail,gasman/wagtail,Tivix/wagtail,takeflight/wagtail,nrsimha/wagtail,iansprice/wagtail,hanpama/wagtail,nealtodd/wagtail,kurtrwall/wagtail,nimasmi/wagtail,nilnvoid/wagtail,kurtw/wagtail,davecranwell/wagtail,kurtw/wagtail,kaedroho/wagtail,davecranwell/wagtail | from wagtail.wagtailimages.models import get_image_model
from wagtail.wagtailimages.formats import get_image_format
class ImageEmbedHandler(object):
"""
ImageEmbedHandler will be invoked whenever we encounter an element in HTML content
with an attribute of data-embedtype="image". The resulting element in the database
representation will be:
<embed embedtype="image" id="42" format="thumb" alt="some custom alt text">
"""
@staticmethod
def get_db_attributes(tag):
"""
Given a tag that we've identified as an image embed (because it has a
data-embedtype="image" attribute), return a dict of the attributes we should
have on the resulting <embed> element.
"""
return {
'id': tag['data-id'],
'format': tag['data-format'],
'alt': tag['data-alt'],
}
@staticmethod
def expand_db_attributes(attrs, for_editor):
"""
Given a dict of attributes from the <embed> tag, return the real HTML
representation.
"""
Image = get_image_model()
try:
image = Image.objects.get(id=attrs['id'])
image_format = get_image_format(attrs['format'])
if for_editor:
try:
return image_format.image_to_editor_html(image, attrs['alt'])
except:
return ''
else:
return image_format.image_to_html(image, attrs['alt'])
except Image.DoesNotExist:
return "<img>"
Refactor try-catch block by limiting code in the try block
Always good to know which line will raise an exception and limit the try block to that statement | from wagtail.wagtailimages.models import get_image_model
from wagtail.wagtailimages.formats import get_image_format
class ImageEmbedHandler(object):
"""
ImageEmbedHandler will be invoked whenever we encounter an element in HTML content
with an attribute of data-embedtype="image". The resulting element in the database
representation will be:
<embed embedtype="image" id="42" format="thumb" alt="some custom alt text">
"""
@staticmethod
def get_db_attributes(tag):
"""
Given a tag that we've identified as an image embed (because it has a
data-embedtype="image" attribute), return a dict of the attributes we should
have on the resulting <embed> element.
"""
return {
'id': tag['data-id'],
'format': tag['data-format'],
'alt': tag['data-alt'],
}
@staticmethod
def expand_db_attributes(attrs, for_editor):
"""
Given a dict of attributes from the <embed> tag, return the real HTML
representation.
"""
Image = get_image_model()
try:
image = Image.objects.get(id=attrs['id'])
except Image.DoesNotExist:
return "<img>"
image_format = get_image_format(attrs['format'])
if for_editor:
try:
return image_format.image_to_editor_html(image, attrs['alt'])
except:
return ''
else:
return image_format.image_to_html(image, attrs['alt'])
| <commit_before>from wagtail.wagtailimages.models import get_image_model
from wagtail.wagtailimages.formats import get_image_format
class ImageEmbedHandler(object):
"""
ImageEmbedHandler will be invoked whenever we encounter an element in HTML content
with an attribute of data-embedtype="image". The resulting element in the database
representation will be:
<embed embedtype="image" id="42" format="thumb" alt="some custom alt text">
"""
@staticmethod
def get_db_attributes(tag):
"""
Given a tag that we've identified as an image embed (because it has a
data-embedtype="image" attribute), return a dict of the attributes we should
have on the resulting <embed> element.
"""
return {
'id': tag['data-id'],
'format': tag['data-format'],
'alt': tag['data-alt'],
}
@staticmethod
def expand_db_attributes(attrs, for_editor):
"""
Given a dict of attributes from the <embed> tag, return the real HTML
representation.
"""
Image = get_image_model()
try:
image = Image.objects.get(id=attrs['id'])
image_format = get_image_format(attrs['format'])
if for_editor:
try:
return image_format.image_to_editor_html(image, attrs['alt'])
except:
return ''
else:
return image_format.image_to_html(image, attrs['alt'])
except Image.DoesNotExist:
return "<img>"
<commit_msg>Refactor try-catch block by limiting code in the try block
Always good to know which line will raise an exception and limit the try block to that statement<commit_after> | from wagtail.wagtailimages.models import get_image_model
from wagtail.wagtailimages.formats import get_image_format
class ImageEmbedHandler(object):
"""
ImageEmbedHandler will be invoked whenever we encounter an element in HTML content
with an attribute of data-embedtype="image". The resulting element in the database
representation will be:
<embed embedtype="image" id="42" format="thumb" alt="some custom alt text">
"""
@staticmethod
def get_db_attributes(tag):
"""
Given a tag that we've identified as an image embed (because it has a
data-embedtype="image" attribute), return a dict of the attributes we should
have on the resulting <embed> element.
"""
return {
'id': tag['data-id'],
'format': tag['data-format'],
'alt': tag['data-alt'],
}
@staticmethod
def expand_db_attributes(attrs, for_editor):
"""
Given a dict of attributes from the <embed> tag, return the real HTML
representation.
"""
Image = get_image_model()
try:
image = Image.objects.get(id=attrs['id'])
except Image.DoesNotExist:
return "<img>"
image_format = get_image_format(attrs['format'])
if for_editor:
try:
return image_format.image_to_editor_html(image, attrs['alt'])
except:
return ''
else:
return image_format.image_to_html(image, attrs['alt'])
| from wagtail.wagtailimages.models import get_image_model
from wagtail.wagtailimages.formats import get_image_format
class ImageEmbedHandler(object):
"""
ImageEmbedHandler will be invoked whenever we encounter an element in HTML content
with an attribute of data-embedtype="image". The resulting element in the database
representation will be:
<embed embedtype="image" id="42" format="thumb" alt="some custom alt text">
"""
@staticmethod
def get_db_attributes(tag):
"""
Given a tag that we've identified as an image embed (because it has a
data-embedtype="image" attribute), return a dict of the attributes we should
have on the resulting <embed> element.
"""
return {
'id': tag['data-id'],
'format': tag['data-format'],
'alt': tag['data-alt'],
}
@staticmethod
def expand_db_attributes(attrs, for_editor):
"""
Given a dict of attributes from the <embed> tag, return the real HTML
representation.
"""
Image = get_image_model()
try:
image = Image.objects.get(id=attrs['id'])
image_format = get_image_format(attrs['format'])
if for_editor:
try:
return image_format.image_to_editor_html(image, attrs['alt'])
except:
return ''
else:
return image_format.image_to_html(image, attrs['alt'])
except Image.DoesNotExist:
return "<img>"
Refactor try-catch block by limiting code in the try block
Always good to know which line will raise an exception and limit the try block to that statementfrom wagtail.wagtailimages.models import get_image_model
from wagtail.wagtailimages.formats import get_image_format
class ImageEmbedHandler(object):
"""
ImageEmbedHandler will be invoked whenever we encounter an element in HTML content
with an attribute of data-embedtype="image". The resulting element in the database
representation will be:
<embed embedtype="image" id="42" format="thumb" alt="some custom alt text">
"""
@staticmethod
def get_db_attributes(tag):
"""
Given a tag that we've identified as an image embed (because it has a
data-embedtype="image" attribute), return a dict of the attributes we should
have on the resulting <embed> element.
"""
return {
'id': tag['data-id'],
'format': tag['data-format'],
'alt': tag['data-alt'],
}
@staticmethod
def expand_db_attributes(attrs, for_editor):
"""
Given a dict of attributes from the <embed> tag, return the real HTML
representation.
"""
Image = get_image_model()
try:
image = Image.objects.get(id=attrs['id'])
except Image.DoesNotExist:
return "<img>"
image_format = get_image_format(attrs['format'])
if for_editor:
try:
return image_format.image_to_editor_html(image, attrs['alt'])
except:
return ''
else:
return image_format.image_to_html(image, attrs['alt'])
| <commit_before>from wagtail.wagtailimages.models import get_image_model
from wagtail.wagtailimages.formats import get_image_format
class ImageEmbedHandler(object):
"""
ImageEmbedHandler will be invoked whenever we encounter an element in HTML content
with an attribute of data-embedtype="image". The resulting element in the database
representation will be:
<embed embedtype="image" id="42" format="thumb" alt="some custom alt text">
"""
@staticmethod
def get_db_attributes(tag):
"""
Given a tag that we've identified as an image embed (because it has a
data-embedtype="image" attribute), return a dict of the attributes we should
have on the resulting <embed> element.
"""
return {
'id': tag['data-id'],
'format': tag['data-format'],
'alt': tag['data-alt'],
}
@staticmethod
def expand_db_attributes(attrs, for_editor):
"""
Given a dict of attributes from the <embed> tag, return the real HTML
representation.
"""
Image = get_image_model()
try:
image = Image.objects.get(id=attrs['id'])
image_format = get_image_format(attrs['format'])
if for_editor:
try:
return image_format.image_to_editor_html(image, attrs['alt'])
except:
return ''
else:
return image_format.image_to_html(image, attrs['alt'])
except Image.DoesNotExist:
return "<img>"
<commit_msg>Refactor try-catch block by limiting code in the try block
Always good to know which line will raise an exception and limit the try block to that statement<commit_after>from wagtail.wagtailimages.models import get_image_model
from wagtail.wagtailimages.formats import get_image_format
class ImageEmbedHandler(object):
"""
ImageEmbedHandler will be invoked whenever we encounter an element in HTML content
with an attribute of data-embedtype="image". The resulting element in the database
representation will be:
<embed embedtype="image" id="42" format="thumb" alt="some custom alt text">
"""
@staticmethod
def get_db_attributes(tag):
"""
Given a tag that we've identified as an image embed (because it has a
data-embedtype="image" attribute), return a dict of the attributes we should
have on the resulting <embed> element.
"""
return {
'id': tag['data-id'],
'format': tag['data-format'],
'alt': tag['data-alt'],
}
@staticmethod
def expand_db_attributes(attrs, for_editor):
"""
Given a dict of attributes from the <embed> tag, return the real HTML
representation.
"""
Image = get_image_model()
try:
image = Image.objects.get(id=attrs['id'])
except Image.DoesNotExist:
return "<img>"
image_format = get_image_format(attrs['format'])
if for_editor:
try:
return image_format.image_to_editor_html(image, attrs['alt'])
except:
return ''
else:
return image_format.image_to_html(image, attrs['alt'])
|
7e1ec1b27d69882005ac5492809c8847c21e2198 | baro.py | baro.py | from datetime import datetime
class Baro:
"""This class represents a Baro item and is initialized with
data in JSON format
"""
def __init__(self, data):
self.config = data['Config']
self.start = datetime.fromtimestamp(data['Activation']['sec'])
self.end = datetime.fromtimestamp(data['Expiry']['sec'])
self.location = data['Node']
self.manifest = data['Manifest']
def __str__(self):
"""Returns a string with all the information about Baro offer
"""
baroItemString = ""
if datetime.now() < self.start:
return "None"
else:
for item in self.manifest:
baroItemString += ('== '+ str(item["ItemType"]) +' ==\n'
'- price: '+ str(item["PrimePrice"]) +' ducats + '+ str(item["RegularPrice"]) +'cr -\n\n' )
return baroItemString
def get_eta_string(self):
"""Returns a string containing the Baro's ETA
"""
seconds = int((self.end - datetime.now()).total_seconds())
return '{} days, {} hrs, {} mins'.format((seconds // 86400), ((seconds % 86400) // 3600),
(seconds % 3600) // 60)
def get_start_string(self):
"""Returns a string containing the Baro's start
"""
seconds = int((self.start - datetime.now()).total_seconds())
return '{} days, {} hrs, {} mins'.format((seconds // 86400), ((seconds % 86400) // 3600),
(seconds % 3600) // 60) | from datetime import datetime
import utils
class Baro:
"""This class contains info about the Void Trader and is initialized with
data in JSON format
"""
def __init__(self, data):
self.config = data['Config']
self.start = datetime.fromtimestamp(data['Activation']['sec'])
self.end = datetime.fromtimestamp(data['Expiry']['sec'])
self.location = data['Node']
self.manifest = data['Manifest']
def __str__(self):
"""Returns a string with all the information about Baro's offers
"""
baroItemString = ""
if datetime.now() < self.start:
return "None"
else:
for item in self.manifest:
baroItemString += ('== '+ str(item["ItemType"]) +' ==\n'
'- price: '+ str(item["PrimePrice"]) +' ducats + '+ str(item["RegularPrice"]) +'cr -\n\n' )
return baroItemString
def get_end_string(self):
"""Returns a string containing Baro's departure time
"""
return timedelta_to_string(self.end - datetime.now())
def get_start_string(self):
"""Returns a string containing Baro's arrival time
"""
return timedelta_to_string(self.start - datetime.now())
| Change class Baro to use timedelta_to_string, some fixes | Change class Baro to use timedelta_to_string, some fixes
| Python | mit | pabletos/Hubot-Warframe,pabletos/Hubot-Warframe | from datetime import datetime
class Baro:
"""This class represents a Baro item and is initialized with
data in JSON format
"""
def __init__(self, data):
self.config = data['Config']
self.start = datetime.fromtimestamp(data['Activation']['sec'])
self.end = datetime.fromtimestamp(data['Expiry']['sec'])
self.location = data['Node']
self.manifest = data['Manifest']
def __str__(self):
"""Returns a string with all the information about Baro offer
"""
baroItemString = ""
if datetime.now() < self.start:
return "None"
else:
for item in self.manifest:
baroItemString += ('== '+ str(item["ItemType"]) +' ==\n'
'- price: '+ str(item["PrimePrice"]) +' ducats + '+ str(item["RegularPrice"]) +'cr -\n\n' )
return baroItemString
def get_eta_string(self):
"""Returns a string containing the Baro's ETA
"""
seconds = int((self.end - datetime.now()).total_seconds())
return '{} days, {} hrs, {} mins'.format((seconds // 86400), ((seconds % 86400) // 3600),
(seconds % 3600) // 60)
def get_start_string(self):
"""Returns a string containing the Baro's start
"""
seconds = int((self.start - datetime.now()).total_seconds())
return '{} days, {} hrs, {} mins'.format((seconds // 86400), ((seconds % 86400) // 3600),
(seconds % 3600) // 60)Change class Baro to use timedelta_to_string, some fixes | from datetime import datetime
import utils
class Baro:
"""This class contains info about the Void Trader and is initialized with
data in JSON format
"""
def __init__(self, data):
self.config = data['Config']
self.start = datetime.fromtimestamp(data['Activation']['sec'])
self.end = datetime.fromtimestamp(data['Expiry']['sec'])
self.location = data['Node']
self.manifest = data['Manifest']
def __str__(self):
"""Returns a string with all the information about Baro's offers
"""
baroItemString = ""
if datetime.now() < self.start:
return "None"
else:
for item in self.manifest:
baroItemString += ('== '+ str(item["ItemType"]) +' ==\n'
'- price: '+ str(item["PrimePrice"]) +' ducats + '+ str(item["RegularPrice"]) +'cr -\n\n' )
return baroItemString
def get_end_string(self):
"""Returns a string containing Baro's departure time
"""
return timedelta_to_string(self.end - datetime.now())
def get_start_string(self):
"""Returns a string containing Baro's arrival time
"""
return timedelta_to_string(self.start - datetime.now())
| <commit_before>from datetime import datetime
class Baro:
"""This class represents a Baro item and is initialized with
data in JSON format
"""
def __init__(self, data):
self.config = data['Config']
self.start = datetime.fromtimestamp(data['Activation']['sec'])
self.end = datetime.fromtimestamp(data['Expiry']['sec'])
self.location = data['Node']
self.manifest = data['Manifest']
def __str__(self):
"""Returns a string with all the information about Baro offer
"""
baroItemString = ""
if datetime.now() < self.start:
return "None"
else:
for item in self.manifest:
baroItemString += ('== '+ str(item["ItemType"]) +' ==\n'
'- price: '+ str(item["PrimePrice"]) +' ducats + '+ str(item["RegularPrice"]) +'cr -\n\n' )
return baroItemString
def get_eta_string(self):
"""Returns a string containing the Baro's ETA
"""
seconds = int((self.end - datetime.now()).total_seconds())
return '{} days, {} hrs, {} mins'.format((seconds // 86400), ((seconds % 86400) // 3600),
(seconds % 3600) // 60)
def get_start_string(self):
"""Returns a string containing the Baro's start
"""
seconds = int((self.start - datetime.now()).total_seconds())
return '{} days, {} hrs, {} mins'.format((seconds // 86400), ((seconds % 86400) // 3600),
(seconds % 3600) // 60)<commit_msg>Change class Baro to use timedelta_to_string, some fixes<commit_after> | from datetime import datetime
import utils
class Baro:
"""This class contains info about the Void Trader and is initialized with
data in JSON format
"""
def __init__(self, data):
self.config = data['Config']
self.start = datetime.fromtimestamp(data['Activation']['sec'])
self.end = datetime.fromtimestamp(data['Expiry']['sec'])
self.location = data['Node']
self.manifest = data['Manifest']
def __str__(self):
"""Returns a string with all the information about Baro's offers
"""
baroItemString = ""
if datetime.now() < self.start:
return "None"
else:
for item in self.manifest:
baroItemString += ('== '+ str(item["ItemType"]) +' ==\n'
'- price: '+ str(item["PrimePrice"]) +' ducats + '+ str(item["RegularPrice"]) +'cr -\n\n' )
return baroItemString
def get_end_string(self):
"""Returns a string containing Baro's departure time
"""
return timedelta_to_string(self.end - datetime.now())
def get_start_string(self):
"""Returns a string containing Baro's arrival time
"""
return timedelta_to_string(self.start - datetime.now())
| from datetime import datetime
class Baro:
"""This class represents a Baro item and is initialized with
data in JSON format
"""
def __init__(self, data):
self.config = data['Config']
self.start = datetime.fromtimestamp(data['Activation']['sec'])
self.end = datetime.fromtimestamp(data['Expiry']['sec'])
self.location = data['Node']
self.manifest = data['Manifest']
def __str__(self):
"""Returns a string with all the information about Baro offer
"""
baroItemString = ""
if datetime.now() < self.start:
return "None"
else:
for item in self.manifest:
baroItemString += ('== '+ str(item["ItemType"]) +' ==\n'
'- price: '+ str(item["PrimePrice"]) +' ducats + '+ str(item["RegularPrice"]) +'cr -\n\n' )
return baroItemString
def get_eta_string(self):
"""Returns a string containing the Baro's ETA
"""
seconds = int((self.end - datetime.now()).total_seconds())
return '{} days, {} hrs, {} mins'.format((seconds // 86400), ((seconds % 86400) // 3600),
(seconds % 3600) // 60)
def get_start_string(self):
"""Returns a string containing the Baro's start
"""
seconds = int((self.start - datetime.now()).total_seconds())
return '{} days, {} hrs, {} mins'.format((seconds // 86400), ((seconds % 86400) // 3600),
(seconds % 3600) // 60)Change class Baro to use timedelta_to_string, some fixesfrom datetime import datetime
import utils
class Baro:
"""This class contains info about the Void Trader and is initialized with
data in JSON format
"""
def __init__(self, data):
self.config = data['Config']
self.start = datetime.fromtimestamp(data['Activation']['sec'])
self.end = datetime.fromtimestamp(data['Expiry']['sec'])
self.location = data['Node']
self.manifest = data['Manifest']
def __str__(self):
"""Returns a string with all the information about Baro's offers
"""
baroItemString = ""
if datetime.now() < self.start:
return "None"
else:
for item in self.manifest:
baroItemString += ('== '+ str(item["ItemType"]) +' ==\n'
'- price: '+ str(item["PrimePrice"]) +' ducats + '+ str(item["RegularPrice"]) +'cr -\n\n' )
return baroItemString
def get_end_string(self):
"""Returns a string containing Baro's departure time
"""
return timedelta_to_string(self.end - datetime.now())
def get_start_string(self):
"""Returns a string containing Baro's arrival time
"""
return timedelta_to_string(self.start - datetime.now())
| <commit_before>from datetime import datetime
class Baro:
"""This class represents a Baro item and is initialized with
data in JSON format
"""
def __init__(self, data):
self.config = data['Config']
self.start = datetime.fromtimestamp(data['Activation']['sec'])
self.end = datetime.fromtimestamp(data['Expiry']['sec'])
self.location = data['Node']
self.manifest = data['Manifest']
def __str__(self):
"""Returns a string with all the information about Baro offer
"""
baroItemString = ""
if datetime.now() < self.start:
return "None"
else:
for item in self.manifest:
baroItemString += ('== '+ str(item["ItemType"]) +' ==\n'
'- price: '+ str(item["PrimePrice"]) +' ducats + '+ str(item["RegularPrice"]) +'cr -\n\n' )
return baroItemString
def get_eta_string(self):
"""Returns a string containing the Baro's ETA
"""
seconds = int((self.end - datetime.now()).total_seconds())
return '{} days, {} hrs, {} mins'.format((seconds // 86400), ((seconds % 86400) // 3600),
(seconds % 3600) // 60)
def get_start_string(self):
"""Returns a string containing the Baro's start
"""
seconds = int((self.start - datetime.now()).total_seconds())
return '{} days, {} hrs, {} mins'.format((seconds // 86400), ((seconds % 86400) // 3600),
(seconds % 3600) // 60)<commit_msg>Change class Baro to use timedelta_to_string, some fixes<commit_after>from datetime import datetime
import utils
class Baro:
"""This class contains info about the Void Trader and is initialized with
data in JSON format
"""
def __init__(self, data):
self.config = data['Config']
self.start = datetime.fromtimestamp(data['Activation']['sec'])
self.end = datetime.fromtimestamp(data['Expiry']['sec'])
self.location = data['Node']
self.manifest = data['Manifest']
def __str__(self):
"""Returns a string with all the information about Baro's offers
"""
baroItemString = ""
if datetime.now() < self.start:
return "None"
else:
for item in self.manifest:
baroItemString += ('== '+ str(item["ItemType"]) +' ==\n'
'- price: '+ str(item["PrimePrice"]) +' ducats + '+ str(item["RegularPrice"]) +'cr -\n\n' )
return baroItemString
def get_end_string(self):
"""Returns a string containing Baro's departure time
"""
return timedelta_to_string(self.end - datetime.now())
def get_start_string(self):
"""Returns a string containing Baro's arrival time
"""
return timedelta_to_string(self.start - datetime.now())
|
12cd87394e09e7481b39ca519f15db4688ab0073 | tmpl/Prompt.py | tmpl/Prompt.py | #--coding:utf-8--
#Platform
class BasePrompt(object):
pass
class ErrPrompt(BasePrompt):
"""
Define some of Err Prompts
Usually print to sys.stderr
"""
def PrintErr(self, content):
import sys
"""
Automous write content to sys.stderr and add '\n' to the end
"""
sys.stderr.write(content + '\n')
sys.stderr.flush()
def IllegalURL(self):
"""
For URL with illegal characters
"""
self.PrintErr("Error: URL include illegal characters!")
def FileExist(self, File = "File"):
"""
return True if want to replace, and False for the other
"""
self.PrintErr("Warning: '%s' already exists, replace?(y/n)" %(Files))
tmp = raw_input()
if tmp == 'y' or tmp == 'Y':
return True
return False
def Exit(self):
self.PrintErr("Info: Terminated")
if __name__ == '__main__':
raise EnvironmentError ("DO NOT DIRECTLY RUN THIS TEMPLATE!")
| #--coding:utf-8--
#Platform
class BasePrompt(object):
pass
class ErrPrompt(BasePrompt):
"""
Define some of Err Prompts
Usually print to sys.stderr
"""
def PrintErr(self, content):
import sys
"""
Automous write content to sys.stderr and add '\n' to the end
"""
sys.stderr.write(content + '\n')
sys.stderr.flush()
def IllegalURL(self):
"""
For URL with illegal characters
"""
self.PrintErr("Error: URL include illegal characters!")
def FileExist(self, File = "File"):
"""
return True if want to replace, and False for the other
"""
self.PrintErr("Warning: '%s' already exists, replace?(y/n)" %(File))
tmp = raw_input()
if tmp == 'y' or tmp == 'Y':
return True
return False
def Exit(self):
self.PrintErr("Info: Terminated")
if __name__ == '__main__':
raise EnvironmentError ("DO NOT DIRECTLY RUN THIS TEMPLATE!")
| Fix Bugs in module when using -O option | Fix Bugs in module when using -O option
| Python | mit | nday-dev/Spider-Framework | #--coding:utf-8--
#Platform
class BasePrompt(object):
pass
class ErrPrompt(BasePrompt):
"""
Define some of Err Prompts
Usually print to sys.stderr
"""
def PrintErr(self, content):
import sys
"""
Automous write content to sys.stderr and add '\n' to the end
"""
sys.stderr.write(content + '\n')
sys.stderr.flush()
def IllegalURL(self):
"""
For URL with illegal characters
"""
self.PrintErr("Error: URL include illegal characters!")
def FileExist(self, File = "File"):
"""
return True if want to replace, and False for the other
"""
self.PrintErr("Warning: '%s' already exists, replace?(y/n)" %(Files))
tmp = raw_input()
if tmp == 'y' or tmp == 'Y':
return True
return False
def Exit(self):
self.PrintErr("Info: Terminated")
if __name__ == '__main__':
raise EnvironmentError ("DO NOT DIRECTLY RUN THIS TEMPLATE!")
Fix Bugs in module when using -O option | #--coding:utf-8--
#Platform
class BasePrompt(object):
pass
class ErrPrompt(BasePrompt):
"""
Define some of Err Prompts
Usually print to sys.stderr
"""
def PrintErr(self, content):
import sys
"""
Automous write content to sys.stderr and add '\n' to the end
"""
sys.stderr.write(content + '\n')
sys.stderr.flush()
def IllegalURL(self):
"""
For URL with illegal characters
"""
self.PrintErr("Error: URL include illegal characters!")
def FileExist(self, File = "File"):
"""
return True if want to replace, and False for the other
"""
self.PrintErr("Warning: '%s' already exists, replace?(y/n)" %(File))
tmp = raw_input()
if tmp == 'y' or tmp == 'Y':
return True
return False
def Exit(self):
self.PrintErr("Info: Terminated")
if __name__ == '__main__':
raise EnvironmentError ("DO NOT DIRECTLY RUN THIS TEMPLATE!")
| <commit_before>#--coding:utf-8--
#Platform
class BasePrompt(object):
pass
class ErrPrompt(BasePrompt):
"""
Define some of Err Prompts
Usually print to sys.stderr
"""
def PrintErr(self, content):
import sys
"""
Automous write content to sys.stderr and add '\n' to the end
"""
sys.stderr.write(content + '\n')
sys.stderr.flush()
def IllegalURL(self):
"""
For URL with illegal characters
"""
self.PrintErr("Error: URL include illegal characters!")
def FileExist(self, File = "File"):
"""
return True if want to replace, and False for the other
"""
self.PrintErr("Warning: '%s' already exists, replace?(y/n)" %(Files))
tmp = raw_input()
if tmp == 'y' or tmp == 'Y':
return True
return False
def Exit(self):
self.PrintErr("Info: Terminated")
if __name__ == '__main__':
raise EnvironmentError ("DO NOT DIRECTLY RUN THIS TEMPLATE!")
<commit_msg>Fix Bugs in module when using -O option<commit_after> | #--coding:utf-8--
#Platform
class BasePrompt(object):
pass
class ErrPrompt(BasePrompt):
"""
Define some of Err Prompts
Usually print to sys.stderr
"""
def PrintErr(self, content):
import sys
"""
Automous write content to sys.stderr and add '\n' to the end
"""
sys.stderr.write(content + '\n')
sys.stderr.flush()
def IllegalURL(self):
"""
For URL with illegal characters
"""
self.PrintErr("Error: URL include illegal characters!")
def FileExist(self, File = "File"):
"""
return True if want to replace, and False for the other
"""
self.PrintErr("Warning: '%s' already exists, replace?(y/n)" %(File))
tmp = raw_input()
if tmp == 'y' or tmp == 'Y':
return True
return False
def Exit(self):
self.PrintErr("Info: Terminated")
if __name__ == '__main__':
raise EnvironmentError ("DO NOT DIRECTLY RUN THIS TEMPLATE!")
| #--coding:utf-8--
#Platform
class BasePrompt(object):
pass
class ErrPrompt(BasePrompt):
"""
Define some of Err Prompts
Usually print to sys.stderr
"""
def PrintErr(self, content):
import sys
"""
Automous write content to sys.stderr and add '\n' to the end
"""
sys.stderr.write(content + '\n')
sys.stderr.flush()
def IllegalURL(self):
"""
For URL with illegal characters
"""
self.PrintErr("Error: URL include illegal characters!")
def FileExist(self, File = "File"):
"""
return True if want to replace, and False for the other
"""
self.PrintErr("Warning: '%s' already exists, replace?(y/n)" %(Files))
tmp = raw_input()
if tmp == 'y' or tmp == 'Y':
return True
return False
def Exit(self):
self.PrintErr("Info: Terminated")
if __name__ == '__main__':
raise EnvironmentError ("DO NOT DIRECTLY RUN THIS TEMPLATE!")
Fix Bugs in module when using -O option#--coding:utf-8--
#Platform
class BasePrompt(object):
pass
class ErrPrompt(BasePrompt):
"""
Define some of Err Prompts
Usually print to sys.stderr
"""
def PrintErr(self, content):
import sys
"""
Automous write content to sys.stderr and add '\n' to the end
"""
sys.stderr.write(content + '\n')
sys.stderr.flush()
def IllegalURL(self):
"""
For URL with illegal characters
"""
self.PrintErr("Error: URL include illegal characters!")
def FileExist(self, File = "File"):
"""
return True if want to replace, and False for the other
"""
self.PrintErr("Warning: '%s' already exists, replace?(y/n)" %(File))
tmp = raw_input()
if tmp == 'y' or tmp == 'Y':
return True
return False
def Exit(self):
self.PrintErr("Info: Terminated")
if __name__ == '__main__':
raise EnvironmentError ("DO NOT DIRECTLY RUN THIS TEMPLATE!")
| <commit_before>#--coding:utf-8--
#Platform
class BasePrompt(object):
pass
class ErrPrompt(BasePrompt):
"""
Define some of Err Prompts
Usually print to sys.stderr
"""
def PrintErr(self, content):
import sys
"""
Automous write content to sys.stderr and add '\n' to the end
"""
sys.stderr.write(content + '\n')
sys.stderr.flush()
def IllegalURL(self):
"""
For URL with illegal characters
"""
self.PrintErr("Error: URL include illegal characters!")
def FileExist(self, File = "File"):
"""
return True if want to replace, and False for the other
"""
self.PrintErr("Warning: '%s' already exists, replace?(y/n)" %(Files))
tmp = raw_input()
if tmp == 'y' or tmp == 'Y':
return True
return False
def Exit(self):
self.PrintErr("Info: Terminated")
if __name__ == '__main__':
raise EnvironmentError ("DO NOT DIRECTLY RUN THIS TEMPLATE!")
<commit_msg>Fix Bugs in module when using -O option<commit_after>#--coding:utf-8--
#Platform
class BasePrompt(object):
pass
class ErrPrompt(BasePrompt):
"""
Define some of Err Prompts
Usually print to sys.stderr
"""
def PrintErr(self, content):
import sys
"""
Automous write content to sys.stderr and add '\n' to the end
"""
sys.stderr.write(content + '\n')
sys.stderr.flush()
def IllegalURL(self):
"""
For URL with illegal characters
"""
self.PrintErr("Error: URL include illegal characters!")
def FileExist(self, File = "File"):
"""
return True if want to replace, and False for the other
"""
self.PrintErr("Warning: '%s' already exists, replace?(y/n)" %(File))
tmp = raw_input()
if tmp == 'y' or tmp == 'Y':
return True
return False
def Exit(self):
self.PrintErr("Info: Terminated")
if __name__ == '__main__':
raise EnvironmentError ("DO NOT DIRECTLY RUN THIS TEMPLATE!")
|
d3078cafd4e64e9c093d9d823df2035b8380d643 | meta-refkit-computervision/recipes-computervision/caffe-bvlc-reference/files/dnn-test.py | meta-refkit-computervision/recipes-computervision/caffe-bvlc-reference/files/dnn-test.py | #!/usr/bin/env python3
# Classify an image using a suitable model. The image conversion magic
# is from
# https://github.com/opencv/opencv_contrib/blob/master/modules/dnn/samples/googlenet_python.py
# (3-clause BSD license).
import numpy as np
import cv2
import sys
if len(sys.argv) != 4:
print("Usage: dnn.py <prototxt> <caffemodel> <image>")
sys.exit(1)
cv2.ocl.setUseOpenCL(False)
# read the image
test_img = cv2.imread(sys.argv[3])
# resize
resized = cv2.resize(test_img, (224,224))
converted = np.moveaxis(resized, 2, 0)
data = np.reshape(converted.astype(np.float32), (-1, 3, 224, 224))
# initialize network
net = cv2.dnn.readNetFromCaffe(sys.argv[1], sys.argv[2])
net.setBlob(".data", data)
# run the network
net.forward()
# print the class
print(str(net.getBlob("prob").argmax()))
| #!/usr/bin/env python3
# Classify an image using a suitable model. The image conversion magic
# is from
# https://github.com/opencv/opencv_contrib/blob/master/modules/dnn/samples/googlenet_python.py
# (3-clause BSD license).
import numpy as np
import cv2
import sys
if len(sys.argv) != 4:
print("Usage: dnn.py <prototxt> <caffemodel> <image>")
sys.exit(1)
cv2.ocl.setUseOpenCL(False)
# read the image
test_img = cv2.imread(sys.argv[3])
# resize
resized = cv2.resize(test_img, (224,224))
converted = np.moveaxis(resized, 2, 0)
data = np.reshape(converted.astype(np.float32), (-1, 3, 224, 224))
# initialize network
net = cv2.dnn.readNetFromCaffe(sys.argv[1], sys.argv[2])
net.setInput(data, "data")
# run the network
result = net.forward("prob")
# print the class
print(str(result.argmax()))
| Fix DNN test to be compatible with OpenCV 3.3. | convnet: Fix DNN test to be compatible with OpenCV 3.3.
OpenCV DNN module API changed with OpenCV 3.3. Fix the tests to use the
new API.
Signed-off-by: Ismo Puustinen <75dda586a9213f0e0695eb79120c94222bb30e60@intel.com>
| Python | mit | intel/intel-iot-refkit,mythi/intel-iot-refkit,mythi/intel-iot-refkit,intel/intel-iot-refkit,intel/intel-iot-refkit,intel/intel-iot-refkit,mythi/intel-iot-refkit,klihub/intel-iot-refkit,klihub/intel-iot-refkit,klihub/intel-iot-refkit,klihub/intel-iot-refkit,klihub/intel-iot-refkit,klihub/intel-iot-refkit,mythi/intel-iot-refkit,mythi/intel-iot-refkit,klihub/intel-iot-refkit,mythi/intel-iot-refkit,intel/intel-iot-refkit,intel/intel-iot-refkit,mythi/intel-iot-refkit,intel/intel-iot-refkit | #!/usr/bin/env python3
# Classify an image using a suitable model. The image conversion magic
# is from
# https://github.com/opencv/opencv_contrib/blob/master/modules/dnn/samples/googlenet_python.py
# (3-clause BSD license).
import numpy as np
import cv2
import sys
if len(sys.argv) != 4:
print("Usage: dnn.py <prototxt> <caffemodel> <image>")
sys.exit(1)
cv2.ocl.setUseOpenCL(False)
# read the image
test_img = cv2.imread(sys.argv[3])
# resize
resized = cv2.resize(test_img, (224,224))
converted = np.moveaxis(resized, 2, 0)
data = np.reshape(converted.astype(np.float32), (-1, 3, 224, 224))
# initialize network
net = cv2.dnn.readNetFromCaffe(sys.argv[1], sys.argv[2])
net.setBlob(".data", data)
# run the network
net.forward()
# print the class
print(str(net.getBlob("prob").argmax()))
convnet: Fix DNN test to be compatible with OpenCV 3.3.
OpenCV DNN module API changed with OpenCV 3.3. Fix the tests to use the
new API.
Signed-off-by: Ismo Puustinen <75dda586a9213f0e0695eb79120c94222bb30e60@intel.com> | #!/usr/bin/env python3
# Classify an image using a suitable model. The image conversion magic
# is from
# https://github.com/opencv/opencv_contrib/blob/master/modules/dnn/samples/googlenet_python.py
# (3-clause BSD license).
import numpy as np
import cv2
import sys
if len(sys.argv) != 4:
print("Usage: dnn.py <prototxt> <caffemodel> <image>")
sys.exit(1)
cv2.ocl.setUseOpenCL(False)
# read the image
test_img = cv2.imread(sys.argv[3])
# resize
resized = cv2.resize(test_img, (224,224))
converted = np.moveaxis(resized, 2, 0)
data = np.reshape(converted.astype(np.float32), (-1, 3, 224, 224))
# initialize network
net = cv2.dnn.readNetFromCaffe(sys.argv[1], sys.argv[2])
net.setInput(data, "data")
# run the network
result = net.forward("prob")
# print the class
print(str(result.argmax()))
| <commit_before>#!/usr/bin/env python3
# Classify an image using a suitable model. The image conversion magic
# is from
# https://github.com/opencv/opencv_contrib/blob/master/modules/dnn/samples/googlenet_python.py
# (3-clause BSD license).
import numpy as np
import cv2
import sys
if len(sys.argv) != 4:
print("Usage: dnn.py <prototxt> <caffemodel> <image>")
sys.exit(1)
cv2.ocl.setUseOpenCL(False)
# read the image
test_img = cv2.imread(sys.argv[3])
# resize
resized = cv2.resize(test_img, (224,224))
converted = np.moveaxis(resized, 2, 0)
data = np.reshape(converted.astype(np.float32), (-1, 3, 224, 224))
# initialize network
net = cv2.dnn.readNetFromCaffe(sys.argv[1], sys.argv[2])
net.setBlob(".data", data)
# run the network
net.forward()
# print the class
print(str(net.getBlob("prob").argmax()))
<commit_msg>convnet: Fix DNN test to be compatible with OpenCV 3.3.
OpenCV DNN module API changed with OpenCV 3.3. Fix the tests to use the
new API.
Signed-off-by: Ismo Puustinen <75dda586a9213f0e0695eb79120c94222bb30e60@intel.com><commit_after> | #!/usr/bin/env python3
# Classify an image using a suitable model. The image conversion magic
# is from
# https://github.com/opencv/opencv_contrib/blob/master/modules/dnn/samples/googlenet_python.py
# (3-clause BSD license).
import numpy as np
import cv2
import sys
if len(sys.argv) != 4:
print("Usage: dnn.py <prototxt> <caffemodel> <image>")
sys.exit(1)
cv2.ocl.setUseOpenCL(False)
# read the image
test_img = cv2.imread(sys.argv[3])
# resize
resized = cv2.resize(test_img, (224,224))
converted = np.moveaxis(resized, 2, 0)
data = np.reshape(converted.astype(np.float32), (-1, 3, 224, 224))
# initialize network
net = cv2.dnn.readNetFromCaffe(sys.argv[1], sys.argv[2])
net.setInput(data, "data")
# run the network
result = net.forward("prob")
# print the class
print(str(result.argmax()))
| #!/usr/bin/env python3
# Classify an image using a suitable model. The image conversion magic
# is from
# https://github.com/opencv/opencv_contrib/blob/master/modules/dnn/samples/googlenet_python.py
# (3-clause BSD license).
import numpy as np
import cv2
import sys
if len(sys.argv) != 4:
print("Usage: dnn.py <prototxt> <caffemodel> <image>")
sys.exit(1)
cv2.ocl.setUseOpenCL(False)
# read the image
test_img = cv2.imread(sys.argv[3])
# resize
resized = cv2.resize(test_img, (224,224))
converted = np.moveaxis(resized, 2, 0)
data = np.reshape(converted.astype(np.float32), (-1, 3, 224, 224))
# initialize network
net = cv2.dnn.readNetFromCaffe(sys.argv[1], sys.argv[2])
net.setBlob(".data", data)
# run the network
net.forward()
# print the class
print(str(net.getBlob("prob").argmax()))
convnet: Fix DNN test to be compatible with OpenCV 3.3.
OpenCV DNN module API changed with OpenCV 3.3. Fix the tests to use the
new API.
Signed-off-by: Ismo Puustinen <75dda586a9213f0e0695eb79120c94222bb30e60@intel.com>#!/usr/bin/env python3
# Classify an image using a suitable model. The image conversion magic
# is from
# https://github.com/opencv/opencv_contrib/blob/master/modules/dnn/samples/googlenet_python.py
# (3-clause BSD license).
import numpy as np
import cv2
import sys
if len(sys.argv) != 4:
print("Usage: dnn.py <prototxt> <caffemodel> <image>")
sys.exit(1)
cv2.ocl.setUseOpenCL(False)
# read the image
test_img = cv2.imread(sys.argv[3])
# resize
resized = cv2.resize(test_img, (224,224))
converted = np.moveaxis(resized, 2, 0)
data = np.reshape(converted.astype(np.float32), (-1, 3, 224, 224))
# initialize network
net = cv2.dnn.readNetFromCaffe(sys.argv[1], sys.argv[2])
net.setInput(data, "data")
# run the network
result = net.forward("prob")
# print the class
print(str(result.argmax()))
| <commit_before>#!/usr/bin/env python3
# Classify an image using a suitable model. The image conversion magic
# is from
# https://github.com/opencv/opencv_contrib/blob/master/modules/dnn/samples/googlenet_python.py
# (3-clause BSD license).
import numpy as np
import cv2
import sys
if len(sys.argv) != 4:
print("Usage: dnn.py <prototxt> <caffemodel> <image>")
sys.exit(1)
cv2.ocl.setUseOpenCL(False)
# read the image
test_img = cv2.imread(sys.argv[3])
# resize
resized = cv2.resize(test_img, (224,224))
converted = np.moveaxis(resized, 2, 0)
data = np.reshape(converted.astype(np.float32), (-1, 3, 224, 224))
# initialize network
net = cv2.dnn.readNetFromCaffe(sys.argv[1], sys.argv[2])
net.setBlob(".data", data)
# run the network
net.forward()
# print the class
print(str(net.getBlob("prob").argmax()))
<commit_msg>convnet: Fix DNN test to be compatible with OpenCV 3.3.
OpenCV DNN module API changed with OpenCV 3.3. Fix the tests to use the
new API.
Signed-off-by: Ismo Puustinen <75dda586a9213f0e0695eb79120c94222bb30e60@intel.com><commit_after>#!/usr/bin/env python3
# Classify an image using a suitable model. The image conversion magic
# is from
# https://github.com/opencv/opencv_contrib/blob/master/modules/dnn/samples/googlenet_python.py
# (3-clause BSD license).
import numpy as np
import cv2
import sys
if len(sys.argv) != 4:
print("Usage: dnn.py <prototxt> <caffemodel> <image>")
sys.exit(1)
cv2.ocl.setUseOpenCL(False)
# read the image
test_img = cv2.imread(sys.argv[3])
# resize
resized = cv2.resize(test_img, (224,224))
converted = np.moveaxis(resized, 2, 0)
data = np.reshape(converted.astype(np.float32), (-1, 3, 224, 224))
# initialize network
net = cv2.dnn.readNetFromCaffe(sys.argv[1], sys.argv[2])
net.setInput(data, "data")
# run the network
result = net.forward("prob")
# print the class
print(str(result.argmax()))
|
210be14772b403e8fb5938e4e2cd391d43275ab1 | tests/test_ot_propagators.py | tests/test_ot_propagators.py | import instana.http_propagator as ihp
import opentracing as ot
from instana import tracer, options, util
from nose.tools import assert_equals
import inspect
def test_basics():
inspect.isclass(ihp.HTTPPropagator)
inject_func = getattr(ihp.HTTPPropagator, "inject", None)
assert inject_func
assert inspect.ismethod(inject_func)
extract_func = getattr(ihp.HTTPPropagator, "extract", None)
assert extract_func
assert inspect.ismethod(extract_func)
def test_inject():
opts = options.Options()
ot.global_tracer = tracer.InstanaTracer(opts)
carrier = {}
span = ot.global_tracer.start_span("nosetests")
ot.global_tracer.inject(span.context, ot.Format.HTTP_HEADERS, carrier)
assert 'X-Instana-T' in carrier
assert_equals(carrier['X-Instana-T'], util.id_to_header(span.context.trace_id))
assert 'X-Instana-S' in carrier
assert_equals(carrier['X-Instana-S'], util.id_to_header(span.context.span_id))
assert 'X-Instana-L' in carrier
assert_equals(carrier['X-Instana-L'], "1")
| import instana.http_propagator as ihp
import opentracing as ot
from instana import tracer, options, util
from nose.tools import assert_equals
import inspect
def test_basics():
inspect.isclass(ihp.HTTPPropagator)
inject_func = getattr(ihp.HTTPPropagator, "inject", None)
assert inject_func
assert callable(inject_func)
extract_func = getattr(ihp.HTTPPropagator, "extract", None)
assert extract_func
assert callable(extract_func)
def test_inject():
opts = options.Options()
ot.global_tracer = tracer.InstanaTracer(opts)
carrier = {}
span = ot.global_tracer.start_span("nosetests")
ot.global_tracer.inject(span.context, ot.Format.HTTP_HEADERS, carrier)
assert 'X-Instana-T' in carrier
assert_equals(carrier['X-Instana-T'], util.id_to_header(span.context.trace_id))
assert 'X-Instana-S' in carrier
assert_equals(carrier['X-Instana-S'], util.id_to_header(span.context.span_id))
assert 'X-Instana-L' in carrier
assert_equals(carrier['X-Instana-L'], "1")
| Fix function test to work on both Py 2 + 3 | Fix function test to work on both Py 2 + 3
| Python | mit | instana/python-sensor,instana/python-sensor | import instana.http_propagator as ihp
import opentracing as ot
from instana import tracer, options, util
from nose.tools import assert_equals
import inspect
def test_basics():
inspect.isclass(ihp.HTTPPropagator)
inject_func = getattr(ihp.HTTPPropagator, "inject", None)
assert inject_func
assert inspect.ismethod(inject_func)
extract_func = getattr(ihp.HTTPPropagator, "extract", None)
assert extract_func
assert inspect.ismethod(extract_func)
def test_inject():
opts = options.Options()
ot.global_tracer = tracer.InstanaTracer(opts)
carrier = {}
span = ot.global_tracer.start_span("nosetests")
ot.global_tracer.inject(span.context, ot.Format.HTTP_HEADERS, carrier)
assert 'X-Instana-T' in carrier
assert_equals(carrier['X-Instana-T'], util.id_to_header(span.context.trace_id))
assert 'X-Instana-S' in carrier
assert_equals(carrier['X-Instana-S'], util.id_to_header(span.context.span_id))
assert 'X-Instana-L' in carrier
assert_equals(carrier['X-Instana-L'], "1")
Fix function test to work on both Py 2 + 3 | import instana.http_propagator as ihp
import opentracing as ot
from instana import tracer, options, util
from nose.tools import assert_equals
import inspect
def test_basics():
inspect.isclass(ihp.HTTPPropagator)
inject_func = getattr(ihp.HTTPPropagator, "inject", None)
assert inject_func
assert callable(inject_func)
extract_func = getattr(ihp.HTTPPropagator, "extract", None)
assert extract_func
assert callable(extract_func)
def test_inject():
opts = options.Options()
ot.global_tracer = tracer.InstanaTracer(opts)
carrier = {}
span = ot.global_tracer.start_span("nosetests")
ot.global_tracer.inject(span.context, ot.Format.HTTP_HEADERS, carrier)
assert 'X-Instana-T' in carrier
assert_equals(carrier['X-Instana-T'], util.id_to_header(span.context.trace_id))
assert 'X-Instana-S' in carrier
assert_equals(carrier['X-Instana-S'], util.id_to_header(span.context.span_id))
assert 'X-Instana-L' in carrier
assert_equals(carrier['X-Instana-L'], "1")
| <commit_before>import instana.http_propagator as ihp
import opentracing as ot
from instana import tracer, options, util
from nose.tools import assert_equals
import inspect
def test_basics():
inspect.isclass(ihp.HTTPPropagator)
inject_func = getattr(ihp.HTTPPropagator, "inject", None)
assert inject_func
assert inspect.ismethod(inject_func)
extract_func = getattr(ihp.HTTPPropagator, "extract", None)
assert extract_func
assert inspect.ismethod(extract_func)
def test_inject():
opts = options.Options()
ot.global_tracer = tracer.InstanaTracer(opts)
carrier = {}
span = ot.global_tracer.start_span("nosetests")
ot.global_tracer.inject(span.context, ot.Format.HTTP_HEADERS, carrier)
assert 'X-Instana-T' in carrier
assert_equals(carrier['X-Instana-T'], util.id_to_header(span.context.trace_id))
assert 'X-Instana-S' in carrier
assert_equals(carrier['X-Instana-S'], util.id_to_header(span.context.span_id))
assert 'X-Instana-L' in carrier
assert_equals(carrier['X-Instana-L'], "1")
<commit_msg>Fix function test to work on both Py 2 + 3<commit_after> | import instana.http_propagator as ihp
import opentracing as ot
from instana import tracer, options, util
from nose.tools import assert_equals
import inspect
def test_basics():
inspect.isclass(ihp.HTTPPropagator)
inject_func = getattr(ihp.HTTPPropagator, "inject", None)
assert inject_func
assert callable(inject_func)
extract_func = getattr(ihp.HTTPPropagator, "extract", None)
assert extract_func
assert callable(extract_func)
def test_inject():
opts = options.Options()
ot.global_tracer = tracer.InstanaTracer(opts)
carrier = {}
span = ot.global_tracer.start_span("nosetests")
ot.global_tracer.inject(span.context, ot.Format.HTTP_HEADERS, carrier)
assert 'X-Instana-T' in carrier
assert_equals(carrier['X-Instana-T'], util.id_to_header(span.context.trace_id))
assert 'X-Instana-S' in carrier
assert_equals(carrier['X-Instana-S'], util.id_to_header(span.context.span_id))
assert 'X-Instana-L' in carrier
assert_equals(carrier['X-Instana-L'], "1")
| import instana.http_propagator as ihp
import opentracing as ot
from instana import tracer, options, util
from nose.tools import assert_equals
import inspect
def test_basics():
inspect.isclass(ihp.HTTPPropagator)
inject_func = getattr(ihp.HTTPPropagator, "inject", None)
assert inject_func
assert inspect.ismethod(inject_func)
extract_func = getattr(ihp.HTTPPropagator, "extract", None)
assert extract_func
assert inspect.ismethod(extract_func)
def test_inject():
opts = options.Options()
ot.global_tracer = tracer.InstanaTracer(opts)
carrier = {}
span = ot.global_tracer.start_span("nosetests")
ot.global_tracer.inject(span.context, ot.Format.HTTP_HEADERS, carrier)
assert 'X-Instana-T' in carrier
assert_equals(carrier['X-Instana-T'], util.id_to_header(span.context.trace_id))
assert 'X-Instana-S' in carrier
assert_equals(carrier['X-Instana-S'], util.id_to_header(span.context.span_id))
assert 'X-Instana-L' in carrier
assert_equals(carrier['X-Instana-L'], "1")
Fix function test to work on both Py 2 + 3import instana.http_propagator as ihp
import opentracing as ot
from instana import tracer, options, util
from nose.tools import assert_equals
import inspect
def test_basics():
inspect.isclass(ihp.HTTPPropagator)
inject_func = getattr(ihp.HTTPPropagator, "inject", None)
assert inject_func
assert callable(inject_func)
extract_func = getattr(ihp.HTTPPropagator, "extract", None)
assert extract_func
assert callable(extract_func)
def test_inject():
opts = options.Options()
ot.global_tracer = tracer.InstanaTracer(opts)
carrier = {}
span = ot.global_tracer.start_span("nosetests")
ot.global_tracer.inject(span.context, ot.Format.HTTP_HEADERS, carrier)
assert 'X-Instana-T' in carrier
assert_equals(carrier['X-Instana-T'], util.id_to_header(span.context.trace_id))
assert 'X-Instana-S' in carrier
assert_equals(carrier['X-Instana-S'], util.id_to_header(span.context.span_id))
assert 'X-Instana-L' in carrier
assert_equals(carrier['X-Instana-L'], "1")
| <commit_before>import instana.http_propagator as ihp
import opentracing as ot
from instana import tracer, options, util
from nose.tools import assert_equals
import inspect
def test_basics():
inspect.isclass(ihp.HTTPPropagator)
inject_func = getattr(ihp.HTTPPropagator, "inject", None)
assert inject_func
assert inspect.ismethod(inject_func)
extract_func = getattr(ihp.HTTPPropagator, "extract", None)
assert extract_func
assert inspect.ismethod(extract_func)
def test_inject():
opts = options.Options()
ot.global_tracer = tracer.InstanaTracer(opts)
carrier = {}
span = ot.global_tracer.start_span("nosetests")
ot.global_tracer.inject(span.context, ot.Format.HTTP_HEADERS, carrier)
assert 'X-Instana-T' in carrier
assert_equals(carrier['X-Instana-T'], util.id_to_header(span.context.trace_id))
assert 'X-Instana-S' in carrier
assert_equals(carrier['X-Instana-S'], util.id_to_header(span.context.span_id))
assert 'X-Instana-L' in carrier
assert_equals(carrier['X-Instana-L'], "1")
<commit_msg>Fix function test to work on both Py 2 + 3<commit_after>import instana.http_propagator as ihp
import opentracing as ot
from instana import tracer, options, util
from nose.tools import assert_equals
import inspect
def test_basics():
inspect.isclass(ihp.HTTPPropagator)
inject_func = getattr(ihp.HTTPPropagator, "inject", None)
assert inject_func
assert callable(inject_func)
extract_func = getattr(ihp.HTTPPropagator, "extract", None)
assert extract_func
assert callable(extract_func)
def test_inject():
opts = options.Options()
ot.global_tracer = tracer.InstanaTracer(opts)
carrier = {}
span = ot.global_tracer.start_span("nosetests")
ot.global_tracer.inject(span.context, ot.Format.HTTP_HEADERS, carrier)
assert 'X-Instana-T' in carrier
assert_equals(carrier['X-Instana-T'], util.id_to_header(span.context.trace_id))
assert 'X-Instana-S' in carrier
assert_equals(carrier['X-Instana-S'], util.id_to_header(span.context.span_id))
assert 'X-Instana-L' in carrier
assert_equals(carrier['X-Instana-L'], "1")
|
ef94948a8ce16d9d80fb69950381e0936a462bb0 | tests/config_tests.py | tests/config_tests.py | from nose.tools import assert_equal
from wunderapi.config import Config
def setup():
return Config(config_file="tests/resources/test_config")
def test_parse_config_with_correct_parms():
pass
def test_parse_config_with_incorrect_parms():
pass
def test_config_created_with_default_parms():
config = setup()
config.parse_config()
assert_equal(config.api_key, 'API Key')
assert_equal(config.location, 'Zipcode')
assert_equal(config.date_format, 'date')
assert_equal(config.units, 'english')
| from nose.tools import assert_equal
from wunderapi.config import Config
def setup():
return Config(config_file="tests/resources/test_config")
def test_parse_config_with_correct_parms():
pass
def test_parse_config_with_incorrect_parms():
pass
def test_config_created_with_default_parms():
config = setup()
config.parse_config()
assert_equal(config.date_format, 'date')
assert_equal(config.units, 'english')
| Update test to get api_key from environment | Update test to get api_key from environment
| Python | mit | paris3200/Weather,paris3200/wunderapi | from nose.tools import assert_equal
from wunderapi.config import Config
def setup():
return Config(config_file="tests/resources/test_config")
def test_parse_config_with_correct_parms():
pass
def test_parse_config_with_incorrect_parms():
pass
def test_config_created_with_default_parms():
config = setup()
config.parse_config()
assert_equal(config.api_key, 'API Key')
assert_equal(config.location, 'Zipcode')
assert_equal(config.date_format, 'date')
assert_equal(config.units, 'english')
Update test to get api_key from environment | from nose.tools import assert_equal
from wunderapi.config import Config
def setup():
return Config(config_file="tests/resources/test_config")
def test_parse_config_with_correct_parms():
pass
def test_parse_config_with_incorrect_parms():
pass
def test_config_created_with_default_parms():
config = setup()
config.parse_config()
assert_equal(config.date_format, 'date')
assert_equal(config.units, 'english')
| <commit_before>from nose.tools import assert_equal
from wunderapi.config import Config
def setup():
return Config(config_file="tests/resources/test_config")
def test_parse_config_with_correct_parms():
pass
def test_parse_config_with_incorrect_parms():
pass
def test_config_created_with_default_parms():
config = setup()
config.parse_config()
assert_equal(config.api_key, 'API Key')
assert_equal(config.location, 'Zipcode')
assert_equal(config.date_format, 'date')
assert_equal(config.units, 'english')
<commit_msg>Update test to get api_key from environment<commit_after> | from nose.tools import assert_equal
from wunderapi.config import Config
def setup():
return Config(config_file="tests/resources/test_config")
def test_parse_config_with_correct_parms():
pass
def test_parse_config_with_incorrect_parms():
pass
def test_config_created_with_default_parms():
config = setup()
config.parse_config()
assert_equal(config.date_format, 'date')
assert_equal(config.units, 'english')
| from nose.tools import assert_equal
from wunderapi.config import Config
def setup():
return Config(config_file="tests/resources/test_config")
def test_parse_config_with_correct_parms():
pass
def test_parse_config_with_incorrect_parms():
pass
def test_config_created_with_default_parms():
config = setup()
config.parse_config()
assert_equal(config.api_key, 'API Key')
assert_equal(config.location, 'Zipcode')
assert_equal(config.date_format, 'date')
assert_equal(config.units, 'english')
Update test to get api_key from environmentfrom nose.tools import assert_equal
from wunderapi.config import Config
def setup():
return Config(config_file="tests/resources/test_config")
def test_parse_config_with_correct_parms():
pass
def test_parse_config_with_incorrect_parms():
pass
def test_config_created_with_default_parms():
config = setup()
config.parse_config()
assert_equal(config.date_format, 'date')
assert_equal(config.units, 'english')
| <commit_before>from nose.tools import assert_equal
from wunderapi.config import Config
def setup():
return Config(config_file="tests/resources/test_config")
def test_parse_config_with_correct_parms():
pass
def test_parse_config_with_incorrect_parms():
pass
def test_config_created_with_default_parms():
config = setup()
config.parse_config()
assert_equal(config.api_key, 'API Key')
assert_equal(config.location, 'Zipcode')
assert_equal(config.date_format, 'date')
assert_equal(config.units, 'english')
<commit_msg>Update test to get api_key from environment<commit_after>from nose.tools import assert_equal
from wunderapi.config import Config
def setup():
return Config(config_file="tests/resources/test_config")
def test_parse_config_with_correct_parms():
pass
def test_parse_config_with_incorrect_parms():
pass
def test_config_created_with_default_parms():
config = setup()
config.parse_config()
assert_equal(config.date_format, 'date')
assert_equal(config.units, 'english')
|
0529c392c8c3e75a03aa312e4fc7b367008fdf27 | tests/test_20_main.py | tests/test_20_main.py |
import click.testing
import pytest
from cfgrib import __main__
def test_main():
runner = click.testing.CliRunner()
res = runner.invoke(__main__.cfgrib_cli, ['selfcheck'])
assert res.exit_code == 0
assert 'Your system is ready.' in res.output
res = runner.invoke(__main__.cfgrib_cli, ['non-existent-command'])
assert res.exit_code == 2
|
import click.testing
from cfgrib import __main__
def test_main():
runner = click.testing.CliRunner()
res = runner.invoke(__main__.cfgrib_cli, ['selfcheck'])
assert res.exit_code == 0
assert 'Your system is ready.' in res.output
res = runner.invoke(__main__.cfgrib_cli, ['non-existent-command'])
assert res.exit_code == 2
| Fix docs and CLI tests. | Fix docs and CLI tests.
| Python | apache-2.0 | ecmwf/cfgrib |
import click.testing
import pytest
from cfgrib import __main__
def test_main():
runner = click.testing.CliRunner()
res = runner.invoke(__main__.cfgrib_cli, ['selfcheck'])
assert res.exit_code == 0
assert 'Your system is ready.' in res.output
res = runner.invoke(__main__.cfgrib_cli, ['non-existent-command'])
assert res.exit_code == 2
Fix docs and CLI tests. |
import click.testing
from cfgrib import __main__
def test_main():
runner = click.testing.CliRunner()
res = runner.invoke(__main__.cfgrib_cli, ['selfcheck'])
assert res.exit_code == 0
assert 'Your system is ready.' in res.output
res = runner.invoke(__main__.cfgrib_cli, ['non-existent-command'])
assert res.exit_code == 2
| <commit_before>
import click.testing
import pytest
from cfgrib import __main__
def test_main():
runner = click.testing.CliRunner()
res = runner.invoke(__main__.cfgrib_cli, ['selfcheck'])
assert res.exit_code == 0
assert 'Your system is ready.' in res.output
res = runner.invoke(__main__.cfgrib_cli, ['non-existent-command'])
assert res.exit_code == 2
<commit_msg>Fix docs and CLI tests.<commit_after> |
import click.testing
from cfgrib import __main__
def test_main():
runner = click.testing.CliRunner()
res = runner.invoke(__main__.cfgrib_cli, ['selfcheck'])
assert res.exit_code == 0
assert 'Your system is ready.' in res.output
res = runner.invoke(__main__.cfgrib_cli, ['non-existent-command'])
assert res.exit_code == 2
|
import click.testing
import pytest
from cfgrib import __main__
def test_main():
runner = click.testing.CliRunner()
res = runner.invoke(__main__.cfgrib_cli, ['selfcheck'])
assert res.exit_code == 0
assert 'Your system is ready.' in res.output
res = runner.invoke(__main__.cfgrib_cli, ['non-existent-command'])
assert res.exit_code == 2
Fix docs and CLI tests.
import click.testing
from cfgrib import __main__
def test_main():
runner = click.testing.CliRunner()
res = runner.invoke(__main__.cfgrib_cli, ['selfcheck'])
assert res.exit_code == 0
assert 'Your system is ready.' in res.output
res = runner.invoke(__main__.cfgrib_cli, ['non-existent-command'])
assert res.exit_code == 2
| <commit_before>
import click.testing
import pytest
from cfgrib import __main__
def test_main():
runner = click.testing.CliRunner()
res = runner.invoke(__main__.cfgrib_cli, ['selfcheck'])
assert res.exit_code == 0
assert 'Your system is ready.' in res.output
res = runner.invoke(__main__.cfgrib_cli, ['non-existent-command'])
assert res.exit_code == 2
<commit_msg>Fix docs and CLI tests.<commit_after>
import click.testing
from cfgrib import __main__
def test_main():
runner = click.testing.CliRunner()
res = runner.invoke(__main__.cfgrib_cli, ['selfcheck'])
assert res.exit_code == 0
assert 'Your system is ready.' in res.output
res = runner.invoke(__main__.cfgrib_cli, ['non-existent-command'])
assert res.exit_code == 2
|
c5a0d0c5bf578a2221322c068a41ce6331b84c9b | tests/test_cattery.py | tests/test_cattery.py | import pytest
from catinabox import cattery
class TestCattery(object):
###########################################################################
# add_cats
###########################################################################
def test__add_cats__succeeds(self):
c = cattery.Cattery()
assert c
###########################################################################
# remove_cat
###########################################################################
def test__remove_cat__succeeds(self):
c = cattery.Cattery()
assert c
def test__remove_cat__no_cats__fails(self):
c = cattery.Cattery()
assert c
def test__remove_cat__cat_not_in_cattery__fails(self):
c = cattery.Cattery()
c.add_cats(["Fluffy"])
with pytest.raises(cattery.CatNotFound):
c.remove_cat("Snookums")
| import pytest
from catinabox import cattery, mccattery
@pytest.fixture(params=[
cattery.Cattery,
mccattery.McCattery
])
def cattery_fixture(request):
return request.param()
###########################################################################
# add_cats
###########################################################################
def test__add_cats__succeeds(cattery_fixture):
cattery_fixture.add_cats(["Fluffy", "Snookums"])
assert cattery_fixture.cats == ["Fluffy", "Snookums"]
assert cattery_fixture.num_cats == 2
###########################################################################
# remove_cat
###########################################################################
def test__remove_cat__succeeds(cattery_fixture):
cattery_fixture.add_cats(["Fluffy", "Junior"])
cattery_fixture.remove_cat("Fluffy")
assert cattery_fixture.cats == ["Junior"]
assert cattery_fixture.num_cats == 1
def test__remove_cat__no_cats__fails(cattery_fixture):
with pytest.raises(cattery.CatNotFound):
cattery_fixture.remove_cat("Fluffles")
def test__remove_cat__cat_not_in_cattery__fails(cattery_fixture):
cattery_fixture.add_cats(["Fluffy"])
with pytest.raises(cattery.CatNotFound):
cattery_fixture.remove_cat("Snookums")
| Add full tests for mccattery and cattery | Add full tests for mccattery and cattery
| Python | mit | keeppythonweird/catinabox,indexOutOfBound5/catinabox | import pytest
from catinabox import cattery
class TestCattery(object):
###########################################################################
# add_cats
###########################################################################
def test__add_cats__succeeds(self):
c = cattery.Cattery()
assert c
###########################################################################
# remove_cat
###########################################################################
def test__remove_cat__succeeds(self):
c = cattery.Cattery()
assert c
def test__remove_cat__no_cats__fails(self):
c = cattery.Cattery()
assert c
def test__remove_cat__cat_not_in_cattery__fails(self):
c = cattery.Cattery()
c.add_cats(["Fluffy"])
with pytest.raises(cattery.CatNotFound):
c.remove_cat("Snookums")
Add full tests for mccattery and cattery | import pytest
from catinabox import cattery, mccattery
@pytest.fixture(params=[
cattery.Cattery,
mccattery.McCattery
])
def cattery_fixture(request):
return request.param()
###########################################################################
# add_cats
###########################################################################
def test__add_cats__succeeds(cattery_fixture):
cattery_fixture.add_cats(["Fluffy", "Snookums"])
assert cattery_fixture.cats == ["Fluffy", "Snookums"]
assert cattery_fixture.num_cats == 2
###########################################################################
# remove_cat
###########################################################################
def test__remove_cat__succeeds(cattery_fixture):
cattery_fixture.add_cats(["Fluffy", "Junior"])
cattery_fixture.remove_cat("Fluffy")
assert cattery_fixture.cats == ["Junior"]
assert cattery_fixture.num_cats == 1
def test__remove_cat__no_cats__fails(cattery_fixture):
with pytest.raises(cattery.CatNotFound):
cattery_fixture.remove_cat("Fluffles")
def test__remove_cat__cat_not_in_cattery__fails(cattery_fixture):
cattery_fixture.add_cats(["Fluffy"])
with pytest.raises(cattery.CatNotFound):
cattery_fixture.remove_cat("Snookums")
| <commit_before>import pytest
from catinabox import cattery
class TestCattery(object):
###########################################################################
# add_cats
###########################################################################
def test__add_cats__succeeds(self):
c = cattery.Cattery()
assert c
###########################################################################
# remove_cat
###########################################################################
def test__remove_cat__succeeds(self):
c = cattery.Cattery()
assert c
def test__remove_cat__no_cats__fails(self):
c = cattery.Cattery()
assert c
def test__remove_cat__cat_not_in_cattery__fails(self):
c = cattery.Cattery()
c.add_cats(["Fluffy"])
with pytest.raises(cattery.CatNotFound):
c.remove_cat("Snookums")
<commit_msg>Add full tests for mccattery and cattery<commit_after> | import pytest
from catinabox import cattery, mccattery
@pytest.fixture(params=[
cattery.Cattery,
mccattery.McCattery
])
def cattery_fixture(request):
return request.param()
###########################################################################
# add_cats
###########################################################################
def test__add_cats__succeeds(cattery_fixture):
cattery_fixture.add_cats(["Fluffy", "Snookums"])
assert cattery_fixture.cats == ["Fluffy", "Snookums"]
assert cattery_fixture.num_cats == 2
###########################################################################
# remove_cat
###########################################################################
def test__remove_cat__succeeds(cattery_fixture):
cattery_fixture.add_cats(["Fluffy", "Junior"])
cattery_fixture.remove_cat("Fluffy")
assert cattery_fixture.cats == ["Junior"]
assert cattery_fixture.num_cats == 1
def test__remove_cat__no_cats__fails(cattery_fixture):
with pytest.raises(cattery.CatNotFound):
cattery_fixture.remove_cat("Fluffles")
def test__remove_cat__cat_not_in_cattery__fails(cattery_fixture):
cattery_fixture.add_cats(["Fluffy"])
with pytest.raises(cattery.CatNotFound):
cattery_fixture.remove_cat("Snookums")
| import pytest
from catinabox import cattery
class TestCattery(object):
###########################################################################
# add_cats
###########################################################################
def test__add_cats__succeeds(self):
c = cattery.Cattery()
assert c
###########################################################################
# remove_cat
###########################################################################
def test__remove_cat__succeeds(self):
c = cattery.Cattery()
assert c
def test__remove_cat__no_cats__fails(self):
c = cattery.Cattery()
assert c
def test__remove_cat__cat_not_in_cattery__fails(self):
c = cattery.Cattery()
c.add_cats(["Fluffy"])
with pytest.raises(cattery.CatNotFound):
c.remove_cat("Snookums")
Add full tests for mccattery and catteryimport pytest
from catinabox import cattery, mccattery
@pytest.fixture(params=[
cattery.Cattery,
mccattery.McCattery
])
def cattery_fixture(request):
return request.param()
###########################################################################
# add_cats
###########################################################################
def test__add_cats__succeeds(cattery_fixture):
cattery_fixture.add_cats(["Fluffy", "Snookums"])
assert cattery_fixture.cats == ["Fluffy", "Snookums"]
assert cattery_fixture.num_cats == 2
###########################################################################
# remove_cat
###########################################################################
def test__remove_cat__succeeds(cattery_fixture):
cattery_fixture.add_cats(["Fluffy", "Junior"])
cattery_fixture.remove_cat("Fluffy")
assert cattery_fixture.cats == ["Junior"]
assert cattery_fixture.num_cats == 1
def test__remove_cat__no_cats__fails(cattery_fixture):
with pytest.raises(cattery.CatNotFound):
cattery_fixture.remove_cat("Fluffles")
def test__remove_cat__cat_not_in_cattery__fails(cattery_fixture):
cattery_fixture.add_cats(["Fluffy"])
with pytest.raises(cattery.CatNotFound):
cattery_fixture.remove_cat("Snookums")
| <commit_before>import pytest
from catinabox import cattery
class TestCattery(object):
###########################################################################
# add_cats
###########################################################################
def test__add_cats__succeeds(self):
c = cattery.Cattery()
assert c
###########################################################################
# remove_cat
###########################################################################
def test__remove_cat__succeeds(self):
c = cattery.Cattery()
assert c
def test__remove_cat__no_cats__fails(self):
c = cattery.Cattery()
assert c
def test__remove_cat__cat_not_in_cattery__fails(self):
c = cattery.Cattery()
c.add_cats(["Fluffy"])
with pytest.raises(cattery.CatNotFound):
c.remove_cat("Snookums")
<commit_msg>Add full tests for mccattery and cattery<commit_after>import pytest
from catinabox import cattery, mccattery
@pytest.fixture(params=[
cattery.Cattery,
mccattery.McCattery
])
def cattery_fixture(request):
return request.param()
###########################################################################
# add_cats
###########################################################################
def test__add_cats__succeeds(cattery_fixture):
cattery_fixture.add_cats(["Fluffy", "Snookums"])
assert cattery_fixture.cats == ["Fluffy", "Snookums"]
assert cattery_fixture.num_cats == 2
###########################################################################
# remove_cat
###########################################################################
def test__remove_cat__succeeds(cattery_fixture):
cattery_fixture.add_cats(["Fluffy", "Junior"])
cattery_fixture.remove_cat("Fluffy")
assert cattery_fixture.cats == ["Junior"]
assert cattery_fixture.num_cats == 1
def test__remove_cat__no_cats__fails(cattery_fixture):
with pytest.raises(cattery.CatNotFound):
cattery_fixture.remove_cat("Fluffles")
def test__remove_cat__cat_not_in_cattery__fails(cattery_fixture):
cattery_fixture.add_cats(["Fluffy"])
with pytest.raises(cattery.CatNotFound):
cattery_fixture.remove_cat("Snookums")
|
e79f87121a955847af92d93fad3e2687fc4f472f | tests/test_gen_sql.py | tests/test_gen_sql.py | #!/usr/bin/env python
class TestGenSql:
def test_gen_drop_statement(self):
pass
def test_create_statement(self):
pass
| #!/usr/bin/env python
import sys
from io import StringIO
from pg_bawler import gen_sql
def test_simple_main(monkeypatch):
stdout = StringIO()
monkeypatch.setattr(sys, 'stdout', stdout)
class Args:
tablename = 'foo'
gen_sql.main(*[Args.tablename])
sql = stdout.getvalue()
assert gen_sql.TRIGGER_FN_FMT.format(args=Args) in sql
assert gen_sql.TRIGGER_FN_FMT.format(args=Args) in sql
| Add simple test for sql full sql generation with only tablename | Add simple test for sql full sql generation with only tablename
| Python | bsd-3-clause | beezz/pg_bawler,beezz/pg_bawler | #!/usr/bin/env python
class TestGenSql:
def test_gen_drop_statement(self):
pass
def test_create_statement(self):
pass
Add simple test for sql full sql generation with only tablename | #!/usr/bin/env python
import sys
from io import StringIO
from pg_bawler import gen_sql
def test_simple_main(monkeypatch):
stdout = StringIO()
monkeypatch.setattr(sys, 'stdout', stdout)
class Args:
tablename = 'foo'
gen_sql.main(*[Args.tablename])
sql = stdout.getvalue()
assert gen_sql.TRIGGER_FN_FMT.format(args=Args) in sql
assert gen_sql.TRIGGER_FN_FMT.format(args=Args) in sql
| <commit_before>#!/usr/bin/env python
class TestGenSql:
def test_gen_drop_statement(self):
pass
def test_create_statement(self):
pass
<commit_msg>Add simple test for sql full sql generation with only tablename<commit_after> | #!/usr/bin/env python
import sys
from io import StringIO
from pg_bawler import gen_sql
def test_simple_main(monkeypatch):
stdout = StringIO()
monkeypatch.setattr(sys, 'stdout', stdout)
class Args:
tablename = 'foo'
gen_sql.main(*[Args.tablename])
sql = stdout.getvalue()
assert gen_sql.TRIGGER_FN_FMT.format(args=Args) in sql
assert gen_sql.TRIGGER_FN_FMT.format(args=Args) in sql
| #!/usr/bin/env python
class TestGenSql:
def test_gen_drop_statement(self):
pass
def test_create_statement(self):
pass
Add simple test for sql full sql generation with only tablename#!/usr/bin/env python
import sys
from io import StringIO
from pg_bawler import gen_sql
def test_simple_main(monkeypatch):
stdout = StringIO()
monkeypatch.setattr(sys, 'stdout', stdout)
class Args:
tablename = 'foo'
gen_sql.main(*[Args.tablename])
sql = stdout.getvalue()
assert gen_sql.TRIGGER_FN_FMT.format(args=Args) in sql
assert gen_sql.TRIGGER_FN_FMT.format(args=Args) in sql
| <commit_before>#!/usr/bin/env python
class TestGenSql:
def test_gen_drop_statement(self):
pass
def test_create_statement(self):
pass
<commit_msg>Add simple test for sql full sql generation with only tablename<commit_after>#!/usr/bin/env python
import sys
from io import StringIO
from pg_bawler import gen_sql
def test_simple_main(monkeypatch):
stdout = StringIO()
monkeypatch.setattr(sys, 'stdout', stdout)
class Args:
tablename = 'foo'
gen_sql.main(*[Args.tablename])
sql = stdout.getvalue()
assert gen_sql.TRIGGER_FN_FMT.format(args=Args) in sql
assert gen_sql.TRIGGER_FN_FMT.format(args=Args) in sql
|
d60116aecbb6935fae508c94905a335fdb0603bb | tests/test_xgboost.py | tests/test_xgboost.py | import unittest
from sklearn import datasets
from xgboost import XGBClassifier
class TestXGBoost(unittest.TestCase):
def test_classifier(self):
boston = datasets.load_boston()
X, y = boston.data, boston.target
xgb1 = XGBClassifier(n_estimators=3)
xgb1.fit(X[0:70],y[0:70])
| import unittest
import xgboost
from distutils.version import StrictVersion
from sklearn import datasets
from xgboost import XGBClassifier
class TestXGBoost(unittest.TestCase):
def test_version(self):
# b/175051617 prevent xgboost version downgrade.
self.assertGreaterEqual(StrictVersion(xgboost.__version__), StrictVersion("1.2.1"))
def test_classifier(self):
boston = datasets.load_boston()
X, y = boston.data, boston.target
xgb1 = XGBClassifier(n_estimators=3)
xgb1.fit(X[0:70],y[0:70])
| Add xgboost version regression test. | Add xgboost version regression test.
BUG=175051617
| Python | apache-2.0 | Kaggle/docker-python,Kaggle/docker-python | import unittest
from sklearn import datasets
from xgboost import XGBClassifier
class TestXGBoost(unittest.TestCase):
def test_classifier(self):
boston = datasets.load_boston()
X, y = boston.data, boston.target
xgb1 = XGBClassifier(n_estimators=3)
xgb1.fit(X[0:70],y[0:70])
Add xgboost version regression test.
BUG=175051617 | import unittest
import xgboost
from distutils.version import StrictVersion
from sklearn import datasets
from xgboost import XGBClassifier
class TestXGBoost(unittest.TestCase):
def test_version(self):
# b/175051617 prevent xgboost version downgrade.
self.assertGreaterEqual(StrictVersion(xgboost.__version__), StrictVersion("1.2.1"))
def test_classifier(self):
boston = datasets.load_boston()
X, y = boston.data, boston.target
xgb1 = XGBClassifier(n_estimators=3)
xgb1.fit(X[0:70],y[0:70])
| <commit_before>import unittest
from sklearn import datasets
from xgboost import XGBClassifier
class TestXGBoost(unittest.TestCase):
def test_classifier(self):
boston = datasets.load_boston()
X, y = boston.data, boston.target
xgb1 = XGBClassifier(n_estimators=3)
xgb1.fit(X[0:70],y[0:70])
<commit_msg>Add xgboost version regression test.
BUG=175051617<commit_after> | import unittest
import xgboost
from distutils.version import StrictVersion
from sklearn import datasets
from xgboost import XGBClassifier
class TestXGBoost(unittest.TestCase):
def test_version(self):
# b/175051617 prevent xgboost version downgrade.
self.assertGreaterEqual(StrictVersion(xgboost.__version__), StrictVersion("1.2.1"))
def test_classifier(self):
boston = datasets.load_boston()
X, y = boston.data, boston.target
xgb1 = XGBClassifier(n_estimators=3)
xgb1.fit(X[0:70],y[0:70])
| import unittest
from sklearn import datasets
from xgboost import XGBClassifier
class TestXGBoost(unittest.TestCase):
def test_classifier(self):
boston = datasets.load_boston()
X, y = boston.data, boston.target
xgb1 = XGBClassifier(n_estimators=3)
xgb1.fit(X[0:70],y[0:70])
Add xgboost version regression test.
BUG=175051617import unittest
import xgboost
from distutils.version import StrictVersion
from sklearn import datasets
from xgboost import XGBClassifier
class TestXGBoost(unittest.TestCase):
def test_version(self):
# b/175051617 prevent xgboost version downgrade.
self.assertGreaterEqual(StrictVersion(xgboost.__version__), StrictVersion("1.2.1"))
def test_classifier(self):
boston = datasets.load_boston()
X, y = boston.data, boston.target
xgb1 = XGBClassifier(n_estimators=3)
xgb1.fit(X[0:70],y[0:70])
| <commit_before>import unittest
from sklearn import datasets
from xgboost import XGBClassifier
class TestXGBoost(unittest.TestCase):
def test_classifier(self):
boston = datasets.load_boston()
X, y = boston.data, boston.target
xgb1 = XGBClassifier(n_estimators=3)
xgb1.fit(X[0:70],y[0:70])
<commit_msg>Add xgboost version regression test.
BUG=175051617<commit_after>import unittest
import xgboost
from distutils.version import StrictVersion
from sklearn import datasets
from xgboost import XGBClassifier
class TestXGBoost(unittest.TestCase):
def test_version(self):
# b/175051617 prevent xgboost version downgrade.
self.assertGreaterEqual(StrictVersion(xgboost.__version__), StrictVersion("1.2.1"))
def test_classifier(self):
boston = datasets.load_boston()
X, y = boston.data, boston.target
xgb1 = XGBClassifier(n_estimators=3)
xgb1.fit(X[0:70],y[0:70])
|
a57e40ea7b0cc55ec67664d9f32658085c24900f | tools/project/check_style.py | tools/project/check_style.py | import subprocess
import sys
git_diff_output = subprocess.check_output(
"git diff --name-only --diff-filter=ACM", universal_newlines=True)
git_diff_lines = git_diff_output.split("\n")
for file_name in git_diff_lines:
if not file_name:
continue
print "Checking style for %s" %file_name
ret_value = subprocess.call(
"python 3rdparty/styleguide/cpplint/cpplint.py --filter=%s %s"
%("-build/include_order", file_name), stderr=subprocess.STDOUT)
if ret_value != 0:
sys.exit(ret_value)
sys.exit(0)
| import subprocess
import sys
git_diff_output = subprocess.check_output(
"git diff --name-only --diff-filter=ACM", universal_newlines=True)
git_diff_lines = git_diff_output.split("\n")
for file_name in git_diff_lines:
if not file_name:
continue
print "Checking style for %s" %file_name
ret_value = subprocess.call(
"python 3rdparty/styleguide/cpplint/cpplint.py --filter="
"-build/include_order,-legal/copyright %s"
%file_name, stderr=subprocess.STDOUT)
if ret_value != 0:
sys.exit(ret_value)
sys.exit(0)
| Update style checker options to ignore copyright. | Update style checker options to ignore copyright.
| Python | mit | damlaren/ogle,damlaren/ogle,damlaren/ogle | import subprocess
import sys
git_diff_output = subprocess.check_output(
"git diff --name-only --diff-filter=ACM", universal_newlines=True)
git_diff_lines = git_diff_output.split("\n")
for file_name in git_diff_lines:
if not file_name:
continue
print "Checking style for %s" %file_name
ret_value = subprocess.call(
"python 3rdparty/styleguide/cpplint/cpplint.py --filter=%s %s"
%("-build/include_order", file_name), stderr=subprocess.STDOUT)
if ret_value != 0:
sys.exit(ret_value)
sys.exit(0)
Update style checker options to ignore copyright. | import subprocess
import sys
git_diff_output = subprocess.check_output(
"git diff --name-only --diff-filter=ACM", universal_newlines=True)
git_diff_lines = git_diff_output.split("\n")
for file_name in git_diff_lines:
if not file_name:
continue
print "Checking style for %s" %file_name
ret_value = subprocess.call(
"python 3rdparty/styleguide/cpplint/cpplint.py --filter="
"-build/include_order,-legal/copyright %s"
%file_name, stderr=subprocess.STDOUT)
if ret_value != 0:
sys.exit(ret_value)
sys.exit(0)
| <commit_before>import subprocess
import sys
git_diff_output = subprocess.check_output(
"git diff --name-only --diff-filter=ACM", universal_newlines=True)
git_diff_lines = git_diff_output.split("\n")
for file_name in git_diff_lines:
if not file_name:
continue
print "Checking style for %s" %file_name
ret_value = subprocess.call(
"python 3rdparty/styleguide/cpplint/cpplint.py --filter=%s %s"
%("-build/include_order", file_name), stderr=subprocess.STDOUT)
if ret_value != 0:
sys.exit(ret_value)
sys.exit(0)
<commit_msg>Update style checker options to ignore copyright.<commit_after> | import subprocess
import sys
git_diff_output = subprocess.check_output(
"git diff --name-only --diff-filter=ACM", universal_newlines=True)
git_diff_lines = git_diff_output.split("\n")
for file_name in git_diff_lines:
if not file_name:
continue
print "Checking style for %s" %file_name
ret_value = subprocess.call(
"python 3rdparty/styleguide/cpplint/cpplint.py --filter="
"-build/include_order,-legal/copyright %s"
%file_name, stderr=subprocess.STDOUT)
if ret_value != 0:
sys.exit(ret_value)
sys.exit(0)
| import subprocess
import sys
git_diff_output = subprocess.check_output(
"git diff --name-only --diff-filter=ACM", universal_newlines=True)
git_diff_lines = git_diff_output.split("\n")
for file_name in git_diff_lines:
if not file_name:
continue
print "Checking style for %s" %file_name
ret_value = subprocess.call(
"python 3rdparty/styleguide/cpplint/cpplint.py --filter=%s %s"
%("-build/include_order", file_name), stderr=subprocess.STDOUT)
if ret_value != 0:
sys.exit(ret_value)
sys.exit(0)
Update style checker options to ignore copyright.import subprocess
import sys
git_diff_output = subprocess.check_output(
"git diff --name-only --diff-filter=ACM", universal_newlines=True)
git_diff_lines = git_diff_output.split("\n")
for file_name in git_diff_lines:
if not file_name:
continue
print "Checking style for %s" %file_name
ret_value = subprocess.call(
"python 3rdparty/styleguide/cpplint/cpplint.py --filter="
"-build/include_order,-legal/copyright %s"
%file_name, stderr=subprocess.STDOUT)
if ret_value != 0:
sys.exit(ret_value)
sys.exit(0)
| <commit_before>import subprocess
import sys
git_diff_output = subprocess.check_output(
"git diff --name-only --diff-filter=ACM", universal_newlines=True)
git_diff_lines = git_diff_output.split("\n")
for file_name in git_diff_lines:
if not file_name:
continue
print "Checking style for %s" %file_name
ret_value = subprocess.call(
"python 3rdparty/styleguide/cpplint/cpplint.py --filter=%s %s"
%("-build/include_order", file_name), stderr=subprocess.STDOUT)
if ret_value != 0:
sys.exit(ret_value)
sys.exit(0)
<commit_msg>Update style checker options to ignore copyright.<commit_after>import subprocess
import sys
git_diff_output = subprocess.check_output(
"git diff --name-only --diff-filter=ACM", universal_newlines=True)
git_diff_lines = git_diff_output.split("\n")
for file_name in git_diff_lines:
if not file_name:
continue
print "Checking style for %s" %file_name
ret_value = subprocess.call(
"python 3rdparty/styleguide/cpplint/cpplint.py --filter="
"-build/include_order,-legal/copyright %s"
%file_name, stderr=subprocess.STDOUT)
if ret_value != 0:
sys.exit(ret_value)
sys.exit(0)
|
800706f5835293ee20dd9505d1d11c28eb38bbb2 | tests/shipane_sdk/matchers/dataframe_matchers.py | tests/shipane_sdk/matchers/dataframe_matchers.py | # -*- coding: utf-8 -*-
import re
from hamcrest.core.base_matcher import BaseMatcher
class HasColumn(BaseMatcher):
def __init__(self, column):
self._column = column
def _matches(self, df):
return self._column in df.columns
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum [{0}]'.format(self._column))
def has_column(column):
return HasColumn(column)
class HasColumnMatches(BaseMatcher):
def __init__(self, column_pattern):
self._column_pattern = re.compile(column_pattern)
def _matches(self, df):
return df.filter(regex=self._column_pattern).columns.size > 0
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum matches [{0}]'.format(self._column_pattern))
def has_column_matches(column_pattern):
return HasColumnMatches(column_pattern)
class HasRow(BaseMatcher):
def __init__(self, row):
self._row = row
def _matches(self, df):
return self._row in df.index
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have row [%s]'.format(self._row))
def has_row(row):
return HasRow(row)
| # -*- coding: utf-8 -*-
import re
from hamcrest.core.base_matcher import BaseMatcher
class HasColumn(BaseMatcher):
def __init__(self, column):
self._column = column
def _matches(self, df):
return self._column in df.columns
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum [{0}]'.format(self._column))
def has_column(column):
return HasColumn(column)
class HasColumnMatches(BaseMatcher):
def __init__(self, column_pattern):
self._column_pattern = re.compile(column_pattern)
def _matches(self, df):
return len(list(filter(self._column_pattern.match, df.columns.values))) > 0
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum matches [{0}]'.format(self._column_pattern))
def has_column_matches(column_pattern):
return HasColumnMatches(column_pattern)
class HasRow(BaseMatcher):
def __init__(self, row):
self._row = row
def _matches(self, df):
return self._row in df.index
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have row [%s]'.format(self._row))
def has_row(row):
return HasRow(row)
| Fix HasColumn matcher for dataframe with duplicated columns | Fix HasColumn matcher for dataframe with duplicated columns
| Python | mit | sinall/ShiPanE-Python-SDK,sinall/ShiPanE-Python-SDK | # -*- coding: utf-8 -*-
import re
from hamcrest.core.base_matcher import BaseMatcher
class HasColumn(BaseMatcher):
def __init__(self, column):
self._column = column
def _matches(self, df):
return self._column in df.columns
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum [{0}]'.format(self._column))
def has_column(column):
return HasColumn(column)
class HasColumnMatches(BaseMatcher):
def __init__(self, column_pattern):
self._column_pattern = re.compile(column_pattern)
def _matches(self, df):
return df.filter(regex=self._column_pattern).columns.size > 0
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum matches [{0}]'.format(self._column_pattern))
def has_column_matches(column_pattern):
return HasColumnMatches(column_pattern)
class HasRow(BaseMatcher):
def __init__(self, row):
self._row = row
def _matches(self, df):
return self._row in df.index
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have row [%s]'.format(self._row))
def has_row(row):
return HasRow(row)
Fix HasColumn matcher for dataframe with duplicated columns | # -*- coding: utf-8 -*-
import re
from hamcrest.core.base_matcher import BaseMatcher
class HasColumn(BaseMatcher):
def __init__(self, column):
self._column = column
def _matches(self, df):
return self._column in df.columns
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum [{0}]'.format(self._column))
def has_column(column):
return HasColumn(column)
class HasColumnMatches(BaseMatcher):
def __init__(self, column_pattern):
self._column_pattern = re.compile(column_pattern)
def _matches(self, df):
return len(list(filter(self._column_pattern.match, df.columns.values))) > 0
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum matches [{0}]'.format(self._column_pattern))
def has_column_matches(column_pattern):
return HasColumnMatches(column_pattern)
class HasRow(BaseMatcher):
def __init__(self, row):
self._row = row
def _matches(self, df):
return self._row in df.index
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have row [%s]'.format(self._row))
def has_row(row):
return HasRow(row)
| <commit_before># -*- coding: utf-8 -*-
import re
from hamcrest.core.base_matcher import BaseMatcher
class HasColumn(BaseMatcher):
def __init__(self, column):
self._column = column
def _matches(self, df):
return self._column in df.columns
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum [{0}]'.format(self._column))
def has_column(column):
return HasColumn(column)
class HasColumnMatches(BaseMatcher):
def __init__(self, column_pattern):
self._column_pattern = re.compile(column_pattern)
def _matches(self, df):
return df.filter(regex=self._column_pattern).columns.size > 0
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum matches [{0}]'.format(self._column_pattern))
def has_column_matches(column_pattern):
return HasColumnMatches(column_pattern)
class HasRow(BaseMatcher):
def __init__(self, row):
self._row = row
def _matches(self, df):
return self._row in df.index
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have row [%s]'.format(self._row))
def has_row(row):
return HasRow(row)
<commit_msg>Fix HasColumn matcher for dataframe with duplicated columns<commit_after> | # -*- coding: utf-8 -*-
import re
from hamcrest.core.base_matcher import BaseMatcher
class HasColumn(BaseMatcher):
def __init__(self, column):
self._column = column
def _matches(self, df):
return self._column in df.columns
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum [{0}]'.format(self._column))
def has_column(column):
return HasColumn(column)
class HasColumnMatches(BaseMatcher):
def __init__(self, column_pattern):
self._column_pattern = re.compile(column_pattern)
def _matches(self, df):
return len(list(filter(self._column_pattern.match, df.columns.values))) > 0
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum matches [{0}]'.format(self._column_pattern))
def has_column_matches(column_pattern):
return HasColumnMatches(column_pattern)
class HasRow(BaseMatcher):
def __init__(self, row):
self._row = row
def _matches(self, df):
return self._row in df.index
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have row [%s]'.format(self._row))
def has_row(row):
return HasRow(row)
| # -*- coding: utf-8 -*-
import re
from hamcrest.core.base_matcher import BaseMatcher
class HasColumn(BaseMatcher):
def __init__(self, column):
self._column = column
def _matches(self, df):
return self._column in df.columns
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum [{0}]'.format(self._column))
def has_column(column):
return HasColumn(column)
class HasColumnMatches(BaseMatcher):
def __init__(self, column_pattern):
self._column_pattern = re.compile(column_pattern)
def _matches(self, df):
return df.filter(regex=self._column_pattern).columns.size > 0
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum matches [{0}]'.format(self._column_pattern))
def has_column_matches(column_pattern):
return HasColumnMatches(column_pattern)
class HasRow(BaseMatcher):
def __init__(self, row):
self._row = row
def _matches(self, df):
return self._row in df.index
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have row [%s]'.format(self._row))
def has_row(row):
return HasRow(row)
Fix HasColumn matcher for dataframe with duplicated columns# -*- coding: utf-8 -*-
import re
from hamcrest.core.base_matcher import BaseMatcher
class HasColumn(BaseMatcher):
def __init__(self, column):
self._column = column
def _matches(self, df):
return self._column in df.columns
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum [{0}]'.format(self._column))
def has_column(column):
return HasColumn(column)
class HasColumnMatches(BaseMatcher):
def __init__(self, column_pattern):
self._column_pattern = re.compile(column_pattern)
def _matches(self, df):
return len(list(filter(self._column_pattern.match, df.columns.values))) > 0
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum matches [{0}]'.format(self._column_pattern))
def has_column_matches(column_pattern):
return HasColumnMatches(column_pattern)
class HasRow(BaseMatcher):
def __init__(self, row):
self._row = row
def _matches(self, df):
return self._row in df.index
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have row [%s]'.format(self._row))
def has_row(row):
return HasRow(row)
| <commit_before># -*- coding: utf-8 -*-
import re
from hamcrest.core.base_matcher import BaseMatcher
class HasColumn(BaseMatcher):
def __init__(self, column):
self._column = column
def _matches(self, df):
return self._column in df.columns
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum [{0}]'.format(self._column))
def has_column(column):
return HasColumn(column)
class HasColumnMatches(BaseMatcher):
def __init__(self, column_pattern):
self._column_pattern = re.compile(column_pattern)
def _matches(self, df):
return df.filter(regex=self._column_pattern).columns.size > 0
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum matches [{0}]'.format(self._column_pattern))
def has_column_matches(column_pattern):
return HasColumnMatches(column_pattern)
class HasRow(BaseMatcher):
def __init__(self, row):
self._row = row
def _matches(self, df):
return self._row in df.index
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have row [%s]'.format(self._row))
def has_row(row):
return HasRow(row)
<commit_msg>Fix HasColumn matcher for dataframe with duplicated columns<commit_after># -*- coding: utf-8 -*-
import re
from hamcrest.core.base_matcher import BaseMatcher
class HasColumn(BaseMatcher):
def __init__(self, column):
self._column = column
def _matches(self, df):
return self._column in df.columns
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum [{0}]'.format(self._column))
def has_column(column):
return HasColumn(column)
class HasColumnMatches(BaseMatcher):
def __init__(self, column_pattern):
self._column_pattern = re.compile(column_pattern)
def _matches(self, df):
return len(list(filter(self._column_pattern.match, df.columns.values))) > 0
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have colum matches [{0}]'.format(self._column_pattern))
def has_column_matches(column_pattern):
return HasColumnMatches(column_pattern)
class HasRow(BaseMatcher):
def __init__(self, row):
self._row = row
def _matches(self, df):
return self._row in df.index
def describe_to(self, description):
description.append_text(u'Dataframe doesn\'t have row [%s]'.format(self._row))
def has_row(row):
return HasRow(row)
|
7d6800c33a525355714e355ec87e989372c293d7 | main.py | main.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from google.appengine.ext import webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.api import xmpp
#from google.appengine.api import urlfetch
import urllib2,md5,logging,csv
############## webapp Models ###################
class MainPage(webapp.RequestHandler):
def get(self):
url = 'http://www.twse.com.tw/ch/trading/exchange/STOCK_DAY_AVG/STOCK_DAY_AVG2.php?STK_NO=2363&myear=2010&mmon=06&type=csv'
cc = urllib2.urlopen(url)
csv_read = csv.reader(cc)
self.response.out.write('Go Ri Stock')
#csv_read.next
for i in csv_read:
print i
self.response.out.write('<br>%s' % type(csv_read))
############## main Models ###################
def main():
""" Start up. """
application = webapp.WSGIApplication(
[
('/', MainPage)
],debug=True)
run_wsgi_app(application)
if __name__ == '__main__':
main()
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from google.appengine.ext import webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.api import xmpp
#from google.appengine.api import urlfetch
import urllib2,md5,logging,csv,re,math
def ckinv(oo):
""" check the value is date or not """
pattern = re.compile(r"[0-9]{2}/[0-9]{2}/[0-9]{2}")
b = re.search(pattern, oo[0])
try:
b.group()
return True
except:
return False
def covstr(s):
""" convert string to int or float. """
try:
ret = int(s)
except ValueError:
ret = float(s)
return ret
############## webapp Models ###################
class MainPage(webapp.RequestHandler):
def get(self):
url = 'http://www.twse.com.tw/ch/trading/exchange/STOCK_DAY_AVG/STOCK_DAY_AVG2.php?STK_NO=2363&myear=2010&mmon=06&type=csv'
cc = urllib2.urlopen(url)
csv_read = csv.reader(cc)
self.response.out.write('Go Ri Stock')
#csv_read.next
getr = []
for i in csv_read:
print i
if ckinv(i):
getr.append(covstr(i[1]))
print getr
print math.fsum(getr)
############## main Models ###################
def main():
""" Start up. """
application = webapp.WSGIApplication(
[
('/', MainPage)
],debug=True)
run_wsgi_app(application)
if __name__ == '__main__':
main()
| Add check the value is date or not. | Add check the value is date or not.
| Python | mit | toomore/goristock | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from google.appengine.ext import webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.api import xmpp
#from google.appengine.api import urlfetch
import urllib2,md5,logging,csv
############## webapp Models ###################
class MainPage(webapp.RequestHandler):
def get(self):
url = 'http://www.twse.com.tw/ch/trading/exchange/STOCK_DAY_AVG/STOCK_DAY_AVG2.php?STK_NO=2363&myear=2010&mmon=06&type=csv'
cc = urllib2.urlopen(url)
csv_read = csv.reader(cc)
self.response.out.write('Go Ri Stock')
#csv_read.next
for i in csv_read:
print i
self.response.out.write('<br>%s' % type(csv_read))
############## main Models ###################
def main():
""" Start up. """
application = webapp.WSGIApplication(
[
('/', MainPage)
],debug=True)
run_wsgi_app(application)
if __name__ == '__main__':
main()
Add check the value is date or not. | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from google.appengine.ext import webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.api import xmpp
#from google.appengine.api import urlfetch
import urllib2,md5,logging,csv,re,math
def ckinv(oo):
""" check the value is date or not """
pattern = re.compile(r"[0-9]{2}/[0-9]{2}/[0-9]{2}")
b = re.search(pattern, oo[0])
try:
b.group()
return True
except:
return False
def covstr(s):
""" convert string to int or float. """
try:
ret = int(s)
except ValueError:
ret = float(s)
return ret
############## webapp Models ###################
class MainPage(webapp.RequestHandler):
def get(self):
url = 'http://www.twse.com.tw/ch/trading/exchange/STOCK_DAY_AVG/STOCK_DAY_AVG2.php?STK_NO=2363&myear=2010&mmon=06&type=csv'
cc = urllib2.urlopen(url)
csv_read = csv.reader(cc)
self.response.out.write('Go Ri Stock')
#csv_read.next
getr = []
for i in csv_read:
print i
if ckinv(i):
getr.append(covstr(i[1]))
print getr
print math.fsum(getr)
############## main Models ###################
def main():
""" Start up. """
application = webapp.WSGIApplication(
[
('/', MainPage)
],debug=True)
run_wsgi_app(application)
if __name__ == '__main__':
main()
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from google.appengine.ext import webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.api import xmpp
#from google.appengine.api import urlfetch
import urllib2,md5,logging,csv
############## webapp Models ###################
class MainPage(webapp.RequestHandler):
def get(self):
url = 'http://www.twse.com.tw/ch/trading/exchange/STOCK_DAY_AVG/STOCK_DAY_AVG2.php?STK_NO=2363&myear=2010&mmon=06&type=csv'
cc = urllib2.urlopen(url)
csv_read = csv.reader(cc)
self.response.out.write('Go Ri Stock')
#csv_read.next
for i in csv_read:
print i
self.response.out.write('<br>%s' % type(csv_read))
############## main Models ###################
def main():
""" Start up. """
application = webapp.WSGIApplication(
[
('/', MainPage)
],debug=True)
run_wsgi_app(application)
if __name__ == '__main__':
main()
<commit_msg>Add check the value is date or not.<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from google.appengine.ext import webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.api import xmpp
#from google.appengine.api import urlfetch
import urllib2,md5,logging,csv,re,math
def ckinv(oo):
""" check the value is date or not """
pattern = re.compile(r"[0-9]{2}/[0-9]{2}/[0-9]{2}")
b = re.search(pattern, oo[0])
try:
b.group()
return True
except:
return False
def covstr(s):
""" convert string to int or float. """
try:
ret = int(s)
except ValueError:
ret = float(s)
return ret
############## webapp Models ###################
class MainPage(webapp.RequestHandler):
def get(self):
url = 'http://www.twse.com.tw/ch/trading/exchange/STOCK_DAY_AVG/STOCK_DAY_AVG2.php?STK_NO=2363&myear=2010&mmon=06&type=csv'
cc = urllib2.urlopen(url)
csv_read = csv.reader(cc)
self.response.out.write('Go Ri Stock')
#csv_read.next
getr = []
for i in csv_read:
print i
if ckinv(i):
getr.append(covstr(i[1]))
print getr
print math.fsum(getr)
############## main Models ###################
def main():
""" Start up. """
application = webapp.WSGIApplication(
[
('/', MainPage)
],debug=True)
run_wsgi_app(application)
if __name__ == '__main__':
main()
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from google.appengine.ext import webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.api import xmpp
#from google.appengine.api import urlfetch
import urllib2,md5,logging,csv
############## webapp Models ###################
class MainPage(webapp.RequestHandler):
def get(self):
url = 'http://www.twse.com.tw/ch/trading/exchange/STOCK_DAY_AVG/STOCK_DAY_AVG2.php?STK_NO=2363&myear=2010&mmon=06&type=csv'
cc = urllib2.urlopen(url)
csv_read = csv.reader(cc)
self.response.out.write('Go Ri Stock')
#csv_read.next
for i in csv_read:
print i
self.response.out.write('<br>%s' % type(csv_read))
############## main Models ###################
def main():
""" Start up. """
application = webapp.WSGIApplication(
[
('/', MainPage)
],debug=True)
run_wsgi_app(application)
if __name__ == '__main__':
main()
Add check the value is date or not.#!/usr/bin/env python
# -*- coding: utf-8 -*-
from google.appengine.ext import webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.api import xmpp
#from google.appengine.api import urlfetch
import urllib2,md5,logging,csv,re,math
def ckinv(oo):
""" check the value is date or not """
pattern = re.compile(r"[0-9]{2}/[0-9]{2}/[0-9]{2}")
b = re.search(pattern, oo[0])
try:
b.group()
return True
except:
return False
def covstr(s):
""" convert string to int or float. """
try:
ret = int(s)
except ValueError:
ret = float(s)
return ret
############## webapp Models ###################
class MainPage(webapp.RequestHandler):
def get(self):
url = 'http://www.twse.com.tw/ch/trading/exchange/STOCK_DAY_AVG/STOCK_DAY_AVG2.php?STK_NO=2363&myear=2010&mmon=06&type=csv'
cc = urllib2.urlopen(url)
csv_read = csv.reader(cc)
self.response.out.write('Go Ri Stock')
#csv_read.next
getr = []
for i in csv_read:
print i
if ckinv(i):
getr.append(covstr(i[1]))
print getr
print math.fsum(getr)
############## main Models ###################
def main():
""" Start up. """
application = webapp.WSGIApplication(
[
('/', MainPage)
],debug=True)
run_wsgi_app(application)
if __name__ == '__main__':
main()
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from google.appengine.ext import webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.api import xmpp
#from google.appengine.api import urlfetch
import urllib2,md5,logging,csv
############## webapp Models ###################
class MainPage(webapp.RequestHandler):
def get(self):
url = 'http://www.twse.com.tw/ch/trading/exchange/STOCK_DAY_AVG/STOCK_DAY_AVG2.php?STK_NO=2363&myear=2010&mmon=06&type=csv'
cc = urllib2.urlopen(url)
csv_read = csv.reader(cc)
self.response.out.write('Go Ri Stock')
#csv_read.next
for i in csv_read:
print i
self.response.out.write('<br>%s' % type(csv_read))
############## main Models ###################
def main():
""" Start up. """
application = webapp.WSGIApplication(
[
('/', MainPage)
],debug=True)
run_wsgi_app(application)
if __name__ == '__main__':
main()
<commit_msg>Add check the value is date or not.<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from google.appengine.ext import webapp
from google.appengine.ext.webapp.util import run_wsgi_app
from google.appengine.api import xmpp
#from google.appengine.api import urlfetch
import urllib2,md5,logging,csv,re,math
def ckinv(oo):
""" check the value is date or not """
pattern = re.compile(r"[0-9]{2}/[0-9]{2}/[0-9]{2}")
b = re.search(pattern, oo[0])
try:
b.group()
return True
except:
return False
def covstr(s):
""" convert string to int or float. """
try:
ret = int(s)
except ValueError:
ret = float(s)
return ret
############## webapp Models ###################
class MainPage(webapp.RequestHandler):
def get(self):
url = 'http://www.twse.com.tw/ch/trading/exchange/STOCK_DAY_AVG/STOCK_DAY_AVG2.php?STK_NO=2363&myear=2010&mmon=06&type=csv'
cc = urllib2.urlopen(url)
csv_read = csv.reader(cc)
self.response.out.write('Go Ri Stock')
#csv_read.next
getr = []
for i in csv_read:
print i
if ckinv(i):
getr.append(covstr(i[1]))
print getr
print math.fsum(getr)
############## main Models ###################
def main():
""" Start up. """
application = webapp.WSGIApplication(
[
('/', MainPage)
],debug=True)
run_wsgi_app(application)
if __name__ == '__main__':
main()
|
f85425a2c74cf15555bbed233287ddbd7ab8b24e | flexget/ui/plugins/log/log.py | flexget/ui/plugins/log/log.py | from __future__ import unicode_literals, division, absolute_import
from flexget.ui import register_plugin, Blueprint, register_menu
log = Blueprint('log', __name__)
register_plugin(log)
log.register_angular_route(
'',
url=log.url_prefix,
template_url='index.html',
controller='LogViewCtrl'
)
log.register_css('log', 'css/log.css', order=99)
log.register_js('log', 'js/log.js')
log.register_js('angular-oboe', 'js/libs/angular-oboe.js')
log.register_js('oboe-browser', 'js/libs/oboe-browser.js')
register_menu(log.url_prefix, 'Log', icon='fa fa-file-text-o')
| from __future__ import unicode_literals, division, absolute_import
from flexget.ui import register_plugin, Blueprint, register_menu
log = Blueprint('log', __name__)
register_plugin(log)
log.register_angular_route(
'',
url=log.url_prefix,
template_url='index.html',
controller='LogViewCtrl'
)
log.register_css('log', 'css/log.css', order=99)
log.register_js('log', 'js/log.js')
log.register_js('angular-oboe', 'libs/oboe/js/angular-oboe.js')
log.register_js('oboe-browser', 'libs/oboe/js/oboe-browser.js')
register_menu(log.url_prefix, 'Log', icon='fa fa-file-text-o')
| Rename libs to keep with standard | Rename libs to keep with standard
| Python | mit | LynxyssCZ/Flexget,qvazzler/Flexget,tobinjt/Flexget,malkavi/Flexget,ZefQ/Flexget,qk4l/Flexget,Flexget/Flexget,tsnoam/Flexget,ianstalk/Flexget,grrr2/Flexget,jacobmetrick/Flexget,qvazzler/Flexget,crawln45/Flexget,poulpito/Flexget,JorisDeRieck/Flexget,gazpachoking/Flexget,tobinjt/Flexget,oxc/Flexget,dsemi/Flexget,jawilson/Flexget,malkavi/Flexget,tsnoam/Flexget,jacobmetrick/Flexget,OmgOhnoes/Flexget,tarzasai/Flexget,jacobmetrick/Flexget,dsemi/Flexget,Danfocus/Flexget,antivirtel/Flexget,Danfocus/Flexget,drwyrm/Flexget,tobinjt/Flexget,JorisDeRieck/Flexget,drwyrm/Flexget,JorisDeRieck/Flexget,antivirtel/Flexget,lildadou/Flexget,tarzasai/Flexget,offbyone/Flexget,grrr2/Flexget,qvazzler/Flexget,Pretagonist/Flexget,crawln45/Flexget,Pretagonist/Flexget,grrr2/Flexget,tsnoam/Flexget,ZefQ/Flexget,dsemi/Flexget,poulpito/Flexget,crawln45/Flexget,gazpachoking/Flexget,sean797/Flexget,Pretagonist/Flexget,lildadou/Flexget,LynxyssCZ/Flexget,JorisDeRieck/Flexget,Flexget/Flexget,OmgOhnoes/Flexget,Flexget/Flexget,Flexget/Flexget,sean797/Flexget,LynxyssCZ/Flexget,jawilson/Flexget,tarzasai/Flexget,malkavi/Flexget,jawilson/Flexget,Danfocus/Flexget,antivirtel/Flexget,ianstalk/Flexget,OmgOhnoes/Flexget,jawilson/Flexget,oxc/Flexget,crawln45/Flexget,cvium/Flexget,qk4l/Flexget,sean797/Flexget,Danfocus/Flexget,lildadou/Flexget,LynxyssCZ/Flexget,ianstalk/Flexget,poulpito/Flexget,offbyone/Flexget,ZefQ/Flexget,tobinjt/Flexget,qk4l/Flexget,cvium/Flexget,cvium/Flexget,drwyrm/Flexget,offbyone/Flexget,malkavi/Flexget,oxc/Flexget | from __future__ import unicode_literals, division, absolute_import
from flexget.ui import register_plugin, Blueprint, register_menu
log = Blueprint('log', __name__)
register_plugin(log)
log.register_angular_route(
'',
url=log.url_prefix,
template_url='index.html',
controller='LogViewCtrl'
)
log.register_css('log', 'css/log.css', order=99)
log.register_js('log', 'js/log.js')
log.register_js('angular-oboe', 'js/libs/angular-oboe.js')
log.register_js('oboe-browser', 'js/libs/oboe-browser.js')
register_menu(log.url_prefix, 'Log', icon='fa fa-file-text-o')
Rename libs to keep with standard | from __future__ import unicode_literals, division, absolute_import
from flexget.ui import register_plugin, Blueprint, register_menu
log = Blueprint('log', __name__)
register_plugin(log)
log.register_angular_route(
'',
url=log.url_prefix,
template_url='index.html',
controller='LogViewCtrl'
)
log.register_css('log', 'css/log.css', order=99)
log.register_js('log', 'js/log.js')
log.register_js('angular-oboe', 'libs/oboe/js/angular-oboe.js')
log.register_js('oboe-browser', 'libs/oboe/js/oboe-browser.js')
register_menu(log.url_prefix, 'Log', icon='fa fa-file-text-o')
| <commit_before>from __future__ import unicode_literals, division, absolute_import
from flexget.ui import register_plugin, Blueprint, register_menu
log = Blueprint('log', __name__)
register_plugin(log)
log.register_angular_route(
'',
url=log.url_prefix,
template_url='index.html',
controller='LogViewCtrl'
)
log.register_css('log', 'css/log.css', order=99)
log.register_js('log', 'js/log.js')
log.register_js('angular-oboe', 'js/libs/angular-oboe.js')
log.register_js('oboe-browser', 'js/libs/oboe-browser.js')
register_menu(log.url_prefix, 'Log', icon='fa fa-file-text-o')
<commit_msg>Rename libs to keep with standard<commit_after> | from __future__ import unicode_literals, division, absolute_import
from flexget.ui import register_plugin, Blueprint, register_menu
log = Blueprint('log', __name__)
register_plugin(log)
log.register_angular_route(
'',
url=log.url_prefix,
template_url='index.html',
controller='LogViewCtrl'
)
log.register_css('log', 'css/log.css', order=99)
log.register_js('log', 'js/log.js')
log.register_js('angular-oboe', 'libs/oboe/js/angular-oboe.js')
log.register_js('oboe-browser', 'libs/oboe/js/oboe-browser.js')
register_menu(log.url_prefix, 'Log', icon='fa fa-file-text-o')
| from __future__ import unicode_literals, division, absolute_import
from flexget.ui import register_plugin, Blueprint, register_menu
log = Blueprint('log', __name__)
register_plugin(log)
log.register_angular_route(
'',
url=log.url_prefix,
template_url='index.html',
controller='LogViewCtrl'
)
log.register_css('log', 'css/log.css', order=99)
log.register_js('log', 'js/log.js')
log.register_js('angular-oboe', 'js/libs/angular-oboe.js')
log.register_js('oboe-browser', 'js/libs/oboe-browser.js')
register_menu(log.url_prefix, 'Log', icon='fa fa-file-text-o')
Rename libs to keep with standardfrom __future__ import unicode_literals, division, absolute_import
from flexget.ui import register_plugin, Blueprint, register_menu
log = Blueprint('log', __name__)
register_plugin(log)
log.register_angular_route(
'',
url=log.url_prefix,
template_url='index.html',
controller='LogViewCtrl'
)
log.register_css('log', 'css/log.css', order=99)
log.register_js('log', 'js/log.js')
log.register_js('angular-oboe', 'libs/oboe/js/angular-oboe.js')
log.register_js('oboe-browser', 'libs/oboe/js/oboe-browser.js')
register_menu(log.url_prefix, 'Log', icon='fa fa-file-text-o')
| <commit_before>from __future__ import unicode_literals, division, absolute_import
from flexget.ui import register_plugin, Blueprint, register_menu
log = Blueprint('log', __name__)
register_plugin(log)
log.register_angular_route(
'',
url=log.url_prefix,
template_url='index.html',
controller='LogViewCtrl'
)
log.register_css('log', 'css/log.css', order=99)
log.register_js('log', 'js/log.js')
log.register_js('angular-oboe', 'js/libs/angular-oboe.js')
log.register_js('oboe-browser', 'js/libs/oboe-browser.js')
register_menu(log.url_prefix, 'Log', icon='fa fa-file-text-o')
<commit_msg>Rename libs to keep with standard<commit_after>from __future__ import unicode_literals, division, absolute_import
from flexget.ui import register_plugin, Blueprint, register_menu
log = Blueprint('log', __name__)
register_plugin(log)
log.register_angular_route(
'',
url=log.url_prefix,
template_url='index.html',
controller='LogViewCtrl'
)
log.register_css('log', 'css/log.css', order=99)
log.register_js('log', 'js/log.js')
log.register_js('angular-oboe', 'libs/oboe/js/angular-oboe.js')
log.register_js('oboe-browser', 'libs/oboe/js/oboe-browser.js')
register_menu(log.url_prefix, 'Log', icon='fa fa-file-text-o')
|
8eb47d151868c8e5906af054749993cd46a73b2d | capstone/player/kerasplayer.py | capstone/player/kerasplayer.py | from keras.models import load_model
from . import Player
from ..utils import normalize_board, utility
class KerasPlayer(Player):
'''
Takes moves based on a Keras neural network model.
'''
name = 'Keras'
def __init__(self, filepath):
self.model = load_model(filepath)
def __str__(self):
return self.name
def __repr__(self):
return self.name
##########
# Player #
##########
def choose_move(self, state):
assert state.cur_player() == 0
best_action = None
best_value = -1000000
for action in state.legal_moves():
s = state.copy()
s = s.make_move(action)
value = self.model.predict(normalize_board(s.board), batch_size=1)
assert value >= -1.0 and value <= 1.0
if value > best_value:
best_action = action
best_value = value
return best_action
| from keras.models import load_model
from . import Player
from ..utils import normalize_board, utility
class KerasPlayer(Player):
'''
Takes moves based on a Keras neural network model.
'''
name = 'Keras'
def __init__(self, filepath):
self.model = load_model(filepath)
def __str__(self):
return self.name
def __repr__(self):
return self.name
##########
# Player #
##########
def choose_move(self, game):
assert game.cur_player() == 0
best_move = None
best_value = -1000000
for move in game.legal_moves():
next_game = game.copy().make_move(move)
value = self.model.predict(normalize_board(next_game.board), batch_size=1)
assert value >= -1.0 and value <= 1.0
if value > best_value:
best_move = move
best_value = value
return best_move
| Rename state to game in KerasPlayer | Rename state to game in KerasPlayer
| Python | mit | davidrobles/mlnd-capstone-code | from keras.models import load_model
from . import Player
from ..utils import normalize_board, utility
class KerasPlayer(Player):
'''
Takes moves based on a Keras neural network model.
'''
name = 'Keras'
def __init__(self, filepath):
self.model = load_model(filepath)
def __str__(self):
return self.name
def __repr__(self):
return self.name
##########
# Player #
##########
def choose_move(self, state):
assert state.cur_player() == 0
best_action = None
best_value = -1000000
for action in state.legal_moves():
s = state.copy()
s = s.make_move(action)
value = self.model.predict(normalize_board(s.board), batch_size=1)
assert value >= -1.0 and value <= 1.0
if value > best_value:
best_action = action
best_value = value
return best_action
Rename state to game in KerasPlayer | from keras.models import load_model
from . import Player
from ..utils import normalize_board, utility
class KerasPlayer(Player):
'''
Takes moves based on a Keras neural network model.
'''
name = 'Keras'
def __init__(self, filepath):
self.model = load_model(filepath)
def __str__(self):
return self.name
def __repr__(self):
return self.name
##########
# Player #
##########
def choose_move(self, game):
assert game.cur_player() == 0
best_move = None
best_value = -1000000
for move in game.legal_moves():
next_game = game.copy().make_move(move)
value = self.model.predict(normalize_board(next_game.board), batch_size=1)
assert value >= -1.0 and value <= 1.0
if value > best_value:
best_move = move
best_value = value
return best_move
| <commit_before>from keras.models import load_model
from . import Player
from ..utils import normalize_board, utility
class KerasPlayer(Player):
'''
Takes moves based on a Keras neural network model.
'''
name = 'Keras'
def __init__(self, filepath):
self.model = load_model(filepath)
def __str__(self):
return self.name
def __repr__(self):
return self.name
##########
# Player #
##########
def choose_move(self, state):
assert state.cur_player() == 0
best_action = None
best_value = -1000000
for action in state.legal_moves():
s = state.copy()
s = s.make_move(action)
value = self.model.predict(normalize_board(s.board), batch_size=1)
assert value >= -1.0 and value <= 1.0
if value > best_value:
best_action = action
best_value = value
return best_action
<commit_msg>Rename state to game in KerasPlayer<commit_after> | from keras.models import load_model
from . import Player
from ..utils import normalize_board, utility
class KerasPlayer(Player):
'''
Takes moves based on a Keras neural network model.
'''
name = 'Keras'
def __init__(self, filepath):
self.model = load_model(filepath)
def __str__(self):
return self.name
def __repr__(self):
return self.name
##########
# Player #
##########
def choose_move(self, game):
assert game.cur_player() == 0
best_move = None
best_value = -1000000
for move in game.legal_moves():
next_game = game.copy().make_move(move)
value = self.model.predict(normalize_board(next_game.board), batch_size=1)
assert value >= -1.0 and value <= 1.0
if value > best_value:
best_move = move
best_value = value
return best_move
| from keras.models import load_model
from . import Player
from ..utils import normalize_board, utility
class KerasPlayer(Player):
'''
Takes moves based on a Keras neural network model.
'''
name = 'Keras'
def __init__(self, filepath):
self.model = load_model(filepath)
def __str__(self):
return self.name
def __repr__(self):
return self.name
##########
# Player #
##########
def choose_move(self, state):
assert state.cur_player() == 0
best_action = None
best_value = -1000000
for action in state.legal_moves():
s = state.copy()
s = s.make_move(action)
value = self.model.predict(normalize_board(s.board), batch_size=1)
assert value >= -1.0 and value <= 1.0
if value > best_value:
best_action = action
best_value = value
return best_action
Rename state to game in KerasPlayerfrom keras.models import load_model
from . import Player
from ..utils import normalize_board, utility
class KerasPlayer(Player):
'''
Takes moves based on a Keras neural network model.
'''
name = 'Keras'
def __init__(self, filepath):
self.model = load_model(filepath)
def __str__(self):
return self.name
def __repr__(self):
return self.name
##########
# Player #
##########
def choose_move(self, game):
assert game.cur_player() == 0
best_move = None
best_value = -1000000
for move in game.legal_moves():
next_game = game.copy().make_move(move)
value = self.model.predict(normalize_board(next_game.board), batch_size=1)
assert value >= -1.0 and value <= 1.0
if value > best_value:
best_move = move
best_value = value
return best_move
| <commit_before>from keras.models import load_model
from . import Player
from ..utils import normalize_board, utility
class KerasPlayer(Player):
'''
Takes moves based on a Keras neural network model.
'''
name = 'Keras'
def __init__(self, filepath):
self.model = load_model(filepath)
def __str__(self):
return self.name
def __repr__(self):
return self.name
##########
# Player #
##########
def choose_move(self, state):
assert state.cur_player() == 0
best_action = None
best_value = -1000000
for action in state.legal_moves():
s = state.copy()
s = s.make_move(action)
value = self.model.predict(normalize_board(s.board), batch_size=1)
assert value >= -1.0 and value <= 1.0
if value > best_value:
best_action = action
best_value = value
return best_action
<commit_msg>Rename state to game in KerasPlayer<commit_after>from keras.models import load_model
from . import Player
from ..utils import normalize_board, utility
class KerasPlayer(Player):
'''
Takes moves based on a Keras neural network model.
'''
name = 'Keras'
def __init__(self, filepath):
self.model = load_model(filepath)
def __str__(self):
return self.name
def __repr__(self):
return self.name
##########
# Player #
##########
def choose_move(self, game):
assert game.cur_player() == 0
best_move = None
best_value = -1000000
for move in game.legal_moves():
next_game = game.copy().make_move(move)
value = self.model.predict(normalize_board(next_game.board), batch_size=1)
assert value >= -1.0 and value <= 1.0
if value > best_value:
best_move = move
best_value = value
return best_move
|
9b032e06156aa011e5d78d0d9ea297420cb33c2e | form_designer/contrib/cms_plugins/form_designer_form/migrations/0001_initial.py | form_designer/contrib/cms_plugins/form_designer_form/migrations/0001_initial.py | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
import cms
from django.db import migrations, models
from pkg_resources import parse_version as V
# Django CMS 3.3.1 is oldest release where the change affects.
# Refs https://github.com/divio/django-cms/commit/871a164
if V(cms.__version__) >= V('3.3.1'):
field_kwargs = {'related_name': 'form_designer_form_cmsformdefinition'}
else:
field_kwargs = {}
class Migration(migrations.Migration):
dependencies = [
('cms', '0001_initial'),
('form_designer', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CMSFormDefinition',
fields=[
('cmsplugin_ptr',
models.OneToOneField(
serialize=False,
auto_created=True,
primary_key=True,
to='cms.CMSPlugin',
parent_link=True,
**field_kwargs)),
('form_definition',
models.ForeignKey(
verbose_name='form',
to='form_designer.FormDefinition')),
],
options={
'abstract': False,
},
bases=(
'cms.cmsplugin',
),
),
]
| # -*- coding: utf-8 -*-
from __future__ import unicode_literals
import cms
from django.db import migrations, models
from pkg_resources import parse_version as V
# Django CMS 3.3.1 is oldest release where the change affects.
# Refs https://github.com/divio/django-cms/commit/871a164
if V(cms.__version__) >= V('3.3.1'):
field_kwargs = {'related_name': 'form_designer_form_cmsformdefinition'}
else:
field_kwargs = {}
class Migration(migrations.Migration):
dependencies = [
('cms', '0001_initial'),
('form_designer', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CMSFormDefinition',
fields=[
('cmsplugin_ptr',
models.OneToOneField(
serialize=False,
auto_created=True,
primary_key=True,
to='cms.CMSPlugin',
parent_link=True,
on_delete=models.CASCADE,
**field_kwargs)),
('form_definition',
models.ForeignKey(
verbose_name='form',
to='form_designer.FormDefinition',
on_delete=models.CASCADE)),
],
options={
'abstract': False,
},
bases=(
'cms.cmsplugin',
),
),
]
| Add on_delete args to CMS plugin migration for Django 2 support | Add on_delete args to CMS plugin migration for Django 2 support
| Python | bsd-3-clause | kcsry/django-form-designer,kcsry/django-form-designer,andersinno/django-form-designer,andersinno/django-form-designer-ai,andersinno/django-form-designer-ai,andersinno/django-form-designer | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
import cms
from django.db import migrations, models
from pkg_resources import parse_version as V
# Django CMS 3.3.1 is oldest release where the change affects.
# Refs https://github.com/divio/django-cms/commit/871a164
if V(cms.__version__) >= V('3.3.1'):
field_kwargs = {'related_name': 'form_designer_form_cmsformdefinition'}
else:
field_kwargs = {}
class Migration(migrations.Migration):
dependencies = [
('cms', '0001_initial'),
('form_designer', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CMSFormDefinition',
fields=[
('cmsplugin_ptr',
models.OneToOneField(
serialize=False,
auto_created=True,
primary_key=True,
to='cms.CMSPlugin',
parent_link=True,
**field_kwargs)),
('form_definition',
models.ForeignKey(
verbose_name='form',
to='form_designer.FormDefinition')),
],
options={
'abstract': False,
},
bases=(
'cms.cmsplugin',
),
),
]
Add on_delete args to CMS plugin migration for Django 2 support | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
import cms
from django.db import migrations, models
from pkg_resources import parse_version as V
# Django CMS 3.3.1 is oldest release where the change affects.
# Refs https://github.com/divio/django-cms/commit/871a164
if V(cms.__version__) >= V('3.3.1'):
field_kwargs = {'related_name': 'form_designer_form_cmsformdefinition'}
else:
field_kwargs = {}
class Migration(migrations.Migration):
dependencies = [
('cms', '0001_initial'),
('form_designer', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CMSFormDefinition',
fields=[
('cmsplugin_ptr',
models.OneToOneField(
serialize=False,
auto_created=True,
primary_key=True,
to='cms.CMSPlugin',
parent_link=True,
on_delete=models.CASCADE,
**field_kwargs)),
('form_definition',
models.ForeignKey(
verbose_name='form',
to='form_designer.FormDefinition',
on_delete=models.CASCADE)),
],
options={
'abstract': False,
},
bases=(
'cms.cmsplugin',
),
),
]
| <commit_before># -*- coding: utf-8 -*-
from __future__ import unicode_literals
import cms
from django.db import migrations, models
from pkg_resources import parse_version as V
# Django CMS 3.3.1 is oldest release where the change affects.
# Refs https://github.com/divio/django-cms/commit/871a164
if V(cms.__version__) >= V('3.3.1'):
field_kwargs = {'related_name': 'form_designer_form_cmsformdefinition'}
else:
field_kwargs = {}
class Migration(migrations.Migration):
dependencies = [
('cms', '0001_initial'),
('form_designer', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CMSFormDefinition',
fields=[
('cmsplugin_ptr',
models.OneToOneField(
serialize=False,
auto_created=True,
primary_key=True,
to='cms.CMSPlugin',
parent_link=True,
**field_kwargs)),
('form_definition',
models.ForeignKey(
verbose_name='form',
to='form_designer.FormDefinition')),
],
options={
'abstract': False,
},
bases=(
'cms.cmsplugin',
),
),
]
<commit_msg>Add on_delete args to CMS plugin migration for Django 2 support<commit_after> | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
import cms
from django.db import migrations, models
from pkg_resources import parse_version as V
# Django CMS 3.3.1 is oldest release where the change affects.
# Refs https://github.com/divio/django-cms/commit/871a164
if V(cms.__version__) >= V('3.3.1'):
field_kwargs = {'related_name': 'form_designer_form_cmsformdefinition'}
else:
field_kwargs = {}
class Migration(migrations.Migration):
dependencies = [
('cms', '0001_initial'),
('form_designer', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CMSFormDefinition',
fields=[
('cmsplugin_ptr',
models.OneToOneField(
serialize=False,
auto_created=True,
primary_key=True,
to='cms.CMSPlugin',
parent_link=True,
on_delete=models.CASCADE,
**field_kwargs)),
('form_definition',
models.ForeignKey(
verbose_name='form',
to='form_designer.FormDefinition',
on_delete=models.CASCADE)),
],
options={
'abstract': False,
},
bases=(
'cms.cmsplugin',
),
),
]
| # -*- coding: utf-8 -*-
from __future__ import unicode_literals
import cms
from django.db import migrations, models
from pkg_resources import parse_version as V
# Django CMS 3.3.1 is oldest release where the change affects.
# Refs https://github.com/divio/django-cms/commit/871a164
if V(cms.__version__) >= V('3.3.1'):
field_kwargs = {'related_name': 'form_designer_form_cmsformdefinition'}
else:
field_kwargs = {}
class Migration(migrations.Migration):
dependencies = [
('cms', '0001_initial'),
('form_designer', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CMSFormDefinition',
fields=[
('cmsplugin_ptr',
models.OneToOneField(
serialize=False,
auto_created=True,
primary_key=True,
to='cms.CMSPlugin',
parent_link=True,
**field_kwargs)),
('form_definition',
models.ForeignKey(
verbose_name='form',
to='form_designer.FormDefinition')),
],
options={
'abstract': False,
},
bases=(
'cms.cmsplugin',
),
),
]
Add on_delete args to CMS plugin migration for Django 2 support# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import cms
from django.db import migrations, models
from pkg_resources import parse_version as V
# Django CMS 3.3.1 is oldest release where the change affects.
# Refs https://github.com/divio/django-cms/commit/871a164
if V(cms.__version__) >= V('3.3.1'):
field_kwargs = {'related_name': 'form_designer_form_cmsformdefinition'}
else:
field_kwargs = {}
class Migration(migrations.Migration):
dependencies = [
('cms', '0001_initial'),
('form_designer', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CMSFormDefinition',
fields=[
('cmsplugin_ptr',
models.OneToOneField(
serialize=False,
auto_created=True,
primary_key=True,
to='cms.CMSPlugin',
parent_link=True,
on_delete=models.CASCADE,
**field_kwargs)),
('form_definition',
models.ForeignKey(
verbose_name='form',
to='form_designer.FormDefinition',
on_delete=models.CASCADE)),
],
options={
'abstract': False,
},
bases=(
'cms.cmsplugin',
),
),
]
| <commit_before># -*- coding: utf-8 -*-
from __future__ import unicode_literals
import cms
from django.db import migrations, models
from pkg_resources import parse_version as V
# Django CMS 3.3.1 is oldest release where the change affects.
# Refs https://github.com/divio/django-cms/commit/871a164
if V(cms.__version__) >= V('3.3.1'):
field_kwargs = {'related_name': 'form_designer_form_cmsformdefinition'}
else:
field_kwargs = {}
class Migration(migrations.Migration):
dependencies = [
('cms', '0001_initial'),
('form_designer', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CMSFormDefinition',
fields=[
('cmsplugin_ptr',
models.OneToOneField(
serialize=False,
auto_created=True,
primary_key=True,
to='cms.CMSPlugin',
parent_link=True,
**field_kwargs)),
('form_definition',
models.ForeignKey(
verbose_name='form',
to='form_designer.FormDefinition')),
],
options={
'abstract': False,
},
bases=(
'cms.cmsplugin',
),
),
]
<commit_msg>Add on_delete args to CMS plugin migration for Django 2 support<commit_after># -*- coding: utf-8 -*-
from __future__ import unicode_literals
import cms
from django.db import migrations, models
from pkg_resources import parse_version as V
# Django CMS 3.3.1 is oldest release where the change affects.
# Refs https://github.com/divio/django-cms/commit/871a164
if V(cms.__version__) >= V('3.3.1'):
field_kwargs = {'related_name': 'form_designer_form_cmsformdefinition'}
else:
field_kwargs = {}
class Migration(migrations.Migration):
dependencies = [
('cms', '0001_initial'),
('form_designer', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='CMSFormDefinition',
fields=[
('cmsplugin_ptr',
models.OneToOneField(
serialize=False,
auto_created=True,
primary_key=True,
to='cms.CMSPlugin',
parent_link=True,
on_delete=models.CASCADE,
**field_kwargs)),
('form_definition',
models.ForeignKey(
verbose_name='form',
to='form_designer.FormDefinition',
on_delete=models.CASCADE)),
],
options={
'abstract': False,
},
bases=(
'cms.cmsplugin',
),
),
]
|
d6787523cb8b58c51fe32d4524389a500e3b7b21 | foliant/cli.py | foliant/cli.py | """Foliant: Markdown to PDF, Docx, and LaTeX generator powered by Pandoc.
Usage:
foliant (build | make) <target> [--path=<project-path>]
foliant (upload | up) <document> [--secret=<client_secret*.json>]
foliant (-h | --help)
foliant --version
Options:
-h --help Show this screen.
-v --version Show version.
-p --path=<project-path> Path to your project [default: .].
-s --secret=<client_secret*.json> Path to Google app's client secret file.
"""
from docopt import docopt
import colorama
from colorama import Fore
from foliant import builder, uploader
from foliant import __version__ as foliant_version
def main():
"""Handles command-line params and runs the respective core function."""
colorama.init(autoreset=True)
args = docopt(__doc__, version="Foliant %s (Python)" % foliant_version)
if args["build"] or args["make"]:
result = builder.build(args["<target>"], args["--path"])
elif args["upload"] or args["up"]:
result = uploader.upload(args["<document>"])
print("---")
print(Fore.GREEN + "Result: %s" % result)
colorama.deinit()
| """Foliant: Markdown to PDF, Docx, and LaTeX generator powered by Pandoc.
Usage:
foliant (build | make) <target> [--path=<project-path>]
foliant (upload | up) <document> [--secret=<client_secret*.json>]
foliant (-h | --help)
foliant --version
Options:
-h --help Show this screen.
-v --version Show version.
-p --path=<project-path> Path to your project [default: .].
-s --secret=<client_secret*.json> Path to Google app's client secret file.
"""
from docopt import docopt
import colorama
from colorama import Fore
from . import builder, uploader
from . import __version__ as foliant_version
def main():
"""Handles command-line params and runs the respective core function."""
colorama.init(autoreset=True)
args = docopt(__doc__, version="Foliant %s (Python)" % foliant_version)
if args["build"] or args["make"]:
result = builder.build(args["<target>"], args["--path"])
elif args["upload"] or args["up"]:
result = uploader.upload(args["<document>"])
print("---")
print(Fore.GREEN + "Result: %s" % result)
colorama.deinit()
| Replace hardcoded package name with ".". | CLI: Replace hardcoded package name with ".".
| Python | mit | foliant-docs/foliant | """Foliant: Markdown to PDF, Docx, and LaTeX generator powered by Pandoc.
Usage:
foliant (build | make) <target> [--path=<project-path>]
foliant (upload | up) <document> [--secret=<client_secret*.json>]
foliant (-h | --help)
foliant --version
Options:
-h --help Show this screen.
-v --version Show version.
-p --path=<project-path> Path to your project [default: .].
-s --secret=<client_secret*.json> Path to Google app's client secret file.
"""
from docopt import docopt
import colorama
from colorama import Fore
from foliant import builder, uploader
from foliant import __version__ as foliant_version
def main():
"""Handles command-line params and runs the respective core function."""
colorama.init(autoreset=True)
args = docopt(__doc__, version="Foliant %s (Python)" % foliant_version)
if args["build"] or args["make"]:
result = builder.build(args["<target>"], args["--path"])
elif args["upload"] or args["up"]:
result = uploader.upload(args["<document>"])
print("---")
print(Fore.GREEN + "Result: %s" % result)
colorama.deinit()
CLI: Replace hardcoded package name with ".". | """Foliant: Markdown to PDF, Docx, and LaTeX generator powered by Pandoc.
Usage:
foliant (build | make) <target> [--path=<project-path>]
foliant (upload | up) <document> [--secret=<client_secret*.json>]
foliant (-h | --help)
foliant --version
Options:
-h --help Show this screen.
-v --version Show version.
-p --path=<project-path> Path to your project [default: .].
-s --secret=<client_secret*.json> Path to Google app's client secret file.
"""
from docopt import docopt
import colorama
from colorama import Fore
from . import builder, uploader
from . import __version__ as foliant_version
def main():
"""Handles command-line params and runs the respective core function."""
colorama.init(autoreset=True)
args = docopt(__doc__, version="Foliant %s (Python)" % foliant_version)
if args["build"] or args["make"]:
result = builder.build(args["<target>"], args["--path"])
elif args["upload"] or args["up"]:
result = uploader.upload(args["<document>"])
print("---")
print(Fore.GREEN + "Result: %s" % result)
colorama.deinit()
| <commit_before>"""Foliant: Markdown to PDF, Docx, and LaTeX generator powered by Pandoc.
Usage:
foliant (build | make) <target> [--path=<project-path>]
foliant (upload | up) <document> [--secret=<client_secret*.json>]
foliant (-h | --help)
foliant --version
Options:
-h --help Show this screen.
-v --version Show version.
-p --path=<project-path> Path to your project [default: .].
-s --secret=<client_secret*.json> Path to Google app's client secret file.
"""
from docopt import docopt
import colorama
from colorama import Fore
from foliant import builder, uploader
from foliant import __version__ as foliant_version
def main():
"""Handles command-line params and runs the respective core function."""
colorama.init(autoreset=True)
args = docopt(__doc__, version="Foliant %s (Python)" % foliant_version)
if args["build"] or args["make"]:
result = builder.build(args["<target>"], args["--path"])
elif args["upload"] or args["up"]:
result = uploader.upload(args["<document>"])
print("---")
print(Fore.GREEN + "Result: %s" % result)
colorama.deinit()
<commit_msg>CLI: Replace hardcoded package name with ".".<commit_after> | """Foliant: Markdown to PDF, Docx, and LaTeX generator powered by Pandoc.
Usage:
foliant (build | make) <target> [--path=<project-path>]
foliant (upload | up) <document> [--secret=<client_secret*.json>]
foliant (-h | --help)
foliant --version
Options:
-h --help Show this screen.
-v --version Show version.
-p --path=<project-path> Path to your project [default: .].
-s --secret=<client_secret*.json> Path to Google app's client secret file.
"""
from docopt import docopt
import colorama
from colorama import Fore
from . import builder, uploader
from . import __version__ as foliant_version
def main():
"""Handles command-line params and runs the respective core function."""
colorama.init(autoreset=True)
args = docopt(__doc__, version="Foliant %s (Python)" % foliant_version)
if args["build"] or args["make"]:
result = builder.build(args["<target>"], args["--path"])
elif args["upload"] or args["up"]:
result = uploader.upload(args["<document>"])
print("---")
print(Fore.GREEN + "Result: %s" % result)
colorama.deinit()
| """Foliant: Markdown to PDF, Docx, and LaTeX generator powered by Pandoc.
Usage:
foliant (build | make) <target> [--path=<project-path>]
foliant (upload | up) <document> [--secret=<client_secret*.json>]
foliant (-h | --help)
foliant --version
Options:
-h --help Show this screen.
-v --version Show version.
-p --path=<project-path> Path to your project [default: .].
-s --secret=<client_secret*.json> Path to Google app's client secret file.
"""
from docopt import docopt
import colorama
from colorama import Fore
from foliant import builder, uploader
from foliant import __version__ as foliant_version
def main():
"""Handles command-line params and runs the respective core function."""
colorama.init(autoreset=True)
args = docopt(__doc__, version="Foliant %s (Python)" % foliant_version)
if args["build"] or args["make"]:
result = builder.build(args["<target>"], args["--path"])
elif args["upload"] or args["up"]:
result = uploader.upload(args["<document>"])
print("---")
print(Fore.GREEN + "Result: %s" % result)
colorama.deinit()
CLI: Replace hardcoded package name with "."."""Foliant: Markdown to PDF, Docx, and LaTeX generator powered by Pandoc.
Usage:
foliant (build | make) <target> [--path=<project-path>]
foliant (upload | up) <document> [--secret=<client_secret*.json>]
foliant (-h | --help)
foliant --version
Options:
-h --help Show this screen.
-v --version Show version.
-p --path=<project-path> Path to your project [default: .].
-s --secret=<client_secret*.json> Path to Google app's client secret file.
"""
from docopt import docopt
import colorama
from colorama import Fore
from . import builder, uploader
from . import __version__ as foliant_version
def main():
"""Handles command-line params and runs the respective core function."""
colorama.init(autoreset=True)
args = docopt(__doc__, version="Foliant %s (Python)" % foliant_version)
if args["build"] or args["make"]:
result = builder.build(args["<target>"], args["--path"])
elif args["upload"] or args["up"]:
result = uploader.upload(args["<document>"])
print("---")
print(Fore.GREEN + "Result: %s" % result)
colorama.deinit()
| <commit_before>"""Foliant: Markdown to PDF, Docx, and LaTeX generator powered by Pandoc.
Usage:
foliant (build | make) <target> [--path=<project-path>]
foliant (upload | up) <document> [--secret=<client_secret*.json>]
foliant (-h | --help)
foliant --version
Options:
-h --help Show this screen.
-v --version Show version.
-p --path=<project-path> Path to your project [default: .].
-s --secret=<client_secret*.json> Path to Google app's client secret file.
"""
from docopt import docopt
import colorama
from colorama import Fore
from foliant import builder, uploader
from foliant import __version__ as foliant_version
def main():
"""Handles command-line params and runs the respective core function."""
colorama.init(autoreset=True)
args = docopt(__doc__, version="Foliant %s (Python)" % foliant_version)
if args["build"] or args["make"]:
result = builder.build(args["<target>"], args["--path"])
elif args["upload"] or args["up"]:
result = uploader.upload(args["<document>"])
print("---")
print(Fore.GREEN + "Result: %s" % result)
colorama.deinit()
<commit_msg>CLI: Replace hardcoded package name with ".".<commit_after>"""Foliant: Markdown to PDF, Docx, and LaTeX generator powered by Pandoc.
Usage:
foliant (build | make) <target> [--path=<project-path>]
foliant (upload | up) <document> [--secret=<client_secret*.json>]
foliant (-h | --help)
foliant --version
Options:
-h --help Show this screen.
-v --version Show version.
-p --path=<project-path> Path to your project [default: .].
-s --secret=<client_secret*.json> Path to Google app's client secret file.
"""
from docopt import docopt
import colorama
from colorama import Fore
from . import builder, uploader
from . import __version__ as foliant_version
def main():
"""Handles command-line params and runs the respective core function."""
colorama.init(autoreset=True)
args = docopt(__doc__, version="Foliant %s (Python)" % foliant_version)
if args["build"] or args["make"]:
result = builder.build(args["<target>"], args["--path"])
elif args["upload"] or args["up"]:
result = uploader.upload(args["<document>"])
print("---")
print(Fore.GREEN + "Result: %s" % result)
colorama.deinit()
|
dd02861cd9fb5b06d42f7a6413b371c52c167ba8 | gcmconsumer.py | gcmconsumer.py | import sys
import fedmsg.consumers
import yaml
class GCMConsumer(fedmsg.consumers.FedmsgConsumer):
topic = 'org.fedoraproject.prod.*'
config_key = 'gcmconsumer'
def __init__(self, *args, **kw):
super(GCMConsumer, self).__init__(*args, **kw)
def get_registration_ids_for_topic(self, topic):
'''Get the Android/GCM registration IDs for all users who subscribe to a
particular topic. We load the config each time so that we can change
it on the fly to add users without having to miss any messages.'''
config = yaml.load(file('/home/ricky/devel/fedora/fedmsg-gcm-demo/config.yaml', 'r'))
return filter(None, [user['registration_id'] if topic in user['topics'] else None for user in config['users']])
def consume(self, msg):
users = self.get_registration_ids_for_topic(msg['topic'])
if users:
print "* SENDING MESSAGE TO USER(S)"
print msg['topic']
| import fedmsg.consumers
import json
import requests
import sys
import yaml
class GCMConsumer(fedmsg.consumers.FedmsgConsumer):
topic = 'org.fedoraproject.prod.*'
config_key = 'gcmconsumer'
def __init__(self, *args, **kw):
self.config_file = '/home/ricky/devel/fedora/fedmsg-gcm-demo/config.yaml'
self.config = yaml.load(file(self.config_file, 'r'))
super(GCMConsumer, self).__init__(*args, **kw)
def _get_registration_ids_for_topic(self, topic):
'''Get the Android/GCM registration IDs for all users who subscribe to a
particular topic. We load the config each time so that we can change
it on the fly to add users without having to miss any messages.'''
self.config = yaml.load(file(self.config_file, 'r'))
return filter(None, [user['registration_id'] if topic in user['topics'] else None for user in self.config['users']])
def _send_gcm(self, data, ids):
'''Send a message to GCM for specific registration IDs.'''
headers = {
'Authorization': 'key=%s' % self.config['api_key'],
'content-type': 'application/json'
}
body = {
'registration_ids': ids,
'data': data
}
request = requests.post(
self.config['post_url'],
data=json.dumps(body),
headers=headers)
return request
def consume(self, msg):
users = self._get_registration_ids_for_topic(msg['topic'])
for user in users:
print "* SENDING MESSAGE TO USER: %s" % user
response = self._send_gcm(
{
'topic': msg['topic']
},
[user])
print response.text
print response.status_code
print response
print msg['topic']
| Handle actually sending out notifications | Handle actually sending out notifications
| Python | apache-2.0 | fedora-infra/fedmsg-gcm-demo | import sys
import fedmsg.consumers
import yaml
class GCMConsumer(fedmsg.consumers.FedmsgConsumer):
topic = 'org.fedoraproject.prod.*'
config_key = 'gcmconsumer'
def __init__(self, *args, **kw):
super(GCMConsumer, self).__init__(*args, **kw)
def get_registration_ids_for_topic(self, topic):
'''Get the Android/GCM registration IDs for all users who subscribe to a
particular topic. We load the config each time so that we can change
it on the fly to add users without having to miss any messages.'''
config = yaml.load(file('/home/ricky/devel/fedora/fedmsg-gcm-demo/config.yaml', 'r'))
return filter(None, [user['registration_id'] if topic in user['topics'] else None for user in config['users']])
def consume(self, msg):
users = self.get_registration_ids_for_topic(msg['topic'])
if users:
print "* SENDING MESSAGE TO USER(S)"
print msg['topic']
Handle actually sending out notifications | import fedmsg.consumers
import json
import requests
import sys
import yaml
class GCMConsumer(fedmsg.consumers.FedmsgConsumer):
topic = 'org.fedoraproject.prod.*'
config_key = 'gcmconsumer'
def __init__(self, *args, **kw):
self.config_file = '/home/ricky/devel/fedora/fedmsg-gcm-demo/config.yaml'
self.config = yaml.load(file(self.config_file, 'r'))
super(GCMConsumer, self).__init__(*args, **kw)
def _get_registration_ids_for_topic(self, topic):
'''Get the Android/GCM registration IDs for all users who subscribe to a
particular topic. We load the config each time so that we can change
it on the fly to add users without having to miss any messages.'''
self.config = yaml.load(file(self.config_file, 'r'))
return filter(None, [user['registration_id'] if topic in user['topics'] else None for user in self.config['users']])
def _send_gcm(self, data, ids):
'''Send a message to GCM for specific registration IDs.'''
headers = {
'Authorization': 'key=%s' % self.config['api_key'],
'content-type': 'application/json'
}
body = {
'registration_ids': ids,
'data': data
}
request = requests.post(
self.config['post_url'],
data=json.dumps(body),
headers=headers)
return request
def consume(self, msg):
users = self._get_registration_ids_for_topic(msg['topic'])
for user in users:
print "* SENDING MESSAGE TO USER: %s" % user
response = self._send_gcm(
{
'topic': msg['topic']
},
[user])
print response.text
print response.status_code
print response
print msg['topic']
| <commit_before>import sys
import fedmsg.consumers
import yaml
class GCMConsumer(fedmsg.consumers.FedmsgConsumer):
topic = 'org.fedoraproject.prod.*'
config_key = 'gcmconsumer'
def __init__(self, *args, **kw):
super(GCMConsumer, self).__init__(*args, **kw)
def get_registration_ids_for_topic(self, topic):
'''Get the Android/GCM registration IDs for all users who subscribe to a
particular topic. We load the config each time so that we can change
it on the fly to add users without having to miss any messages.'''
config = yaml.load(file('/home/ricky/devel/fedora/fedmsg-gcm-demo/config.yaml', 'r'))
return filter(None, [user['registration_id'] if topic in user['topics'] else None for user in config['users']])
def consume(self, msg):
users = self.get_registration_ids_for_topic(msg['topic'])
if users:
print "* SENDING MESSAGE TO USER(S)"
print msg['topic']
<commit_msg>Handle actually sending out notifications<commit_after> | import fedmsg.consumers
import json
import requests
import sys
import yaml
class GCMConsumer(fedmsg.consumers.FedmsgConsumer):
topic = 'org.fedoraproject.prod.*'
config_key = 'gcmconsumer'
def __init__(self, *args, **kw):
self.config_file = '/home/ricky/devel/fedora/fedmsg-gcm-demo/config.yaml'
self.config = yaml.load(file(self.config_file, 'r'))
super(GCMConsumer, self).__init__(*args, **kw)
def _get_registration_ids_for_topic(self, topic):
'''Get the Android/GCM registration IDs for all users who subscribe to a
particular topic. We load the config each time so that we can change
it on the fly to add users without having to miss any messages.'''
self.config = yaml.load(file(self.config_file, 'r'))
return filter(None, [user['registration_id'] if topic in user['topics'] else None for user in self.config['users']])
def _send_gcm(self, data, ids):
'''Send a message to GCM for specific registration IDs.'''
headers = {
'Authorization': 'key=%s' % self.config['api_key'],
'content-type': 'application/json'
}
body = {
'registration_ids': ids,
'data': data
}
request = requests.post(
self.config['post_url'],
data=json.dumps(body),
headers=headers)
return request
def consume(self, msg):
users = self._get_registration_ids_for_topic(msg['topic'])
for user in users:
print "* SENDING MESSAGE TO USER: %s" % user
response = self._send_gcm(
{
'topic': msg['topic']
},
[user])
print response.text
print response.status_code
print response
print msg['topic']
| import sys
import fedmsg.consumers
import yaml
class GCMConsumer(fedmsg.consumers.FedmsgConsumer):
topic = 'org.fedoraproject.prod.*'
config_key = 'gcmconsumer'
def __init__(self, *args, **kw):
super(GCMConsumer, self).__init__(*args, **kw)
def get_registration_ids_for_topic(self, topic):
'''Get the Android/GCM registration IDs for all users who subscribe to a
particular topic. We load the config each time so that we can change
it on the fly to add users without having to miss any messages.'''
config = yaml.load(file('/home/ricky/devel/fedora/fedmsg-gcm-demo/config.yaml', 'r'))
return filter(None, [user['registration_id'] if topic in user['topics'] else None for user in config['users']])
def consume(self, msg):
users = self.get_registration_ids_for_topic(msg['topic'])
if users:
print "* SENDING MESSAGE TO USER(S)"
print msg['topic']
Handle actually sending out notificationsimport fedmsg.consumers
import json
import requests
import sys
import yaml
class GCMConsumer(fedmsg.consumers.FedmsgConsumer):
topic = 'org.fedoraproject.prod.*'
config_key = 'gcmconsumer'
def __init__(self, *args, **kw):
self.config_file = '/home/ricky/devel/fedora/fedmsg-gcm-demo/config.yaml'
self.config = yaml.load(file(self.config_file, 'r'))
super(GCMConsumer, self).__init__(*args, **kw)
def _get_registration_ids_for_topic(self, topic):
'''Get the Android/GCM registration IDs for all users who subscribe to a
particular topic. We load the config each time so that we can change
it on the fly to add users without having to miss any messages.'''
self.config = yaml.load(file(self.config_file, 'r'))
return filter(None, [user['registration_id'] if topic in user['topics'] else None for user in self.config['users']])
def _send_gcm(self, data, ids):
'''Send a message to GCM for specific registration IDs.'''
headers = {
'Authorization': 'key=%s' % self.config['api_key'],
'content-type': 'application/json'
}
body = {
'registration_ids': ids,
'data': data
}
request = requests.post(
self.config['post_url'],
data=json.dumps(body),
headers=headers)
return request
def consume(self, msg):
users = self._get_registration_ids_for_topic(msg['topic'])
for user in users:
print "* SENDING MESSAGE TO USER: %s" % user
response = self._send_gcm(
{
'topic': msg['topic']
},
[user])
print response.text
print response.status_code
print response
print msg['topic']
| <commit_before>import sys
import fedmsg.consumers
import yaml
class GCMConsumer(fedmsg.consumers.FedmsgConsumer):
topic = 'org.fedoraproject.prod.*'
config_key = 'gcmconsumer'
def __init__(self, *args, **kw):
super(GCMConsumer, self).__init__(*args, **kw)
def get_registration_ids_for_topic(self, topic):
'''Get the Android/GCM registration IDs for all users who subscribe to a
particular topic. We load the config each time so that we can change
it on the fly to add users without having to miss any messages.'''
config = yaml.load(file('/home/ricky/devel/fedora/fedmsg-gcm-demo/config.yaml', 'r'))
return filter(None, [user['registration_id'] if topic in user['topics'] else None for user in config['users']])
def consume(self, msg):
users = self.get_registration_ids_for_topic(msg['topic'])
if users:
print "* SENDING MESSAGE TO USER(S)"
print msg['topic']
<commit_msg>Handle actually sending out notifications<commit_after>import fedmsg.consumers
import json
import requests
import sys
import yaml
class GCMConsumer(fedmsg.consumers.FedmsgConsumer):
topic = 'org.fedoraproject.prod.*'
config_key = 'gcmconsumer'
def __init__(self, *args, **kw):
self.config_file = '/home/ricky/devel/fedora/fedmsg-gcm-demo/config.yaml'
self.config = yaml.load(file(self.config_file, 'r'))
super(GCMConsumer, self).__init__(*args, **kw)
def _get_registration_ids_for_topic(self, topic):
'''Get the Android/GCM registration IDs for all users who subscribe to a
particular topic. We load the config each time so that we can change
it on the fly to add users without having to miss any messages.'''
self.config = yaml.load(file(self.config_file, 'r'))
return filter(None, [user['registration_id'] if topic in user['topics'] else None for user in self.config['users']])
def _send_gcm(self, data, ids):
'''Send a message to GCM for specific registration IDs.'''
headers = {
'Authorization': 'key=%s' % self.config['api_key'],
'content-type': 'application/json'
}
body = {
'registration_ids': ids,
'data': data
}
request = requests.post(
self.config['post_url'],
data=json.dumps(body),
headers=headers)
return request
def consume(self, msg):
users = self._get_registration_ids_for_topic(msg['topic'])
for user in users:
print "* SENDING MESSAGE TO USER: %s" % user
response = self._send_gcm(
{
'topic': msg['topic']
},
[user])
print response.text
print response.status_code
print response
print msg['topic']
|
81e7d7d45e71f96d85468737708a31aef939091b | grip/default_config.py | grip/default_config.py | """\
Default Configuration
Do NOT change the values here for risk of accidentally committing them.
Override them using command-line arguments or with a local_config.py instead.
"""
HOST = 'localhost'
PORT = 5000
DEBUG = True
DEBUG_GRIP = False
STYLE_URLS = []
STYLE_URL_SOURCE = 'https://github.com/joeyespo/grip'
STYLE_URL_RE = '<link.+href=[\'"]?([^\'" >]+)[\'"]?.+media=[\'"]?screen[\'"]?.+rel=[\'"]?stylesheet[\'"]?.+/>'
| """\
Default Configuration
Do NOT change the values here for risk of accidentally committing them.
Override them using command-line arguments or with a local_config.py instead.
"""
HOST = 'localhost'
PORT = 5000
DEBUG = True
DEBUG_GRIP = False
STYLE_URLS = []
STYLE_URL_SOURCE = 'https://github.com/joeyespo/grip'
STYLE_URL_RE = '<link.+href=[\'"]?([^\'" >]+)[\'"]?.+media=[\'"]?(?:screen|all)[\'"]?.+rel=[\'"]?stylesheet[\'"]?.+/>'
| Update GitHub CSS file regex. | Update GitHub CSS file regex.
| Python | mit | mgoddard-pivotal/grip,mgoddard-pivotal/grip,ssundarraj/grip,joeyespo/grip,ssundarraj/grip,joeyespo/grip,jbarreras/grip,jbarreras/grip | """\
Default Configuration
Do NOT change the values here for risk of accidentally committing them.
Override them using command-line arguments or with a local_config.py instead.
"""
HOST = 'localhost'
PORT = 5000
DEBUG = True
DEBUG_GRIP = False
STYLE_URLS = []
STYLE_URL_SOURCE = 'https://github.com/joeyespo/grip'
STYLE_URL_RE = '<link.+href=[\'"]?([^\'" >]+)[\'"]?.+media=[\'"]?screen[\'"]?.+rel=[\'"]?stylesheet[\'"]?.+/>'
Update GitHub CSS file regex. | """\
Default Configuration
Do NOT change the values here for risk of accidentally committing them.
Override them using command-line arguments or with a local_config.py instead.
"""
HOST = 'localhost'
PORT = 5000
DEBUG = True
DEBUG_GRIP = False
STYLE_URLS = []
STYLE_URL_SOURCE = 'https://github.com/joeyespo/grip'
STYLE_URL_RE = '<link.+href=[\'"]?([^\'" >]+)[\'"]?.+media=[\'"]?(?:screen|all)[\'"]?.+rel=[\'"]?stylesheet[\'"]?.+/>'
| <commit_before>"""\
Default Configuration
Do NOT change the values here for risk of accidentally committing them.
Override them using command-line arguments or with a local_config.py instead.
"""
HOST = 'localhost'
PORT = 5000
DEBUG = True
DEBUG_GRIP = False
STYLE_URLS = []
STYLE_URL_SOURCE = 'https://github.com/joeyespo/grip'
STYLE_URL_RE = '<link.+href=[\'"]?([^\'" >]+)[\'"]?.+media=[\'"]?screen[\'"]?.+rel=[\'"]?stylesheet[\'"]?.+/>'
<commit_msg>Update GitHub CSS file regex.<commit_after> | """\
Default Configuration
Do NOT change the values here for risk of accidentally committing them.
Override them using command-line arguments or with a local_config.py instead.
"""
HOST = 'localhost'
PORT = 5000
DEBUG = True
DEBUG_GRIP = False
STYLE_URLS = []
STYLE_URL_SOURCE = 'https://github.com/joeyespo/grip'
STYLE_URL_RE = '<link.+href=[\'"]?([^\'" >]+)[\'"]?.+media=[\'"]?(?:screen|all)[\'"]?.+rel=[\'"]?stylesheet[\'"]?.+/>'
| """\
Default Configuration
Do NOT change the values here for risk of accidentally committing them.
Override them using command-line arguments or with a local_config.py instead.
"""
HOST = 'localhost'
PORT = 5000
DEBUG = True
DEBUG_GRIP = False
STYLE_URLS = []
STYLE_URL_SOURCE = 'https://github.com/joeyespo/grip'
STYLE_URL_RE = '<link.+href=[\'"]?([^\'" >]+)[\'"]?.+media=[\'"]?screen[\'"]?.+rel=[\'"]?stylesheet[\'"]?.+/>'
Update GitHub CSS file regex."""\
Default Configuration
Do NOT change the values here for risk of accidentally committing them.
Override them using command-line arguments or with a local_config.py instead.
"""
HOST = 'localhost'
PORT = 5000
DEBUG = True
DEBUG_GRIP = False
STYLE_URLS = []
STYLE_URL_SOURCE = 'https://github.com/joeyespo/grip'
STYLE_URL_RE = '<link.+href=[\'"]?([^\'" >]+)[\'"]?.+media=[\'"]?(?:screen|all)[\'"]?.+rel=[\'"]?stylesheet[\'"]?.+/>'
| <commit_before>"""\
Default Configuration
Do NOT change the values here for risk of accidentally committing them.
Override them using command-line arguments or with a local_config.py instead.
"""
HOST = 'localhost'
PORT = 5000
DEBUG = True
DEBUG_GRIP = False
STYLE_URLS = []
STYLE_URL_SOURCE = 'https://github.com/joeyespo/grip'
STYLE_URL_RE = '<link.+href=[\'"]?([^\'" >]+)[\'"]?.+media=[\'"]?screen[\'"]?.+rel=[\'"]?stylesheet[\'"]?.+/>'
<commit_msg>Update GitHub CSS file regex.<commit_after>"""\
Default Configuration
Do NOT change the values here for risk of accidentally committing them.
Override them using command-line arguments or with a local_config.py instead.
"""
HOST = 'localhost'
PORT = 5000
DEBUG = True
DEBUG_GRIP = False
STYLE_URLS = []
STYLE_URL_SOURCE = 'https://github.com/joeyespo/grip'
STYLE_URL_RE = '<link.+href=[\'"]?([^\'" >]+)[\'"]?.+media=[\'"]?(?:screen|all)[\'"]?.+rel=[\'"]?stylesheet[\'"]?.+/>'
|
019c91a8cd32fe1a4034837ed75dcc849d9033e5 | format_json.py | format_json.py | #! /usr/bin/env python3
import sys
import json
for filepath in sys.argv[1:]:
with open(filepath) as f:
try:
oyster = json.load(f)
except ValueError:
sys.stderr.write("In file: {}\n".format(filepath))
raise
with open(filepath, 'w') as f:
json.dump(oyster, f, ensure_ascii=False, indent=4, separators=(',', ': '), sort_keys=True)
f.write('\n') # add a trailing newline.
| #! /usr/bin/env python3
import sys
import json
import argparse
def format_json(fp):
try:
data = json.load(fp)
except ValueError:
sys.stderr.write("In file: {}\n".format(fp.name))
raise
# Jump back to the beginning of the file before overwriting it.
fp.seek(0)
json.dump(data, fp, ensure_ascii=False, indent=4, separators=(',', ': '), sort_keys=True)
fp.write('\n') # add a trailing newline.
fp.close()
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Format JSON files in place.'
)
parser.add_argument(
'files',
type=argparse.FileType('r+'),
help='JSON filepaths',
nargs='+'
)
args = parser.parse_args()
for json_file in args.files:
format_json(json_file)
| Make this a proper argparse script. | Make this a proper argparse script.
| Python | mit | nbeaver/cmd-oysters,nbeaver/cmd-oysters | #! /usr/bin/env python3
import sys
import json
for filepath in sys.argv[1:]:
with open(filepath) as f:
try:
oyster = json.load(f)
except ValueError:
sys.stderr.write("In file: {}\n".format(filepath))
raise
with open(filepath, 'w') as f:
json.dump(oyster, f, ensure_ascii=False, indent=4, separators=(',', ': '), sort_keys=True)
f.write('\n') # add a trailing newline.
Make this a proper argparse script. | #! /usr/bin/env python3
import sys
import json
import argparse
def format_json(fp):
try:
data = json.load(fp)
except ValueError:
sys.stderr.write("In file: {}\n".format(fp.name))
raise
# Jump back to the beginning of the file before overwriting it.
fp.seek(0)
json.dump(data, fp, ensure_ascii=False, indent=4, separators=(',', ': '), sort_keys=True)
fp.write('\n') # add a trailing newline.
fp.close()
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Format JSON files in place.'
)
parser.add_argument(
'files',
type=argparse.FileType('r+'),
help='JSON filepaths',
nargs='+'
)
args = parser.parse_args()
for json_file in args.files:
format_json(json_file)
| <commit_before>#! /usr/bin/env python3
import sys
import json
for filepath in sys.argv[1:]:
with open(filepath) as f:
try:
oyster = json.load(f)
except ValueError:
sys.stderr.write("In file: {}\n".format(filepath))
raise
with open(filepath, 'w') as f:
json.dump(oyster, f, ensure_ascii=False, indent=4, separators=(',', ': '), sort_keys=True)
f.write('\n') # add a trailing newline.
<commit_msg>Make this a proper argparse script.<commit_after> | #! /usr/bin/env python3
import sys
import json
import argparse
def format_json(fp):
try:
data = json.load(fp)
except ValueError:
sys.stderr.write("In file: {}\n".format(fp.name))
raise
# Jump back to the beginning of the file before overwriting it.
fp.seek(0)
json.dump(data, fp, ensure_ascii=False, indent=4, separators=(',', ': '), sort_keys=True)
fp.write('\n') # add a trailing newline.
fp.close()
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Format JSON files in place.'
)
parser.add_argument(
'files',
type=argparse.FileType('r+'),
help='JSON filepaths',
nargs='+'
)
args = parser.parse_args()
for json_file in args.files:
format_json(json_file)
| #! /usr/bin/env python3
import sys
import json
for filepath in sys.argv[1:]:
with open(filepath) as f:
try:
oyster = json.load(f)
except ValueError:
sys.stderr.write("In file: {}\n".format(filepath))
raise
with open(filepath, 'w') as f:
json.dump(oyster, f, ensure_ascii=False, indent=4, separators=(',', ': '), sort_keys=True)
f.write('\n') # add a trailing newline.
Make this a proper argparse script.#! /usr/bin/env python3
import sys
import json
import argparse
def format_json(fp):
try:
data = json.load(fp)
except ValueError:
sys.stderr.write("In file: {}\n".format(fp.name))
raise
# Jump back to the beginning of the file before overwriting it.
fp.seek(0)
json.dump(data, fp, ensure_ascii=False, indent=4, separators=(',', ': '), sort_keys=True)
fp.write('\n') # add a trailing newline.
fp.close()
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Format JSON files in place.'
)
parser.add_argument(
'files',
type=argparse.FileType('r+'),
help='JSON filepaths',
nargs='+'
)
args = parser.parse_args()
for json_file in args.files:
format_json(json_file)
| <commit_before>#! /usr/bin/env python3
import sys
import json
for filepath in sys.argv[1:]:
with open(filepath) as f:
try:
oyster = json.load(f)
except ValueError:
sys.stderr.write("In file: {}\n".format(filepath))
raise
with open(filepath, 'w') as f:
json.dump(oyster, f, ensure_ascii=False, indent=4, separators=(',', ': '), sort_keys=True)
f.write('\n') # add a trailing newline.
<commit_msg>Make this a proper argparse script.<commit_after>#! /usr/bin/env python3
import sys
import json
import argparse
def format_json(fp):
try:
data = json.load(fp)
except ValueError:
sys.stderr.write("In file: {}\n".format(fp.name))
raise
# Jump back to the beginning of the file before overwriting it.
fp.seek(0)
json.dump(data, fp, ensure_ascii=False, indent=4, separators=(',', ': '), sort_keys=True)
fp.write('\n') # add a trailing newline.
fp.close()
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Format JSON files in place.'
)
parser.add_argument(
'files',
type=argparse.FileType('r+'),
help='JSON filepaths',
nargs='+'
)
args = parser.parse_args()
for json_file in args.files:
format_json(json_file)
|
2a242bb6984fae5e32f117fa5ae68118621f3c95 | pycroft/model/alembic/versions/fb8d553a7268_add_account_pattern.py | pycroft/model/alembic/versions/fb8d553a7268_add_account_pattern.py | """add account_pattern
Revision ID: fb8d553a7268
Revises: 28e56bf6f62c
Create Date: 2021-04-26 22:16:41.772282
"""
from alembic import op
import sqlalchemy as sa
import pycroft
# revision identifiers, used by Alembic.
revision = 'fb8d553a7268'
down_revision = '0b69e80a9388'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('account_pattern',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('pattern', sa.String(), nullable=False),
sa.Column('account_id', sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(['account_id'], ['account.id'], ondelete='CASCADE'),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('account_pattern')
# ### end Alembic commands ###
| """add account_pattern
Revision ID: fb8d553a7268
Revises: 28e56bf6f62c
Create Date: 2021-04-26 22:16:41.772282
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = 'fb8d553a7268'
down_revision = '0b69e80a9388'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('account_pattern',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('pattern', sa.String(), nullable=False),
sa.Column('account_id', sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(['account_id'], ['account.id'], ondelete='CASCADE'),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('account_pattern')
# ### end Alembic commands ###
| Remove unnecessary pycroft import in migration | Remove unnecessary pycroft import in migration
| Python | apache-2.0 | agdsn/pycroft,agdsn/pycroft,agdsn/pycroft,agdsn/pycroft,agdsn/pycroft | """add account_pattern
Revision ID: fb8d553a7268
Revises: 28e56bf6f62c
Create Date: 2021-04-26 22:16:41.772282
"""
from alembic import op
import sqlalchemy as sa
import pycroft
# revision identifiers, used by Alembic.
revision = 'fb8d553a7268'
down_revision = '0b69e80a9388'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('account_pattern',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('pattern', sa.String(), nullable=False),
sa.Column('account_id', sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(['account_id'], ['account.id'], ondelete='CASCADE'),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('account_pattern')
# ### end Alembic commands ###
Remove unnecessary pycroft import in migration | """add account_pattern
Revision ID: fb8d553a7268
Revises: 28e56bf6f62c
Create Date: 2021-04-26 22:16:41.772282
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = 'fb8d553a7268'
down_revision = '0b69e80a9388'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('account_pattern',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('pattern', sa.String(), nullable=False),
sa.Column('account_id', sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(['account_id'], ['account.id'], ondelete='CASCADE'),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('account_pattern')
# ### end Alembic commands ###
| <commit_before>"""add account_pattern
Revision ID: fb8d553a7268
Revises: 28e56bf6f62c
Create Date: 2021-04-26 22:16:41.772282
"""
from alembic import op
import sqlalchemy as sa
import pycroft
# revision identifiers, used by Alembic.
revision = 'fb8d553a7268'
down_revision = '0b69e80a9388'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('account_pattern',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('pattern', sa.String(), nullable=False),
sa.Column('account_id', sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(['account_id'], ['account.id'], ondelete='CASCADE'),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('account_pattern')
# ### end Alembic commands ###
<commit_msg>Remove unnecessary pycroft import in migration<commit_after> | """add account_pattern
Revision ID: fb8d553a7268
Revises: 28e56bf6f62c
Create Date: 2021-04-26 22:16:41.772282
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = 'fb8d553a7268'
down_revision = '0b69e80a9388'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('account_pattern',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('pattern', sa.String(), nullable=False),
sa.Column('account_id', sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(['account_id'], ['account.id'], ondelete='CASCADE'),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('account_pattern')
# ### end Alembic commands ###
| """add account_pattern
Revision ID: fb8d553a7268
Revises: 28e56bf6f62c
Create Date: 2021-04-26 22:16:41.772282
"""
from alembic import op
import sqlalchemy as sa
import pycroft
# revision identifiers, used by Alembic.
revision = 'fb8d553a7268'
down_revision = '0b69e80a9388'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('account_pattern',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('pattern', sa.String(), nullable=False),
sa.Column('account_id', sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(['account_id'], ['account.id'], ondelete='CASCADE'),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('account_pattern')
# ### end Alembic commands ###
Remove unnecessary pycroft import in migration"""add account_pattern
Revision ID: fb8d553a7268
Revises: 28e56bf6f62c
Create Date: 2021-04-26 22:16:41.772282
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = 'fb8d553a7268'
down_revision = '0b69e80a9388'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('account_pattern',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('pattern', sa.String(), nullable=False),
sa.Column('account_id', sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(['account_id'], ['account.id'], ondelete='CASCADE'),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('account_pattern')
# ### end Alembic commands ###
| <commit_before>"""add account_pattern
Revision ID: fb8d553a7268
Revises: 28e56bf6f62c
Create Date: 2021-04-26 22:16:41.772282
"""
from alembic import op
import sqlalchemy as sa
import pycroft
# revision identifiers, used by Alembic.
revision = 'fb8d553a7268'
down_revision = '0b69e80a9388'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('account_pattern',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('pattern', sa.String(), nullable=False),
sa.Column('account_id', sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(['account_id'], ['account.id'], ondelete='CASCADE'),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('account_pattern')
# ### end Alembic commands ###
<commit_msg>Remove unnecessary pycroft import in migration<commit_after>"""add account_pattern
Revision ID: fb8d553a7268
Revises: 28e56bf6f62c
Create Date: 2021-04-26 22:16:41.772282
"""
from alembic import op
import sqlalchemy as sa
# revision identifiers, used by Alembic.
revision = 'fb8d553a7268'
down_revision = '0b69e80a9388'
branch_labels = None
depends_on = None
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table('account_pattern',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('pattern', sa.String(), nullable=False),
sa.Column('account_id', sa.Integer(), nullable=False),
sa.ForeignKeyConstraint(['account_id'], ['account.id'], ondelete='CASCADE'),
sa.PrimaryKeyConstraint('id')
)
# ### end Alembic commands ###
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table('account_pattern')
# ### end Alembic commands ###
|
569e180b99be2ec67f360a7081bbd54020d78a25 | grum/models.py | grum/models.py | import bcrypt
from grum import db
class User(db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), unique=True)
password = db.Column(db.String(128))
display_name = db.Column(db.String(128))
def __init__(self, username=None, password=None):
if username:
self.username = username
if password:
self.set_password(password)
def set_password(self, plaintext_password):
self.password = bcrypt.hashpw(plaintext_password.encode('utf-8'), bcrypt.gensalt())
def validate_password(self, plaintext_password):
hashed = bcrypt.hashpw(plaintext_password.encode('utf-8'), bytes(self.password.encode('utf-8')))
return hashed == bytes(self.password.encode('utf-8'))
class EmailAccount(db.Model):
address = db.Column(db.String(128), primary_key=True)
owner_id = db.Column(db.Integer, db.ForeignKey('user.id'))
mg_api = db.Column(db.String(64)) | import bcrypt
from grum import db
class User(db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), unique=True)
password = db.Column(db.String(128))
display_name = db.Column(db.String(128))
def __init__(self, username=None, display_name=None, password=None):
if username:
self.username = username
if display_name:
self.display_name = display_name
if password:
self.set_password(password)
def set_password(self, plaintext_password):
self.password = bcrypt.hashpw(plaintext_password.encode('utf-8'), bcrypt.gensalt())
def validate_password(self, plaintext_password):
hashed = bcrypt.hashpw(plaintext_password.encode('utf-8'), bytes(self.password.encode('utf-8')))
return hashed == bytes(self.password.encode('utf-8'))
class EmailAccount(db.Model):
address = db.Column(db.String(128), primary_key=True)
owner_id = db.Column(db.Integer, db.ForeignKey('user.id'))
mg_api = db.Column(db.String(64)) | Add display name to the constructor for User | Add display name to the constructor for User
| Python | mit | Grum-Hackdee/grum-web,Grum-Hackdee/grum-web,Grum-Hackdee/grum-web,Grum-Hackdee/grum-web | import bcrypt
from grum import db
class User(db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), unique=True)
password = db.Column(db.String(128))
display_name = db.Column(db.String(128))
def __init__(self, username=None, password=None):
if username:
self.username = username
if password:
self.set_password(password)
def set_password(self, plaintext_password):
self.password = bcrypt.hashpw(plaintext_password.encode('utf-8'), bcrypt.gensalt())
def validate_password(self, plaintext_password):
hashed = bcrypt.hashpw(plaintext_password.encode('utf-8'), bytes(self.password.encode('utf-8')))
return hashed == bytes(self.password.encode('utf-8'))
class EmailAccount(db.Model):
address = db.Column(db.String(128), primary_key=True)
owner_id = db.Column(db.Integer, db.ForeignKey('user.id'))
mg_api = db.Column(db.String(64))Add display name to the constructor for User | import bcrypt
from grum import db
class User(db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), unique=True)
password = db.Column(db.String(128))
display_name = db.Column(db.String(128))
def __init__(self, username=None, display_name=None, password=None):
if username:
self.username = username
if display_name:
self.display_name = display_name
if password:
self.set_password(password)
def set_password(self, plaintext_password):
self.password = bcrypt.hashpw(plaintext_password.encode('utf-8'), bcrypt.gensalt())
def validate_password(self, plaintext_password):
hashed = bcrypt.hashpw(plaintext_password.encode('utf-8'), bytes(self.password.encode('utf-8')))
return hashed == bytes(self.password.encode('utf-8'))
class EmailAccount(db.Model):
address = db.Column(db.String(128), primary_key=True)
owner_id = db.Column(db.Integer, db.ForeignKey('user.id'))
mg_api = db.Column(db.String(64)) | <commit_before>import bcrypt
from grum import db
class User(db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), unique=True)
password = db.Column(db.String(128))
display_name = db.Column(db.String(128))
def __init__(self, username=None, password=None):
if username:
self.username = username
if password:
self.set_password(password)
def set_password(self, plaintext_password):
self.password = bcrypt.hashpw(plaintext_password.encode('utf-8'), bcrypt.gensalt())
def validate_password(self, plaintext_password):
hashed = bcrypt.hashpw(plaintext_password.encode('utf-8'), bytes(self.password.encode('utf-8')))
return hashed == bytes(self.password.encode('utf-8'))
class EmailAccount(db.Model):
address = db.Column(db.String(128), primary_key=True)
owner_id = db.Column(db.Integer, db.ForeignKey('user.id'))
mg_api = db.Column(db.String(64))<commit_msg>Add display name to the constructor for User<commit_after> | import bcrypt
from grum import db
class User(db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), unique=True)
password = db.Column(db.String(128))
display_name = db.Column(db.String(128))
def __init__(self, username=None, display_name=None, password=None):
if username:
self.username = username
if display_name:
self.display_name = display_name
if password:
self.set_password(password)
def set_password(self, plaintext_password):
self.password = bcrypt.hashpw(plaintext_password.encode('utf-8'), bcrypt.gensalt())
def validate_password(self, plaintext_password):
hashed = bcrypt.hashpw(plaintext_password.encode('utf-8'), bytes(self.password.encode('utf-8')))
return hashed == bytes(self.password.encode('utf-8'))
class EmailAccount(db.Model):
address = db.Column(db.String(128), primary_key=True)
owner_id = db.Column(db.Integer, db.ForeignKey('user.id'))
mg_api = db.Column(db.String(64)) | import bcrypt
from grum import db
class User(db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), unique=True)
password = db.Column(db.String(128))
display_name = db.Column(db.String(128))
def __init__(self, username=None, password=None):
if username:
self.username = username
if password:
self.set_password(password)
def set_password(self, plaintext_password):
self.password = bcrypt.hashpw(plaintext_password.encode('utf-8'), bcrypt.gensalt())
def validate_password(self, plaintext_password):
hashed = bcrypt.hashpw(plaintext_password.encode('utf-8'), bytes(self.password.encode('utf-8')))
return hashed == bytes(self.password.encode('utf-8'))
class EmailAccount(db.Model):
address = db.Column(db.String(128), primary_key=True)
owner_id = db.Column(db.Integer, db.ForeignKey('user.id'))
mg_api = db.Column(db.String(64))Add display name to the constructor for Userimport bcrypt
from grum import db
class User(db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), unique=True)
password = db.Column(db.String(128))
display_name = db.Column(db.String(128))
def __init__(self, username=None, display_name=None, password=None):
if username:
self.username = username
if display_name:
self.display_name = display_name
if password:
self.set_password(password)
def set_password(self, plaintext_password):
self.password = bcrypt.hashpw(plaintext_password.encode('utf-8'), bcrypt.gensalt())
def validate_password(self, plaintext_password):
hashed = bcrypt.hashpw(plaintext_password.encode('utf-8'), bytes(self.password.encode('utf-8')))
return hashed == bytes(self.password.encode('utf-8'))
class EmailAccount(db.Model):
address = db.Column(db.String(128), primary_key=True)
owner_id = db.Column(db.Integer, db.ForeignKey('user.id'))
mg_api = db.Column(db.String(64)) | <commit_before>import bcrypt
from grum import db
class User(db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), unique=True)
password = db.Column(db.String(128))
display_name = db.Column(db.String(128))
def __init__(self, username=None, password=None):
if username:
self.username = username
if password:
self.set_password(password)
def set_password(self, plaintext_password):
self.password = bcrypt.hashpw(plaintext_password.encode('utf-8'), bcrypt.gensalt())
def validate_password(self, plaintext_password):
hashed = bcrypt.hashpw(plaintext_password.encode('utf-8'), bytes(self.password.encode('utf-8')))
return hashed == bytes(self.password.encode('utf-8'))
class EmailAccount(db.Model):
address = db.Column(db.String(128), primary_key=True)
owner_id = db.Column(db.Integer, db.ForeignKey('user.id'))
mg_api = db.Column(db.String(64))<commit_msg>Add display name to the constructor for User<commit_after>import bcrypt
from grum import db
class User(db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), unique=True)
password = db.Column(db.String(128))
display_name = db.Column(db.String(128))
def __init__(self, username=None, display_name=None, password=None):
if username:
self.username = username
if display_name:
self.display_name = display_name
if password:
self.set_password(password)
def set_password(self, plaintext_password):
self.password = bcrypt.hashpw(plaintext_password.encode('utf-8'), bcrypt.gensalt())
def validate_password(self, plaintext_password):
hashed = bcrypt.hashpw(plaintext_password.encode('utf-8'), bytes(self.password.encode('utf-8')))
return hashed == bytes(self.password.encode('utf-8'))
class EmailAccount(db.Model):
address = db.Column(db.String(128), primary_key=True)
owner_id = db.Column(db.Integer, db.ForeignKey('user.id'))
mg_api = db.Column(db.String(64)) |
94861438189537b88deaf8d04cc9942192038d8c | user_messages/views.py | user_messages/views.py | from django.contrib.auth.decorators import login_required
from django.db.models import Q
from django.shortcuts import get_object_or_404
from django.template import RequestContext
from user_messages.models import Thread, Message
@login_required
def inbox(request, template_name='user_messages/inbox.html'):
threads = list(Thread.objects.inbox(request.user))
threads.sort(key=lambda o: o.latest_message.sent_at, reversed=True)
return render_to_response(template_name, {'threads': threads}, context_instance=RequestContext(request))
@login_required
def thread_detail(request, thread_id,
template_name='user_messages/thread_detail.html'):
qs = Thread.objects.filter(Q(to_user=request.user) | Q(from_user=request.user))
thread = get_object_or_404(qs, pk=thread_id)
return render_to_response(template_name, {'thread': thread}, context_instance=RequestContext(request))
| from django.contrib.auth.decorators import login_required
from django.db.models import Q
from django.shortcuts import get_object_or_404
from django.template import RequestContext
from user_messages.models import Thread, Message
@login_required
def inbox(request, template_name='user_messages/inbox.html'):
threads = list(Thread.objects.inbox(request.user))
threads.sort(key=lambda o: o.latest_message.sent_at, reversed=True)
return render_to_response(template_name, {'threads': threads}, context_instance=RequestContext(request))
@login_required
def thread_detail(request, thread_id,
template_name='user_messages/thread_detail.html'):
qs = Thread.objects.filter(Q(to_user=request.user) | Q(from_user=request.user))
thread = get_object_or_404(qs, pk=thread_id)
if request.user == thread.to_user:
thread.to_user_unread = False
else:
thread.from_user_unread = False
thread.save()
return render_to_response(template_name, {'thread': thread}, context_instance=RequestContext(request))
| Update the read status of a thread when it's viewed | Update the read status of a thread when it's viewed
| Python | mit | eldarion/user_messages,eldarion/user_messages,pinax/pinax-messages,arthur-wsw/pinax-messages,pinax/pinax-messages,arthur-wsw/pinax-messages | from django.contrib.auth.decorators import login_required
from django.db.models import Q
from django.shortcuts import get_object_or_404
from django.template import RequestContext
from user_messages.models import Thread, Message
@login_required
def inbox(request, template_name='user_messages/inbox.html'):
threads = list(Thread.objects.inbox(request.user))
threads.sort(key=lambda o: o.latest_message.sent_at, reversed=True)
return render_to_response(template_name, {'threads': threads}, context_instance=RequestContext(request))
@login_required
def thread_detail(request, thread_id,
template_name='user_messages/thread_detail.html'):
qs = Thread.objects.filter(Q(to_user=request.user) | Q(from_user=request.user))
thread = get_object_or_404(qs, pk=thread_id)
return render_to_response(template_name, {'thread': thread}, context_instance=RequestContext(request))
Update the read status of a thread when it's viewed | from django.contrib.auth.decorators import login_required
from django.db.models import Q
from django.shortcuts import get_object_or_404
from django.template import RequestContext
from user_messages.models import Thread, Message
@login_required
def inbox(request, template_name='user_messages/inbox.html'):
threads = list(Thread.objects.inbox(request.user))
threads.sort(key=lambda o: o.latest_message.sent_at, reversed=True)
return render_to_response(template_name, {'threads': threads}, context_instance=RequestContext(request))
@login_required
def thread_detail(request, thread_id,
template_name='user_messages/thread_detail.html'):
qs = Thread.objects.filter(Q(to_user=request.user) | Q(from_user=request.user))
thread = get_object_or_404(qs, pk=thread_id)
if request.user == thread.to_user:
thread.to_user_unread = False
else:
thread.from_user_unread = False
thread.save()
return render_to_response(template_name, {'thread': thread}, context_instance=RequestContext(request))
| <commit_before>from django.contrib.auth.decorators import login_required
from django.db.models import Q
from django.shortcuts import get_object_or_404
from django.template import RequestContext
from user_messages.models import Thread, Message
@login_required
def inbox(request, template_name='user_messages/inbox.html'):
threads = list(Thread.objects.inbox(request.user))
threads.sort(key=lambda o: o.latest_message.sent_at, reversed=True)
return render_to_response(template_name, {'threads': threads}, context_instance=RequestContext(request))
@login_required
def thread_detail(request, thread_id,
template_name='user_messages/thread_detail.html'):
qs = Thread.objects.filter(Q(to_user=request.user) | Q(from_user=request.user))
thread = get_object_or_404(qs, pk=thread_id)
return render_to_response(template_name, {'thread': thread}, context_instance=RequestContext(request))
<commit_msg>Update the read status of a thread when it's viewed<commit_after> | from django.contrib.auth.decorators import login_required
from django.db.models import Q
from django.shortcuts import get_object_or_404
from django.template import RequestContext
from user_messages.models import Thread, Message
@login_required
def inbox(request, template_name='user_messages/inbox.html'):
threads = list(Thread.objects.inbox(request.user))
threads.sort(key=lambda o: o.latest_message.sent_at, reversed=True)
return render_to_response(template_name, {'threads': threads}, context_instance=RequestContext(request))
@login_required
def thread_detail(request, thread_id,
template_name='user_messages/thread_detail.html'):
qs = Thread.objects.filter(Q(to_user=request.user) | Q(from_user=request.user))
thread = get_object_or_404(qs, pk=thread_id)
if request.user == thread.to_user:
thread.to_user_unread = False
else:
thread.from_user_unread = False
thread.save()
return render_to_response(template_name, {'thread': thread}, context_instance=RequestContext(request))
| from django.contrib.auth.decorators import login_required
from django.db.models import Q
from django.shortcuts import get_object_or_404
from django.template import RequestContext
from user_messages.models import Thread, Message
@login_required
def inbox(request, template_name='user_messages/inbox.html'):
threads = list(Thread.objects.inbox(request.user))
threads.sort(key=lambda o: o.latest_message.sent_at, reversed=True)
return render_to_response(template_name, {'threads': threads}, context_instance=RequestContext(request))
@login_required
def thread_detail(request, thread_id,
template_name='user_messages/thread_detail.html'):
qs = Thread.objects.filter(Q(to_user=request.user) | Q(from_user=request.user))
thread = get_object_or_404(qs, pk=thread_id)
return render_to_response(template_name, {'thread': thread}, context_instance=RequestContext(request))
Update the read status of a thread when it's viewedfrom django.contrib.auth.decorators import login_required
from django.db.models import Q
from django.shortcuts import get_object_or_404
from django.template import RequestContext
from user_messages.models import Thread, Message
@login_required
def inbox(request, template_name='user_messages/inbox.html'):
threads = list(Thread.objects.inbox(request.user))
threads.sort(key=lambda o: o.latest_message.sent_at, reversed=True)
return render_to_response(template_name, {'threads': threads}, context_instance=RequestContext(request))
@login_required
def thread_detail(request, thread_id,
template_name='user_messages/thread_detail.html'):
qs = Thread.objects.filter(Q(to_user=request.user) | Q(from_user=request.user))
thread = get_object_or_404(qs, pk=thread_id)
if request.user == thread.to_user:
thread.to_user_unread = False
else:
thread.from_user_unread = False
thread.save()
return render_to_response(template_name, {'thread': thread}, context_instance=RequestContext(request))
| <commit_before>from django.contrib.auth.decorators import login_required
from django.db.models import Q
from django.shortcuts import get_object_or_404
from django.template import RequestContext
from user_messages.models import Thread, Message
@login_required
def inbox(request, template_name='user_messages/inbox.html'):
threads = list(Thread.objects.inbox(request.user))
threads.sort(key=lambda o: o.latest_message.sent_at, reversed=True)
return render_to_response(template_name, {'threads': threads}, context_instance=RequestContext(request))
@login_required
def thread_detail(request, thread_id,
template_name='user_messages/thread_detail.html'):
qs = Thread.objects.filter(Q(to_user=request.user) | Q(from_user=request.user))
thread = get_object_or_404(qs, pk=thread_id)
return render_to_response(template_name, {'thread': thread}, context_instance=RequestContext(request))
<commit_msg>Update the read status of a thread when it's viewed<commit_after>from django.contrib.auth.decorators import login_required
from django.db.models import Q
from django.shortcuts import get_object_or_404
from django.template import RequestContext
from user_messages.models import Thread, Message
@login_required
def inbox(request, template_name='user_messages/inbox.html'):
threads = list(Thread.objects.inbox(request.user))
threads.sort(key=lambda o: o.latest_message.sent_at, reversed=True)
return render_to_response(template_name, {'threads': threads}, context_instance=RequestContext(request))
@login_required
def thread_detail(request, thread_id,
template_name='user_messages/thread_detail.html'):
qs = Thread.objects.filter(Q(to_user=request.user) | Q(from_user=request.user))
thread = get_object_or_404(qs, pk=thread_id)
if request.user == thread.to_user:
thread.to_user_unread = False
else:
thread.from_user_unread = False
thread.save()
return render_to_response(template_name, {'thread': thread}, context_instance=RequestContext(request))
|
7897423ea3c8e418b405ce2d09318ad9b1526a22 | tests/test_urls.py | tests/test_urls.py | from django.conf.urls import patterns, url
from djproxy.urls import generate_routes
from test_views import LocalProxy, index
urlpatterns = patterns(
'',
url(r'^some/content/.*$', index, name='index'),
url(r'^local_proxy/(?P<url>.*)$', LocalProxy.as_view(), name='proxy')
) + generate_routes({
'service_one': {
'base_url': 'https://yahoo.com/',
'prefix': 'yahoo/'
},
'service_two': {
'base_url': 'https://google.com/',
'prefix': 'google/'
}
})
| from django.conf.urls import patterns, url
from djproxy.urls import generate_routes
from test_views import LocalProxy, index
urlpatterns = patterns(
'',
url(r'^some/content/.*$', index, name='index'),
url(r'^local_proxy/(?P<url>.*)$', LocalProxy.as_view(), name='proxy')
) + generate_routes({
'service_one': {
'base_url': 'https://yahoo.com/',
'prefix': 'yahoo/'
},
'service_two': {
'base_url': 'https://www.google.com/',
'prefix': 'google/'
}
})
| Add www to test google config | Add www to test google config
This makes it work right for google.com sub pages when testing via a browser.
| Python | mit | thomasw/djproxy | from django.conf.urls import patterns, url
from djproxy.urls import generate_routes
from test_views import LocalProxy, index
urlpatterns = patterns(
'',
url(r'^some/content/.*$', index, name='index'),
url(r'^local_proxy/(?P<url>.*)$', LocalProxy.as_view(), name='proxy')
) + generate_routes({
'service_one': {
'base_url': 'https://yahoo.com/',
'prefix': 'yahoo/'
},
'service_two': {
'base_url': 'https://google.com/',
'prefix': 'google/'
}
})
Add www to test google config
This makes it work right for google.com sub pages when testing via a browser. | from django.conf.urls import patterns, url
from djproxy.urls import generate_routes
from test_views import LocalProxy, index
urlpatterns = patterns(
'',
url(r'^some/content/.*$', index, name='index'),
url(r'^local_proxy/(?P<url>.*)$', LocalProxy.as_view(), name='proxy')
) + generate_routes({
'service_one': {
'base_url': 'https://yahoo.com/',
'prefix': 'yahoo/'
},
'service_two': {
'base_url': 'https://www.google.com/',
'prefix': 'google/'
}
})
| <commit_before>from django.conf.urls import patterns, url
from djproxy.urls import generate_routes
from test_views import LocalProxy, index
urlpatterns = patterns(
'',
url(r'^some/content/.*$', index, name='index'),
url(r'^local_proxy/(?P<url>.*)$', LocalProxy.as_view(), name='proxy')
) + generate_routes({
'service_one': {
'base_url': 'https://yahoo.com/',
'prefix': 'yahoo/'
},
'service_two': {
'base_url': 'https://google.com/',
'prefix': 'google/'
}
})
<commit_msg>Add www to test google config
This makes it work right for google.com sub pages when testing via a browser.<commit_after> | from django.conf.urls import patterns, url
from djproxy.urls import generate_routes
from test_views import LocalProxy, index
urlpatterns = patterns(
'',
url(r'^some/content/.*$', index, name='index'),
url(r'^local_proxy/(?P<url>.*)$', LocalProxy.as_view(), name='proxy')
) + generate_routes({
'service_one': {
'base_url': 'https://yahoo.com/',
'prefix': 'yahoo/'
},
'service_two': {
'base_url': 'https://www.google.com/',
'prefix': 'google/'
}
})
| from django.conf.urls import patterns, url
from djproxy.urls import generate_routes
from test_views import LocalProxy, index
urlpatterns = patterns(
'',
url(r'^some/content/.*$', index, name='index'),
url(r'^local_proxy/(?P<url>.*)$', LocalProxy.as_view(), name='proxy')
) + generate_routes({
'service_one': {
'base_url': 'https://yahoo.com/',
'prefix': 'yahoo/'
},
'service_two': {
'base_url': 'https://google.com/',
'prefix': 'google/'
}
})
Add www to test google config
This makes it work right for google.com sub pages when testing via a browser.from django.conf.urls import patterns, url
from djproxy.urls import generate_routes
from test_views import LocalProxy, index
urlpatterns = patterns(
'',
url(r'^some/content/.*$', index, name='index'),
url(r'^local_proxy/(?P<url>.*)$', LocalProxy.as_view(), name='proxy')
) + generate_routes({
'service_one': {
'base_url': 'https://yahoo.com/',
'prefix': 'yahoo/'
},
'service_two': {
'base_url': 'https://www.google.com/',
'prefix': 'google/'
}
})
| <commit_before>from django.conf.urls import patterns, url
from djproxy.urls import generate_routes
from test_views import LocalProxy, index
urlpatterns = patterns(
'',
url(r'^some/content/.*$', index, name='index'),
url(r'^local_proxy/(?P<url>.*)$', LocalProxy.as_view(), name='proxy')
) + generate_routes({
'service_one': {
'base_url': 'https://yahoo.com/',
'prefix': 'yahoo/'
},
'service_two': {
'base_url': 'https://google.com/',
'prefix': 'google/'
}
})
<commit_msg>Add www to test google config
This makes it work right for google.com sub pages when testing via a browser.<commit_after>from django.conf.urls import patterns, url
from djproxy.urls import generate_routes
from test_views import LocalProxy, index
urlpatterns = patterns(
'',
url(r'^some/content/.*$', index, name='index'),
url(r'^local_proxy/(?P<url>.*)$', LocalProxy.as_view(), name='proxy')
) + generate_routes({
'service_one': {
'base_url': 'https://yahoo.com/',
'prefix': 'yahoo/'
},
'service_two': {
'base_url': 'https://www.google.com/',
'prefix': 'google/'
}
})
|
64f9ef6fcc71ef09e113161711369fe4d9781a18 | shorpypaper.py | shorpypaper.py | #!/usr/bin/python
from pyquery import PyQuery as pq
import requests
import subprocess
APPLESCRIPT = """/usr/bin/osascript<<END
tell application "Finder"
set desktop picture to POSIX file "%s"
end tell
END"""
def main():
# Load main site.
root = 'http://www.shorpy.com'
r = requests.get(root)
j = pq(r.content)
# Load first photo.
first_photo = root + j('div.node div.content a').eq(1).attr('href')
r = requests.get(first_photo)
j = pq(r.content)
image = j('img').eq(0).attr('src')
with open('/tmp/dailyshorpy.jpg', 'wb') as handle:
# To reset the cached dailyshorpy.jpg.
subprocess.Popen(APPLESCRIPT % '/Library/Desktop Pictures/Frog.jpg', shell=True)
request = requests.get(image, stream=True)
for block in request.iter_content(1024):
if not block:
break
handle.write(block)
subprocess.Popen(APPLESCRIPT % '/tmp/dailyshorpy.jpg', shell=True)
if __name__ == '__main__':
main()
| #!/usr/bin/python
from pyquery import PyQuery as pq
import requests
import subprocess
APPLESCRIPT = """/usr/bin/osascript<<END
tell application "Finder"
set desktop picture to POSIX file "%s"
end tell
END"""
def main():
# Load main site.
root = 'http://www.shorpy.com'
r = requests.get(root)
j = pq(r.content)
# Load first photo.
first_photo = root + j('div.node div.content a').eq(1).attr('href')
r = requests.get(first_photo)
j = pq(r.content)
image = j('img').eq(0).attr('src')
with open('/tmp/dailyshorpy.jpg', 'wb') as handle:
# To reset the cached dailyshorpy.jpg.
subprocess.Popen(APPLESCRIPT % '/Library/Desktop Pictures/Solid Colors/Solid Gray Light.png', shell=True)
request = requests.get(image, stream=True)
for block in request.iter_content(1024):
if not block:
break
handle.write(block)
subprocess.Popen(APPLESCRIPT % '/tmp/dailyshorpy.jpg', shell=True)
if __name__ == '__main__':
main()
| Use a grey solid instead of the damn frog. | Use a grey solid instead of the damn frog.
| Python | mit | nicksergeant/shorpypaper | #!/usr/bin/python
from pyquery import PyQuery as pq
import requests
import subprocess
APPLESCRIPT = """/usr/bin/osascript<<END
tell application "Finder"
set desktop picture to POSIX file "%s"
end tell
END"""
def main():
# Load main site.
root = 'http://www.shorpy.com'
r = requests.get(root)
j = pq(r.content)
# Load first photo.
first_photo = root + j('div.node div.content a').eq(1).attr('href')
r = requests.get(first_photo)
j = pq(r.content)
image = j('img').eq(0).attr('src')
with open('/tmp/dailyshorpy.jpg', 'wb') as handle:
# To reset the cached dailyshorpy.jpg.
subprocess.Popen(APPLESCRIPT % '/Library/Desktop Pictures/Frog.jpg', shell=True)
request = requests.get(image, stream=True)
for block in request.iter_content(1024):
if not block:
break
handle.write(block)
subprocess.Popen(APPLESCRIPT % '/tmp/dailyshorpy.jpg', shell=True)
if __name__ == '__main__':
main()
Use a grey solid instead of the damn frog. | #!/usr/bin/python
from pyquery import PyQuery as pq
import requests
import subprocess
APPLESCRIPT = """/usr/bin/osascript<<END
tell application "Finder"
set desktop picture to POSIX file "%s"
end tell
END"""
def main():
# Load main site.
root = 'http://www.shorpy.com'
r = requests.get(root)
j = pq(r.content)
# Load first photo.
first_photo = root + j('div.node div.content a').eq(1).attr('href')
r = requests.get(first_photo)
j = pq(r.content)
image = j('img').eq(0).attr('src')
with open('/tmp/dailyshorpy.jpg', 'wb') as handle:
# To reset the cached dailyshorpy.jpg.
subprocess.Popen(APPLESCRIPT % '/Library/Desktop Pictures/Solid Colors/Solid Gray Light.png', shell=True)
request = requests.get(image, stream=True)
for block in request.iter_content(1024):
if not block:
break
handle.write(block)
subprocess.Popen(APPLESCRIPT % '/tmp/dailyshorpy.jpg', shell=True)
if __name__ == '__main__':
main()
| <commit_before>#!/usr/bin/python
from pyquery import PyQuery as pq
import requests
import subprocess
APPLESCRIPT = """/usr/bin/osascript<<END
tell application "Finder"
set desktop picture to POSIX file "%s"
end tell
END"""
def main():
# Load main site.
root = 'http://www.shorpy.com'
r = requests.get(root)
j = pq(r.content)
# Load first photo.
first_photo = root + j('div.node div.content a').eq(1).attr('href')
r = requests.get(first_photo)
j = pq(r.content)
image = j('img').eq(0).attr('src')
with open('/tmp/dailyshorpy.jpg', 'wb') as handle:
# To reset the cached dailyshorpy.jpg.
subprocess.Popen(APPLESCRIPT % '/Library/Desktop Pictures/Frog.jpg', shell=True)
request = requests.get(image, stream=True)
for block in request.iter_content(1024):
if not block:
break
handle.write(block)
subprocess.Popen(APPLESCRIPT % '/tmp/dailyshorpy.jpg', shell=True)
if __name__ == '__main__':
main()
<commit_msg>Use a grey solid instead of the damn frog.<commit_after> | #!/usr/bin/python
from pyquery import PyQuery as pq
import requests
import subprocess
APPLESCRIPT = """/usr/bin/osascript<<END
tell application "Finder"
set desktop picture to POSIX file "%s"
end tell
END"""
def main():
# Load main site.
root = 'http://www.shorpy.com'
r = requests.get(root)
j = pq(r.content)
# Load first photo.
first_photo = root + j('div.node div.content a').eq(1).attr('href')
r = requests.get(first_photo)
j = pq(r.content)
image = j('img').eq(0).attr('src')
with open('/tmp/dailyshorpy.jpg', 'wb') as handle:
# To reset the cached dailyshorpy.jpg.
subprocess.Popen(APPLESCRIPT % '/Library/Desktop Pictures/Solid Colors/Solid Gray Light.png', shell=True)
request = requests.get(image, stream=True)
for block in request.iter_content(1024):
if not block:
break
handle.write(block)
subprocess.Popen(APPLESCRIPT % '/tmp/dailyshorpy.jpg', shell=True)
if __name__ == '__main__':
main()
| #!/usr/bin/python
from pyquery import PyQuery as pq
import requests
import subprocess
APPLESCRIPT = """/usr/bin/osascript<<END
tell application "Finder"
set desktop picture to POSIX file "%s"
end tell
END"""
def main():
# Load main site.
root = 'http://www.shorpy.com'
r = requests.get(root)
j = pq(r.content)
# Load first photo.
first_photo = root + j('div.node div.content a').eq(1).attr('href')
r = requests.get(first_photo)
j = pq(r.content)
image = j('img').eq(0).attr('src')
with open('/tmp/dailyshorpy.jpg', 'wb') as handle:
# To reset the cached dailyshorpy.jpg.
subprocess.Popen(APPLESCRIPT % '/Library/Desktop Pictures/Frog.jpg', shell=True)
request = requests.get(image, stream=True)
for block in request.iter_content(1024):
if not block:
break
handle.write(block)
subprocess.Popen(APPLESCRIPT % '/tmp/dailyshorpy.jpg', shell=True)
if __name__ == '__main__':
main()
Use a grey solid instead of the damn frog.#!/usr/bin/python
from pyquery import PyQuery as pq
import requests
import subprocess
APPLESCRIPT = """/usr/bin/osascript<<END
tell application "Finder"
set desktop picture to POSIX file "%s"
end tell
END"""
def main():
# Load main site.
root = 'http://www.shorpy.com'
r = requests.get(root)
j = pq(r.content)
# Load first photo.
first_photo = root + j('div.node div.content a').eq(1).attr('href')
r = requests.get(first_photo)
j = pq(r.content)
image = j('img').eq(0).attr('src')
with open('/tmp/dailyshorpy.jpg', 'wb') as handle:
# To reset the cached dailyshorpy.jpg.
subprocess.Popen(APPLESCRIPT % '/Library/Desktop Pictures/Solid Colors/Solid Gray Light.png', shell=True)
request = requests.get(image, stream=True)
for block in request.iter_content(1024):
if not block:
break
handle.write(block)
subprocess.Popen(APPLESCRIPT % '/tmp/dailyshorpy.jpg', shell=True)
if __name__ == '__main__':
main()
| <commit_before>#!/usr/bin/python
from pyquery import PyQuery as pq
import requests
import subprocess
APPLESCRIPT = """/usr/bin/osascript<<END
tell application "Finder"
set desktop picture to POSIX file "%s"
end tell
END"""
def main():
# Load main site.
root = 'http://www.shorpy.com'
r = requests.get(root)
j = pq(r.content)
# Load first photo.
first_photo = root + j('div.node div.content a').eq(1).attr('href')
r = requests.get(first_photo)
j = pq(r.content)
image = j('img').eq(0).attr('src')
with open('/tmp/dailyshorpy.jpg', 'wb') as handle:
# To reset the cached dailyshorpy.jpg.
subprocess.Popen(APPLESCRIPT % '/Library/Desktop Pictures/Frog.jpg', shell=True)
request = requests.get(image, stream=True)
for block in request.iter_content(1024):
if not block:
break
handle.write(block)
subprocess.Popen(APPLESCRIPT % '/tmp/dailyshorpy.jpg', shell=True)
if __name__ == '__main__':
main()
<commit_msg>Use a grey solid instead of the damn frog.<commit_after>#!/usr/bin/python
from pyquery import PyQuery as pq
import requests
import subprocess
APPLESCRIPT = """/usr/bin/osascript<<END
tell application "Finder"
set desktop picture to POSIX file "%s"
end tell
END"""
def main():
# Load main site.
root = 'http://www.shorpy.com'
r = requests.get(root)
j = pq(r.content)
# Load first photo.
first_photo = root + j('div.node div.content a').eq(1).attr('href')
r = requests.get(first_photo)
j = pq(r.content)
image = j('img').eq(0).attr('src')
with open('/tmp/dailyshorpy.jpg', 'wb') as handle:
# To reset the cached dailyshorpy.jpg.
subprocess.Popen(APPLESCRIPT % '/Library/Desktop Pictures/Solid Colors/Solid Gray Light.png', shell=True)
request = requests.get(image, stream=True)
for block in request.iter_content(1024):
if not block:
break
handle.write(block)
subprocess.Popen(APPLESCRIPT % '/tmp/dailyshorpy.jpg', shell=True)
if __name__ == '__main__':
main()
|
c09f586bfa36f4ff66ae3b8a82fd7b4eeb8ea5d7 | windpowerlib/tools.py | windpowerlib/tools.py | """
The ``tools`` module contains a collection of functions used in the
windpowerlib.
"""
__copyright__ = "Copyright oemof developer group"
__license__ = "GPLv3"
| """
The ``tools`` module contains a collection of functions used in the
windpowerlib.
"""
__copyright__ = "Copyright oemof developer group"
__license__ = "GPLv3"
import collections
def select_closer_value(value_1, value_2, comp_value, corresp_1, corresp_2):
r"""
Selects the value with the smaller difference to a comparative value.
Additionally returns a corresponding value. This function is for example
used in :py:func:`~.modelchain.v_wind_hub` of the
:class:`~.modelchain.ModelChain` to choose the wind speed data that is
close to the hub height of the examined wind turbine. In this case
`value_1` and `value_2` are the heights of the corresponding wind speed
data sets `corresp_1` and `corresp_2`.
Parameters
----------
value_1 : float
First value of which the difference to `comp_value` will be
compared with the difference to `comp_value` of `value_2`.
value_2 : float
Second value for comparison.
comp_value : float
Comparative value.
corresp_1 : float
Corresponding value to `value_1`.
corresp_2 : float
Corresponding value to `value_2`.
Returns
-------
Tuple(float, float, string)
Value closer to comparing value as float, corresponding value as
float and a string for logging.debug.
"""
if (value_2 is not None and corresp_2 is not None):
if value_1 == comp_value:
closest_value = value_1
logging_string = '(at hub height).'
elif value_2 == comp_value:
closest_value = value_2
logging_string = '(2) (at hub height).'
elif abs(value_1 - comp_value) <= abs(value_2 - comp_value):
closest_value = value_1
logging_string = None
else:
closest_value = value_2
logging_string = None
else:
closest_value = value_1
if value_1 == comp_value:
logging_string = '(at hub height).'
# Select correponding value
if closest_value == value_1:
corresp_value = corresp_1
else:
corresp_value = corresp_2
# Store values in a named tuple
return_tuple = collections.namedtuple('selected_values',
['closest_value',
'corresp_value', 'logging_string'])
return return_tuple(closest_value, corresp_value, logging_string)
| Add function for selection of value closer to comparative value | Add function for selection of value closer to comparative value
| Python | mit | wind-python/windpowerlib | """
The ``tools`` module contains a collection of functions used in the
windpowerlib.
"""
__copyright__ = "Copyright oemof developer group"
__license__ = "GPLv3"
Add function for selection of value closer to comparative value | """
The ``tools`` module contains a collection of functions used in the
windpowerlib.
"""
__copyright__ = "Copyright oemof developer group"
__license__ = "GPLv3"
import collections
def select_closer_value(value_1, value_2, comp_value, corresp_1, corresp_2):
r"""
Selects the value with the smaller difference to a comparative value.
Additionally returns a corresponding value. This function is for example
used in :py:func:`~.modelchain.v_wind_hub` of the
:class:`~.modelchain.ModelChain` to choose the wind speed data that is
close to the hub height of the examined wind turbine. In this case
`value_1` and `value_2` are the heights of the corresponding wind speed
data sets `corresp_1` and `corresp_2`.
Parameters
----------
value_1 : float
First value of which the difference to `comp_value` will be
compared with the difference to `comp_value` of `value_2`.
value_2 : float
Second value for comparison.
comp_value : float
Comparative value.
corresp_1 : float
Corresponding value to `value_1`.
corresp_2 : float
Corresponding value to `value_2`.
Returns
-------
Tuple(float, float, string)
Value closer to comparing value as float, corresponding value as
float and a string for logging.debug.
"""
if (value_2 is not None and corresp_2 is not None):
if value_1 == comp_value:
closest_value = value_1
logging_string = '(at hub height).'
elif value_2 == comp_value:
closest_value = value_2
logging_string = '(2) (at hub height).'
elif abs(value_1 - comp_value) <= abs(value_2 - comp_value):
closest_value = value_1
logging_string = None
else:
closest_value = value_2
logging_string = None
else:
closest_value = value_1
if value_1 == comp_value:
logging_string = '(at hub height).'
# Select correponding value
if closest_value == value_1:
corresp_value = corresp_1
else:
corresp_value = corresp_2
# Store values in a named tuple
return_tuple = collections.namedtuple('selected_values',
['closest_value',
'corresp_value', 'logging_string'])
return return_tuple(closest_value, corresp_value, logging_string)
| <commit_before>"""
The ``tools`` module contains a collection of functions used in the
windpowerlib.
"""
__copyright__ = "Copyright oemof developer group"
__license__ = "GPLv3"
<commit_msg>Add function for selection of value closer to comparative value<commit_after> | """
The ``tools`` module contains a collection of functions used in the
windpowerlib.
"""
__copyright__ = "Copyright oemof developer group"
__license__ = "GPLv3"
import collections
def select_closer_value(value_1, value_2, comp_value, corresp_1, corresp_2):
r"""
Selects the value with the smaller difference to a comparative value.
Additionally returns a corresponding value. This function is for example
used in :py:func:`~.modelchain.v_wind_hub` of the
:class:`~.modelchain.ModelChain` to choose the wind speed data that is
close to the hub height of the examined wind turbine. In this case
`value_1` and `value_2` are the heights of the corresponding wind speed
data sets `corresp_1` and `corresp_2`.
Parameters
----------
value_1 : float
First value of which the difference to `comp_value` will be
compared with the difference to `comp_value` of `value_2`.
value_2 : float
Second value for comparison.
comp_value : float
Comparative value.
corresp_1 : float
Corresponding value to `value_1`.
corresp_2 : float
Corresponding value to `value_2`.
Returns
-------
Tuple(float, float, string)
Value closer to comparing value as float, corresponding value as
float and a string for logging.debug.
"""
if (value_2 is not None and corresp_2 is not None):
if value_1 == comp_value:
closest_value = value_1
logging_string = '(at hub height).'
elif value_2 == comp_value:
closest_value = value_2
logging_string = '(2) (at hub height).'
elif abs(value_1 - comp_value) <= abs(value_2 - comp_value):
closest_value = value_1
logging_string = None
else:
closest_value = value_2
logging_string = None
else:
closest_value = value_1
if value_1 == comp_value:
logging_string = '(at hub height).'
# Select correponding value
if closest_value == value_1:
corresp_value = corresp_1
else:
corresp_value = corresp_2
# Store values in a named tuple
return_tuple = collections.namedtuple('selected_values',
['closest_value',
'corresp_value', 'logging_string'])
return return_tuple(closest_value, corresp_value, logging_string)
| """
The ``tools`` module contains a collection of functions used in the
windpowerlib.
"""
__copyright__ = "Copyright oemof developer group"
__license__ = "GPLv3"
Add function for selection of value closer to comparative value"""
The ``tools`` module contains a collection of functions used in the
windpowerlib.
"""
__copyright__ = "Copyright oemof developer group"
__license__ = "GPLv3"
import collections
def select_closer_value(value_1, value_2, comp_value, corresp_1, corresp_2):
r"""
Selects the value with the smaller difference to a comparative value.
Additionally returns a corresponding value. This function is for example
used in :py:func:`~.modelchain.v_wind_hub` of the
:class:`~.modelchain.ModelChain` to choose the wind speed data that is
close to the hub height of the examined wind turbine. In this case
`value_1` and `value_2` are the heights of the corresponding wind speed
data sets `corresp_1` and `corresp_2`.
Parameters
----------
value_1 : float
First value of which the difference to `comp_value` will be
compared with the difference to `comp_value` of `value_2`.
value_2 : float
Second value for comparison.
comp_value : float
Comparative value.
corresp_1 : float
Corresponding value to `value_1`.
corresp_2 : float
Corresponding value to `value_2`.
Returns
-------
Tuple(float, float, string)
Value closer to comparing value as float, corresponding value as
float and a string for logging.debug.
"""
if (value_2 is not None and corresp_2 is not None):
if value_1 == comp_value:
closest_value = value_1
logging_string = '(at hub height).'
elif value_2 == comp_value:
closest_value = value_2
logging_string = '(2) (at hub height).'
elif abs(value_1 - comp_value) <= abs(value_2 - comp_value):
closest_value = value_1
logging_string = None
else:
closest_value = value_2
logging_string = None
else:
closest_value = value_1
if value_1 == comp_value:
logging_string = '(at hub height).'
# Select correponding value
if closest_value == value_1:
corresp_value = corresp_1
else:
corresp_value = corresp_2
# Store values in a named tuple
return_tuple = collections.namedtuple('selected_values',
['closest_value',
'corresp_value', 'logging_string'])
return return_tuple(closest_value, corresp_value, logging_string)
| <commit_before>"""
The ``tools`` module contains a collection of functions used in the
windpowerlib.
"""
__copyright__ = "Copyright oemof developer group"
__license__ = "GPLv3"
<commit_msg>Add function for selection of value closer to comparative value<commit_after>"""
The ``tools`` module contains a collection of functions used in the
windpowerlib.
"""
__copyright__ = "Copyright oemof developer group"
__license__ = "GPLv3"
import collections
def select_closer_value(value_1, value_2, comp_value, corresp_1, corresp_2):
r"""
Selects the value with the smaller difference to a comparative value.
Additionally returns a corresponding value. This function is for example
used in :py:func:`~.modelchain.v_wind_hub` of the
:class:`~.modelchain.ModelChain` to choose the wind speed data that is
close to the hub height of the examined wind turbine. In this case
`value_1` and `value_2` are the heights of the corresponding wind speed
data sets `corresp_1` and `corresp_2`.
Parameters
----------
value_1 : float
First value of which the difference to `comp_value` will be
compared with the difference to `comp_value` of `value_2`.
value_2 : float
Second value for comparison.
comp_value : float
Comparative value.
corresp_1 : float
Corresponding value to `value_1`.
corresp_2 : float
Corresponding value to `value_2`.
Returns
-------
Tuple(float, float, string)
Value closer to comparing value as float, corresponding value as
float and a string for logging.debug.
"""
if (value_2 is not None and corresp_2 is not None):
if value_1 == comp_value:
closest_value = value_1
logging_string = '(at hub height).'
elif value_2 == comp_value:
closest_value = value_2
logging_string = '(2) (at hub height).'
elif abs(value_1 - comp_value) <= abs(value_2 - comp_value):
closest_value = value_1
logging_string = None
else:
closest_value = value_2
logging_string = None
else:
closest_value = value_1
if value_1 == comp_value:
logging_string = '(at hub height).'
# Select correponding value
if closest_value == value_1:
corresp_value = corresp_1
else:
corresp_value = corresp_2
# Store values in a named tuple
return_tuple = collections.namedtuple('selected_values',
['closest_value',
'corresp_value', 'logging_string'])
return return_tuple(closest_value, corresp_value, logging_string)
|
9f6d6509b1f3f4a5f3fd20919bcc465475fc1ce3 | app/composer.py | app/composer.py | import os
from app.configuration import get_value
from app.helper import php
def initialization():
checker_dir = get_value('checker-dir')
if not os.path.isfile(checker_dir+'bin/composer'):
download(checker_dir)
if not os.path.isfile(checker_dir+'bin/phpcs'):
php('bin/composer install')
def download(checker_dir):
php_bin = get_value('php')
if not os.path.exists(checker_dir+'bin'):
os.makedirs(checker_dir+'bin')
print('>>> Download composer')
os.system('curl -sS https://getcomposer.org/installer | '+php_bin+' -- --install-dir='+checker_dir+'bin --filename=composer')
def update():
php('bin/composer self-update')
php('bin/composer update')
def project_installation():
code = php('bin/composer install --optimize-autoloader')
if code != 0:
raise SystemExit('The composer install command for the project failed with the code '+str(code))
| import os
from app.configuration import get_value
from app.helper import php
def initialization():
checker_dir = get_value('checker-dir')
if not os.path.isfile(checker_dir+'bin/composer'):
download(checker_dir)
if not os.path.isfile(checker_dir+'bin/phpcs'):
composer('install')
def download(checker_dir):
php_bin = get_value('php')
if not os.path.exists(checker_dir+'bin'):
os.makedirs(checker_dir+'bin')
print('>>> Download composer')
os.system('curl -sS https://getcomposer.org/installer | '+php_bin+' -- --install-dir='+checker_dir+'bin --filename=composer')
def update():
composer('self-update')
composer('update')
def project_installation():
code = php('bin/composer install --optimize-autoloader')
if code != 0:
raise SystemExit('The composer install command for the project failed with the code '+str(code))
def composer(command):
base_dir = os.getcwd()
os.chdir(get_value('checker-dir'))
php('bin/composer '+command)
os.chdir(base_dir)
| Add execution path for internal update | Add execution path for internal update
| Python | mit | mi-schi/php-code-checker | import os
from app.configuration import get_value
from app.helper import php
def initialization():
checker_dir = get_value('checker-dir')
if not os.path.isfile(checker_dir+'bin/composer'):
download(checker_dir)
if not os.path.isfile(checker_dir+'bin/phpcs'):
php('bin/composer install')
def download(checker_dir):
php_bin = get_value('php')
if not os.path.exists(checker_dir+'bin'):
os.makedirs(checker_dir+'bin')
print('>>> Download composer')
os.system('curl -sS https://getcomposer.org/installer | '+php_bin+' -- --install-dir='+checker_dir+'bin --filename=composer')
def update():
php('bin/composer self-update')
php('bin/composer update')
def project_installation():
code = php('bin/composer install --optimize-autoloader')
if code != 0:
raise SystemExit('The composer install command for the project failed with the code '+str(code))
Add execution path for internal update | import os
from app.configuration import get_value
from app.helper import php
def initialization():
checker_dir = get_value('checker-dir')
if not os.path.isfile(checker_dir+'bin/composer'):
download(checker_dir)
if not os.path.isfile(checker_dir+'bin/phpcs'):
composer('install')
def download(checker_dir):
php_bin = get_value('php')
if not os.path.exists(checker_dir+'bin'):
os.makedirs(checker_dir+'bin')
print('>>> Download composer')
os.system('curl -sS https://getcomposer.org/installer | '+php_bin+' -- --install-dir='+checker_dir+'bin --filename=composer')
def update():
composer('self-update')
composer('update')
def project_installation():
code = php('bin/composer install --optimize-autoloader')
if code != 0:
raise SystemExit('The composer install command for the project failed with the code '+str(code))
def composer(command):
base_dir = os.getcwd()
os.chdir(get_value('checker-dir'))
php('bin/composer '+command)
os.chdir(base_dir)
| <commit_before>import os
from app.configuration import get_value
from app.helper import php
def initialization():
checker_dir = get_value('checker-dir')
if not os.path.isfile(checker_dir+'bin/composer'):
download(checker_dir)
if not os.path.isfile(checker_dir+'bin/phpcs'):
php('bin/composer install')
def download(checker_dir):
php_bin = get_value('php')
if not os.path.exists(checker_dir+'bin'):
os.makedirs(checker_dir+'bin')
print('>>> Download composer')
os.system('curl -sS https://getcomposer.org/installer | '+php_bin+' -- --install-dir='+checker_dir+'bin --filename=composer')
def update():
php('bin/composer self-update')
php('bin/composer update')
def project_installation():
code = php('bin/composer install --optimize-autoloader')
if code != 0:
raise SystemExit('The composer install command for the project failed with the code '+str(code))
<commit_msg>Add execution path for internal update<commit_after> | import os
from app.configuration import get_value
from app.helper import php
def initialization():
checker_dir = get_value('checker-dir')
if not os.path.isfile(checker_dir+'bin/composer'):
download(checker_dir)
if not os.path.isfile(checker_dir+'bin/phpcs'):
composer('install')
def download(checker_dir):
php_bin = get_value('php')
if not os.path.exists(checker_dir+'bin'):
os.makedirs(checker_dir+'bin')
print('>>> Download composer')
os.system('curl -sS https://getcomposer.org/installer | '+php_bin+' -- --install-dir='+checker_dir+'bin --filename=composer')
def update():
composer('self-update')
composer('update')
def project_installation():
code = php('bin/composer install --optimize-autoloader')
if code != 0:
raise SystemExit('The composer install command for the project failed with the code '+str(code))
def composer(command):
base_dir = os.getcwd()
os.chdir(get_value('checker-dir'))
php('bin/composer '+command)
os.chdir(base_dir)
| import os
from app.configuration import get_value
from app.helper import php
def initialization():
checker_dir = get_value('checker-dir')
if not os.path.isfile(checker_dir+'bin/composer'):
download(checker_dir)
if not os.path.isfile(checker_dir+'bin/phpcs'):
php('bin/composer install')
def download(checker_dir):
php_bin = get_value('php')
if not os.path.exists(checker_dir+'bin'):
os.makedirs(checker_dir+'bin')
print('>>> Download composer')
os.system('curl -sS https://getcomposer.org/installer | '+php_bin+' -- --install-dir='+checker_dir+'bin --filename=composer')
def update():
php('bin/composer self-update')
php('bin/composer update')
def project_installation():
code = php('bin/composer install --optimize-autoloader')
if code != 0:
raise SystemExit('The composer install command for the project failed with the code '+str(code))
Add execution path for internal updateimport os
from app.configuration import get_value
from app.helper import php
def initialization():
checker_dir = get_value('checker-dir')
if not os.path.isfile(checker_dir+'bin/composer'):
download(checker_dir)
if not os.path.isfile(checker_dir+'bin/phpcs'):
composer('install')
def download(checker_dir):
php_bin = get_value('php')
if not os.path.exists(checker_dir+'bin'):
os.makedirs(checker_dir+'bin')
print('>>> Download composer')
os.system('curl -sS https://getcomposer.org/installer | '+php_bin+' -- --install-dir='+checker_dir+'bin --filename=composer')
def update():
composer('self-update')
composer('update')
def project_installation():
code = php('bin/composer install --optimize-autoloader')
if code != 0:
raise SystemExit('The composer install command for the project failed with the code '+str(code))
def composer(command):
base_dir = os.getcwd()
os.chdir(get_value('checker-dir'))
php('bin/composer '+command)
os.chdir(base_dir)
| <commit_before>import os
from app.configuration import get_value
from app.helper import php
def initialization():
checker_dir = get_value('checker-dir')
if not os.path.isfile(checker_dir+'bin/composer'):
download(checker_dir)
if not os.path.isfile(checker_dir+'bin/phpcs'):
php('bin/composer install')
def download(checker_dir):
php_bin = get_value('php')
if not os.path.exists(checker_dir+'bin'):
os.makedirs(checker_dir+'bin')
print('>>> Download composer')
os.system('curl -sS https://getcomposer.org/installer | '+php_bin+' -- --install-dir='+checker_dir+'bin --filename=composer')
def update():
php('bin/composer self-update')
php('bin/composer update')
def project_installation():
code = php('bin/composer install --optimize-autoloader')
if code != 0:
raise SystemExit('The composer install command for the project failed with the code '+str(code))
<commit_msg>Add execution path for internal update<commit_after>import os
from app.configuration import get_value
from app.helper import php
def initialization():
checker_dir = get_value('checker-dir')
if not os.path.isfile(checker_dir+'bin/composer'):
download(checker_dir)
if not os.path.isfile(checker_dir+'bin/phpcs'):
composer('install')
def download(checker_dir):
php_bin = get_value('php')
if not os.path.exists(checker_dir+'bin'):
os.makedirs(checker_dir+'bin')
print('>>> Download composer')
os.system('curl -sS https://getcomposer.org/installer | '+php_bin+' -- --install-dir='+checker_dir+'bin --filename=composer')
def update():
composer('self-update')
composer('update')
def project_installation():
code = php('bin/composer install --optimize-autoloader')
if code != 0:
raise SystemExit('The composer install command for the project failed with the code '+str(code))
def composer(command):
base_dir = os.getcwd()
os.chdir(get_value('checker-dir'))
php('bin/composer '+command)
os.chdir(base_dir)
|
618245ab759cbf47fb53946b4c6149efdca7e1e0 | troposphere/sqs.py | troposphere/sqs.py | # Copyright (c) 2012-2013, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
from .validators import integer
try:
from awacs.aws import Policy
policytypes = (dict, Policy)
except ImportError:
policytypes = dict,
class Queue(AWSObject):
type = "AWS::SQS::Queue"
props = {
'DelaySeconds': (integer, False),
'MaximumMessageSize': (integer, False),
'MessageRetentionPeriod': (integer, False),
'QueueName': (basestring, False),
'ReceiveMessageWaitTimeSeconds': (integer, False),
'VisibilityTimeout': (integer, False),
}
class QueuePolicy(AWSObject):
type = "AWS::SQS::QueuePolicy"
props = {
'PolicyDocument': (policytypes, False),
'Queues': (list, True),
}
| # Copyright (c) 2012-2013, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject, AWSProperty
from .validators import integer
try:
from awacs.aws import Policy
policytypes = (dict, Policy)
except ImportError:
policytypes = dict,
class RedrivePolicy(AWSProperty):
props = {
'deadLetterTargetArn': (basestring, False),
'maxReceiveCount': (integer, False),
}
class Queue(AWSObject):
type = "AWS::SQS::Queue"
props = {
'DelaySeconds': (integer, False),
'MaximumMessageSize': (integer, False),
'MessageRetentionPeriod': (integer, False),
'QueueName': (basestring, False),
'ReceiveMessageWaitTimeSeconds': (integer, False),
'RedrivePolicy': (RedrivePolicy, False),
'VisibilityTimeout': (integer, False),
}
class QueuePolicy(AWSObject):
type = "AWS::SQS::QueuePolicy"
props = {
'PolicyDocument': (policytypes, False),
'Queues': (list, True),
}
| Add SQS dead letter queue from CloudFormation release 2014-01-29 | Add SQS dead letter queue from CloudFormation release 2014-01-29
| Python | bsd-2-clause | cloudtools/troposphere,mhahn/troposphere,johnctitus/troposphere,ikben/troposphere,ikben/troposphere,garnaat/troposphere,pas256/troposphere,alonsodomin/troposphere,Yipit/troposphere,ptoraskar/troposphere,mannytoledo/troposphere,dmm92/troposphere,craigbruce/troposphere,jantman/troposphere,micahhausler/troposphere,jdc0589/troposphere,wangqiang8511/troposphere,cryptickp/troposphere,cloudtools/troposphere,unravelin/troposphere,7digital/troposphere,amosshapira/troposphere,LouTheBrew/troposphere,inetCatapult/troposphere,johnctitus/troposphere,pas256/troposphere,DualSpark/troposphere,horacio3/troposphere,horacio3/troposphere,Hons/troposphere,ccortezb/troposphere,alonsodomin/troposphere,yxd-hde/troposphere,xxxVxxx/troposphere,WeAreCloudar/troposphere,iblazevic/troposphere,dmm92/troposphere,kid/troposphere,7digital/troposphere,samcrang/troposphere,nicolaka/troposphere | # Copyright (c) 2012-2013, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
from .validators import integer
try:
from awacs.aws import Policy
policytypes = (dict, Policy)
except ImportError:
policytypes = dict,
class Queue(AWSObject):
type = "AWS::SQS::Queue"
props = {
'DelaySeconds': (integer, False),
'MaximumMessageSize': (integer, False),
'MessageRetentionPeriod': (integer, False),
'QueueName': (basestring, False),
'ReceiveMessageWaitTimeSeconds': (integer, False),
'VisibilityTimeout': (integer, False),
}
class QueuePolicy(AWSObject):
type = "AWS::SQS::QueuePolicy"
props = {
'PolicyDocument': (policytypes, False),
'Queues': (list, True),
}
Add SQS dead letter queue from CloudFormation release 2014-01-29 | # Copyright (c) 2012-2013, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject, AWSProperty
from .validators import integer
try:
from awacs.aws import Policy
policytypes = (dict, Policy)
except ImportError:
policytypes = dict,
class RedrivePolicy(AWSProperty):
props = {
'deadLetterTargetArn': (basestring, False),
'maxReceiveCount': (integer, False),
}
class Queue(AWSObject):
type = "AWS::SQS::Queue"
props = {
'DelaySeconds': (integer, False),
'MaximumMessageSize': (integer, False),
'MessageRetentionPeriod': (integer, False),
'QueueName': (basestring, False),
'ReceiveMessageWaitTimeSeconds': (integer, False),
'RedrivePolicy': (RedrivePolicy, False),
'VisibilityTimeout': (integer, False),
}
class QueuePolicy(AWSObject):
type = "AWS::SQS::QueuePolicy"
props = {
'PolicyDocument': (policytypes, False),
'Queues': (list, True),
}
| <commit_before># Copyright (c) 2012-2013, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
from .validators import integer
try:
from awacs.aws import Policy
policytypes = (dict, Policy)
except ImportError:
policytypes = dict,
class Queue(AWSObject):
type = "AWS::SQS::Queue"
props = {
'DelaySeconds': (integer, False),
'MaximumMessageSize': (integer, False),
'MessageRetentionPeriod': (integer, False),
'QueueName': (basestring, False),
'ReceiveMessageWaitTimeSeconds': (integer, False),
'VisibilityTimeout': (integer, False),
}
class QueuePolicy(AWSObject):
type = "AWS::SQS::QueuePolicy"
props = {
'PolicyDocument': (policytypes, False),
'Queues': (list, True),
}
<commit_msg>Add SQS dead letter queue from CloudFormation release 2014-01-29<commit_after> | # Copyright (c) 2012-2013, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject, AWSProperty
from .validators import integer
try:
from awacs.aws import Policy
policytypes = (dict, Policy)
except ImportError:
policytypes = dict,
class RedrivePolicy(AWSProperty):
props = {
'deadLetterTargetArn': (basestring, False),
'maxReceiveCount': (integer, False),
}
class Queue(AWSObject):
type = "AWS::SQS::Queue"
props = {
'DelaySeconds': (integer, False),
'MaximumMessageSize': (integer, False),
'MessageRetentionPeriod': (integer, False),
'QueueName': (basestring, False),
'ReceiveMessageWaitTimeSeconds': (integer, False),
'RedrivePolicy': (RedrivePolicy, False),
'VisibilityTimeout': (integer, False),
}
class QueuePolicy(AWSObject):
type = "AWS::SQS::QueuePolicy"
props = {
'PolicyDocument': (policytypes, False),
'Queues': (list, True),
}
| # Copyright (c) 2012-2013, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
from .validators import integer
try:
from awacs.aws import Policy
policytypes = (dict, Policy)
except ImportError:
policytypes = dict,
class Queue(AWSObject):
type = "AWS::SQS::Queue"
props = {
'DelaySeconds': (integer, False),
'MaximumMessageSize': (integer, False),
'MessageRetentionPeriod': (integer, False),
'QueueName': (basestring, False),
'ReceiveMessageWaitTimeSeconds': (integer, False),
'VisibilityTimeout': (integer, False),
}
class QueuePolicy(AWSObject):
type = "AWS::SQS::QueuePolicy"
props = {
'PolicyDocument': (policytypes, False),
'Queues': (list, True),
}
Add SQS dead letter queue from CloudFormation release 2014-01-29# Copyright (c) 2012-2013, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject, AWSProperty
from .validators import integer
try:
from awacs.aws import Policy
policytypes = (dict, Policy)
except ImportError:
policytypes = dict,
class RedrivePolicy(AWSProperty):
props = {
'deadLetterTargetArn': (basestring, False),
'maxReceiveCount': (integer, False),
}
class Queue(AWSObject):
type = "AWS::SQS::Queue"
props = {
'DelaySeconds': (integer, False),
'MaximumMessageSize': (integer, False),
'MessageRetentionPeriod': (integer, False),
'QueueName': (basestring, False),
'ReceiveMessageWaitTimeSeconds': (integer, False),
'RedrivePolicy': (RedrivePolicy, False),
'VisibilityTimeout': (integer, False),
}
class QueuePolicy(AWSObject):
type = "AWS::SQS::QueuePolicy"
props = {
'PolicyDocument': (policytypes, False),
'Queues': (list, True),
}
| <commit_before># Copyright (c) 2012-2013, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
from .validators import integer
try:
from awacs.aws import Policy
policytypes = (dict, Policy)
except ImportError:
policytypes = dict,
class Queue(AWSObject):
type = "AWS::SQS::Queue"
props = {
'DelaySeconds': (integer, False),
'MaximumMessageSize': (integer, False),
'MessageRetentionPeriod': (integer, False),
'QueueName': (basestring, False),
'ReceiveMessageWaitTimeSeconds': (integer, False),
'VisibilityTimeout': (integer, False),
}
class QueuePolicy(AWSObject):
type = "AWS::SQS::QueuePolicy"
props = {
'PolicyDocument': (policytypes, False),
'Queues': (list, True),
}
<commit_msg>Add SQS dead letter queue from CloudFormation release 2014-01-29<commit_after># Copyright (c) 2012-2013, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject, AWSProperty
from .validators import integer
try:
from awacs.aws import Policy
policytypes = (dict, Policy)
except ImportError:
policytypes = dict,
class RedrivePolicy(AWSProperty):
props = {
'deadLetterTargetArn': (basestring, False),
'maxReceiveCount': (integer, False),
}
class Queue(AWSObject):
type = "AWS::SQS::Queue"
props = {
'DelaySeconds': (integer, False),
'MaximumMessageSize': (integer, False),
'MessageRetentionPeriod': (integer, False),
'QueueName': (basestring, False),
'ReceiveMessageWaitTimeSeconds': (integer, False),
'RedrivePolicy': (RedrivePolicy, False),
'VisibilityTimeout': (integer, False),
}
class QueuePolicy(AWSObject):
type = "AWS::SQS::QueuePolicy"
props = {
'PolicyDocument': (policytypes, False),
'Queues': (list, True),
}
|
1026581107668e15db91912302ae3fd577140008 | builder.py | builder.py | import ratebeer
import string
def strip_brewery_name(brewery_name, beer_name):
brewery_word_list = brewery_name.split()
for word in brewery_word_list:
beer_name = beer_name.replace(word, "")
return beer_name.strip()
categories = []
categories.append("0-9")
for letter in string.ascii_uppercase:
categories.append(letter)
rb = ratebeer.RateBeer()
with open("eng.user_words",'w') as f:
for category in categories:
brewery_list = rb.brewers_by_alpha(category)
for brewery in brewery_list:
beer_list = brewery.get_beers()
for beer in beer_list:
#index the beer name without the bewery too
beer_name_without_brewery = strip_brewery_name(brewery.name, beer.name)
f.writelines(beer_name_without_brewery.encode('utf8') + "\n")
| import ratebeer
import string
import csv
from io import BytesIO
def strip_brewery_name(brewery_name, beer_name):
brewery_word_list = brewery_name.split()
for word in brewery_word_list:
beer_name = beer_name.replace(word, "")
return beer_name.strip()
def brewery_name_field(brewery):
val = getattr(brewery, 'name', 'RateBeer does not have this field filled out for this brewery')
return val.encode('utf8')
def beer_description_field(beer):
val = getattr(beer, 'description', 'no description is available')
return val.encode('utf8')
categories = []
categories.append("0-9")
for letter in string.ascii_uppercase:
categories.append(letter)
rb = ratebeer.RateBeer()
mapping = open('mapping.csv','w')
kv = csv.writer(mapping, quoting=csv.QUOTE_NONNUMERIC)
kv.writerow( ('beer name','url','description','full name') )
with open("eng.user_words",'w') as f:
for category in categories:
brewery_list = rb.brewers_by_alpha(categories)
for brewery in brewery_list:
beer_list = brewery.get_beers()
for beer in beer_list:
#index the beer name without the bewery too
beer_name_without_brewery = strip_brewery_name(brewery.name, beer.name)
f.writelines(beer_name_without_brewery.encode('utf8') + "\n")
kv.writerow( (beer_name_without_brewery.encode('utf8'),beer.url,beer_description_field(beer),brewery_name_field(brewery)) )
mapping.close()
| Add a csv export for caching the list of beers for display in the app | Add a csv export for caching the list of beers for display in the app
| Python | mit | jwrubel/initial_dictionary | import ratebeer
import string
def strip_brewery_name(brewery_name, beer_name):
brewery_word_list = brewery_name.split()
for word in brewery_word_list:
beer_name = beer_name.replace(word, "")
return beer_name.strip()
categories = []
categories.append("0-9")
for letter in string.ascii_uppercase:
categories.append(letter)
rb = ratebeer.RateBeer()
with open("eng.user_words",'w') as f:
for category in categories:
brewery_list = rb.brewers_by_alpha(category)
for brewery in brewery_list:
beer_list = brewery.get_beers()
for beer in beer_list:
#index the beer name without the bewery too
beer_name_without_brewery = strip_brewery_name(brewery.name, beer.name)
f.writelines(beer_name_without_brewery.encode('utf8') + "\n")
Add a csv export for caching the list of beers for display in the app | import ratebeer
import string
import csv
from io import BytesIO
def strip_brewery_name(brewery_name, beer_name):
brewery_word_list = brewery_name.split()
for word in brewery_word_list:
beer_name = beer_name.replace(word, "")
return beer_name.strip()
def brewery_name_field(brewery):
val = getattr(brewery, 'name', 'RateBeer does not have this field filled out for this brewery')
return val.encode('utf8')
def beer_description_field(beer):
val = getattr(beer, 'description', 'no description is available')
return val.encode('utf8')
categories = []
categories.append("0-9")
for letter in string.ascii_uppercase:
categories.append(letter)
rb = ratebeer.RateBeer()
mapping = open('mapping.csv','w')
kv = csv.writer(mapping, quoting=csv.QUOTE_NONNUMERIC)
kv.writerow( ('beer name','url','description','full name') )
with open("eng.user_words",'w') as f:
for category in categories:
brewery_list = rb.brewers_by_alpha(categories)
for brewery in brewery_list:
beer_list = brewery.get_beers()
for beer in beer_list:
#index the beer name without the bewery too
beer_name_without_brewery = strip_brewery_name(brewery.name, beer.name)
f.writelines(beer_name_without_brewery.encode('utf8') + "\n")
kv.writerow( (beer_name_without_brewery.encode('utf8'),beer.url,beer_description_field(beer),brewery_name_field(brewery)) )
mapping.close()
| <commit_before>import ratebeer
import string
def strip_brewery_name(brewery_name, beer_name):
brewery_word_list = brewery_name.split()
for word in brewery_word_list:
beer_name = beer_name.replace(word, "")
return beer_name.strip()
categories = []
categories.append("0-9")
for letter in string.ascii_uppercase:
categories.append(letter)
rb = ratebeer.RateBeer()
with open("eng.user_words",'w') as f:
for category in categories:
brewery_list = rb.brewers_by_alpha(category)
for brewery in brewery_list:
beer_list = brewery.get_beers()
for beer in beer_list:
#index the beer name without the bewery too
beer_name_without_brewery = strip_brewery_name(brewery.name, beer.name)
f.writelines(beer_name_without_brewery.encode('utf8') + "\n")
<commit_msg>Add a csv export for caching the list of beers for display in the app<commit_after> | import ratebeer
import string
import csv
from io import BytesIO
def strip_brewery_name(brewery_name, beer_name):
brewery_word_list = brewery_name.split()
for word in brewery_word_list:
beer_name = beer_name.replace(word, "")
return beer_name.strip()
def brewery_name_field(brewery):
val = getattr(brewery, 'name', 'RateBeer does not have this field filled out for this brewery')
return val.encode('utf8')
def beer_description_field(beer):
val = getattr(beer, 'description', 'no description is available')
return val.encode('utf8')
categories = []
categories.append("0-9")
for letter in string.ascii_uppercase:
categories.append(letter)
rb = ratebeer.RateBeer()
mapping = open('mapping.csv','w')
kv = csv.writer(mapping, quoting=csv.QUOTE_NONNUMERIC)
kv.writerow( ('beer name','url','description','full name') )
with open("eng.user_words",'w') as f:
for category in categories:
brewery_list = rb.brewers_by_alpha(categories)
for brewery in brewery_list:
beer_list = brewery.get_beers()
for beer in beer_list:
#index the beer name without the bewery too
beer_name_without_brewery = strip_brewery_name(brewery.name, beer.name)
f.writelines(beer_name_without_brewery.encode('utf8') + "\n")
kv.writerow( (beer_name_without_brewery.encode('utf8'),beer.url,beer_description_field(beer),brewery_name_field(brewery)) )
mapping.close()
| import ratebeer
import string
def strip_brewery_name(brewery_name, beer_name):
brewery_word_list = brewery_name.split()
for word in brewery_word_list:
beer_name = beer_name.replace(word, "")
return beer_name.strip()
categories = []
categories.append("0-9")
for letter in string.ascii_uppercase:
categories.append(letter)
rb = ratebeer.RateBeer()
with open("eng.user_words",'w') as f:
for category in categories:
brewery_list = rb.brewers_by_alpha(category)
for brewery in brewery_list:
beer_list = brewery.get_beers()
for beer in beer_list:
#index the beer name without the bewery too
beer_name_without_brewery = strip_brewery_name(brewery.name, beer.name)
f.writelines(beer_name_without_brewery.encode('utf8') + "\n")
Add a csv export for caching the list of beers for display in the appimport ratebeer
import string
import csv
from io import BytesIO
def strip_brewery_name(brewery_name, beer_name):
brewery_word_list = brewery_name.split()
for word in brewery_word_list:
beer_name = beer_name.replace(word, "")
return beer_name.strip()
def brewery_name_field(brewery):
val = getattr(brewery, 'name', 'RateBeer does not have this field filled out for this brewery')
return val.encode('utf8')
def beer_description_field(beer):
val = getattr(beer, 'description', 'no description is available')
return val.encode('utf8')
categories = []
categories.append("0-9")
for letter in string.ascii_uppercase:
categories.append(letter)
rb = ratebeer.RateBeer()
mapping = open('mapping.csv','w')
kv = csv.writer(mapping, quoting=csv.QUOTE_NONNUMERIC)
kv.writerow( ('beer name','url','description','full name') )
with open("eng.user_words",'w') as f:
for category in categories:
brewery_list = rb.brewers_by_alpha(categories)
for brewery in brewery_list:
beer_list = brewery.get_beers()
for beer in beer_list:
#index the beer name without the bewery too
beer_name_without_brewery = strip_brewery_name(brewery.name, beer.name)
f.writelines(beer_name_without_brewery.encode('utf8') + "\n")
kv.writerow( (beer_name_without_brewery.encode('utf8'),beer.url,beer_description_field(beer),brewery_name_field(brewery)) )
mapping.close()
| <commit_before>import ratebeer
import string
def strip_brewery_name(brewery_name, beer_name):
brewery_word_list = brewery_name.split()
for word in brewery_word_list:
beer_name = beer_name.replace(word, "")
return beer_name.strip()
categories = []
categories.append("0-9")
for letter in string.ascii_uppercase:
categories.append(letter)
rb = ratebeer.RateBeer()
with open("eng.user_words",'w') as f:
for category in categories:
brewery_list = rb.brewers_by_alpha(category)
for brewery in brewery_list:
beer_list = brewery.get_beers()
for beer in beer_list:
#index the beer name without the bewery too
beer_name_without_brewery = strip_brewery_name(brewery.name, beer.name)
f.writelines(beer_name_without_brewery.encode('utf8') + "\n")
<commit_msg>Add a csv export for caching the list of beers for display in the app<commit_after>import ratebeer
import string
import csv
from io import BytesIO
def strip_brewery_name(brewery_name, beer_name):
brewery_word_list = brewery_name.split()
for word in brewery_word_list:
beer_name = beer_name.replace(word, "")
return beer_name.strip()
def brewery_name_field(brewery):
val = getattr(brewery, 'name', 'RateBeer does not have this field filled out for this brewery')
return val.encode('utf8')
def beer_description_field(beer):
val = getattr(beer, 'description', 'no description is available')
return val.encode('utf8')
categories = []
categories.append("0-9")
for letter in string.ascii_uppercase:
categories.append(letter)
rb = ratebeer.RateBeer()
mapping = open('mapping.csv','w')
kv = csv.writer(mapping, quoting=csv.QUOTE_NONNUMERIC)
kv.writerow( ('beer name','url','description','full name') )
with open("eng.user_words",'w') as f:
for category in categories:
brewery_list = rb.brewers_by_alpha(categories)
for brewery in brewery_list:
beer_list = brewery.get_beers()
for beer in beer_list:
#index the beer name without the bewery too
beer_name_without_brewery = strip_brewery_name(brewery.name, beer.name)
f.writelines(beer_name_without_brewery.encode('utf8') + "\n")
kv.writerow( (beer_name_without_brewery.encode('utf8'),beer.url,beer_description_field(beer),brewery_name_field(brewery)) )
mapping.close()
|
7e36568d5b8aeaf2c77e4643a793fdc13cb9ba51 | spacy/about.py | spacy/about.py | # inspired from:
# https://python-packaging-user-guide.readthedocs.org/en/latest/single_source_version/
# https://github.com/pypa/warehouse/blob/master/warehouse/__about__.py
__title__ = 'spaCy'
__version__ = '1.6.0'
__summary__ = 'Industrial-strength Natural Language Processing (NLP) with Python and Cython'
__uri__ = 'https://spacy.io'
__author__ = 'Matthew Honnibal'
__email__ = 'matt@explosion.ai'
__license__ = 'MIT'
__models__ = {
'en': 'en>=1.1.0,<1.2.0',
'de': 'de>=1.0.0,<1.1.0',
}
| # inspired from:
# https://python-packaging-user-guide.readthedocs.org/en/latest/single_source_version/
# https://github.com/pypa/warehouse/blob/master/warehouse/__about__.py
__title__ = 'spacy'
__version__ = '1.6.0'
__summary__ = 'Industrial-strength Natural Language Processing (NLP) with Python and Cython'
__uri__ = 'https://spacy.io'
__author__ = 'Matthew Honnibal'
__email__ = 'matt@explosion.ai'
__license__ = 'MIT'
__models__ = {
'en': 'en>=1.1.0,<1.2.0',
'de': 'de>=1.0.0,<1.1.0',
}
| Fix title to accommodate sputnik | Fix title to accommodate sputnik | Python | mit | Gregory-Howard/spaCy,spacy-io/spaCy,honnibal/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,raphael0202/spaCy,aikramer2/spaCy,banglakit/spaCy,Gregory-Howard/spaCy,spacy-io/spaCy,aikramer2/spaCy,aikramer2/spaCy,explosion/spaCy,explosion/spaCy,honnibal/spaCy,spacy-io/spaCy,explosion/spaCy,Gregory-Howard/spaCy,raphael0202/spaCy,recognai/spaCy,raphael0202/spaCy,spacy-io/spaCy,Gregory-Howard/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,oroszgy/spaCy.hu,banglakit/spaCy,explosion/spaCy,recognai/spaCy,spacy-io/spaCy,banglakit/spaCy,raphael0202/spaCy,aikramer2/spaCy,aikramer2/spaCy,honnibal/spaCy,recognai/spaCy,banglakit/spaCy,banglakit/spaCy,spacy-io/spaCy,recognai/spaCy,recognai/spaCy,explosion/spaCy,oroszgy/spaCy.hu,oroszgy/spaCy.hu,recognai/spaCy,oroszgy/spaCy.hu,honnibal/spaCy,oroszgy/spaCy.hu,banglakit/spaCy,Gregory-Howard/spaCy,explosion/spaCy,raphael0202/spaCy | # inspired from:
# https://python-packaging-user-guide.readthedocs.org/en/latest/single_source_version/
# https://github.com/pypa/warehouse/blob/master/warehouse/__about__.py
__title__ = 'spaCy'
__version__ = '1.6.0'
__summary__ = 'Industrial-strength Natural Language Processing (NLP) with Python and Cython'
__uri__ = 'https://spacy.io'
__author__ = 'Matthew Honnibal'
__email__ = 'matt@explosion.ai'
__license__ = 'MIT'
__models__ = {
'en': 'en>=1.1.0,<1.2.0',
'de': 'de>=1.0.0,<1.1.0',
}
Fix title to accommodate sputnik | # inspired from:
# https://python-packaging-user-guide.readthedocs.org/en/latest/single_source_version/
# https://github.com/pypa/warehouse/blob/master/warehouse/__about__.py
__title__ = 'spacy'
__version__ = '1.6.0'
__summary__ = 'Industrial-strength Natural Language Processing (NLP) with Python and Cython'
__uri__ = 'https://spacy.io'
__author__ = 'Matthew Honnibal'
__email__ = 'matt@explosion.ai'
__license__ = 'MIT'
__models__ = {
'en': 'en>=1.1.0,<1.2.0',
'de': 'de>=1.0.0,<1.1.0',
}
| <commit_before># inspired from:
# https://python-packaging-user-guide.readthedocs.org/en/latest/single_source_version/
# https://github.com/pypa/warehouse/blob/master/warehouse/__about__.py
__title__ = 'spaCy'
__version__ = '1.6.0'
__summary__ = 'Industrial-strength Natural Language Processing (NLP) with Python and Cython'
__uri__ = 'https://spacy.io'
__author__ = 'Matthew Honnibal'
__email__ = 'matt@explosion.ai'
__license__ = 'MIT'
__models__ = {
'en': 'en>=1.1.0,<1.2.0',
'de': 'de>=1.0.0,<1.1.0',
}
<commit_msg>Fix title to accommodate sputnik<commit_after> | # inspired from:
# https://python-packaging-user-guide.readthedocs.org/en/latest/single_source_version/
# https://github.com/pypa/warehouse/blob/master/warehouse/__about__.py
__title__ = 'spacy'
__version__ = '1.6.0'
__summary__ = 'Industrial-strength Natural Language Processing (NLP) with Python and Cython'
__uri__ = 'https://spacy.io'
__author__ = 'Matthew Honnibal'
__email__ = 'matt@explosion.ai'
__license__ = 'MIT'
__models__ = {
'en': 'en>=1.1.0,<1.2.0',
'de': 'de>=1.0.0,<1.1.0',
}
| # inspired from:
# https://python-packaging-user-guide.readthedocs.org/en/latest/single_source_version/
# https://github.com/pypa/warehouse/blob/master/warehouse/__about__.py
__title__ = 'spaCy'
__version__ = '1.6.0'
__summary__ = 'Industrial-strength Natural Language Processing (NLP) with Python and Cython'
__uri__ = 'https://spacy.io'
__author__ = 'Matthew Honnibal'
__email__ = 'matt@explosion.ai'
__license__ = 'MIT'
__models__ = {
'en': 'en>=1.1.0,<1.2.0',
'de': 'de>=1.0.0,<1.1.0',
}
Fix title to accommodate sputnik# inspired from:
# https://python-packaging-user-guide.readthedocs.org/en/latest/single_source_version/
# https://github.com/pypa/warehouse/blob/master/warehouse/__about__.py
__title__ = 'spacy'
__version__ = '1.6.0'
__summary__ = 'Industrial-strength Natural Language Processing (NLP) with Python and Cython'
__uri__ = 'https://spacy.io'
__author__ = 'Matthew Honnibal'
__email__ = 'matt@explosion.ai'
__license__ = 'MIT'
__models__ = {
'en': 'en>=1.1.0,<1.2.0',
'de': 'de>=1.0.0,<1.1.0',
}
| <commit_before># inspired from:
# https://python-packaging-user-guide.readthedocs.org/en/latest/single_source_version/
# https://github.com/pypa/warehouse/blob/master/warehouse/__about__.py
__title__ = 'spaCy'
__version__ = '1.6.0'
__summary__ = 'Industrial-strength Natural Language Processing (NLP) with Python and Cython'
__uri__ = 'https://spacy.io'
__author__ = 'Matthew Honnibal'
__email__ = 'matt@explosion.ai'
__license__ = 'MIT'
__models__ = {
'en': 'en>=1.1.0,<1.2.0',
'de': 'de>=1.0.0,<1.1.0',
}
<commit_msg>Fix title to accommodate sputnik<commit_after># inspired from:
# https://python-packaging-user-guide.readthedocs.org/en/latest/single_source_version/
# https://github.com/pypa/warehouse/blob/master/warehouse/__about__.py
__title__ = 'spacy'
__version__ = '1.6.0'
__summary__ = 'Industrial-strength Natural Language Processing (NLP) with Python and Cython'
__uri__ = 'https://spacy.io'
__author__ = 'Matthew Honnibal'
__email__ = 'matt@explosion.ai'
__license__ = 'MIT'
__models__ = {
'en': 'en>=1.1.0,<1.2.0',
'de': 'de>=1.0.0,<1.1.0',
}
|
054b0bf9cacef4e55fb8167fb5f2611e2ce39b43 | hw3/hw3_2a.py | hw3/hw3_2a.py | import sympy
x1, x2 = sympy.symbols('x1 x2')
f = 100*(x2 - x1**2)**2 + (1-x1)**2
df_dx1 = sympy.diff(f,x1)
df_dx2 = sympy.diff(f,x2)
H = sympy.hessian(f, (x1, x2))
xs = sympy.solve([df_dx1, df_dx2], [x1, x2])
H_xs = H.subs([(x1,xs[0][0]), (x2,xs[0][1])])
flag = True
for i in H_xs.eigenvals().keys():
if i.evalf() <= 0:
flag = False
break
if flag:
print 'Stationary point'
else:
print 'Saddle point'
| import sympy
x1, x2 = sympy.symbols('x1 x2')
f = 100*(x2 - x1**2)**2 + (1-x1)**2
df_dx1 = sympy.diff(f,x1)
df_dx2 = sympy.diff(f,x2)
H = sympy.hessian(f, (x1, x2))
xs = sympy.solve([df_dx1, df_dx2], [x1, x2])
H_xs = H.subs([(x1,xs[0][0]), (x2,xs[0][1])])
lambda_xs = H_xs.eigenvals()
count = 0
for i in lambda_xs.keys():
if i.evalf() <= 0:
count += 1
if count == 0:
print 'Local minima'
elif count == len(lambda_xs.keys()):
print 'Lacal maxima'
else:
print 'Saddle point'
| Fix decision about minima, maxima and saddle point | Fix decision about minima, maxima and saddle point
| Python | bsd-2-clause | escorciav/amcs211,escorciav/amcs211 | import sympy
x1, x2 = sympy.symbols('x1 x2')
f = 100*(x2 - x1**2)**2 + (1-x1)**2
df_dx1 = sympy.diff(f,x1)
df_dx2 = sympy.diff(f,x2)
H = sympy.hessian(f, (x1, x2))
xs = sympy.solve([df_dx1, df_dx2], [x1, x2])
H_xs = H.subs([(x1,xs[0][0]), (x2,xs[0][1])])
flag = True
for i in H_xs.eigenvals().keys():
if i.evalf() <= 0:
flag = False
break
if flag:
print 'Stationary point'
else:
print 'Saddle point'
Fix decision about minima, maxima and saddle point | import sympy
x1, x2 = sympy.symbols('x1 x2')
f = 100*(x2 - x1**2)**2 + (1-x1)**2
df_dx1 = sympy.diff(f,x1)
df_dx2 = sympy.diff(f,x2)
H = sympy.hessian(f, (x1, x2))
xs = sympy.solve([df_dx1, df_dx2], [x1, x2])
H_xs = H.subs([(x1,xs[0][0]), (x2,xs[0][1])])
lambda_xs = H_xs.eigenvals()
count = 0
for i in lambda_xs.keys():
if i.evalf() <= 0:
count += 1
if count == 0:
print 'Local minima'
elif count == len(lambda_xs.keys()):
print 'Lacal maxima'
else:
print 'Saddle point'
| <commit_before>import sympy
x1, x2 = sympy.symbols('x1 x2')
f = 100*(x2 - x1**2)**2 + (1-x1)**2
df_dx1 = sympy.diff(f,x1)
df_dx2 = sympy.diff(f,x2)
H = sympy.hessian(f, (x1, x2))
xs = sympy.solve([df_dx1, df_dx2], [x1, x2])
H_xs = H.subs([(x1,xs[0][0]), (x2,xs[0][1])])
flag = True
for i in H_xs.eigenvals().keys():
if i.evalf() <= 0:
flag = False
break
if flag:
print 'Stationary point'
else:
print 'Saddle point'
<commit_msg>Fix decision about minima, maxima and saddle point<commit_after> | import sympy
x1, x2 = sympy.symbols('x1 x2')
f = 100*(x2 - x1**2)**2 + (1-x1)**2
df_dx1 = sympy.diff(f,x1)
df_dx2 = sympy.diff(f,x2)
H = sympy.hessian(f, (x1, x2))
xs = sympy.solve([df_dx1, df_dx2], [x1, x2])
H_xs = H.subs([(x1,xs[0][0]), (x2,xs[0][1])])
lambda_xs = H_xs.eigenvals()
count = 0
for i in lambda_xs.keys():
if i.evalf() <= 0:
count += 1
if count == 0:
print 'Local minima'
elif count == len(lambda_xs.keys()):
print 'Lacal maxima'
else:
print 'Saddle point'
| import sympy
x1, x2 = sympy.symbols('x1 x2')
f = 100*(x2 - x1**2)**2 + (1-x1)**2
df_dx1 = sympy.diff(f,x1)
df_dx2 = sympy.diff(f,x2)
H = sympy.hessian(f, (x1, x2))
xs = sympy.solve([df_dx1, df_dx2], [x1, x2])
H_xs = H.subs([(x1,xs[0][0]), (x2,xs[0][1])])
flag = True
for i in H_xs.eigenvals().keys():
if i.evalf() <= 0:
flag = False
break
if flag:
print 'Stationary point'
else:
print 'Saddle point'
Fix decision about minima, maxima and saddle pointimport sympy
x1, x2 = sympy.symbols('x1 x2')
f = 100*(x2 - x1**2)**2 + (1-x1)**2
df_dx1 = sympy.diff(f,x1)
df_dx2 = sympy.diff(f,x2)
H = sympy.hessian(f, (x1, x2))
xs = sympy.solve([df_dx1, df_dx2], [x1, x2])
H_xs = H.subs([(x1,xs[0][0]), (x2,xs[0][1])])
lambda_xs = H_xs.eigenvals()
count = 0
for i in lambda_xs.keys():
if i.evalf() <= 0:
count += 1
if count == 0:
print 'Local minima'
elif count == len(lambda_xs.keys()):
print 'Lacal maxima'
else:
print 'Saddle point'
| <commit_before>import sympy
x1, x2 = sympy.symbols('x1 x2')
f = 100*(x2 - x1**2)**2 + (1-x1)**2
df_dx1 = sympy.diff(f,x1)
df_dx2 = sympy.diff(f,x2)
H = sympy.hessian(f, (x1, x2))
xs = sympy.solve([df_dx1, df_dx2], [x1, x2])
H_xs = H.subs([(x1,xs[0][0]), (x2,xs[0][1])])
flag = True
for i in H_xs.eigenvals().keys():
if i.evalf() <= 0:
flag = False
break
if flag:
print 'Stationary point'
else:
print 'Saddle point'
<commit_msg>Fix decision about minima, maxima and saddle point<commit_after>import sympy
x1, x2 = sympy.symbols('x1 x2')
f = 100*(x2 - x1**2)**2 + (1-x1)**2
df_dx1 = sympy.diff(f,x1)
df_dx2 = sympy.diff(f,x2)
H = sympy.hessian(f, (x1, x2))
xs = sympy.solve([df_dx1, df_dx2], [x1, x2])
H_xs = H.subs([(x1,xs[0][0]), (x2,xs[0][1])])
lambda_xs = H_xs.eigenvals()
count = 0
for i in lambda_xs.keys():
if i.evalf() <= 0:
count += 1
if count == 0:
print 'Local minima'
elif count == len(lambda_xs.keys()):
print 'Lacal maxima'
else:
print 'Saddle point'
|
210581cfef3d54b055ec9f9b1dc6d19b757a4d6e | cli/cli.py | cli/cli.py | import argparse
parser = argparse.ArgumentParser(prog='moocx', description='EdX MOOC Data Anaylysis')
parser.add_argument('-v', '--version', action='version', version='0.1.0')
parser.parse_args()
| Add cmd for getting version | Add cmd for getting version
| Python | mit | McGillX/edx_data_research,McGillX/edx_data_research,McGillX/edx_data_research | Add cmd for getting version | import argparse
parser = argparse.ArgumentParser(prog='moocx', description='EdX MOOC Data Anaylysis')
parser.add_argument('-v', '--version', action='version', version='0.1.0')
parser.parse_args()
| <commit_before><commit_msg>Add cmd for getting version<commit_after> | import argparse
parser = argparse.ArgumentParser(prog='moocx', description='EdX MOOC Data Anaylysis')
parser.add_argument('-v', '--version', action='version', version='0.1.0')
parser.parse_args()
| Add cmd for getting versionimport argparse
parser = argparse.ArgumentParser(prog='moocx', description='EdX MOOC Data Anaylysis')
parser.add_argument('-v', '--version', action='version', version='0.1.0')
parser.parse_args()
| <commit_before><commit_msg>Add cmd for getting version<commit_after>import argparse
parser = argparse.ArgumentParser(prog='moocx', description='EdX MOOC Data Anaylysis')
parser.add_argument('-v', '--version', action='version', version='0.1.0')
parser.parse_args()
| |
ddbcd88bb086d1978c9196833d126ded18db97f8 | airflow/migrations/versions/211e584da130_add_ti_state_index.py | airflow/migrations/versions/211e584da130_add_ti_state_index.py | """add TI state index
Revision ID: 211e584da130
Revises: 2e82aab8ef20
Create Date: 2016-06-30 10:54:24.323588
"""
# revision identifiers, used by Alembic.
revision = '211e584da130'
down_revision = '2e82aab8ef20'
branch_labels = None
depends_on = None
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index('ti_state', 'task_instance', ['state'], unique=False)
def downgrade():
op.drop_index('ti_state', table_name='task_instance')
| # -*- coding: utf-8 -*-
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""add TI state index
Revision ID: 211e584da130
Revises: 2e82aab8ef20
Create Date: 2016-06-30 10:54:24.323588
"""
# revision identifiers, used by Alembic.
revision = '211e584da130'
down_revision = '2e82aab8ef20'
branch_labels = None
depends_on = None
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index('ti_state', 'task_instance', ['state'], unique=False)
def downgrade():
op.drop_index('ti_state', table_name='task_instance')
| Add license to migration file | Add license to migration file
| Python | apache-2.0 | dhuang/incubator-airflow,Twistbioscience/incubator-airflow,kerzhner/airflow,vineet-rh/incubator-airflow,zodiac/incubator-airflow,cademarkegard/airflow,MetrodataTeam/incubator-airflow,sekikn/incubator-airflow,sergiohgz/incubator-airflow,wooga/airflow,RealImpactAnalytics/airflow,CloverHealth/airflow,apache/incubator-airflow,r39132/airflow,MortalViews/incubator-airflow,RealImpactAnalytics/airflow,sergiohgz/incubator-airflow,KL-WLCR/incubator-airflow,apache/incubator-airflow,skudriashev/incubator-airflow,ronfung/incubator-airflow,AllisonWang/incubator-airflow,cademarkegard/airflow,caseyching/incubator-airflow,alexvanboxel/airflow,DEVELByte/incubator-airflow,dgies/incubator-airflow,owlabs/incubator-airflow,bolkedebruin/airflow,andyxhadji/incubator-airflow,bolkedebruin/airflow,nathanielvarona/airflow,apache/airflow,preete-dixit-ck/incubator-airflow,lxneng/incubator-airflow,mrkm4ntr/incubator-airflow,vineet-rh/incubator-airflow,DinoCow/airflow,OpringaoDoTurno/airflow,plypaul/airflow,DEVELByte/incubator-airflow,malmiron/incubator-airflow,jesusfcr/airflow,dgies/incubator-airflow,sid88in/incubator-airflow,cfei18/incubator-airflow,spektom/incubator-airflow,gritlogic/incubator-airflow,Acehaidrey/incubator-airflow,vijaysbhat/incubator-airflow,easytaxibr/airflow,easytaxibr/airflow,edgarRd/incubator-airflow,cademarkegard/airflow,ronfung/incubator-airflow,adrpar/incubator-airflow,jlowin/airflow,alexvanboxel/airflow,jhsenjaliya/incubator-airflow,akosel/incubator-airflow,ty707/airflow,alexvanboxel/airflow,ty707/airflow,kerzhner/airflow,akosel/incubator-airflow,jesusfcr/airflow,jesusfcr/airflow,ProstoMaxim/incubator-airflow,asnir/airflow,fenglu-g/incubator-airflow,stverhae/incubator-airflow,easytaxibr/airflow,lyft/incubator-airflow,adamhaney/airflow,ledsusop/airflow,criccomini/airflow,mrkm4ntr/incubator-airflow,yk5/incubator-airflow,MortalViews/incubator-airflow,wooga/airflow,CloverHealth/airflow,Fokko/incubator-airflow,yk5/incubator-airflow,plypaul/airflow,OpringaoDoTurno/airflow,cjqian/incubator-airflow,Acehaidrey/incubator-airflow,sid88in/incubator-airflow,spektom/incubator-airflow,zack3241/incubator-airflow,alexvanboxel/airflow,artwr/airflow,preete-dixit-ck/incubator-airflow,janczak10/incubator-airflow,hamedhsn/incubator-airflow,preete-dixit-ck/incubator-airflow,zoyahav/incubator-airflow,mrares/incubator-airflow,wolfier/incubator-airflow,btallman/incubator-airflow,asnir/airflow,jgao54/airflow,wooga/airflow,criccomini/airflow,apache/incubator-airflow,aminghadersohi/airflow,mrkm4ntr/incubator-airflow,caseyching/incubator-airflow,yati-sagade/incubator-airflow,vijaysbhat/incubator-airflow,nathanielvarona/airflow,dmitry-r/incubator-airflow,andrewmchen/incubator-airflow,vineet-rh/incubator-airflow,ty707/airflow,artwr/airflow,jfantom/incubator-airflow,danielvdende/incubator-airflow,sdiazb/airflow,d-lee/airflow,cjqian/incubator-airflow,gtoonstra/airflow,cjqian/incubator-airflow,adamhaney/airflow,yati-sagade/incubator-airflow,rishibarve/incubator-airflow,jiwang576/incubator-airflow,kerzhner/airflow,NielsZeilemaker/incubator-airflow,subodhchhabra/airflow,apache/incubator-airflow,sekikn/incubator-airflow,DinoCow/airflow,ronfung/incubator-airflow,fenglu-g/incubator-airflow,artwr/airflow,Twistbioscience/incubator-airflow,airbnb/airflow,nathanielvarona/airflow,jiwang576/incubator-airflow,Fokko/incubator-airflow,bolkedebruin/airflow,MetrodataTeam/incubator-airflow,andrewmchen/incubator-airflow,subodhchhabra/airflow,zack3241/incubator-airflow,mistercrunch/airflow,wooga/airflow,ty707/airflow,sdiazb/airflow,vijaysbhat/incubator-airflow,cfei18/incubator-airflow,jbhsieh/incubator-airflow,DEVELByte/incubator-airflow,yiqingj/airflow,CloverHealth/airflow,Twistbioscience/incubator-airflow,mistercrunch/airflow,mtdewulf/incubator-airflow,N3da/incubator-airflow,mattuuh7/incubator-airflow,rishibarve/incubator-airflow,Fokko/incubator-airflow,apache/airflow,NielsZeilemaker/incubator-airflow,dmitry-r/incubator-airflow,zoyahav/incubator-airflow,AllisonWang/incubator-airflow,mattuuh7/incubator-airflow,ledsusop/airflow,ronfung/incubator-airflow,mylons/incubator-airflow,wileeam/airflow,vineet-rh/incubator-airflow,fenglu-g/incubator-airflow,forevernull/incubator-airflow,lyft/incubator-airflow,lxneng/incubator-airflow,caseyching/incubator-airflow,wileeam/airflow,hamedhsn/incubator-airflow,mrares/incubator-airflow,KL-WLCR/incubator-airflow,jlowin/airflow,nathanielvarona/airflow,owlabs/incubator-airflow,btallman/incubator-airflow,spektom/incubator-airflow,malmiron/incubator-airflow,adrpar/incubator-airflow,brandsoulmates/incubator-airflow,juvoinc/airflow,juvoinc/airflow,jfantom/incubator-airflow,hgrif/incubator-airflow,dmitry-r/incubator-airflow,zack3241/incubator-airflow,asnir/airflow,bolkedebruin/airflow,malmiron/incubator-airflow,mtdewulf/incubator-airflow,OpringaoDoTurno/airflow,subodhchhabra/airflow,wolfier/incubator-airflow,apache/airflow,andrewmchen/incubator-airflow,jbhsieh/incubator-airflow,adrpar/incubator-airflow,hamedhsn/incubator-airflow,dgies/incubator-airflow,airbnb/airflow,stverhae/incubator-airflow,sergiohgz/incubator-airflow,jbhsieh/incubator-airflow,zack3241/incubator-airflow,Twistbioscience/incubator-airflow,jesusfcr/airflow,mtagle/airflow,wndhydrnt/airflow,ledsusop/airflow,andyxhadji/incubator-airflow,d-lee/airflow,cfei18/incubator-airflow,mrkm4ntr/incubator-airflow,jhsenjaliya/incubator-airflow,caseyching/incubator-airflow,danielvdende/incubator-airflow,forevernull/incubator-airflow,lyft/incubator-airflow,skudriashev/incubator-airflow,MetrodataTeam/incubator-airflow,jhsenjaliya/incubator-airflow,sekikn/incubator-airflow,r39132/airflow,cfei18/incubator-airflow,Acehaidrey/incubator-airflow,zodiac/incubator-airflow,dhuang/incubator-airflow,easytaxibr/airflow,DEVELByte/incubator-airflow,sid88in/incubator-airflow,apache/airflow,Tagar/incubator-airflow,janczak10/incubator-airflow,forevernull/incubator-airflow,jlowin/airflow,r39132/airflow,brandsoulmates/incubator-airflow,AllisonWang/incubator-airflow,NielsZeilemaker/incubator-airflow,hgrif/incubator-airflow,cjqian/incubator-airflow,aminghadersohi/airflow,mtagle/airflow,adamhaney/airflow,yati-sagade/incubator-airflow,gtoonstra/airflow,andyxhadji/incubator-airflow,yati-sagade/incubator-airflow,preete-dixit-ck/incubator-airflow,cfei18/incubator-airflow,bolkedebruin/airflow,saguziel/incubator-airflow,janczak10/incubator-airflow,criccomini/airflow,KL-WLCR/incubator-airflow,d-lee/airflow,mrares/incubator-airflow,DinoCow/airflow,criccomini/airflow,MortalViews/incubator-airflow,mtagle/airflow,andyxhadji/incubator-airflow,vijaysbhat/incubator-airflow,mattuuh7/incubator-airflow,btallman/incubator-airflow,stverhae/incubator-airflow,janczak10/incubator-airflow,jgao54/airflow,hamedhsn/incubator-airflow,saguziel/incubator-airflow,gilt/incubator-airflow,Fokko/incubator-airflow,yk5/incubator-airflow,DinoCow/airflow,danielvdende/incubator-airflow,MortalViews/incubator-airflow,ProstoMaxim/incubator-airflow,dhuang/incubator-airflow,N3da/incubator-airflow,zodiac/incubator-airflow,mylons/incubator-airflow,edgarRd/incubator-airflow,mtdewulf/incubator-airflow,jbhsieh/incubator-airflow,forevernull/incubator-airflow,edgarRd/incubator-airflow,gilt/incubator-airflow,mylons/incubator-airflow,owlabs/incubator-airflow,sekikn/incubator-airflow,ledsusop/airflow,apache/airflow,juvoinc/airflow,jfantom/incubator-airflow,akosel/incubator-airflow,gritlogic/incubator-airflow,hgrif/incubator-airflow,saguziel/incubator-airflow,lxneng/incubator-airflow,skudriashev/incubator-airflow,ProstoMaxim/incubator-airflow,saguziel/incubator-airflow,gtoonstra/airflow,dmitry-r/incubator-airflow,nathanielvarona/airflow,yiqingj/airflow,jlowin/airflow,yiqingj/airflow,skudriashev/incubator-airflow,wileeam/airflow,airbnb/airflow,gilt/incubator-airflow,mtdewulf/incubator-airflow,zoyahav/incubator-airflow,aminghadersohi/airflow,lxneng/incubator-airflow,jiwang576/incubator-airflow,Tagar/incubator-airflow,adrpar/incubator-airflow,mrares/incubator-airflow,brandsoulmates/incubator-airflow,hgrif/incubator-airflow,wndhydrnt/airflow,MetrodataTeam/incubator-airflow,nathanielvarona/airflow,mtagle/airflow,holygits/incubator-airflow,rishibarve/incubator-airflow,Acehaidrey/incubator-airflow,Acehaidrey/incubator-airflow,stverhae/incubator-airflow,Acehaidrey/incubator-airflow,OpringaoDoTurno/airflow,kerzhner/airflow,jgao54/airflow,juvoinc/airflow,ProstoMaxim/incubator-airflow,brandsoulmates/incubator-airflow,yk5/incubator-airflow,danielvdende/incubator-airflow,subodhchhabra/airflow,gritlogic/incubator-airflow,Tagar/incubator-airflow,adamhaney/airflow,KL-WLCR/incubator-airflow,wolfier/incubator-airflow,danielvdende/incubator-airflow,jhsenjaliya/incubator-airflow,apache/airflow,sdiazb/airflow,plypaul/airflow,dhuang/incubator-airflow,airbnb/airflow,N3da/incubator-airflow,holygits/incubator-airflow,gtoonstra/airflow,CloverHealth/airflow,mistercrunch/airflow,plypaul/airflow,rishibarve/incubator-airflow,mattuuh7/incubator-airflow,Tagar/incubator-airflow,d-lee/airflow,dgies/incubator-airflow,jiwang576/incubator-airflow,akosel/incubator-airflow,sdiazb/airflow,edgarRd/incubator-airflow,AllisonWang/incubator-airflow,yiqingj/airflow,malmiron/incubator-airflow,btallman/incubator-airflow,wndhydrnt/airflow,jfantom/incubator-airflow,holygits/incubator-airflow,r39132/airflow,mistercrunch/airflow,wndhydrnt/airflow,gilt/incubator-airflow,mylons/incubator-airflow,spektom/incubator-airflow,fenglu-g/incubator-airflow,jgao54/airflow,RealImpactAnalytics/airflow,holygits/incubator-airflow,RealImpactAnalytics/airflow,lyft/incubator-airflow,cfei18/incubator-airflow,asnir/airflow,andrewmchen/incubator-airflow,zoyahav/incubator-airflow,zodiac/incubator-airflow,owlabs/incubator-airflow,wolfier/incubator-airflow,sergiohgz/incubator-airflow,danielvdende/incubator-airflow,artwr/airflow,NielsZeilemaker/incubator-airflow,sid88in/incubator-airflow,gritlogic/incubator-airflow,cademarkegard/airflow,aminghadersohi/airflow,wileeam/airflow,N3da/incubator-airflow | """add TI state index
Revision ID: 211e584da130
Revises: 2e82aab8ef20
Create Date: 2016-06-30 10:54:24.323588
"""
# revision identifiers, used by Alembic.
revision = '211e584da130'
down_revision = '2e82aab8ef20'
branch_labels = None
depends_on = None
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index('ti_state', 'task_instance', ['state'], unique=False)
def downgrade():
op.drop_index('ti_state', table_name='task_instance')
Add license to migration file | # -*- coding: utf-8 -*-
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""add TI state index
Revision ID: 211e584da130
Revises: 2e82aab8ef20
Create Date: 2016-06-30 10:54:24.323588
"""
# revision identifiers, used by Alembic.
revision = '211e584da130'
down_revision = '2e82aab8ef20'
branch_labels = None
depends_on = None
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index('ti_state', 'task_instance', ['state'], unique=False)
def downgrade():
op.drop_index('ti_state', table_name='task_instance')
| <commit_before>"""add TI state index
Revision ID: 211e584da130
Revises: 2e82aab8ef20
Create Date: 2016-06-30 10:54:24.323588
"""
# revision identifiers, used by Alembic.
revision = '211e584da130'
down_revision = '2e82aab8ef20'
branch_labels = None
depends_on = None
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index('ti_state', 'task_instance', ['state'], unique=False)
def downgrade():
op.drop_index('ti_state', table_name='task_instance')
<commit_msg>Add license to migration file<commit_after> | # -*- coding: utf-8 -*-
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""add TI state index
Revision ID: 211e584da130
Revises: 2e82aab8ef20
Create Date: 2016-06-30 10:54:24.323588
"""
# revision identifiers, used by Alembic.
revision = '211e584da130'
down_revision = '2e82aab8ef20'
branch_labels = None
depends_on = None
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index('ti_state', 'task_instance', ['state'], unique=False)
def downgrade():
op.drop_index('ti_state', table_name='task_instance')
| """add TI state index
Revision ID: 211e584da130
Revises: 2e82aab8ef20
Create Date: 2016-06-30 10:54:24.323588
"""
# revision identifiers, used by Alembic.
revision = '211e584da130'
down_revision = '2e82aab8ef20'
branch_labels = None
depends_on = None
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index('ti_state', 'task_instance', ['state'], unique=False)
def downgrade():
op.drop_index('ti_state', table_name='task_instance')
Add license to migration file# -*- coding: utf-8 -*-
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""add TI state index
Revision ID: 211e584da130
Revises: 2e82aab8ef20
Create Date: 2016-06-30 10:54:24.323588
"""
# revision identifiers, used by Alembic.
revision = '211e584da130'
down_revision = '2e82aab8ef20'
branch_labels = None
depends_on = None
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index('ti_state', 'task_instance', ['state'], unique=False)
def downgrade():
op.drop_index('ti_state', table_name='task_instance')
| <commit_before>"""add TI state index
Revision ID: 211e584da130
Revises: 2e82aab8ef20
Create Date: 2016-06-30 10:54:24.323588
"""
# revision identifiers, used by Alembic.
revision = '211e584da130'
down_revision = '2e82aab8ef20'
branch_labels = None
depends_on = None
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index('ti_state', 'task_instance', ['state'], unique=False)
def downgrade():
op.drop_index('ti_state', table_name='task_instance')
<commit_msg>Add license to migration file<commit_after># -*- coding: utf-8 -*-
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""add TI state index
Revision ID: 211e584da130
Revises: 2e82aab8ef20
Create Date: 2016-06-30 10:54:24.323588
"""
# revision identifiers, used by Alembic.
revision = '211e584da130'
down_revision = '2e82aab8ef20'
branch_labels = None
depends_on = None
from alembic import op
import sqlalchemy as sa
def upgrade():
op.create_index('ti_state', 'task_instance', ['state'], unique=False)
def downgrade():
op.drop_index('ti_state', table_name='task_instance')
|
0469f8f707ba542b0e1af2915d8a46a0107d9d62 | calexicon/tests/test_dates.py | calexicon/tests/test_dates.py | import unittest
from datetime import date
from calexicon.dates import DateWithCalendar
class TestDateWithCalendar(unittest.TestCase):
def setUp(self):
date_dt = date(2010, 8, 1)
self.date_wc = DateWithCalendar(None, date_dt)
def test_comparisons(self):
self.assertTrue(self.date_wc < date(2010, 8, 2))
self.assertFalse(self.date_wc < date(2010, 7, 31))
self.assertTrue(self.date_wc > date(2010, 7, 2))
self.assertFalse(self.date_wc > date(2010, 8, 31))
def test_nonstrict_comparisons(self):
self.assertTrue(self.date_wc <= date(2010, 8, 2))
self.assertFalse(self.date_wc <= date(2010, 7, 31))
self.assertTrue(self.date_wc >= date(2010, 7, 2))
self.assertFalse(self.date_wc >= date(2010, 8, 31))
self.assertTrue(self.date_wc <= date(2010, 8, 1))
self.assertTrue(self.date_wc >= date(2010, 8, 1))
| import unittest
from datetime import date, timedelta
from calexicon.dates import DateWithCalendar
class TestDateWithCalendar(unittest.TestCase):
def setUp(self):
date_dt = date(2010, 8, 1)
self.date_wc = DateWithCalendar(None, date_dt)
def test_comparisons(self):
self.assertTrue(self.date_wc < date(2010, 8, 2))
self.assertFalse(self.date_wc < date(2010, 7, 31))
self.assertTrue(self.date_wc > date(2010, 7, 2))
self.assertFalse(self.date_wc > date(2010, 8, 31))
def test_nonstrict_comparisons(self):
self.assertTrue(self.date_wc <= date(2010, 8, 2))
self.assertFalse(self.date_wc <= date(2010, 7, 31))
self.assertTrue(self.date_wc >= date(2010, 7, 2))
self.assertFalse(self.date_wc >= date(2010, 8, 31))
self.assertTrue(self.date_wc <= date(2010, 8, 1))
self.assertTrue(self.date_wc >= date(2010, 8, 1))
def test_subtraction(self):
self.assertEqual(self.date_wc - date(2012, 10, 30), timedelta(days=-821))
| Add a test for __sub__ between a DateWith and a vanilla date. | Add a test for __sub__ between a DateWith and a vanilla date.
| Python | apache-2.0 | jwg4/calexicon,jwg4/qual | import unittest
from datetime import date
from calexicon.dates import DateWithCalendar
class TestDateWithCalendar(unittest.TestCase):
def setUp(self):
date_dt = date(2010, 8, 1)
self.date_wc = DateWithCalendar(None, date_dt)
def test_comparisons(self):
self.assertTrue(self.date_wc < date(2010, 8, 2))
self.assertFalse(self.date_wc < date(2010, 7, 31))
self.assertTrue(self.date_wc > date(2010, 7, 2))
self.assertFalse(self.date_wc > date(2010, 8, 31))
def test_nonstrict_comparisons(self):
self.assertTrue(self.date_wc <= date(2010, 8, 2))
self.assertFalse(self.date_wc <= date(2010, 7, 31))
self.assertTrue(self.date_wc >= date(2010, 7, 2))
self.assertFalse(self.date_wc >= date(2010, 8, 31))
self.assertTrue(self.date_wc <= date(2010, 8, 1))
self.assertTrue(self.date_wc >= date(2010, 8, 1))
Add a test for __sub__ between a DateWith and a vanilla date. | import unittest
from datetime import date, timedelta
from calexicon.dates import DateWithCalendar
class TestDateWithCalendar(unittest.TestCase):
def setUp(self):
date_dt = date(2010, 8, 1)
self.date_wc = DateWithCalendar(None, date_dt)
def test_comparisons(self):
self.assertTrue(self.date_wc < date(2010, 8, 2))
self.assertFalse(self.date_wc < date(2010, 7, 31))
self.assertTrue(self.date_wc > date(2010, 7, 2))
self.assertFalse(self.date_wc > date(2010, 8, 31))
def test_nonstrict_comparisons(self):
self.assertTrue(self.date_wc <= date(2010, 8, 2))
self.assertFalse(self.date_wc <= date(2010, 7, 31))
self.assertTrue(self.date_wc >= date(2010, 7, 2))
self.assertFalse(self.date_wc >= date(2010, 8, 31))
self.assertTrue(self.date_wc <= date(2010, 8, 1))
self.assertTrue(self.date_wc >= date(2010, 8, 1))
def test_subtraction(self):
self.assertEqual(self.date_wc - date(2012, 10, 30), timedelta(days=-821))
| <commit_before>import unittest
from datetime import date
from calexicon.dates import DateWithCalendar
class TestDateWithCalendar(unittest.TestCase):
def setUp(self):
date_dt = date(2010, 8, 1)
self.date_wc = DateWithCalendar(None, date_dt)
def test_comparisons(self):
self.assertTrue(self.date_wc < date(2010, 8, 2))
self.assertFalse(self.date_wc < date(2010, 7, 31))
self.assertTrue(self.date_wc > date(2010, 7, 2))
self.assertFalse(self.date_wc > date(2010, 8, 31))
def test_nonstrict_comparisons(self):
self.assertTrue(self.date_wc <= date(2010, 8, 2))
self.assertFalse(self.date_wc <= date(2010, 7, 31))
self.assertTrue(self.date_wc >= date(2010, 7, 2))
self.assertFalse(self.date_wc >= date(2010, 8, 31))
self.assertTrue(self.date_wc <= date(2010, 8, 1))
self.assertTrue(self.date_wc >= date(2010, 8, 1))
<commit_msg>Add a test for __sub__ between a DateWith and a vanilla date.<commit_after> | import unittest
from datetime import date, timedelta
from calexicon.dates import DateWithCalendar
class TestDateWithCalendar(unittest.TestCase):
def setUp(self):
date_dt = date(2010, 8, 1)
self.date_wc = DateWithCalendar(None, date_dt)
def test_comparisons(self):
self.assertTrue(self.date_wc < date(2010, 8, 2))
self.assertFalse(self.date_wc < date(2010, 7, 31))
self.assertTrue(self.date_wc > date(2010, 7, 2))
self.assertFalse(self.date_wc > date(2010, 8, 31))
def test_nonstrict_comparisons(self):
self.assertTrue(self.date_wc <= date(2010, 8, 2))
self.assertFalse(self.date_wc <= date(2010, 7, 31))
self.assertTrue(self.date_wc >= date(2010, 7, 2))
self.assertFalse(self.date_wc >= date(2010, 8, 31))
self.assertTrue(self.date_wc <= date(2010, 8, 1))
self.assertTrue(self.date_wc >= date(2010, 8, 1))
def test_subtraction(self):
self.assertEqual(self.date_wc - date(2012, 10, 30), timedelta(days=-821))
| import unittest
from datetime import date
from calexicon.dates import DateWithCalendar
class TestDateWithCalendar(unittest.TestCase):
def setUp(self):
date_dt = date(2010, 8, 1)
self.date_wc = DateWithCalendar(None, date_dt)
def test_comparisons(self):
self.assertTrue(self.date_wc < date(2010, 8, 2))
self.assertFalse(self.date_wc < date(2010, 7, 31))
self.assertTrue(self.date_wc > date(2010, 7, 2))
self.assertFalse(self.date_wc > date(2010, 8, 31))
def test_nonstrict_comparisons(self):
self.assertTrue(self.date_wc <= date(2010, 8, 2))
self.assertFalse(self.date_wc <= date(2010, 7, 31))
self.assertTrue(self.date_wc >= date(2010, 7, 2))
self.assertFalse(self.date_wc >= date(2010, 8, 31))
self.assertTrue(self.date_wc <= date(2010, 8, 1))
self.assertTrue(self.date_wc >= date(2010, 8, 1))
Add a test for __sub__ between a DateWith and a vanilla date.import unittest
from datetime import date, timedelta
from calexicon.dates import DateWithCalendar
class TestDateWithCalendar(unittest.TestCase):
def setUp(self):
date_dt = date(2010, 8, 1)
self.date_wc = DateWithCalendar(None, date_dt)
def test_comparisons(self):
self.assertTrue(self.date_wc < date(2010, 8, 2))
self.assertFalse(self.date_wc < date(2010, 7, 31))
self.assertTrue(self.date_wc > date(2010, 7, 2))
self.assertFalse(self.date_wc > date(2010, 8, 31))
def test_nonstrict_comparisons(self):
self.assertTrue(self.date_wc <= date(2010, 8, 2))
self.assertFalse(self.date_wc <= date(2010, 7, 31))
self.assertTrue(self.date_wc >= date(2010, 7, 2))
self.assertFalse(self.date_wc >= date(2010, 8, 31))
self.assertTrue(self.date_wc <= date(2010, 8, 1))
self.assertTrue(self.date_wc >= date(2010, 8, 1))
def test_subtraction(self):
self.assertEqual(self.date_wc - date(2012, 10, 30), timedelta(days=-821))
| <commit_before>import unittest
from datetime import date
from calexicon.dates import DateWithCalendar
class TestDateWithCalendar(unittest.TestCase):
def setUp(self):
date_dt = date(2010, 8, 1)
self.date_wc = DateWithCalendar(None, date_dt)
def test_comparisons(self):
self.assertTrue(self.date_wc < date(2010, 8, 2))
self.assertFalse(self.date_wc < date(2010, 7, 31))
self.assertTrue(self.date_wc > date(2010, 7, 2))
self.assertFalse(self.date_wc > date(2010, 8, 31))
def test_nonstrict_comparisons(self):
self.assertTrue(self.date_wc <= date(2010, 8, 2))
self.assertFalse(self.date_wc <= date(2010, 7, 31))
self.assertTrue(self.date_wc >= date(2010, 7, 2))
self.assertFalse(self.date_wc >= date(2010, 8, 31))
self.assertTrue(self.date_wc <= date(2010, 8, 1))
self.assertTrue(self.date_wc >= date(2010, 8, 1))
<commit_msg>Add a test for __sub__ between a DateWith and a vanilla date.<commit_after>import unittest
from datetime import date, timedelta
from calexicon.dates import DateWithCalendar
class TestDateWithCalendar(unittest.TestCase):
def setUp(self):
date_dt = date(2010, 8, 1)
self.date_wc = DateWithCalendar(None, date_dt)
def test_comparisons(self):
self.assertTrue(self.date_wc < date(2010, 8, 2))
self.assertFalse(self.date_wc < date(2010, 7, 31))
self.assertTrue(self.date_wc > date(2010, 7, 2))
self.assertFalse(self.date_wc > date(2010, 8, 31))
def test_nonstrict_comparisons(self):
self.assertTrue(self.date_wc <= date(2010, 8, 2))
self.assertFalse(self.date_wc <= date(2010, 7, 31))
self.assertTrue(self.date_wc >= date(2010, 7, 2))
self.assertFalse(self.date_wc >= date(2010, 8, 31))
self.assertTrue(self.date_wc <= date(2010, 8, 1))
self.assertTrue(self.date_wc >= date(2010, 8, 1))
def test_subtraction(self):
self.assertEqual(self.date_wc - date(2012, 10, 30), timedelta(days=-821))
|
f3e6bc366ea77468772905c0094c9b4305c49fed | jsonpickle/handlers.py | jsonpickle/handlers.py |
class TypeRegistered(type):
"""
As classes of this metaclass are created, they keep a registry in the
base class of all handler referenced by the keys in cls._handles.
"""
def __init__(cls, name, bases, namespace):
super(TypeRegistered, cls).__init__(name, bases, namespace)
if not hasattr(cls, '_registry'):
cls._registry = {}
types_handled = getattr(cls, '_handles', [])
cls._registry.update((type_, cls) for type_ in types_handled)
class BaseHandler(object):
"""
Abstract base class for handlers.
"""
__metaclass__ = TypeRegistered
def __init__(self, base):
"""
Initialize a new handler to handle `type`.
:Parameters:
- `base`: reference to pickler/unpickler
"""
self._base = base
def flatten(self, obj, data):
"""
Flatten `obj` into a json-friendly form.
:Parameters:
- `obj`: object of `type`
"""
raise NotImplementedError("Abstract method.")
def restore(self, obj):
"""
Restores the `obj` to `type`
:Parameters:
- `object`: json-friendly object
"""
raise NotImplementedError("Abstract method.")
|
class TypeRegistered(type):
"""
As classes of this metaclass are created, they keep a registry in the
base class of all handler referenced by the keys in cls._handles.
"""
def __init__(cls, name, bases, namespace):
super(TypeRegistered, cls).__init__(name, bases, namespace)
if not hasattr(cls, '_registry'):
cls._registry = {}
types_handled = getattr(cls, '_handles', [])
cls._registry.update((type_, cls) for type_ in types_handled)
class BaseHandler(object):
"""
Abstract base class for handlers.
"""
__metaclass__ = TypeRegistered
def __init__(self, base):
"""
Initialize a new handler to handle `type`.
:Parameters:
- `base`: reference to pickler/unpickler
"""
self._base = base
def flatten(self, obj, data):
"""
Flatten `obj` into a json-friendly form.
:Parameters:
- `obj`: object of `type`
"""
raise NotImplementedError("Abstract method.")
def restore(self, obj):
"""
Restores the `obj` to `type`
:Parameters:
- `object`: json-friendly object
"""
raise NotImplementedError("Abstract method.")
# for backward compatibility, provide 'registry'
# jsonpickle 0.4 clients will call it with something like:
# registry.register(handled_type, handler_class)
class registry:
@staticmethod
def register(handled_type, handler_class):
pass
| Add a backward-compatibility shim to lessen the burden upgrading from 0.4 to 0.5 | Add a backward-compatibility shim to lessen the burden upgrading from 0.4 to 0.5
| Python | bsd-3-clause | mandx/jsonpickle,dongguangming/jsonpickle,mandx/jsonpickle,mandx/jsonpickle,dongguangming/jsonpickle,dongguangming/jsonpickle,mandx/jsonpickle,dongguangming/jsonpickle |
class TypeRegistered(type):
"""
As classes of this metaclass are created, they keep a registry in the
base class of all handler referenced by the keys in cls._handles.
"""
def __init__(cls, name, bases, namespace):
super(TypeRegistered, cls).__init__(name, bases, namespace)
if not hasattr(cls, '_registry'):
cls._registry = {}
types_handled = getattr(cls, '_handles', [])
cls._registry.update((type_, cls) for type_ in types_handled)
class BaseHandler(object):
"""
Abstract base class for handlers.
"""
__metaclass__ = TypeRegistered
def __init__(self, base):
"""
Initialize a new handler to handle `type`.
:Parameters:
- `base`: reference to pickler/unpickler
"""
self._base = base
def flatten(self, obj, data):
"""
Flatten `obj` into a json-friendly form.
:Parameters:
- `obj`: object of `type`
"""
raise NotImplementedError("Abstract method.")
def restore(self, obj):
"""
Restores the `obj` to `type`
:Parameters:
- `object`: json-friendly object
"""
raise NotImplementedError("Abstract method.")
Add a backward-compatibility shim to lessen the burden upgrading from 0.4 to 0.5 |
class TypeRegistered(type):
"""
As classes of this metaclass are created, they keep a registry in the
base class of all handler referenced by the keys in cls._handles.
"""
def __init__(cls, name, bases, namespace):
super(TypeRegistered, cls).__init__(name, bases, namespace)
if not hasattr(cls, '_registry'):
cls._registry = {}
types_handled = getattr(cls, '_handles', [])
cls._registry.update((type_, cls) for type_ in types_handled)
class BaseHandler(object):
"""
Abstract base class for handlers.
"""
__metaclass__ = TypeRegistered
def __init__(self, base):
"""
Initialize a new handler to handle `type`.
:Parameters:
- `base`: reference to pickler/unpickler
"""
self._base = base
def flatten(self, obj, data):
"""
Flatten `obj` into a json-friendly form.
:Parameters:
- `obj`: object of `type`
"""
raise NotImplementedError("Abstract method.")
def restore(self, obj):
"""
Restores the `obj` to `type`
:Parameters:
- `object`: json-friendly object
"""
raise NotImplementedError("Abstract method.")
# for backward compatibility, provide 'registry'
# jsonpickle 0.4 clients will call it with something like:
# registry.register(handled_type, handler_class)
class registry:
@staticmethod
def register(handled_type, handler_class):
pass
| <commit_before>
class TypeRegistered(type):
"""
As classes of this metaclass are created, they keep a registry in the
base class of all handler referenced by the keys in cls._handles.
"""
def __init__(cls, name, bases, namespace):
super(TypeRegistered, cls).__init__(name, bases, namespace)
if not hasattr(cls, '_registry'):
cls._registry = {}
types_handled = getattr(cls, '_handles', [])
cls._registry.update((type_, cls) for type_ in types_handled)
class BaseHandler(object):
"""
Abstract base class for handlers.
"""
__metaclass__ = TypeRegistered
def __init__(self, base):
"""
Initialize a new handler to handle `type`.
:Parameters:
- `base`: reference to pickler/unpickler
"""
self._base = base
def flatten(self, obj, data):
"""
Flatten `obj` into a json-friendly form.
:Parameters:
- `obj`: object of `type`
"""
raise NotImplementedError("Abstract method.")
def restore(self, obj):
"""
Restores the `obj` to `type`
:Parameters:
- `object`: json-friendly object
"""
raise NotImplementedError("Abstract method.")
<commit_msg>Add a backward-compatibility shim to lessen the burden upgrading from 0.4 to 0.5<commit_after> |
class TypeRegistered(type):
"""
As classes of this metaclass are created, they keep a registry in the
base class of all handler referenced by the keys in cls._handles.
"""
def __init__(cls, name, bases, namespace):
super(TypeRegistered, cls).__init__(name, bases, namespace)
if not hasattr(cls, '_registry'):
cls._registry = {}
types_handled = getattr(cls, '_handles', [])
cls._registry.update((type_, cls) for type_ in types_handled)
class BaseHandler(object):
"""
Abstract base class for handlers.
"""
__metaclass__ = TypeRegistered
def __init__(self, base):
"""
Initialize a new handler to handle `type`.
:Parameters:
- `base`: reference to pickler/unpickler
"""
self._base = base
def flatten(self, obj, data):
"""
Flatten `obj` into a json-friendly form.
:Parameters:
- `obj`: object of `type`
"""
raise NotImplementedError("Abstract method.")
def restore(self, obj):
"""
Restores the `obj` to `type`
:Parameters:
- `object`: json-friendly object
"""
raise NotImplementedError("Abstract method.")
# for backward compatibility, provide 'registry'
# jsonpickle 0.4 clients will call it with something like:
# registry.register(handled_type, handler_class)
class registry:
@staticmethod
def register(handled_type, handler_class):
pass
|
class TypeRegistered(type):
"""
As classes of this metaclass are created, they keep a registry in the
base class of all handler referenced by the keys in cls._handles.
"""
def __init__(cls, name, bases, namespace):
super(TypeRegistered, cls).__init__(name, bases, namespace)
if not hasattr(cls, '_registry'):
cls._registry = {}
types_handled = getattr(cls, '_handles', [])
cls._registry.update((type_, cls) for type_ in types_handled)
class BaseHandler(object):
"""
Abstract base class for handlers.
"""
__metaclass__ = TypeRegistered
def __init__(self, base):
"""
Initialize a new handler to handle `type`.
:Parameters:
- `base`: reference to pickler/unpickler
"""
self._base = base
def flatten(self, obj, data):
"""
Flatten `obj` into a json-friendly form.
:Parameters:
- `obj`: object of `type`
"""
raise NotImplementedError("Abstract method.")
def restore(self, obj):
"""
Restores the `obj` to `type`
:Parameters:
- `object`: json-friendly object
"""
raise NotImplementedError("Abstract method.")
Add a backward-compatibility shim to lessen the burden upgrading from 0.4 to 0.5
class TypeRegistered(type):
"""
As classes of this metaclass are created, they keep a registry in the
base class of all handler referenced by the keys in cls._handles.
"""
def __init__(cls, name, bases, namespace):
super(TypeRegistered, cls).__init__(name, bases, namespace)
if not hasattr(cls, '_registry'):
cls._registry = {}
types_handled = getattr(cls, '_handles', [])
cls._registry.update((type_, cls) for type_ in types_handled)
class BaseHandler(object):
"""
Abstract base class for handlers.
"""
__metaclass__ = TypeRegistered
def __init__(self, base):
"""
Initialize a new handler to handle `type`.
:Parameters:
- `base`: reference to pickler/unpickler
"""
self._base = base
def flatten(self, obj, data):
"""
Flatten `obj` into a json-friendly form.
:Parameters:
- `obj`: object of `type`
"""
raise NotImplementedError("Abstract method.")
def restore(self, obj):
"""
Restores the `obj` to `type`
:Parameters:
- `object`: json-friendly object
"""
raise NotImplementedError("Abstract method.")
# for backward compatibility, provide 'registry'
# jsonpickle 0.4 clients will call it with something like:
# registry.register(handled_type, handler_class)
class registry:
@staticmethod
def register(handled_type, handler_class):
pass
| <commit_before>
class TypeRegistered(type):
"""
As classes of this metaclass are created, they keep a registry in the
base class of all handler referenced by the keys in cls._handles.
"""
def __init__(cls, name, bases, namespace):
super(TypeRegistered, cls).__init__(name, bases, namespace)
if not hasattr(cls, '_registry'):
cls._registry = {}
types_handled = getattr(cls, '_handles', [])
cls._registry.update((type_, cls) for type_ in types_handled)
class BaseHandler(object):
"""
Abstract base class for handlers.
"""
__metaclass__ = TypeRegistered
def __init__(self, base):
"""
Initialize a new handler to handle `type`.
:Parameters:
- `base`: reference to pickler/unpickler
"""
self._base = base
def flatten(self, obj, data):
"""
Flatten `obj` into a json-friendly form.
:Parameters:
- `obj`: object of `type`
"""
raise NotImplementedError("Abstract method.")
def restore(self, obj):
"""
Restores the `obj` to `type`
:Parameters:
- `object`: json-friendly object
"""
raise NotImplementedError("Abstract method.")
<commit_msg>Add a backward-compatibility shim to lessen the burden upgrading from 0.4 to 0.5<commit_after>
class TypeRegistered(type):
"""
As classes of this metaclass are created, they keep a registry in the
base class of all handler referenced by the keys in cls._handles.
"""
def __init__(cls, name, bases, namespace):
super(TypeRegistered, cls).__init__(name, bases, namespace)
if not hasattr(cls, '_registry'):
cls._registry = {}
types_handled = getattr(cls, '_handles', [])
cls._registry.update((type_, cls) for type_ in types_handled)
class BaseHandler(object):
"""
Abstract base class for handlers.
"""
__metaclass__ = TypeRegistered
def __init__(self, base):
"""
Initialize a new handler to handle `type`.
:Parameters:
- `base`: reference to pickler/unpickler
"""
self._base = base
def flatten(self, obj, data):
"""
Flatten `obj` into a json-friendly form.
:Parameters:
- `obj`: object of `type`
"""
raise NotImplementedError("Abstract method.")
def restore(self, obj):
"""
Restores the `obj` to `type`
:Parameters:
- `object`: json-friendly object
"""
raise NotImplementedError("Abstract method.")
# for backward compatibility, provide 'registry'
# jsonpickle 0.4 clients will call it with something like:
# registry.register(handled_type, handler_class)
class registry:
@staticmethod
def register(handled_type, handler_class):
pass
|
08eb1f9e510b85e77d401ca4e13b7ad5354f4ecf | ingestors/email/outlookpst.py | ingestors/email/outlookpst.py | import logging
from followthemoney import model
from ingestors.ingestor import Ingestor
from ingestors.support.temp import TempFileSupport
from ingestors.support.shell import ShellSupport
from ingestors.support.ole import OLESupport
from ingestors.directory import DirectoryIngestor
log = logging.getLogger(__name__)
class OutlookPSTIngestor(Ingestor, TempFileSupport, OLESupport, ShellSupport):
MIME_TYPES = ['application/vnd.ms-outlook']
EXTENSIONS = ['pst', 'ost', 'pab']
BASE_SCORE = 5
COMMAND_TIMEOUT = 12 * 60 * 60
def ingest(self, file_path, entity):
entity.schema = model.get('Package')
self.extract_ole_metadata(file_path, entity)
temp_dir = self.make_empty_directory()
try:
self.exec_command('readpst',
'-e', # make subfolders, files per message
'-D', # include deleted
'-r', # recursive structure
'-8', # utf-8 where possible
'-b',
'-q', # quiet
'-o', temp_dir,
file_path)
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
except Exception:
log.exception("Failed to unpack PST.")
# Handle partially extracted archives.
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
raise
| import logging
from followthemoney import model
from ingestors.ingestor import Ingestor
from ingestors.support.temp import TempFileSupport
from ingestors.support.shell import ShellSupport
from ingestors.support.ole import OLESupport
from ingestors.directory import DirectoryIngestor
log = logging.getLogger(__name__)
class OutlookPSTIngestor(Ingestor, TempFileSupport, OLESupport, ShellSupport):
MIME_TYPES = ['application/vnd.ms-outlook']
EXTENSIONS = ['pst', 'ost', 'pab']
BASE_SCORE = 5
COMMAND_TIMEOUT = 12 * 60 * 60
def ingest(self, file_path, entity):
entity.schema = model.get('Package')
self.extract_ole_metadata(file_path, entity)
temp_dir = self.make_empty_directory()
try:
self.exec_command('readpst',
'-e', # make subfolders, files per message
'-S', # single files
'-D', # include deleted
# '-r', # recursive structure
'-8', # utf-8 where possible
'-cv', # export vcards
# '-q', # quiet
'-o', temp_dir,
file_path)
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
except Exception:
log.exception("Failed to unpack PST.")
# Handle partially extracted archives.
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
raise
| Make outlook emit single files | Make outlook emit single files
| Python | mit | alephdata/ingestors | import logging
from followthemoney import model
from ingestors.ingestor import Ingestor
from ingestors.support.temp import TempFileSupport
from ingestors.support.shell import ShellSupport
from ingestors.support.ole import OLESupport
from ingestors.directory import DirectoryIngestor
log = logging.getLogger(__name__)
class OutlookPSTIngestor(Ingestor, TempFileSupport, OLESupport, ShellSupport):
MIME_TYPES = ['application/vnd.ms-outlook']
EXTENSIONS = ['pst', 'ost', 'pab']
BASE_SCORE = 5
COMMAND_TIMEOUT = 12 * 60 * 60
def ingest(self, file_path, entity):
entity.schema = model.get('Package')
self.extract_ole_metadata(file_path, entity)
temp_dir = self.make_empty_directory()
try:
self.exec_command('readpst',
'-e', # make subfolders, files per message
'-D', # include deleted
'-r', # recursive structure
'-8', # utf-8 where possible
'-b',
'-q', # quiet
'-o', temp_dir,
file_path)
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
except Exception:
log.exception("Failed to unpack PST.")
# Handle partially extracted archives.
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
raise
Make outlook emit single files | import logging
from followthemoney import model
from ingestors.ingestor import Ingestor
from ingestors.support.temp import TempFileSupport
from ingestors.support.shell import ShellSupport
from ingestors.support.ole import OLESupport
from ingestors.directory import DirectoryIngestor
log = logging.getLogger(__name__)
class OutlookPSTIngestor(Ingestor, TempFileSupport, OLESupport, ShellSupport):
MIME_TYPES = ['application/vnd.ms-outlook']
EXTENSIONS = ['pst', 'ost', 'pab']
BASE_SCORE = 5
COMMAND_TIMEOUT = 12 * 60 * 60
def ingest(self, file_path, entity):
entity.schema = model.get('Package')
self.extract_ole_metadata(file_path, entity)
temp_dir = self.make_empty_directory()
try:
self.exec_command('readpst',
'-e', # make subfolders, files per message
'-S', # single files
'-D', # include deleted
# '-r', # recursive structure
'-8', # utf-8 where possible
'-cv', # export vcards
# '-q', # quiet
'-o', temp_dir,
file_path)
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
except Exception:
log.exception("Failed to unpack PST.")
# Handle partially extracted archives.
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
raise
| <commit_before>import logging
from followthemoney import model
from ingestors.ingestor import Ingestor
from ingestors.support.temp import TempFileSupport
from ingestors.support.shell import ShellSupport
from ingestors.support.ole import OLESupport
from ingestors.directory import DirectoryIngestor
log = logging.getLogger(__name__)
class OutlookPSTIngestor(Ingestor, TempFileSupport, OLESupport, ShellSupport):
MIME_TYPES = ['application/vnd.ms-outlook']
EXTENSIONS = ['pst', 'ost', 'pab']
BASE_SCORE = 5
COMMAND_TIMEOUT = 12 * 60 * 60
def ingest(self, file_path, entity):
entity.schema = model.get('Package')
self.extract_ole_metadata(file_path, entity)
temp_dir = self.make_empty_directory()
try:
self.exec_command('readpst',
'-e', # make subfolders, files per message
'-D', # include deleted
'-r', # recursive structure
'-8', # utf-8 where possible
'-b',
'-q', # quiet
'-o', temp_dir,
file_path)
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
except Exception:
log.exception("Failed to unpack PST.")
# Handle partially extracted archives.
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
raise
<commit_msg>Make outlook emit single files<commit_after> | import logging
from followthemoney import model
from ingestors.ingestor import Ingestor
from ingestors.support.temp import TempFileSupport
from ingestors.support.shell import ShellSupport
from ingestors.support.ole import OLESupport
from ingestors.directory import DirectoryIngestor
log = logging.getLogger(__name__)
class OutlookPSTIngestor(Ingestor, TempFileSupport, OLESupport, ShellSupport):
MIME_TYPES = ['application/vnd.ms-outlook']
EXTENSIONS = ['pst', 'ost', 'pab']
BASE_SCORE = 5
COMMAND_TIMEOUT = 12 * 60 * 60
def ingest(self, file_path, entity):
entity.schema = model.get('Package')
self.extract_ole_metadata(file_path, entity)
temp_dir = self.make_empty_directory()
try:
self.exec_command('readpst',
'-e', # make subfolders, files per message
'-S', # single files
'-D', # include deleted
# '-r', # recursive structure
'-8', # utf-8 where possible
'-cv', # export vcards
# '-q', # quiet
'-o', temp_dir,
file_path)
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
except Exception:
log.exception("Failed to unpack PST.")
# Handle partially extracted archives.
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
raise
| import logging
from followthemoney import model
from ingestors.ingestor import Ingestor
from ingestors.support.temp import TempFileSupport
from ingestors.support.shell import ShellSupport
from ingestors.support.ole import OLESupport
from ingestors.directory import DirectoryIngestor
log = logging.getLogger(__name__)
class OutlookPSTIngestor(Ingestor, TempFileSupport, OLESupport, ShellSupport):
MIME_TYPES = ['application/vnd.ms-outlook']
EXTENSIONS = ['pst', 'ost', 'pab']
BASE_SCORE = 5
COMMAND_TIMEOUT = 12 * 60 * 60
def ingest(self, file_path, entity):
entity.schema = model.get('Package')
self.extract_ole_metadata(file_path, entity)
temp_dir = self.make_empty_directory()
try:
self.exec_command('readpst',
'-e', # make subfolders, files per message
'-D', # include deleted
'-r', # recursive structure
'-8', # utf-8 where possible
'-b',
'-q', # quiet
'-o', temp_dir,
file_path)
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
except Exception:
log.exception("Failed to unpack PST.")
# Handle partially extracted archives.
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
raise
Make outlook emit single filesimport logging
from followthemoney import model
from ingestors.ingestor import Ingestor
from ingestors.support.temp import TempFileSupport
from ingestors.support.shell import ShellSupport
from ingestors.support.ole import OLESupport
from ingestors.directory import DirectoryIngestor
log = logging.getLogger(__name__)
class OutlookPSTIngestor(Ingestor, TempFileSupport, OLESupport, ShellSupport):
MIME_TYPES = ['application/vnd.ms-outlook']
EXTENSIONS = ['pst', 'ost', 'pab']
BASE_SCORE = 5
COMMAND_TIMEOUT = 12 * 60 * 60
def ingest(self, file_path, entity):
entity.schema = model.get('Package')
self.extract_ole_metadata(file_path, entity)
temp_dir = self.make_empty_directory()
try:
self.exec_command('readpst',
'-e', # make subfolders, files per message
'-S', # single files
'-D', # include deleted
# '-r', # recursive structure
'-8', # utf-8 where possible
'-cv', # export vcards
# '-q', # quiet
'-o', temp_dir,
file_path)
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
except Exception:
log.exception("Failed to unpack PST.")
# Handle partially extracted archives.
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
raise
| <commit_before>import logging
from followthemoney import model
from ingestors.ingestor import Ingestor
from ingestors.support.temp import TempFileSupport
from ingestors.support.shell import ShellSupport
from ingestors.support.ole import OLESupport
from ingestors.directory import DirectoryIngestor
log = logging.getLogger(__name__)
class OutlookPSTIngestor(Ingestor, TempFileSupport, OLESupport, ShellSupport):
MIME_TYPES = ['application/vnd.ms-outlook']
EXTENSIONS = ['pst', 'ost', 'pab']
BASE_SCORE = 5
COMMAND_TIMEOUT = 12 * 60 * 60
def ingest(self, file_path, entity):
entity.schema = model.get('Package')
self.extract_ole_metadata(file_path, entity)
temp_dir = self.make_empty_directory()
try:
self.exec_command('readpst',
'-e', # make subfolders, files per message
'-D', # include deleted
'-r', # recursive structure
'-8', # utf-8 where possible
'-b',
'-q', # quiet
'-o', temp_dir,
file_path)
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
except Exception:
log.exception("Failed to unpack PST.")
# Handle partially extracted archives.
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
raise
<commit_msg>Make outlook emit single files<commit_after>import logging
from followthemoney import model
from ingestors.ingestor import Ingestor
from ingestors.support.temp import TempFileSupport
from ingestors.support.shell import ShellSupport
from ingestors.support.ole import OLESupport
from ingestors.directory import DirectoryIngestor
log = logging.getLogger(__name__)
class OutlookPSTIngestor(Ingestor, TempFileSupport, OLESupport, ShellSupport):
MIME_TYPES = ['application/vnd.ms-outlook']
EXTENSIONS = ['pst', 'ost', 'pab']
BASE_SCORE = 5
COMMAND_TIMEOUT = 12 * 60 * 60
def ingest(self, file_path, entity):
entity.schema = model.get('Package')
self.extract_ole_metadata(file_path, entity)
temp_dir = self.make_empty_directory()
try:
self.exec_command('readpst',
'-e', # make subfolders, files per message
'-S', # single files
'-D', # include deleted
# '-r', # recursive structure
'-8', # utf-8 where possible
'-cv', # export vcards
# '-q', # quiet
'-o', temp_dir,
file_path)
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
except Exception:
log.exception("Failed to unpack PST.")
# Handle partially extracted archives.
self.manager.delegate(DirectoryIngestor, temp_dir, entity)
raise
|
a2c8ade4d73b6756fef2829c0e656acbe60f2b03 | fabfile.py | fabfile.py | from fabric.api import local
from fabric.api import warn_only
CMD_MANAGE = "python manage.py "
def auto_schema():
with warn_only():
schema('rockit.foundation.core')
schema('rockit.plugins.mailout')
schema('rockit.plugins.razberry')
def build():
migrate('rockit.foundation.core')
migrate('rockit.plugins.mailout')
migrate('rockit.plugins.razberry')
load_data('rockit/foundation/core/fixtures/settings.json')
load_data('rockit/plugins/mailout/fixtures/servers.json')
test()
def load_data(path):
local(CMD_MANAGE + 'loaddata %s' % path)
def migrate(app):
local(CMD_MANAGE + 'migrate %s' % app)
def runserver(localonly=True):
if localonly:
local(CMD_MANAGE + 'runserver')
else:
local(CMD_MANAGE + 'runserver 0.0.0.0')
def schema(app):
local(CMD_MANAGE + 'schemamigration %s --auto' % app)
def setup(environment):
local('pip install -r requirements/%s' % environment)
def test():
local(CMD_MANAGE + 'test') | from fabric.api import local
CMD_MANAGE = "python manage.py "
def auto_schema():
schema('rockit.foundation.core')
schema('rockit.plugins.mailout')
schema('rockit.plugins.razberry')
def build():
migrate('rockit.foundation.core')
migrate('rockit.plugins.mailout')
migrate('rockit.plugins.razberry')
load_data('rockit/foundation/core/fixtures/settings.json')
load_data('rockit/plugins/mailout/fixtures/servers.json')
test()
def load_data(path):
local(CMD_MANAGE + 'loaddata %s' % path)
def migrate(app):
local(CMD_MANAGE + 'migrate %s' % app)
def runserver(localonly=True):
if localonly:
local(CMD_MANAGE + 'runserver')
else:
local(CMD_MANAGE + 'runserver 0.0.0.0')
def schema(app):
local(CMD_MANAGE + 'schemamigration %s --auto' % app)
def setup(environment):
local('pip install -r requirements/%s' % environment)
def test():
local(CMD_MANAGE + 'test') | Remove warn only from fabric file | Remove warn only from fabric file
| Python | mit | acreations/rockit-server,acreations/rockit-server,acreations/rockit-server,acreations/rockit-server | from fabric.api import local
from fabric.api import warn_only
CMD_MANAGE = "python manage.py "
def auto_schema():
with warn_only():
schema('rockit.foundation.core')
schema('rockit.plugins.mailout')
schema('rockit.plugins.razberry')
def build():
migrate('rockit.foundation.core')
migrate('rockit.plugins.mailout')
migrate('rockit.plugins.razberry')
load_data('rockit/foundation/core/fixtures/settings.json')
load_data('rockit/plugins/mailout/fixtures/servers.json')
test()
def load_data(path):
local(CMD_MANAGE + 'loaddata %s' % path)
def migrate(app):
local(CMD_MANAGE + 'migrate %s' % app)
def runserver(localonly=True):
if localonly:
local(CMD_MANAGE + 'runserver')
else:
local(CMD_MANAGE + 'runserver 0.0.0.0')
def schema(app):
local(CMD_MANAGE + 'schemamigration %s --auto' % app)
def setup(environment):
local('pip install -r requirements/%s' % environment)
def test():
local(CMD_MANAGE + 'test')Remove warn only from fabric file | from fabric.api import local
CMD_MANAGE = "python manage.py "
def auto_schema():
schema('rockit.foundation.core')
schema('rockit.plugins.mailout')
schema('rockit.plugins.razberry')
def build():
migrate('rockit.foundation.core')
migrate('rockit.plugins.mailout')
migrate('rockit.plugins.razberry')
load_data('rockit/foundation/core/fixtures/settings.json')
load_data('rockit/plugins/mailout/fixtures/servers.json')
test()
def load_data(path):
local(CMD_MANAGE + 'loaddata %s' % path)
def migrate(app):
local(CMD_MANAGE + 'migrate %s' % app)
def runserver(localonly=True):
if localonly:
local(CMD_MANAGE + 'runserver')
else:
local(CMD_MANAGE + 'runserver 0.0.0.0')
def schema(app):
local(CMD_MANAGE + 'schemamigration %s --auto' % app)
def setup(environment):
local('pip install -r requirements/%s' % environment)
def test():
local(CMD_MANAGE + 'test') | <commit_before>from fabric.api import local
from fabric.api import warn_only
CMD_MANAGE = "python manage.py "
def auto_schema():
with warn_only():
schema('rockit.foundation.core')
schema('rockit.plugins.mailout')
schema('rockit.plugins.razberry')
def build():
migrate('rockit.foundation.core')
migrate('rockit.plugins.mailout')
migrate('rockit.plugins.razberry')
load_data('rockit/foundation/core/fixtures/settings.json')
load_data('rockit/plugins/mailout/fixtures/servers.json')
test()
def load_data(path):
local(CMD_MANAGE + 'loaddata %s' % path)
def migrate(app):
local(CMD_MANAGE + 'migrate %s' % app)
def runserver(localonly=True):
if localonly:
local(CMD_MANAGE + 'runserver')
else:
local(CMD_MANAGE + 'runserver 0.0.0.0')
def schema(app):
local(CMD_MANAGE + 'schemamigration %s --auto' % app)
def setup(environment):
local('pip install -r requirements/%s' % environment)
def test():
local(CMD_MANAGE + 'test')<commit_msg>Remove warn only from fabric file<commit_after> | from fabric.api import local
CMD_MANAGE = "python manage.py "
def auto_schema():
schema('rockit.foundation.core')
schema('rockit.plugins.mailout')
schema('rockit.plugins.razberry')
def build():
migrate('rockit.foundation.core')
migrate('rockit.plugins.mailout')
migrate('rockit.plugins.razberry')
load_data('rockit/foundation/core/fixtures/settings.json')
load_data('rockit/plugins/mailout/fixtures/servers.json')
test()
def load_data(path):
local(CMD_MANAGE + 'loaddata %s' % path)
def migrate(app):
local(CMD_MANAGE + 'migrate %s' % app)
def runserver(localonly=True):
if localonly:
local(CMD_MANAGE + 'runserver')
else:
local(CMD_MANAGE + 'runserver 0.0.0.0')
def schema(app):
local(CMD_MANAGE + 'schemamigration %s --auto' % app)
def setup(environment):
local('pip install -r requirements/%s' % environment)
def test():
local(CMD_MANAGE + 'test') | from fabric.api import local
from fabric.api import warn_only
CMD_MANAGE = "python manage.py "
def auto_schema():
with warn_only():
schema('rockit.foundation.core')
schema('rockit.plugins.mailout')
schema('rockit.plugins.razberry')
def build():
migrate('rockit.foundation.core')
migrate('rockit.plugins.mailout')
migrate('rockit.plugins.razberry')
load_data('rockit/foundation/core/fixtures/settings.json')
load_data('rockit/plugins/mailout/fixtures/servers.json')
test()
def load_data(path):
local(CMD_MANAGE + 'loaddata %s' % path)
def migrate(app):
local(CMD_MANAGE + 'migrate %s' % app)
def runserver(localonly=True):
if localonly:
local(CMD_MANAGE + 'runserver')
else:
local(CMD_MANAGE + 'runserver 0.0.0.0')
def schema(app):
local(CMD_MANAGE + 'schemamigration %s --auto' % app)
def setup(environment):
local('pip install -r requirements/%s' % environment)
def test():
local(CMD_MANAGE + 'test')Remove warn only from fabric filefrom fabric.api import local
CMD_MANAGE = "python manage.py "
def auto_schema():
schema('rockit.foundation.core')
schema('rockit.plugins.mailout')
schema('rockit.plugins.razberry')
def build():
migrate('rockit.foundation.core')
migrate('rockit.plugins.mailout')
migrate('rockit.plugins.razberry')
load_data('rockit/foundation/core/fixtures/settings.json')
load_data('rockit/plugins/mailout/fixtures/servers.json')
test()
def load_data(path):
local(CMD_MANAGE + 'loaddata %s' % path)
def migrate(app):
local(CMD_MANAGE + 'migrate %s' % app)
def runserver(localonly=True):
if localonly:
local(CMD_MANAGE + 'runserver')
else:
local(CMD_MANAGE + 'runserver 0.0.0.0')
def schema(app):
local(CMD_MANAGE + 'schemamigration %s --auto' % app)
def setup(environment):
local('pip install -r requirements/%s' % environment)
def test():
local(CMD_MANAGE + 'test') | <commit_before>from fabric.api import local
from fabric.api import warn_only
CMD_MANAGE = "python manage.py "
def auto_schema():
with warn_only():
schema('rockit.foundation.core')
schema('rockit.plugins.mailout')
schema('rockit.plugins.razberry')
def build():
migrate('rockit.foundation.core')
migrate('rockit.plugins.mailout')
migrate('rockit.plugins.razberry')
load_data('rockit/foundation/core/fixtures/settings.json')
load_data('rockit/plugins/mailout/fixtures/servers.json')
test()
def load_data(path):
local(CMD_MANAGE + 'loaddata %s' % path)
def migrate(app):
local(CMD_MANAGE + 'migrate %s' % app)
def runserver(localonly=True):
if localonly:
local(CMD_MANAGE + 'runserver')
else:
local(CMD_MANAGE + 'runserver 0.0.0.0')
def schema(app):
local(CMD_MANAGE + 'schemamigration %s --auto' % app)
def setup(environment):
local('pip install -r requirements/%s' % environment)
def test():
local(CMD_MANAGE + 'test')<commit_msg>Remove warn only from fabric file<commit_after>from fabric.api import local
CMD_MANAGE = "python manage.py "
def auto_schema():
schema('rockit.foundation.core')
schema('rockit.plugins.mailout')
schema('rockit.plugins.razberry')
def build():
migrate('rockit.foundation.core')
migrate('rockit.plugins.mailout')
migrate('rockit.plugins.razberry')
load_data('rockit/foundation/core/fixtures/settings.json')
load_data('rockit/plugins/mailout/fixtures/servers.json')
test()
def load_data(path):
local(CMD_MANAGE + 'loaddata %s' % path)
def migrate(app):
local(CMD_MANAGE + 'migrate %s' % app)
def runserver(localonly=True):
if localonly:
local(CMD_MANAGE + 'runserver')
else:
local(CMD_MANAGE + 'runserver 0.0.0.0')
def schema(app):
local(CMD_MANAGE + 'schemamigration %s --auto' % app)
def setup(environment):
local('pip install -r requirements/%s' % environment)
def test():
local(CMD_MANAGE + 'test') |
79911105899c95bf3fdb27c1aa61e8ff08ebef14 | bokeh/models/component.py | bokeh/models/component.py | from __future__ import absolute_import
from ..model import Model
from ..core.properties import abstract
from ..core.properties import Bool
from ..embed import notebook_div
@abstract
class Component(Model):
""" A base class for all embeddable models, i.e. plots, layouts and widgets.
"""
disabled = Bool(False, help="""
Whether the widget will be disabled when rendered. If ``True``,
the widget will be greyed-out, and not respond to UI events.
""")
# TODO: (mp) Not yet, because it breaks plotting/notebook examples.
# Rename to _repr_html_ if we decide to enable this by default.
def __repr_html__(self):
return notebook_div(self)
@property
def html(self):
from IPython.core.display import HTML
return HTML(self.__repr_html__())
| from __future__ import absolute_import
from ..model import Model
from ..core.properties import abstract
from ..core.properties import Bool
from ..embed import notebook_div
@abstract
class Component(Model):
""" A base class for all for all DOM-level components, i.e. plots, layouts
and widgets.
"""
disabled = Bool(False, help="""
Whether the widget will be disabled when rendered. If ``True``,
the widget will be greyed-out, and not respond to UI events.
""")
# TODO: (mp) Not yet, because it breaks plotting/notebook examples.
# Rename to _repr_html_ if we decide to enable this by default.
def __repr_html__(self):
return notebook_div(self)
@property
def html(self):
from IPython.core.display import HTML
return HTML(self.__repr_html__())
| Change Component docstring to mention DOM-level models | Change Component docstring to mention DOM-level models
| Python | bsd-3-clause | aavanian/bokeh,bokeh/bokeh,ericmjl/bokeh,schoolie/bokeh,ericmjl/bokeh,dennisobrien/bokeh,azjps/bokeh,mindriot101/bokeh,jakirkham/bokeh,aiguofer/bokeh,dennisobrien/bokeh,ptitjano/bokeh,msarahan/bokeh,azjps/bokeh,aiguofer/bokeh,ericmjl/bokeh,mindriot101/bokeh,aavanian/bokeh,phobson/bokeh,bokeh/bokeh,philippjfr/bokeh,KasperPRasmussen/bokeh,KasperPRasmussen/bokeh,msarahan/bokeh,msarahan/bokeh,bokeh/bokeh,DuCorey/bokeh,clairetang6/bokeh,jakirkham/bokeh,philippjfr/bokeh,clairetang6/bokeh,ericmjl/bokeh,mindriot101/bokeh,justacec/bokeh,stonebig/bokeh,azjps/bokeh,draperjames/bokeh,bokeh/bokeh,timsnyder/bokeh,draperjames/bokeh,timsnyder/bokeh,percyfal/bokeh,justacec/bokeh,Karel-van-de-Plassche/bokeh,draperjames/bokeh,KasperPRasmussen/bokeh,aavanian/bokeh,rs2/bokeh,Karel-van-de-Plassche/bokeh,draperjames/bokeh,rs2/bokeh,aavanian/bokeh,clairetang6/bokeh,ptitjano/bokeh,DuCorey/bokeh,dennisobrien/bokeh,percyfal/bokeh,DuCorey/bokeh,percyfal/bokeh,jakirkham/bokeh,azjps/bokeh,aiguofer/bokeh,msarahan/bokeh,rs2/bokeh,quasiben/bokeh,DuCorey/bokeh,stonebig/bokeh,Karel-van-de-Plassche/bokeh,azjps/bokeh,dennisobrien/bokeh,phobson/bokeh,schoolie/bokeh,schoolie/bokeh,philippjfr/bokeh,ptitjano/bokeh,DuCorey/bokeh,quasiben/bokeh,clairetang6/bokeh,KasperPRasmussen/bokeh,percyfal/bokeh,jakirkham/bokeh,phobson/bokeh,Karel-van-de-Plassche/bokeh,stonebig/bokeh,justacec/bokeh,mindriot101/bokeh,philippjfr/bokeh,aiguofer/bokeh,percyfal/bokeh,rs2/bokeh,Karel-van-de-Plassche/bokeh,stonebig/bokeh,timsnyder/bokeh,rs2/bokeh,justacec/bokeh,ptitjano/bokeh,schoolie/bokeh,schoolie/bokeh,philippjfr/bokeh,phobson/bokeh,jakirkham/bokeh,bokeh/bokeh,timsnyder/bokeh,quasiben/bokeh,timsnyder/bokeh,ericmjl/bokeh,aavanian/bokeh,phobson/bokeh,draperjames/bokeh,aiguofer/bokeh,KasperPRasmussen/bokeh,dennisobrien/bokeh,ptitjano/bokeh | from __future__ import absolute_import
from ..model import Model
from ..core.properties import abstract
from ..core.properties import Bool
from ..embed import notebook_div
@abstract
class Component(Model):
""" A base class for all embeddable models, i.e. plots, layouts and widgets.
"""
disabled = Bool(False, help="""
Whether the widget will be disabled when rendered. If ``True``,
the widget will be greyed-out, and not respond to UI events.
""")
# TODO: (mp) Not yet, because it breaks plotting/notebook examples.
# Rename to _repr_html_ if we decide to enable this by default.
def __repr_html__(self):
return notebook_div(self)
@property
def html(self):
from IPython.core.display import HTML
return HTML(self.__repr_html__())
Change Component docstring to mention DOM-level models | from __future__ import absolute_import
from ..model import Model
from ..core.properties import abstract
from ..core.properties import Bool
from ..embed import notebook_div
@abstract
class Component(Model):
""" A base class for all for all DOM-level components, i.e. plots, layouts
and widgets.
"""
disabled = Bool(False, help="""
Whether the widget will be disabled when rendered. If ``True``,
the widget will be greyed-out, and not respond to UI events.
""")
# TODO: (mp) Not yet, because it breaks plotting/notebook examples.
# Rename to _repr_html_ if we decide to enable this by default.
def __repr_html__(self):
return notebook_div(self)
@property
def html(self):
from IPython.core.display import HTML
return HTML(self.__repr_html__())
| <commit_before>from __future__ import absolute_import
from ..model import Model
from ..core.properties import abstract
from ..core.properties import Bool
from ..embed import notebook_div
@abstract
class Component(Model):
""" A base class for all embeddable models, i.e. plots, layouts and widgets.
"""
disabled = Bool(False, help="""
Whether the widget will be disabled when rendered. If ``True``,
the widget will be greyed-out, and not respond to UI events.
""")
# TODO: (mp) Not yet, because it breaks plotting/notebook examples.
# Rename to _repr_html_ if we decide to enable this by default.
def __repr_html__(self):
return notebook_div(self)
@property
def html(self):
from IPython.core.display import HTML
return HTML(self.__repr_html__())
<commit_msg>Change Component docstring to mention DOM-level models<commit_after> | from __future__ import absolute_import
from ..model import Model
from ..core.properties import abstract
from ..core.properties import Bool
from ..embed import notebook_div
@abstract
class Component(Model):
""" A base class for all for all DOM-level components, i.e. plots, layouts
and widgets.
"""
disabled = Bool(False, help="""
Whether the widget will be disabled when rendered. If ``True``,
the widget will be greyed-out, and not respond to UI events.
""")
# TODO: (mp) Not yet, because it breaks plotting/notebook examples.
# Rename to _repr_html_ if we decide to enable this by default.
def __repr_html__(self):
return notebook_div(self)
@property
def html(self):
from IPython.core.display import HTML
return HTML(self.__repr_html__())
| from __future__ import absolute_import
from ..model import Model
from ..core.properties import abstract
from ..core.properties import Bool
from ..embed import notebook_div
@abstract
class Component(Model):
""" A base class for all embeddable models, i.e. plots, layouts and widgets.
"""
disabled = Bool(False, help="""
Whether the widget will be disabled when rendered. If ``True``,
the widget will be greyed-out, and not respond to UI events.
""")
# TODO: (mp) Not yet, because it breaks plotting/notebook examples.
# Rename to _repr_html_ if we decide to enable this by default.
def __repr_html__(self):
return notebook_div(self)
@property
def html(self):
from IPython.core.display import HTML
return HTML(self.__repr_html__())
Change Component docstring to mention DOM-level modelsfrom __future__ import absolute_import
from ..model import Model
from ..core.properties import abstract
from ..core.properties import Bool
from ..embed import notebook_div
@abstract
class Component(Model):
""" A base class for all for all DOM-level components, i.e. plots, layouts
and widgets.
"""
disabled = Bool(False, help="""
Whether the widget will be disabled when rendered. If ``True``,
the widget will be greyed-out, and not respond to UI events.
""")
# TODO: (mp) Not yet, because it breaks plotting/notebook examples.
# Rename to _repr_html_ if we decide to enable this by default.
def __repr_html__(self):
return notebook_div(self)
@property
def html(self):
from IPython.core.display import HTML
return HTML(self.__repr_html__())
| <commit_before>from __future__ import absolute_import
from ..model import Model
from ..core.properties import abstract
from ..core.properties import Bool
from ..embed import notebook_div
@abstract
class Component(Model):
""" A base class for all embeddable models, i.e. plots, layouts and widgets.
"""
disabled = Bool(False, help="""
Whether the widget will be disabled when rendered. If ``True``,
the widget will be greyed-out, and not respond to UI events.
""")
# TODO: (mp) Not yet, because it breaks plotting/notebook examples.
# Rename to _repr_html_ if we decide to enable this by default.
def __repr_html__(self):
return notebook_div(self)
@property
def html(self):
from IPython.core.display import HTML
return HTML(self.__repr_html__())
<commit_msg>Change Component docstring to mention DOM-level models<commit_after>from __future__ import absolute_import
from ..model import Model
from ..core.properties import abstract
from ..core.properties import Bool
from ..embed import notebook_div
@abstract
class Component(Model):
""" A base class for all for all DOM-level components, i.e. plots, layouts
and widgets.
"""
disabled = Bool(False, help="""
Whether the widget will be disabled when rendered. If ``True``,
the widget will be greyed-out, and not respond to UI events.
""")
# TODO: (mp) Not yet, because it breaks plotting/notebook examples.
# Rename to _repr_html_ if we decide to enable this by default.
def __repr_html__(self):
return notebook_div(self)
@property
def html(self):
from IPython.core.display import HTML
return HTML(self.__repr_html__())
|
5847e9db8f316fdee6493fefc9cbc64a1e6a28de | km_api/know_me/serializers/subscription_serializers.py | km_api/know_me/serializers/subscription_serializers.py | import hashlib
import logging
from django.utils.translation import ugettext
from rest_framework import serializers
from know_me import models, subscriptions
logger = logging.getLogger(__name__)
class AppleSubscriptionSerializer(serializers.ModelSerializer):
"""
Serializer for an Apple subscription.
"""
class Meta:
fields = (
"id",
"time_created",
"time_updated",
"expiration_time",
"receipt_data",
)
model = models.SubscriptionAppleData
def validate(self, data):
"""
Ensure the provided receipt data corresponds to a valid Apple
receipt.
Returns:
The validated data.
"""
validated_data = data.copy()
receipt_data = validated_data["receipt_data"]
data_hash = hashlib.sha256(receipt_data.encode()).hexdigest()
if models.SubscriptionAppleData.objects.filter(
receipt_data_hash=data_hash
).exists():
logger.warning(
"Duplicate Apple receipt submitted with hash: %s", data_hash
)
raise serializers.ValidationError(
{
"receipt_data": ugettext(
"This receipt has already been used."
)
}
)
try:
receipt = subscriptions.validate_apple_receipt(receipt_data)
except subscriptions.ReceiptException as e:
raise serializers.ValidationError(
code=e.code, detail={"receipt_data": e.msg}
)
validated_data["expiration_time"] = receipt.expires_date
return validated_data
| import hashlib
import logging
from django.utils.translation import ugettext
from rest_framework import serializers
from know_me import models, subscriptions
logger = logging.getLogger(__name__)
class AppleSubscriptionSerializer(serializers.ModelSerializer):
"""
Serializer for an Apple subscription.
"""
class Meta:
fields = (
"id",
"time_created",
"time_updated",
"expiration_time",
"receipt_data",
)
model = models.SubscriptionAppleData
read_only_fields = ("expiration_time",)
def validate(self, data):
"""
Ensure the provided receipt data corresponds to a valid Apple
receipt.
Returns:
The validated data.
"""
validated_data = data.copy()
receipt_data = validated_data["receipt_data"]
data_hash = hashlib.sha256(receipt_data.encode()).hexdigest()
if models.SubscriptionAppleData.objects.filter(
receipt_data_hash=data_hash
).exists():
logger.warning(
"Duplicate Apple receipt submitted with hash: %s", data_hash
)
raise serializers.ValidationError(
{
"receipt_data": ugettext(
"This receipt has already been used."
)
}
)
try:
receipt = subscriptions.validate_apple_receipt(receipt_data)
except subscriptions.ReceiptException as e:
raise serializers.ValidationError(
code=e.code, detail={"receipt_data": e.msg}
)
validated_data["expiration_time"] = receipt.expires_date
return validated_data
| Mark apple receipt expiration time as read only. | Mark apple receipt expiration time as read only.
| Python | apache-2.0 | knowmetools/km-api,knowmetools/km-api,knowmetools/km-api,knowmetools/km-api | import hashlib
import logging
from django.utils.translation import ugettext
from rest_framework import serializers
from know_me import models, subscriptions
logger = logging.getLogger(__name__)
class AppleSubscriptionSerializer(serializers.ModelSerializer):
"""
Serializer for an Apple subscription.
"""
class Meta:
fields = (
"id",
"time_created",
"time_updated",
"expiration_time",
"receipt_data",
)
model = models.SubscriptionAppleData
def validate(self, data):
"""
Ensure the provided receipt data corresponds to a valid Apple
receipt.
Returns:
The validated data.
"""
validated_data = data.copy()
receipt_data = validated_data["receipt_data"]
data_hash = hashlib.sha256(receipt_data.encode()).hexdigest()
if models.SubscriptionAppleData.objects.filter(
receipt_data_hash=data_hash
).exists():
logger.warning(
"Duplicate Apple receipt submitted with hash: %s", data_hash
)
raise serializers.ValidationError(
{
"receipt_data": ugettext(
"This receipt has already been used."
)
}
)
try:
receipt = subscriptions.validate_apple_receipt(receipt_data)
except subscriptions.ReceiptException as e:
raise serializers.ValidationError(
code=e.code, detail={"receipt_data": e.msg}
)
validated_data["expiration_time"] = receipt.expires_date
return validated_data
Mark apple receipt expiration time as read only. | import hashlib
import logging
from django.utils.translation import ugettext
from rest_framework import serializers
from know_me import models, subscriptions
logger = logging.getLogger(__name__)
class AppleSubscriptionSerializer(serializers.ModelSerializer):
"""
Serializer for an Apple subscription.
"""
class Meta:
fields = (
"id",
"time_created",
"time_updated",
"expiration_time",
"receipt_data",
)
model = models.SubscriptionAppleData
read_only_fields = ("expiration_time",)
def validate(self, data):
"""
Ensure the provided receipt data corresponds to a valid Apple
receipt.
Returns:
The validated data.
"""
validated_data = data.copy()
receipt_data = validated_data["receipt_data"]
data_hash = hashlib.sha256(receipt_data.encode()).hexdigest()
if models.SubscriptionAppleData.objects.filter(
receipt_data_hash=data_hash
).exists():
logger.warning(
"Duplicate Apple receipt submitted with hash: %s", data_hash
)
raise serializers.ValidationError(
{
"receipt_data": ugettext(
"This receipt has already been used."
)
}
)
try:
receipt = subscriptions.validate_apple_receipt(receipt_data)
except subscriptions.ReceiptException as e:
raise serializers.ValidationError(
code=e.code, detail={"receipt_data": e.msg}
)
validated_data["expiration_time"] = receipt.expires_date
return validated_data
| <commit_before>import hashlib
import logging
from django.utils.translation import ugettext
from rest_framework import serializers
from know_me import models, subscriptions
logger = logging.getLogger(__name__)
class AppleSubscriptionSerializer(serializers.ModelSerializer):
"""
Serializer for an Apple subscription.
"""
class Meta:
fields = (
"id",
"time_created",
"time_updated",
"expiration_time",
"receipt_data",
)
model = models.SubscriptionAppleData
def validate(self, data):
"""
Ensure the provided receipt data corresponds to a valid Apple
receipt.
Returns:
The validated data.
"""
validated_data = data.copy()
receipt_data = validated_data["receipt_data"]
data_hash = hashlib.sha256(receipt_data.encode()).hexdigest()
if models.SubscriptionAppleData.objects.filter(
receipt_data_hash=data_hash
).exists():
logger.warning(
"Duplicate Apple receipt submitted with hash: %s", data_hash
)
raise serializers.ValidationError(
{
"receipt_data": ugettext(
"This receipt has already been used."
)
}
)
try:
receipt = subscriptions.validate_apple_receipt(receipt_data)
except subscriptions.ReceiptException as e:
raise serializers.ValidationError(
code=e.code, detail={"receipt_data": e.msg}
)
validated_data["expiration_time"] = receipt.expires_date
return validated_data
<commit_msg>Mark apple receipt expiration time as read only.<commit_after> | import hashlib
import logging
from django.utils.translation import ugettext
from rest_framework import serializers
from know_me import models, subscriptions
logger = logging.getLogger(__name__)
class AppleSubscriptionSerializer(serializers.ModelSerializer):
"""
Serializer for an Apple subscription.
"""
class Meta:
fields = (
"id",
"time_created",
"time_updated",
"expiration_time",
"receipt_data",
)
model = models.SubscriptionAppleData
read_only_fields = ("expiration_time",)
def validate(self, data):
"""
Ensure the provided receipt data corresponds to a valid Apple
receipt.
Returns:
The validated data.
"""
validated_data = data.copy()
receipt_data = validated_data["receipt_data"]
data_hash = hashlib.sha256(receipt_data.encode()).hexdigest()
if models.SubscriptionAppleData.objects.filter(
receipt_data_hash=data_hash
).exists():
logger.warning(
"Duplicate Apple receipt submitted with hash: %s", data_hash
)
raise serializers.ValidationError(
{
"receipt_data": ugettext(
"This receipt has already been used."
)
}
)
try:
receipt = subscriptions.validate_apple_receipt(receipt_data)
except subscriptions.ReceiptException as e:
raise serializers.ValidationError(
code=e.code, detail={"receipt_data": e.msg}
)
validated_data["expiration_time"] = receipt.expires_date
return validated_data
| import hashlib
import logging
from django.utils.translation import ugettext
from rest_framework import serializers
from know_me import models, subscriptions
logger = logging.getLogger(__name__)
class AppleSubscriptionSerializer(serializers.ModelSerializer):
"""
Serializer for an Apple subscription.
"""
class Meta:
fields = (
"id",
"time_created",
"time_updated",
"expiration_time",
"receipt_data",
)
model = models.SubscriptionAppleData
def validate(self, data):
"""
Ensure the provided receipt data corresponds to a valid Apple
receipt.
Returns:
The validated data.
"""
validated_data = data.copy()
receipt_data = validated_data["receipt_data"]
data_hash = hashlib.sha256(receipt_data.encode()).hexdigest()
if models.SubscriptionAppleData.objects.filter(
receipt_data_hash=data_hash
).exists():
logger.warning(
"Duplicate Apple receipt submitted with hash: %s", data_hash
)
raise serializers.ValidationError(
{
"receipt_data": ugettext(
"This receipt has already been used."
)
}
)
try:
receipt = subscriptions.validate_apple_receipt(receipt_data)
except subscriptions.ReceiptException as e:
raise serializers.ValidationError(
code=e.code, detail={"receipt_data": e.msg}
)
validated_data["expiration_time"] = receipt.expires_date
return validated_data
Mark apple receipt expiration time as read only.import hashlib
import logging
from django.utils.translation import ugettext
from rest_framework import serializers
from know_me import models, subscriptions
logger = logging.getLogger(__name__)
class AppleSubscriptionSerializer(serializers.ModelSerializer):
"""
Serializer for an Apple subscription.
"""
class Meta:
fields = (
"id",
"time_created",
"time_updated",
"expiration_time",
"receipt_data",
)
model = models.SubscriptionAppleData
read_only_fields = ("expiration_time",)
def validate(self, data):
"""
Ensure the provided receipt data corresponds to a valid Apple
receipt.
Returns:
The validated data.
"""
validated_data = data.copy()
receipt_data = validated_data["receipt_data"]
data_hash = hashlib.sha256(receipt_data.encode()).hexdigest()
if models.SubscriptionAppleData.objects.filter(
receipt_data_hash=data_hash
).exists():
logger.warning(
"Duplicate Apple receipt submitted with hash: %s", data_hash
)
raise serializers.ValidationError(
{
"receipt_data": ugettext(
"This receipt has already been used."
)
}
)
try:
receipt = subscriptions.validate_apple_receipt(receipt_data)
except subscriptions.ReceiptException as e:
raise serializers.ValidationError(
code=e.code, detail={"receipt_data": e.msg}
)
validated_data["expiration_time"] = receipt.expires_date
return validated_data
| <commit_before>import hashlib
import logging
from django.utils.translation import ugettext
from rest_framework import serializers
from know_me import models, subscriptions
logger = logging.getLogger(__name__)
class AppleSubscriptionSerializer(serializers.ModelSerializer):
"""
Serializer for an Apple subscription.
"""
class Meta:
fields = (
"id",
"time_created",
"time_updated",
"expiration_time",
"receipt_data",
)
model = models.SubscriptionAppleData
def validate(self, data):
"""
Ensure the provided receipt data corresponds to a valid Apple
receipt.
Returns:
The validated data.
"""
validated_data = data.copy()
receipt_data = validated_data["receipt_data"]
data_hash = hashlib.sha256(receipt_data.encode()).hexdigest()
if models.SubscriptionAppleData.objects.filter(
receipt_data_hash=data_hash
).exists():
logger.warning(
"Duplicate Apple receipt submitted with hash: %s", data_hash
)
raise serializers.ValidationError(
{
"receipt_data": ugettext(
"This receipt has already been used."
)
}
)
try:
receipt = subscriptions.validate_apple_receipt(receipt_data)
except subscriptions.ReceiptException as e:
raise serializers.ValidationError(
code=e.code, detail={"receipt_data": e.msg}
)
validated_data["expiration_time"] = receipt.expires_date
return validated_data
<commit_msg>Mark apple receipt expiration time as read only.<commit_after>import hashlib
import logging
from django.utils.translation import ugettext
from rest_framework import serializers
from know_me import models, subscriptions
logger = logging.getLogger(__name__)
class AppleSubscriptionSerializer(serializers.ModelSerializer):
"""
Serializer for an Apple subscription.
"""
class Meta:
fields = (
"id",
"time_created",
"time_updated",
"expiration_time",
"receipt_data",
)
model = models.SubscriptionAppleData
read_only_fields = ("expiration_time",)
def validate(self, data):
"""
Ensure the provided receipt data corresponds to a valid Apple
receipt.
Returns:
The validated data.
"""
validated_data = data.copy()
receipt_data = validated_data["receipt_data"]
data_hash = hashlib.sha256(receipt_data.encode()).hexdigest()
if models.SubscriptionAppleData.objects.filter(
receipt_data_hash=data_hash
).exists():
logger.warning(
"Duplicate Apple receipt submitted with hash: %s", data_hash
)
raise serializers.ValidationError(
{
"receipt_data": ugettext(
"This receipt has already been used."
)
}
)
try:
receipt = subscriptions.validate_apple_receipt(receipt_data)
except subscriptions.ReceiptException as e:
raise serializers.ValidationError(
code=e.code, detail={"receipt_data": e.msg}
)
validated_data["expiration_time"] = receipt.expires_date
return validated_data
|
fa279ca1f8e4c8e6b4094840d3ab40c0ac637eff | ocradmin/ocrpresets/models.py | ocradmin/ocrpresets/models.py | from django.db import models
from django.contrib.auth.models import User
from picklefield import fields
from tagging.fields import TagField
import tagging
class OcrPreset(models.Model):
user = models.ForeignKey(User)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
public = models.BooleanField(default=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True, auto_now=True)
type = models.CharField(max_length=20,
choices=[("segment", "Segment"), ("binarize", "Binarize")])
data = fields.PickledObjectField()
def __unicode__(self):
"""
String representation.
"""
return self.name
| from django.db import models
from django.contrib.auth.models import User
from picklefield import fields
from tagging.fields import TagField
import tagging
class OcrPreset(models.Model):
user = models.ForeignKey(User)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
public = models.BooleanField(default=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True, auto_now=True)
type = models.CharField(max_length=20,
choices=[("segment", "Segment"), ("binarize", "Binarize")])
data = fields.PickledObjectField()
def __unicode__(self):
"""
String representation.
"""
return "<%s: %s>" % (self.__class__.__name__, self.name)
| Improve unicode method. Whitespace cleanup | Improve unicode method. Whitespace cleanup
| Python | apache-2.0 | vitorio/ocropodium,vitorio/ocropodium,vitorio/ocropodium,vitorio/ocropodium | from django.db import models
from django.contrib.auth.models import User
from picklefield import fields
from tagging.fields import TagField
import tagging
class OcrPreset(models.Model):
user = models.ForeignKey(User)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
public = models.BooleanField(default=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True, auto_now=True)
type = models.CharField(max_length=20,
choices=[("segment", "Segment"), ("binarize", "Binarize")])
data = fields.PickledObjectField()
def __unicode__(self):
"""
String representation.
"""
return self.name
Improve unicode method. Whitespace cleanup | from django.db import models
from django.contrib.auth.models import User
from picklefield import fields
from tagging.fields import TagField
import tagging
class OcrPreset(models.Model):
user = models.ForeignKey(User)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
public = models.BooleanField(default=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True, auto_now=True)
type = models.CharField(max_length=20,
choices=[("segment", "Segment"), ("binarize", "Binarize")])
data = fields.PickledObjectField()
def __unicode__(self):
"""
String representation.
"""
return "<%s: %s>" % (self.__class__.__name__, self.name)
| <commit_before>from django.db import models
from django.contrib.auth.models import User
from picklefield import fields
from tagging.fields import TagField
import tagging
class OcrPreset(models.Model):
user = models.ForeignKey(User)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
public = models.BooleanField(default=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True, auto_now=True)
type = models.CharField(max_length=20,
choices=[("segment", "Segment"), ("binarize", "Binarize")])
data = fields.PickledObjectField()
def __unicode__(self):
"""
String representation.
"""
return self.name
<commit_msg>Improve unicode method. Whitespace cleanup<commit_after> | from django.db import models
from django.contrib.auth.models import User
from picklefield import fields
from tagging.fields import TagField
import tagging
class OcrPreset(models.Model):
user = models.ForeignKey(User)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
public = models.BooleanField(default=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True, auto_now=True)
type = models.CharField(max_length=20,
choices=[("segment", "Segment"), ("binarize", "Binarize")])
data = fields.PickledObjectField()
def __unicode__(self):
"""
String representation.
"""
return "<%s: %s>" % (self.__class__.__name__, self.name)
| from django.db import models
from django.contrib.auth.models import User
from picklefield import fields
from tagging.fields import TagField
import tagging
class OcrPreset(models.Model):
user = models.ForeignKey(User)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
public = models.BooleanField(default=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True, auto_now=True)
type = models.CharField(max_length=20,
choices=[("segment", "Segment"), ("binarize", "Binarize")])
data = fields.PickledObjectField()
def __unicode__(self):
"""
String representation.
"""
return self.name
Improve unicode method. Whitespace cleanupfrom django.db import models
from django.contrib.auth.models import User
from picklefield import fields
from tagging.fields import TagField
import tagging
class OcrPreset(models.Model):
user = models.ForeignKey(User)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
public = models.BooleanField(default=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True, auto_now=True)
type = models.CharField(max_length=20,
choices=[("segment", "Segment"), ("binarize", "Binarize")])
data = fields.PickledObjectField()
def __unicode__(self):
"""
String representation.
"""
return "<%s: %s>" % (self.__class__.__name__, self.name)
| <commit_before>from django.db import models
from django.contrib.auth.models import User
from picklefield import fields
from tagging.fields import TagField
import tagging
class OcrPreset(models.Model):
user = models.ForeignKey(User)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
public = models.BooleanField(default=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True, auto_now=True)
type = models.CharField(max_length=20,
choices=[("segment", "Segment"), ("binarize", "Binarize")])
data = fields.PickledObjectField()
def __unicode__(self):
"""
String representation.
"""
return self.name
<commit_msg>Improve unicode method. Whitespace cleanup<commit_after>from django.db import models
from django.contrib.auth.models import User
from picklefield import fields
from tagging.fields import TagField
import tagging
class OcrPreset(models.Model):
user = models.ForeignKey(User)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
public = models.BooleanField(default=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True, auto_now=True)
type = models.CharField(max_length=20,
choices=[("segment", "Segment"), ("binarize", "Binarize")])
data = fields.PickledObjectField()
def __unicode__(self):
"""
String representation.
"""
return "<%s: %s>" % (self.__class__.__name__, self.name)
|
ebdc3a1d00ddd8c15aaa64c436e43f3815317923 | pythainlp/segment/pyicu.py | pythainlp/segment/pyicu.py | from __future__ import absolute_import,print_function
from itertools import groupby
import PyICU
import six
# ตัดคำภาษาไทย
def segment(txt):
"""รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU"""
bd = PyICU.BreakIterator.createWordInstance(PyICU.Locale("th"))
bd.setText(six.u(txt))
breaks = list(bd)
return [txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)]
if __name__ == "__main__":
print(segment('ทดสอบระบบตัดคำด้วยไอซียู'))
print(segment('ผมชอบพูดไทยคำ English คำ'))
print(segment('ผมชอบพูดไทยคำEnglishคำ')) | from __future__ import absolute_import,print_function
from itertools import groupby
import PyICU
# ตัดคำภาษาไทย
def segment(txt):
"""รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU"""
bd = PyICU.BreakIterator.createWordInstance(PyICU.Locale("th"))
bd.setText(six.u(txt))
breaks = list(bd)
return [txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)]
if __name__ == "__main__":
print(segment('ทดสอบระบบตัดคำด้วยไอซียู'))
print(segment('ผมชอบพูดไทยคำ English คำ'))
print(segment('ผมชอบพูดไทยคำEnglishคำ')) | Revert "fix bug import six" | Revert "fix bug import six"
This reverts commit a80c1d7c80d68f72d435dbb7ac5c48a6114716fb.
| Python | apache-2.0 | PyThaiNLP/pythainlp | from __future__ import absolute_import,print_function
from itertools import groupby
import PyICU
import six
# ตัดคำภาษาไทย
def segment(txt):
"""รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU"""
bd = PyICU.BreakIterator.createWordInstance(PyICU.Locale("th"))
bd.setText(six.u(txt))
breaks = list(bd)
return [txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)]
if __name__ == "__main__":
print(segment('ทดสอบระบบตัดคำด้วยไอซียู'))
print(segment('ผมชอบพูดไทยคำ English คำ'))
print(segment('ผมชอบพูดไทยคำEnglishคำ'))Revert "fix bug import six"
This reverts commit a80c1d7c80d68f72d435dbb7ac5c48a6114716fb. | from __future__ import absolute_import,print_function
from itertools import groupby
import PyICU
# ตัดคำภาษาไทย
def segment(txt):
"""รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU"""
bd = PyICU.BreakIterator.createWordInstance(PyICU.Locale("th"))
bd.setText(six.u(txt))
breaks = list(bd)
return [txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)]
if __name__ == "__main__":
print(segment('ทดสอบระบบตัดคำด้วยไอซียู'))
print(segment('ผมชอบพูดไทยคำ English คำ'))
print(segment('ผมชอบพูดไทยคำEnglishคำ')) | <commit_before>from __future__ import absolute_import,print_function
from itertools import groupby
import PyICU
import six
# ตัดคำภาษาไทย
def segment(txt):
"""รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU"""
bd = PyICU.BreakIterator.createWordInstance(PyICU.Locale("th"))
bd.setText(six.u(txt))
breaks = list(bd)
return [txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)]
if __name__ == "__main__":
print(segment('ทดสอบระบบตัดคำด้วยไอซียู'))
print(segment('ผมชอบพูดไทยคำ English คำ'))
print(segment('ผมชอบพูดไทยคำEnglishคำ'))<commit_msg>Revert "fix bug import six"
This reverts commit a80c1d7c80d68f72d435dbb7ac5c48a6114716fb.<commit_after> | from __future__ import absolute_import,print_function
from itertools import groupby
import PyICU
# ตัดคำภาษาไทย
def segment(txt):
"""รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU"""
bd = PyICU.BreakIterator.createWordInstance(PyICU.Locale("th"))
bd.setText(six.u(txt))
breaks = list(bd)
return [txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)]
if __name__ == "__main__":
print(segment('ทดสอบระบบตัดคำด้วยไอซียู'))
print(segment('ผมชอบพูดไทยคำ English คำ'))
print(segment('ผมชอบพูดไทยคำEnglishคำ')) | from __future__ import absolute_import,print_function
from itertools import groupby
import PyICU
import six
# ตัดคำภาษาไทย
def segment(txt):
"""รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU"""
bd = PyICU.BreakIterator.createWordInstance(PyICU.Locale("th"))
bd.setText(six.u(txt))
breaks = list(bd)
return [txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)]
if __name__ == "__main__":
print(segment('ทดสอบระบบตัดคำด้วยไอซียู'))
print(segment('ผมชอบพูดไทยคำ English คำ'))
print(segment('ผมชอบพูดไทยคำEnglishคำ'))Revert "fix bug import six"
This reverts commit a80c1d7c80d68f72d435dbb7ac5c48a6114716fb.from __future__ import absolute_import,print_function
from itertools import groupby
import PyICU
# ตัดคำภาษาไทย
def segment(txt):
"""รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU"""
bd = PyICU.BreakIterator.createWordInstance(PyICU.Locale("th"))
bd.setText(six.u(txt))
breaks = list(bd)
return [txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)]
if __name__ == "__main__":
print(segment('ทดสอบระบบตัดคำด้วยไอซียู'))
print(segment('ผมชอบพูดไทยคำ English คำ'))
print(segment('ผมชอบพูดไทยคำEnglishคำ')) | <commit_before>from __future__ import absolute_import,print_function
from itertools import groupby
import PyICU
import six
# ตัดคำภาษาไทย
def segment(txt):
"""รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU"""
bd = PyICU.BreakIterator.createWordInstance(PyICU.Locale("th"))
bd.setText(six.u(txt))
breaks = list(bd)
return [txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)]
if __name__ == "__main__":
print(segment('ทดสอบระบบตัดคำด้วยไอซียู'))
print(segment('ผมชอบพูดไทยคำ English คำ'))
print(segment('ผมชอบพูดไทยคำEnglishคำ'))<commit_msg>Revert "fix bug import six"
This reverts commit a80c1d7c80d68f72d435dbb7ac5c48a6114716fb.<commit_after>from __future__ import absolute_import,print_function
from itertools import groupby
import PyICU
# ตัดคำภาษาไทย
def segment(txt):
"""รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU"""
bd = PyICU.BreakIterator.createWordInstance(PyICU.Locale("th"))
bd.setText(six.u(txt))
breaks = list(bd)
return [txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)]
if __name__ == "__main__":
print(segment('ทดสอบระบบตัดคำด้วยไอซียู'))
print(segment('ผมชอบพูดไทยคำ English คำ'))
print(segment('ผมชอบพูดไทยคำEnglishคำ')) |
f29665f853d1a33bcf08d1a9298460d0be11d610 | molly/apps/places/__init__.py | molly/apps/places/__init__.py | from flask import Blueprint
from flask.ext.babel import lazy_gettext as _
from molly.apps.common.app import BaseApp
from molly.apps.places.endpoints import PointOfInterestEndpoint
from molly.apps.places.services import PointsOfInterest
class App(BaseApp):
module = 'http://mollyproject.org/apps/places'
human_name = _('Places')
def __init__(self, instance_name, config, providers, services):
self.instance_name = instance_name
poi_service = PointsOfInterest(instance_name, services['kv'].db[instance_name])
for provider in providers:
provider.poi_service = poi_service
self._register_provider_as_importer(provider, services)
self._poi_endpoint = PointOfInterestEndpoint(instance_name, poi_service)
self.blueprint = Blueprint(self.instance_name, __name__)
self.blueprint.add_url_rule('/<slug>', 'poi', self._poi_endpoint.get)
self.links = []
| from flask import Blueprint
from flask.ext.babel import lazy_gettext as _
from molly.apps.common.app import BaseApp
from molly.apps.places.endpoints import PointOfInterestEndpoint
from molly.apps.places.services import PointsOfInterest
class App(BaseApp):
module = 'http://mollyproject.org/apps/places'
human_name = _('Places')
def __init__(self, instance_name, config, providers, services):
self.instance_name = instance_name
poi_service = PointsOfInterest(instance_name, services['kv'].db[instance_name])
for provider in providers:
provider.poi_service = poi_service
self._register_provider_as_importer(provider, services)
self._poi_endpoint = PointOfInterestEndpoint(instance_name, poi_service)
self.blueprint = Blueprint(self.instance_name, __name__)
self.blueprint.add_url_rule('/<slug>/', 'poi', self._poi_endpoint.get)
self.links = []
| Update URL rules to match Molly 1.x | Update URL rules to match Molly 1.x
| Python | apache-2.0 | ManchesterIO/mollyproject-next,ManchesterIO/mollyproject-next,ManchesterIO/mollyproject-next | from flask import Blueprint
from flask.ext.babel import lazy_gettext as _
from molly.apps.common.app import BaseApp
from molly.apps.places.endpoints import PointOfInterestEndpoint
from molly.apps.places.services import PointsOfInterest
class App(BaseApp):
module = 'http://mollyproject.org/apps/places'
human_name = _('Places')
def __init__(self, instance_name, config, providers, services):
self.instance_name = instance_name
poi_service = PointsOfInterest(instance_name, services['kv'].db[instance_name])
for provider in providers:
provider.poi_service = poi_service
self._register_provider_as_importer(provider, services)
self._poi_endpoint = PointOfInterestEndpoint(instance_name, poi_service)
self.blueprint = Blueprint(self.instance_name, __name__)
self.blueprint.add_url_rule('/<slug>', 'poi', self._poi_endpoint.get)
self.links = []
Update URL rules to match Molly 1.x | from flask import Blueprint
from flask.ext.babel import lazy_gettext as _
from molly.apps.common.app import BaseApp
from molly.apps.places.endpoints import PointOfInterestEndpoint
from molly.apps.places.services import PointsOfInterest
class App(BaseApp):
module = 'http://mollyproject.org/apps/places'
human_name = _('Places')
def __init__(self, instance_name, config, providers, services):
self.instance_name = instance_name
poi_service = PointsOfInterest(instance_name, services['kv'].db[instance_name])
for provider in providers:
provider.poi_service = poi_service
self._register_provider_as_importer(provider, services)
self._poi_endpoint = PointOfInterestEndpoint(instance_name, poi_service)
self.blueprint = Blueprint(self.instance_name, __name__)
self.blueprint.add_url_rule('/<slug>/', 'poi', self._poi_endpoint.get)
self.links = []
| <commit_before>from flask import Blueprint
from flask.ext.babel import lazy_gettext as _
from molly.apps.common.app import BaseApp
from molly.apps.places.endpoints import PointOfInterestEndpoint
from molly.apps.places.services import PointsOfInterest
class App(BaseApp):
module = 'http://mollyproject.org/apps/places'
human_name = _('Places')
def __init__(self, instance_name, config, providers, services):
self.instance_name = instance_name
poi_service = PointsOfInterest(instance_name, services['kv'].db[instance_name])
for provider in providers:
provider.poi_service = poi_service
self._register_provider_as_importer(provider, services)
self._poi_endpoint = PointOfInterestEndpoint(instance_name, poi_service)
self.blueprint = Blueprint(self.instance_name, __name__)
self.blueprint.add_url_rule('/<slug>', 'poi', self._poi_endpoint.get)
self.links = []
<commit_msg>Update URL rules to match Molly 1.x<commit_after> | from flask import Blueprint
from flask.ext.babel import lazy_gettext as _
from molly.apps.common.app import BaseApp
from molly.apps.places.endpoints import PointOfInterestEndpoint
from molly.apps.places.services import PointsOfInterest
class App(BaseApp):
module = 'http://mollyproject.org/apps/places'
human_name = _('Places')
def __init__(self, instance_name, config, providers, services):
self.instance_name = instance_name
poi_service = PointsOfInterest(instance_name, services['kv'].db[instance_name])
for provider in providers:
provider.poi_service = poi_service
self._register_provider_as_importer(provider, services)
self._poi_endpoint = PointOfInterestEndpoint(instance_name, poi_service)
self.blueprint = Blueprint(self.instance_name, __name__)
self.blueprint.add_url_rule('/<slug>/', 'poi', self._poi_endpoint.get)
self.links = []
| from flask import Blueprint
from flask.ext.babel import lazy_gettext as _
from molly.apps.common.app import BaseApp
from molly.apps.places.endpoints import PointOfInterestEndpoint
from molly.apps.places.services import PointsOfInterest
class App(BaseApp):
module = 'http://mollyproject.org/apps/places'
human_name = _('Places')
def __init__(self, instance_name, config, providers, services):
self.instance_name = instance_name
poi_service = PointsOfInterest(instance_name, services['kv'].db[instance_name])
for provider in providers:
provider.poi_service = poi_service
self._register_provider_as_importer(provider, services)
self._poi_endpoint = PointOfInterestEndpoint(instance_name, poi_service)
self.blueprint = Blueprint(self.instance_name, __name__)
self.blueprint.add_url_rule('/<slug>', 'poi', self._poi_endpoint.get)
self.links = []
Update URL rules to match Molly 1.xfrom flask import Blueprint
from flask.ext.babel import lazy_gettext as _
from molly.apps.common.app import BaseApp
from molly.apps.places.endpoints import PointOfInterestEndpoint
from molly.apps.places.services import PointsOfInterest
class App(BaseApp):
module = 'http://mollyproject.org/apps/places'
human_name = _('Places')
def __init__(self, instance_name, config, providers, services):
self.instance_name = instance_name
poi_service = PointsOfInterest(instance_name, services['kv'].db[instance_name])
for provider in providers:
provider.poi_service = poi_service
self._register_provider_as_importer(provider, services)
self._poi_endpoint = PointOfInterestEndpoint(instance_name, poi_service)
self.blueprint = Blueprint(self.instance_name, __name__)
self.blueprint.add_url_rule('/<slug>/', 'poi', self._poi_endpoint.get)
self.links = []
| <commit_before>from flask import Blueprint
from flask.ext.babel import lazy_gettext as _
from molly.apps.common.app import BaseApp
from molly.apps.places.endpoints import PointOfInterestEndpoint
from molly.apps.places.services import PointsOfInterest
class App(BaseApp):
module = 'http://mollyproject.org/apps/places'
human_name = _('Places')
def __init__(self, instance_name, config, providers, services):
self.instance_name = instance_name
poi_service = PointsOfInterest(instance_name, services['kv'].db[instance_name])
for provider in providers:
provider.poi_service = poi_service
self._register_provider_as_importer(provider, services)
self._poi_endpoint = PointOfInterestEndpoint(instance_name, poi_service)
self.blueprint = Blueprint(self.instance_name, __name__)
self.blueprint.add_url_rule('/<slug>', 'poi', self._poi_endpoint.get)
self.links = []
<commit_msg>Update URL rules to match Molly 1.x<commit_after>from flask import Blueprint
from flask.ext.babel import lazy_gettext as _
from molly.apps.common.app import BaseApp
from molly.apps.places.endpoints import PointOfInterestEndpoint
from molly.apps.places.services import PointsOfInterest
class App(BaseApp):
module = 'http://mollyproject.org/apps/places'
human_name = _('Places')
def __init__(self, instance_name, config, providers, services):
self.instance_name = instance_name
poi_service = PointsOfInterest(instance_name, services['kv'].db[instance_name])
for provider in providers:
provider.poi_service = poi_service
self._register_provider_as_importer(provider, services)
self._poi_endpoint = PointOfInterestEndpoint(instance_name, poi_service)
self.blueprint = Blueprint(self.instance_name, __name__)
self.blueprint.add_url_rule('/<slug>/', 'poi', self._poi_endpoint.get)
self.links = []
|
c7455da1b0092e926ed9dafe5ac5ae1335401dba | admin.py | admin.py | from django.contrib import admin
from django.contrib.sites.models import Site
from .models import Church
admin.site.site_header = "Churches of Bridlington Administration"
admin.site.register(Church)
admin.site.unregister(Site)
| from django.contrib import admin
from django.contrib.sites.models import Site
from .models import Church
admin.site.site_header = "Churches of Bridlington Administration"
admin.site.register(Church)
| Undo deregistration of Site object | Undo deregistration of Site object
This will now be controlled by restricting permissions in the
admin.
| Python | mit | bm424/churchmanager,bm424/churchmanager | from django.contrib import admin
from django.contrib.sites.models import Site
from .models import Church
admin.site.site_header = "Churches of Bridlington Administration"
admin.site.register(Church)
admin.site.unregister(Site)
Undo deregistration of Site object
This will now be controlled by restricting permissions in the
admin. | from django.contrib import admin
from django.contrib.sites.models import Site
from .models import Church
admin.site.site_header = "Churches of Bridlington Administration"
admin.site.register(Church)
| <commit_before>from django.contrib import admin
from django.contrib.sites.models import Site
from .models import Church
admin.site.site_header = "Churches of Bridlington Administration"
admin.site.register(Church)
admin.site.unregister(Site)
<commit_msg>Undo deregistration of Site object
This will now be controlled by restricting permissions in the
admin.<commit_after> | from django.contrib import admin
from django.contrib.sites.models import Site
from .models import Church
admin.site.site_header = "Churches of Bridlington Administration"
admin.site.register(Church)
| from django.contrib import admin
from django.contrib.sites.models import Site
from .models import Church
admin.site.site_header = "Churches of Bridlington Administration"
admin.site.register(Church)
admin.site.unregister(Site)
Undo deregistration of Site object
This will now be controlled by restricting permissions in the
admin.from django.contrib import admin
from django.contrib.sites.models import Site
from .models import Church
admin.site.site_header = "Churches of Bridlington Administration"
admin.site.register(Church)
| <commit_before>from django.contrib import admin
from django.contrib.sites.models import Site
from .models import Church
admin.site.site_header = "Churches of Bridlington Administration"
admin.site.register(Church)
admin.site.unregister(Site)
<commit_msg>Undo deregistration of Site object
This will now be controlled by restricting permissions in the
admin.<commit_after>from django.contrib import admin
from django.contrib.sites.models import Site
from .models import Church
admin.site.site_header = "Churches of Bridlington Administration"
admin.site.register(Church)
|
dee0b3764259ee7f4916e8e5e303c48afb3e5edd | api/base/urls.py | api/base/urls.py | from django.conf import settings
from django.conf.urls import include, url
# from django.contrib import admin
from django.conf.urls.static import static
from . import views
urlpatterns = [
### API ###
url(r'^$', views.root),
url(r'^nodes/', include('api.nodes.urls', namespace='nodes')),
url(r'^users/', include('api.users.urls', namespace='users')),
url(r'^docs/', include('rest_framework_swagger.urls')),
] + static('/static/', document_root=settings.STATIC_ROOT) | from django.conf import settings
from django.conf.urls import include, url, patterns
# from django.contrib import admin
from django.conf.urls.static import static
from . import views
urlpatterns = [
### API ###
url(r'^v2/', include(patterns('',
url(r'^$', views.root),
url(r'^nodes/', include('api.nodes.urls', namespace='nodes')),
url(r'^users/', include('api.users.urls', namespace='users')),
url(r'^docs/', include('rest_framework_swagger.urls')),
)))] + static('/static/', document_root=settings.STATIC_ROOT) | Change API url prefix to 'v2' | Change API url prefix to 'v2'
| Python | apache-2.0 | laurenrevere/osf.io,Johnetordoff/osf.io,CenterForOpenScience/osf.io,cosenal/osf.io,reinaH/osf.io,binoculars/osf.io,CenterForOpenScience/osf.io,mluo613/osf.io,aaxelb/osf.io,MerlinZhang/osf.io,TomHeatwole/osf.io,felliott/osf.io,sloria/osf.io,dplorimer/osf,wearpants/osf.io,chrisseto/osf.io,zamattiac/osf.io,wearpants/osf.io,brianjgeiger/osf.io,KAsante95/osf.io,SSJohns/osf.io,cwisecarver/osf.io,cslzchen/osf.io,HarryRybacki/osf.io,TomBaxter/osf.io,cldershem/osf.io,alexschiller/osf.io,chrisseto/osf.io,rdhyee/osf.io,SSJohns/osf.io,jolene-esposito/osf.io,saradbowman/osf.io,alexschiller/osf.io,emetsger/osf.io,Ghalko/osf.io,MerlinZhang/osf.io,haoyuchen1992/osf.io,pattisdr/osf.io,jmcarp/osf.io,billyhunt/osf.io,GageGaskins/osf.io,hmoco/osf.io,acshi/osf.io,acshi/osf.io,CenterForOpenScience/osf.io,kch8qx/osf.io,monikagrabowska/osf.io,amyshi188/osf.io,samchrisinger/osf.io,ZobairAlijan/osf.io,jmcarp/osf.io,cosenal/osf.io,njantrania/osf.io,erinspace/osf.io,icereval/osf.io,hmoco/osf.io,Johnetordoff/osf.io,fabianvf/osf.io,samanehsan/osf.io,brianjgeiger/osf.io,bdyetton/prettychart,petermalcolm/osf.io,acshi/osf.io,zachjanicki/osf.io,caseyrygt/osf.io,chrisseto/osf.io,DanielSBrown/osf.io,caseyrygt/osf.io,CenterForOpenScience/osf.io,mattclark/osf.io,kch8qx/osf.io,barbour-em/osf.io,KAsante95/osf.io,TomBaxter/osf.io,baylee-d/osf.io,MerlinZhang/osf.io,doublebits/osf.io,felliott/osf.io,mluo613/osf.io,alexschiller/osf.io,bdyetton/prettychart,leb2dg/osf.io,doublebits/osf.io,lyndsysimon/osf.io,adlius/osf.io,zamattiac/osf.io,jolene-esposito/osf.io,brandonPurvis/osf.io,TomHeatwole/osf.io,felliott/osf.io,danielneis/osf.io,wearpants/osf.io,KAsante95/osf.io,icereval/osf.io,TomBaxter/osf.io,samchrisinger/osf.io,adlius/osf.io,jnayak1/osf.io,cldershem/osf.io,mluke93/osf.io,caneruguz/osf.io,leb2dg/osf.io,icereval/osf.io,jeffreyliu3230/osf.io,sbt9uc/osf.io,felliott/osf.io,njantrania/osf.io,HalcyonChimera/osf.io,ckc6cz/osf.io,Johnetordoff/osf.io,barbour-em/osf.io,kch8qx/osf.io,haoyuchen1992/osf.io,cslzchen/osf.io,rdhyee/osf.io,chennan47/osf.io,ckc6cz/osf.io,jnayak1/osf.io,lyndsysimon/osf.io,zachjanicki/osf.io,mfraezz/osf.io,abought/osf.io,caneruguz/osf.io,sbt9uc/osf.io,samanehsan/osf.io,jinluyuan/osf.io,pattisdr/osf.io,mfraezz/osf.io,DanielSBrown/osf.io,lyndsysimon/osf.io,alexschiller/osf.io,abought/osf.io,samanehsan/osf.io,kch8qx/osf.io,ticklemepierce/osf.io,ckc6cz/osf.io,mfraezz/osf.io,cosenal/osf.io,haoyuchen1992/osf.io,asanfilippo7/osf.io,doublebits/osf.io,mfraezz/osf.io,reinaH/osf.io,ZobairAlijan/osf.io,barbour-em/osf.io,caseyrygt/osf.io,cwisecarver/osf.io,cldershem/osf.io,petermalcolm/osf.io,adlius/osf.io,adlius/osf.io,lyndsysimon/osf.io,mattclark/osf.io,chrisseto/osf.io,alexschiller/osf.io,HalcyonChimera/osf.io,brandonPurvis/osf.io,crcresearch/osf.io,cslzchen/osf.io,mluo613/osf.io,acshi/osf.io,danielneis/osf.io,chennan47/osf.io,fabianvf/osf.io,wearpants/osf.io,jnayak1/osf.io,petermalcolm/osf.io,Nesiehr/osf.io,billyhunt/osf.io,TomHeatwole/osf.io,jolene-esposito/osf.io,njantrania/osf.io,laurenrevere/osf.io,HarryRybacki/osf.io,mluo613/osf.io,doublebits/osf.io,rdhyee/osf.io,rdhyee/osf.io,fabianvf/osf.io,monikagrabowska/osf.io,jinluyuan/osf.io,monikagrabowska/osf.io,Ghalko/osf.io,ticklemepierce/osf.io,HarryRybacki/osf.io,erinspace/osf.io,asanfilippo7/osf.io,SSJohns/osf.io,petermalcolm/osf.io,asanfilippo7/osf.io,arpitar/osf.io,jeffreyliu3230/osf.io,RomanZWang/osf.io,kwierman/osf.io,zachjanicki/osf.io,arpitar/osf.io,samanehsan/osf.io,cosenal/osf.io,MerlinZhang/osf.io,leb2dg/osf.io,arpitar/osf.io,kch8qx/osf.io,cslzchen/osf.io,mluke93/osf.io,sloria/osf.io,GageGaskins/osf.io,caseyrollins/osf.io,danielneis/osf.io,pattisdr/osf.io,SSJohns/osf.io,njantrania/osf.io,sbt9uc/osf.io,kwierman/osf.io,asanfilippo7/osf.io,amyshi188/osf.io,emetsger/osf.io,abought/osf.io,crcresearch/osf.io,jmcarp/osf.io,TomHeatwole/osf.io,binoculars/osf.io,emetsger/osf.io,amyshi188/osf.io,cwisecarver/osf.io,ZobairAlijan/osf.io,laurenrevere/osf.io,RomanZWang/osf.io,kwierman/osf.io,billyhunt/osf.io,billyhunt/osf.io,doublebits/osf.io,jinluyuan/osf.io,brandonPurvis/osf.io,reinaH/osf.io,haoyuchen1992/osf.io,GageGaskins/osf.io,aaxelb/osf.io,fabianvf/osf.io,hmoco/osf.io,caseyrollins/osf.io,brandonPurvis/osf.io,mluo613/osf.io,danielneis/osf.io,jeffreyliu3230/osf.io,bdyetton/prettychart,jmcarp/osf.io,chennan47/osf.io,sloria/osf.io,caneruguz/osf.io,brianjgeiger/osf.io,bdyetton/prettychart,monikagrabowska/osf.io,Nesiehr/osf.io,HarryRybacki/osf.io,Ghalko/osf.io,GageGaskins/osf.io,aaxelb/osf.io,RomanZWang/osf.io,zamattiac/osf.io,brandonPurvis/osf.io,erinspace/osf.io,emetsger/osf.io,jinluyuan/osf.io,caseyrygt/osf.io,RomanZWang/osf.io,cwisecarver/osf.io,kwierman/osf.io,mluke93/osf.io,zamattiac/osf.io,acshi/osf.io,samchrisinger/osf.io,RomanZWang/osf.io,reinaH/osf.io,KAsante95/osf.io,leb2dg/osf.io,HalcyonChimera/osf.io,Johnetordoff/osf.io,samchrisinger/osf.io,cldershem/osf.io,Ghalko/osf.io,caneruguz/osf.io,baylee-d/osf.io,crcresearch/osf.io,binoculars/osf.io,zachjanicki/osf.io,KAsante95/osf.io,barbour-em/osf.io,amyshi188/osf.io,arpitar/osf.io,DanielSBrown/osf.io,dplorimer/osf,ckc6cz/osf.io,jeffreyliu3230/osf.io,ticklemepierce/osf.io,aaxelb/osf.io,Nesiehr/osf.io,mluke93/osf.io,brianjgeiger/osf.io,sbt9uc/osf.io,ZobairAlijan/osf.io,mattclark/osf.io,jolene-esposito/osf.io,Nesiehr/osf.io,GageGaskins/osf.io,abought/osf.io,caseyrollins/osf.io,DanielSBrown/osf.io,dplorimer/osf,baylee-d/osf.io,jnayak1/osf.io,ticklemepierce/osf.io,billyhunt/osf.io,monikagrabowska/osf.io,HalcyonChimera/osf.io,dplorimer/osf,saradbowman/osf.io,hmoco/osf.io | from django.conf import settings
from django.conf.urls import include, url
# from django.contrib import admin
from django.conf.urls.static import static
from . import views
urlpatterns = [
### API ###
url(r'^$', views.root),
url(r'^nodes/', include('api.nodes.urls', namespace='nodes')),
url(r'^users/', include('api.users.urls', namespace='users')),
url(r'^docs/', include('rest_framework_swagger.urls')),
] + static('/static/', document_root=settings.STATIC_ROOT)Change API url prefix to 'v2' | from django.conf import settings
from django.conf.urls import include, url, patterns
# from django.contrib import admin
from django.conf.urls.static import static
from . import views
urlpatterns = [
### API ###
url(r'^v2/', include(patterns('',
url(r'^$', views.root),
url(r'^nodes/', include('api.nodes.urls', namespace='nodes')),
url(r'^users/', include('api.users.urls', namespace='users')),
url(r'^docs/', include('rest_framework_swagger.urls')),
)))] + static('/static/', document_root=settings.STATIC_ROOT) | <commit_before>from django.conf import settings
from django.conf.urls import include, url
# from django.contrib import admin
from django.conf.urls.static import static
from . import views
urlpatterns = [
### API ###
url(r'^$', views.root),
url(r'^nodes/', include('api.nodes.urls', namespace='nodes')),
url(r'^users/', include('api.users.urls', namespace='users')),
url(r'^docs/', include('rest_framework_swagger.urls')),
] + static('/static/', document_root=settings.STATIC_ROOT)<commit_msg>Change API url prefix to 'v2'<commit_after> | from django.conf import settings
from django.conf.urls import include, url, patterns
# from django.contrib import admin
from django.conf.urls.static import static
from . import views
urlpatterns = [
### API ###
url(r'^v2/', include(patterns('',
url(r'^$', views.root),
url(r'^nodes/', include('api.nodes.urls', namespace='nodes')),
url(r'^users/', include('api.users.urls', namespace='users')),
url(r'^docs/', include('rest_framework_swagger.urls')),
)))] + static('/static/', document_root=settings.STATIC_ROOT) | from django.conf import settings
from django.conf.urls import include, url
# from django.contrib import admin
from django.conf.urls.static import static
from . import views
urlpatterns = [
### API ###
url(r'^$', views.root),
url(r'^nodes/', include('api.nodes.urls', namespace='nodes')),
url(r'^users/', include('api.users.urls', namespace='users')),
url(r'^docs/', include('rest_framework_swagger.urls')),
] + static('/static/', document_root=settings.STATIC_ROOT)Change API url prefix to 'v2'from django.conf import settings
from django.conf.urls import include, url, patterns
# from django.contrib import admin
from django.conf.urls.static import static
from . import views
urlpatterns = [
### API ###
url(r'^v2/', include(patterns('',
url(r'^$', views.root),
url(r'^nodes/', include('api.nodes.urls', namespace='nodes')),
url(r'^users/', include('api.users.urls', namespace='users')),
url(r'^docs/', include('rest_framework_swagger.urls')),
)))] + static('/static/', document_root=settings.STATIC_ROOT) | <commit_before>from django.conf import settings
from django.conf.urls import include, url
# from django.contrib import admin
from django.conf.urls.static import static
from . import views
urlpatterns = [
### API ###
url(r'^$', views.root),
url(r'^nodes/', include('api.nodes.urls', namespace='nodes')),
url(r'^users/', include('api.users.urls', namespace='users')),
url(r'^docs/', include('rest_framework_swagger.urls')),
] + static('/static/', document_root=settings.STATIC_ROOT)<commit_msg>Change API url prefix to 'v2'<commit_after>from django.conf import settings
from django.conf.urls import include, url, patterns
# from django.contrib import admin
from django.conf.urls.static import static
from . import views
urlpatterns = [
### API ###
url(r'^v2/', include(patterns('',
url(r'^$', views.root),
url(r'^nodes/', include('api.nodes.urls', namespace='nodes')),
url(r'^users/', include('api.users.urls', namespace='users')),
url(r'^docs/', include('rest_framework_swagger.urls')),
)))] + static('/static/', document_root=settings.STATIC_ROOT) |
f70574c38140c9a5493981f5baf72bab82be8c60 | opps/articles/tests/models.py | opps/articles/tests/models.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def setUP(self):
self.post = Post.objects.get(id=1)
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertTrue(post[0], self.post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
self.assertTrue(self.post.child_class)
self.assertEqual(self.post.child_class, 'Post')
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def setUp(self):
self.post = Post.objects.get(id=1)
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertTrue(post[0], self.post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
self.assertTrue(self.post.child_class)
self.assertEqual(self.post.child_class, 'Post')
def test_get_absolute_url(self):
self.assertEqual(self.post.get_absolute_url(),
u'/channel-01/test-post-application')
self.assertEqual(self.post.get_absolute_url(),
"/{0}/{1}".format(self.post.channel.long_slug,
self.post.slug))
| Add test articles, post get absolute url | Add test articles, post get absolute url
| Python | mit | YACOWS/opps,YACOWS/opps,YACOWS/opps,jeanmask/opps,williamroot/opps,opps/opps,jeanmask/opps,williamroot/opps,YACOWS/opps,opps/opps,williamroot/opps,jeanmask/opps,opps/opps,jeanmask/opps,opps/opps,williamroot/opps | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def setUP(self):
self.post = Post.objects.get(id=1)
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertTrue(post[0], self.post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
self.assertTrue(self.post.child_class)
self.assertEqual(self.post.child_class, 'Post')
Add test articles, post get absolute url | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def setUp(self):
self.post = Post.objects.get(id=1)
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertTrue(post[0], self.post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
self.assertTrue(self.post.child_class)
self.assertEqual(self.post.child_class, 'Post')
def test_get_absolute_url(self):
self.assertEqual(self.post.get_absolute_url(),
u'/channel-01/test-post-application')
self.assertEqual(self.post.get_absolute_url(),
"/{0}/{1}".format(self.post.channel.long_slug,
self.post.slug))
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def setUP(self):
self.post = Post.objects.get(id=1)
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertTrue(post[0], self.post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
self.assertTrue(self.post.child_class)
self.assertEqual(self.post.child_class, 'Post')
<commit_msg>Add test articles, post get absolute url<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def setUp(self):
self.post = Post.objects.get(id=1)
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertTrue(post[0], self.post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
self.assertTrue(self.post.child_class)
self.assertEqual(self.post.child_class, 'Post')
def test_get_absolute_url(self):
self.assertEqual(self.post.get_absolute_url(),
u'/channel-01/test-post-application')
self.assertEqual(self.post.get_absolute_url(),
"/{0}/{1}".format(self.post.channel.long_slug,
self.post.slug))
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def setUP(self):
self.post = Post.objects.get(id=1)
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertTrue(post[0], self.post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
self.assertTrue(self.post.child_class)
self.assertEqual(self.post.child_class, 'Post')
Add test articles, post get absolute url#!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def setUp(self):
self.post = Post.objects.get(id=1)
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertTrue(post[0], self.post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
self.assertTrue(self.post.child_class)
self.assertEqual(self.post.child_class, 'Post')
def test_get_absolute_url(self):
self.assertEqual(self.post.get_absolute_url(),
u'/channel-01/test-post-application')
self.assertEqual(self.post.get_absolute_url(),
"/{0}/{1}".format(self.post.channel.long_slug,
self.post.slug))
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def setUP(self):
self.post = Post.objects.get(id=1)
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertTrue(post[0], self.post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
self.assertTrue(self.post.child_class)
self.assertEqual(self.post.child_class, 'Post')
<commit_msg>Add test articles, post get absolute url<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def setUp(self):
self.post = Post.objects.get(id=1)
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertTrue(post[0], self.post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
self.assertTrue(self.post.child_class)
self.assertEqual(self.post.child_class, 'Post')
def test_get_absolute_url(self):
self.assertEqual(self.post.get_absolute_url(),
u'/channel-01/test-post-application')
self.assertEqual(self.post.get_absolute_url(),
"/{0}/{1}".format(self.post.channel.long_slug,
self.post.slug))
|
3eee55236a709e2929ffab7f15b8e50d541ed9a7 | utils/graph.py | utils/graph.py | """
This module serves as an interface to
matplotlib.
"""
from utils import config
OFFSET = 2 # offset = max_x/stepsize * OFFSET
def init(output):
import matplotlib
config.mpl(matplotlib, bool(output))
from matplotlib import pyplot
globals()['plt'] = pyplot
def line_plot(xs, ys, color='red'):
plt.plot(
xs,
ys,
color=color,
linewidth=2.0
)
def legend(*args):
plt.legend(args, loc='best')
def scatter_plot(x, y, color='red'):
plt.scatter(x, y, color=color)
def scale_plot(max_x, stepsize):
offset = max_x/stepsize * OFFSET
plt.axis(xmin=-offset, xmax=max_x+offset, ymin=0)
def prepare_plot(xlabel, ylabel, title):
plt.xlabel(xlabel)
plt.ylabel(ylabel)
plt.title(title)
plt.grid(True)
def display_plot(output):
if output:
if output == 'stdout':
plt.savefig(sys.stdout, format='png')
else:
plt.savefig(output)
else:
plt.show()
| """
This module serves as an interface to
matplotlib.
"""
from utils import config
OFFSET = 2 # offset = max_x/stepsize * OFFSET
def init(output):
import matplotlib
config.mpl(matplotlib, bool(output))
from matplotlib import pyplot
globals()['plt'] = pyplot
def line_plot(xs, ys, color='red'):
plt.plot(
xs,
ys,
color=color,
linewidth=2.0
)
def legend(*args):
plt.legend(args, loc='best')
def scatter_plot(x, y, color='blue'):
plt.scatter(x, y, color=color)
def scale_plot(max_x, stepsize):
offset = max_x/stepsize * OFFSET
plt.axis(xmin=-offset, xmax=max_x+offset, ymin=0)
def prepare_plot(xlabel, ylabel, title):
plt.xlabel(xlabel)
plt.ylabel(ylabel)
plt.title(title)
plt.grid(True)
def display_plot(output):
if output:
if output == 'stdout':
plt.savefig(sys.stdout, format='png')
else:
plt.savefig(output)
else:
plt.show()
| Add a different color for scatter plots to differentiate from line. | Add a different color for scatter plots to differentiate from line.
| Python | mit | wei2912/bce-simulation,wei2912/bce-simulation,wei2912/bce-simulation,wei2912/bce-simulation | """
This module serves as an interface to
matplotlib.
"""
from utils import config
OFFSET = 2 # offset = max_x/stepsize * OFFSET
def init(output):
import matplotlib
config.mpl(matplotlib, bool(output))
from matplotlib import pyplot
globals()['plt'] = pyplot
def line_plot(xs, ys, color='red'):
plt.plot(
xs,
ys,
color=color,
linewidth=2.0
)
def legend(*args):
plt.legend(args, loc='best')
def scatter_plot(x, y, color='red'):
plt.scatter(x, y, color=color)
def scale_plot(max_x, stepsize):
offset = max_x/stepsize * OFFSET
plt.axis(xmin=-offset, xmax=max_x+offset, ymin=0)
def prepare_plot(xlabel, ylabel, title):
plt.xlabel(xlabel)
plt.ylabel(ylabel)
plt.title(title)
plt.grid(True)
def display_plot(output):
if output:
if output == 'stdout':
plt.savefig(sys.stdout, format='png')
else:
plt.savefig(output)
else:
plt.show()
Add a different color for scatter plots to differentiate from line. | """
This module serves as an interface to
matplotlib.
"""
from utils import config
OFFSET = 2 # offset = max_x/stepsize * OFFSET
def init(output):
import matplotlib
config.mpl(matplotlib, bool(output))
from matplotlib import pyplot
globals()['plt'] = pyplot
def line_plot(xs, ys, color='red'):
plt.plot(
xs,
ys,
color=color,
linewidth=2.0
)
def legend(*args):
plt.legend(args, loc='best')
def scatter_plot(x, y, color='blue'):
plt.scatter(x, y, color=color)
def scale_plot(max_x, stepsize):
offset = max_x/stepsize * OFFSET
plt.axis(xmin=-offset, xmax=max_x+offset, ymin=0)
def prepare_plot(xlabel, ylabel, title):
plt.xlabel(xlabel)
plt.ylabel(ylabel)
plt.title(title)
plt.grid(True)
def display_plot(output):
if output:
if output == 'stdout':
plt.savefig(sys.stdout, format='png')
else:
plt.savefig(output)
else:
plt.show()
| <commit_before>"""
This module serves as an interface to
matplotlib.
"""
from utils import config
OFFSET = 2 # offset = max_x/stepsize * OFFSET
def init(output):
import matplotlib
config.mpl(matplotlib, bool(output))
from matplotlib import pyplot
globals()['plt'] = pyplot
def line_plot(xs, ys, color='red'):
plt.plot(
xs,
ys,
color=color,
linewidth=2.0
)
def legend(*args):
plt.legend(args, loc='best')
def scatter_plot(x, y, color='red'):
plt.scatter(x, y, color=color)
def scale_plot(max_x, stepsize):
offset = max_x/stepsize * OFFSET
plt.axis(xmin=-offset, xmax=max_x+offset, ymin=0)
def prepare_plot(xlabel, ylabel, title):
plt.xlabel(xlabel)
plt.ylabel(ylabel)
plt.title(title)
plt.grid(True)
def display_plot(output):
if output:
if output == 'stdout':
plt.savefig(sys.stdout, format='png')
else:
plt.savefig(output)
else:
plt.show()
<commit_msg>Add a different color for scatter plots to differentiate from line.<commit_after> | """
This module serves as an interface to
matplotlib.
"""
from utils import config
OFFSET = 2 # offset = max_x/stepsize * OFFSET
def init(output):
import matplotlib
config.mpl(matplotlib, bool(output))
from matplotlib import pyplot
globals()['plt'] = pyplot
def line_plot(xs, ys, color='red'):
plt.plot(
xs,
ys,
color=color,
linewidth=2.0
)
def legend(*args):
plt.legend(args, loc='best')
def scatter_plot(x, y, color='blue'):
plt.scatter(x, y, color=color)
def scale_plot(max_x, stepsize):
offset = max_x/stepsize * OFFSET
plt.axis(xmin=-offset, xmax=max_x+offset, ymin=0)
def prepare_plot(xlabel, ylabel, title):
plt.xlabel(xlabel)
plt.ylabel(ylabel)
plt.title(title)
plt.grid(True)
def display_plot(output):
if output:
if output == 'stdout':
plt.savefig(sys.stdout, format='png')
else:
plt.savefig(output)
else:
plt.show()
| """
This module serves as an interface to
matplotlib.
"""
from utils import config
OFFSET = 2 # offset = max_x/stepsize * OFFSET
def init(output):
import matplotlib
config.mpl(matplotlib, bool(output))
from matplotlib import pyplot
globals()['plt'] = pyplot
def line_plot(xs, ys, color='red'):
plt.plot(
xs,
ys,
color=color,
linewidth=2.0
)
def legend(*args):
plt.legend(args, loc='best')
def scatter_plot(x, y, color='red'):
plt.scatter(x, y, color=color)
def scale_plot(max_x, stepsize):
offset = max_x/stepsize * OFFSET
plt.axis(xmin=-offset, xmax=max_x+offset, ymin=0)
def prepare_plot(xlabel, ylabel, title):
plt.xlabel(xlabel)
plt.ylabel(ylabel)
plt.title(title)
plt.grid(True)
def display_plot(output):
if output:
if output == 'stdout':
plt.savefig(sys.stdout, format='png')
else:
plt.savefig(output)
else:
plt.show()
Add a different color for scatter plots to differentiate from line."""
This module serves as an interface to
matplotlib.
"""
from utils import config
OFFSET = 2 # offset = max_x/stepsize * OFFSET
def init(output):
import matplotlib
config.mpl(matplotlib, bool(output))
from matplotlib import pyplot
globals()['plt'] = pyplot
def line_plot(xs, ys, color='red'):
plt.plot(
xs,
ys,
color=color,
linewidth=2.0
)
def legend(*args):
plt.legend(args, loc='best')
def scatter_plot(x, y, color='blue'):
plt.scatter(x, y, color=color)
def scale_plot(max_x, stepsize):
offset = max_x/stepsize * OFFSET
plt.axis(xmin=-offset, xmax=max_x+offset, ymin=0)
def prepare_plot(xlabel, ylabel, title):
plt.xlabel(xlabel)
plt.ylabel(ylabel)
plt.title(title)
plt.grid(True)
def display_plot(output):
if output:
if output == 'stdout':
plt.savefig(sys.stdout, format='png')
else:
plt.savefig(output)
else:
plt.show()
| <commit_before>"""
This module serves as an interface to
matplotlib.
"""
from utils import config
OFFSET = 2 # offset = max_x/stepsize * OFFSET
def init(output):
import matplotlib
config.mpl(matplotlib, bool(output))
from matplotlib import pyplot
globals()['plt'] = pyplot
def line_plot(xs, ys, color='red'):
plt.plot(
xs,
ys,
color=color,
linewidth=2.0
)
def legend(*args):
plt.legend(args, loc='best')
def scatter_plot(x, y, color='red'):
plt.scatter(x, y, color=color)
def scale_plot(max_x, stepsize):
offset = max_x/stepsize * OFFSET
plt.axis(xmin=-offset, xmax=max_x+offset, ymin=0)
def prepare_plot(xlabel, ylabel, title):
plt.xlabel(xlabel)
plt.ylabel(ylabel)
plt.title(title)
plt.grid(True)
def display_plot(output):
if output:
if output == 'stdout':
plt.savefig(sys.stdout, format='png')
else:
plt.savefig(output)
else:
plt.show()
<commit_msg>Add a different color for scatter plots to differentiate from line.<commit_after>"""
This module serves as an interface to
matplotlib.
"""
from utils import config
OFFSET = 2 # offset = max_x/stepsize * OFFSET
def init(output):
import matplotlib
config.mpl(matplotlib, bool(output))
from matplotlib import pyplot
globals()['plt'] = pyplot
def line_plot(xs, ys, color='red'):
plt.plot(
xs,
ys,
color=color,
linewidth=2.0
)
def legend(*args):
plt.legend(args, loc='best')
def scatter_plot(x, y, color='blue'):
plt.scatter(x, y, color=color)
def scale_plot(max_x, stepsize):
offset = max_x/stepsize * OFFSET
plt.axis(xmin=-offset, xmax=max_x+offset, ymin=0)
def prepare_plot(xlabel, ylabel, title):
plt.xlabel(xlabel)
plt.ylabel(ylabel)
plt.title(title)
plt.grid(True)
def display_plot(output):
if output:
if output == 'stdout':
plt.savefig(sys.stdout, format='png')
else:
plt.savefig(output)
else:
plt.show()
|
dcc810f3181ebe358481c30c2248d25511aab26c | npz_to_my5c.py | npz_to_my5c.py | import numpy as np
import argparse
import sys
import pandas as pd
parser = argparse.ArgumentParser(description='Calculate coverage over different window sizes for a list of bam files.')
parser.add_argument('-n', '--npz_frequencies_file', required=True, help='An npz file containing co-segregation frequencies to convert to correlations')
args = parser.parse_args()
def open_npz(fp):
handle = np.load(fp)
return handle['windows'], handle['scores']
windows, data = open_npz(args.npz_frequencies_file)
names = [ '{}:{}-{}'.format(*i) for i in windows ]
pd.DataFrame(data, index=names, columns=names).to_csv(sys.stdout, sep='\t')
| import numpy as np
import argparse
import sys
import pandas as pd
parser = argparse.ArgumentParser(description='Calculate coverage over different window sizes for a list of bam files.')
parser.add_argument('-n', '--npz_frequencies_file', required=True, help='An npz file containing co-segregation frequencies to convert to correlations')
args = parser.parse_args()
def open_npz(fp):
handle = np.load(fp)
return handle['windows'], handle['scores']
windows, data = open_npz(args.npz_frequencies_file)
names = [ '{}:{}-{}'.format(*i) for i in windows ]
pd.DataFrame(data, index=names, columns=names).to_csv(sys.stdout, sep='\t', na_rep="NaN")
| Handle NaNs properly when flattening matrices. | Handle NaNs properly when flattening matrices.
| Python | apache-2.0 | pombo-lab/gamtools,pombo-lab/gamtools | import numpy as np
import argparse
import sys
import pandas as pd
parser = argparse.ArgumentParser(description='Calculate coverage over different window sizes for a list of bam files.')
parser.add_argument('-n', '--npz_frequencies_file', required=True, help='An npz file containing co-segregation frequencies to convert to correlations')
args = parser.parse_args()
def open_npz(fp):
handle = np.load(fp)
return handle['windows'], handle['scores']
windows, data = open_npz(args.npz_frequencies_file)
names = [ '{}:{}-{}'.format(*i) for i in windows ]
pd.DataFrame(data, index=names, columns=names).to_csv(sys.stdout, sep='\t')
Handle NaNs properly when flattening matrices. | import numpy as np
import argparse
import sys
import pandas as pd
parser = argparse.ArgumentParser(description='Calculate coverage over different window sizes for a list of bam files.')
parser.add_argument('-n', '--npz_frequencies_file', required=True, help='An npz file containing co-segregation frequencies to convert to correlations')
args = parser.parse_args()
def open_npz(fp):
handle = np.load(fp)
return handle['windows'], handle['scores']
windows, data = open_npz(args.npz_frequencies_file)
names = [ '{}:{}-{}'.format(*i) for i in windows ]
pd.DataFrame(data, index=names, columns=names).to_csv(sys.stdout, sep='\t', na_rep="NaN")
| <commit_before>import numpy as np
import argparse
import sys
import pandas as pd
parser = argparse.ArgumentParser(description='Calculate coverage over different window sizes for a list of bam files.')
parser.add_argument('-n', '--npz_frequencies_file', required=True, help='An npz file containing co-segregation frequencies to convert to correlations')
args = parser.parse_args()
def open_npz(fp):
handle = np.load(fp)
return handle['windows'], handle['scores']
windows, data = open_npz(args.npz_frequencies_file)
names = [ '{}:{}-{}'.format(*i) for i in windows ]
pd.DataFrame(data, index=names, columns=names).to_csv(sys.stdout, sep='\t')
<commit_msg>Handle NaNs properly when flattening matrices.<commit_after> | import numpy as np
import argparse
import sys
import pandas as pd
parser = argparse.ArgumentParser(description='Calculate coverage over different window sizes for a list of bam files.')
parser.add_argument('-n', '--npz_frequencies_file', required=True, help='An npz file containing co-segregation frequencies to convert to correlations')
args = parser.parse_args()
def open_npz(fp):
handle = np.load(fp)
return handle['windows'], handle['scores']
windows, data = open_npz(args.npz_frequencies_file)
names = [ '{}:{}-{}'.format(*i) for i in windows ]
pd.DataFrame(data, index=names, columns=names).to_csv(sys.stdout, sep='\t', na_rep="NaN")
| import numpy as np
import argparse
import sys
import pandas as pd
parser = argparse.ArgumentParser(description='Calculate coverage over different window sizes for a list of bam files.')
parser.add_argument('-n', '--npz_frequencies_file', required=True, help='An npz file containing co-segregation frequencies to convert to correlations')
args = parser.parse_args()
def open_npz(fp):
handle = np.load(fp)
return handle['windows'], handle['scores']
windows, data = open_npz(args.npz_frequencies_file)
names = [ '{}:{}-{}'.format(*i) for i in windows ]
pd.DataFrame(data, index=names, columns=names).to_csv(sys.stdout, sep='\t')
Handle NaNs properly when flattening matrices.import numpy as np
import argparse
import sys
import pandas as pd
parser = argparse.ArgumentParser(description='Calculate coverage over different window sizes for a list of bam files.')
parser.add_argument('-n', '--npz_frequencies_file', required=True, help='An npz file containing co-segregation frequencies to convert to correlations')
args = parser.parse_args()
def open_npz(fp):
handle = np.load(fp)
return handle['windows'], handle['scores']
windows, data = open_npz(args.npz_frequencies_file)
names = [ '{}:{}-{}'.format(*i) for i in windows ]
pd.DataFrame(data, index=names, columns=names).to_csv(sys.stdout, sep='\t', na_rep="NaN")
| <commit_before>import numpy as np
import argparse
import sys
import pandas as pd
parser = argparse.ArgumentParser(description='Calculate coverage over different window sizes for a list of bam files.')
parser.add_argument('-n', '--npz_frequencies_file', required=True, help='An npz file containing co-segregation frequencies to convert to correlations')
args = parser.parse_args()
def open_npz(fp):
handle = np.load(fp)
return handle['windows'], handle['scores']
windows, data = open_npz(args.npz_frequencies_file)
names = [ '{}:{}-{}'.format(*i) for i in windows ]
pd.DataFrame(data, index=names, columns=names).to_csv(sys.stdout, sep='\t')
<commit_msg>Handle NaNs properly when flattening matrices.<commit_after>import numpy as np
import argparse
import sys
import pandas as pd
parser = argparse.ArgumentParser(description='Calculate coverage over different window sizes for a list of bam files.')
parser.add_argument('-n', '--npz_frequencies_file', required=True, help='An npz file containing co-segregation frequencies to convert to correlations')
args = parser.parse_args()
def open_npz(fp):
handle = np.load(fp)
return handle['windows'], handle['scores']
windows, data = open_npz(args.npz_frequencies_file)
names = [ '{}:{}-{}'.format(*i) for i in windows ]
pd.DataFrame(data, index=names, columns=names).to_csv(sys.stdout, sep='\t', na_rep="NaN")
|
8ad95ada5e57ad941b1333cea8f8b81ce739a245 | knights/defaultfilters.py | knights/defaultfilters.py |
from .library import Library
from .filters import Filter
register = Library()
@register.filter(name='title')
class TitleFilter(Filter):
def __rshift__(self, other):
return str(other).title()
|
from .library import Library
register = Library()
@register.filter
def title(val):
return str(val).title()
| Convert to new style filter | Convert to new style filter
| Python | mit | funkybob/knights-templater,funkybob/knights-templater |
from .library import Library
from .filters import Filter
register = Library()
@register.filter(name='title')
class TitleFilter(Filter):
def __rshift__(self, other):
return str(other).title()
Convert to new style filter |
from .library import Library
register = Library()
@register.filter
def title(val):
return str(val).title()
| <commit_before>
from .library import Library
from .filters import Filter
register = Library()
@register.filter(name='title')
class TitleFilter(Filter):
def __rshift__(self, other):
return str(other).title()
<commit_msg>Convert to new style filter<commit_after> |
from .library import Library
register = Library()
@register.filter
def title(val):
return str(val).title()
|
from .library import Library
from .filters import Filter
register = Library()
@register.filter(name='title')
class TitleFilter(Filter):
def __rshift__(self, other):
return str(other).title()
Convert to new style filter
from .library import Library
register = Library()
@register.filter
def title(val):
return str(val).title()
| <commit_before>
from .library import Library
from .filters import Filter
register = Library()
@register.filter(name='title')
class TitleFilter(Filter):
def __rshift__(self, other):
return str(other).title()
<commit_msg>Convert to new style filter<commit_after>
from .library import Library
register = Library()
@register.filter
def title(val):
return str(val).title()
|
0512534c4067b6c36d68241d1ccc7de349a3bbe8 | betfairlightweight/__init__.py | betfairlightweight/__init__.py | from .apiclient import APIClient
from .exceptions import BetfairError
from .streaming import StreamListener
from .filters import MarketFilter, StreamingMarketFilter, StreamingMarketDataFilter
__title__ = 'betfairlightweight'
__version__ = '0.9.9'
__author__ = 'Liam Pauling'
| import logging
from .apiclient import APIClient
from .exceptions import BetfairError
from .filters import MarketFilter, StreamingMarketFilter, StreamingMarketDataFilter
from .streaming import StreamListener
__title__ = 'betfairlightweight'
__version__ = '0.9.9'
__author__ = 'Liam Pauling'
# Set default logging handler to avoid "No handler found" warnings.
try: # Python 2.7+
from logging import NullHandler
except ImportError:
class NullHandler(logging.Handler):
def emit(self, record):
pass
logging.getLogger(__name__).addHandler(NullHandler())
| Add NullHandler to top level package logger | Add NullHandler to top level package logger
| Python | mit | liampauling/betfair,liampauling/betfairlightweight | from .apiclient import APIClient
from .exceptions import BetfairError
from .streaming import StreamListener
from .filters import MarketFilter, StreamingMarketFilter, StreamingMarketDataFilter
__title__ = 'betfairlightweight'
__version__ = '0.9.9'
__author__ = 'Liam Pauling'
Add NullHandler to top level package logger | import logging
from .apiclient import APIClient
from .exceptions import BetfairError
from .filters import MarketFilter, StreamingMarketFilter, StreamingMarketDataFilter
from .streaming import StreamListener
__title__ = 'betfairlightweight'
__version__ = '0.9.9'
__author__ = 'Liam Pauling'
# Set default logging handler to avoid "No handler found" warnings.
try: # Python 2.7+
from logging import NullHandler
except ImportError:
class NullHandler(logging.Handler):
def emit(self, record):
pass
logging.getLogger(__name__).addHandler(NullHandler())
| <commit_before>from .apiclient import APIClient
from .exceptions import BetfairError
from .streaming import StreamListener
from .filters import MarketFilter, StreamingMarketFilter, StreamingMarketDataFilter
__title__ = 'betfairlightweight'
__version__ = '0.9.9'
__author__ = 'Liam Pauling'
<commit_msg>Add NullHandler to top level package logger<commit_after> | import logging
from .apiclient import APIClient
from .exceptions import BetfairError
from .filters import MarketFilter, StreamingMarketFilter, StreamingMarketDataFilter
from .streaming import StreamListener
__title__ = 'betfairlightweight'
__version__ = '0.9.9'
__author__ = 'Liam Pauling'
# Set default logging handler to avoid "No handler found" warnings.
try: # Python 2.7+
from logging import NullHandler
except ImportError:
class NullHandler(logging.Handler):
def emit(self, record):
pass
logging.getLogger(__name__).addHandler(NullHandler())
| from .apiclient import APIClient
from .exceptions import BetfairError
from .streaming import StreamListener
from .filters import MarketFilter, StreamingMarketFilter, StreamingMarketDataFilter
__title__ = 'betfairlightweight'
__version__ = '0.9.9'
__author__ = 'Liam Pauling'
Add NullHandler to top level package loggerimport logging
from .apiclient import APIClient
from .exceptions import BetfairError
from .filters import MarketFilter, StreamingMarketFilter, StreamingMarketDataFilter
from .streaming import StreamListener
__title__ = 'betfairlightweight'
__version__ = '0.9.9'
__author__ = 'Liam Pauling'
# Set default logging handler to avoid "No handler found" warnings.
try: # Python 2.7+
from logging import NullHandler
except ImportError:
class NullHandler(logging.Handler):
def emit(self, record):
pass
logging.getLogger(__name__).addHandler(NullHandler())
| <commit_before>from .apiclient import APIClient
from .exceptions import BetfairError
from .streaming import StreamListener
from .filters import MarketFilter, StreamingMarketFilter, StreamingMarketDataFilter
__title__ = 'betfairlightweight'
__version__ = '0.9.9'
__author__ = 'Liam Pauling'
<commit_msg>Add NullHandler to top level package logger<commit_after>import logging
from .apiclient import APIClient
from .exceptions import BetfairError
from .filters import MarketFilter, StreamingMarketFilter, StreamingMarketDataFilter
from .streaming import StreamListener
__title__ = 'betfairlightweight'
__version__ = '0.9.9'
__author__ = 'Liam Pauling'
# Set default logging handler to avoid "No handler found" warnings.
try: # Python 2.7+
from logging import NullHandler
except ImportError:
class NullHandler(logging.Handler):
def emit(self, record):
pass
logging.getLogger(__name__).addHandler(NullHandler())
|
9be232ab83a4c482eaf56ea99f7b1be81412c517 | Bookie/fabfile/development.py | Bookie/fabfile/development.py | """Fabric commands useful for working on developing Bookie are loaded here"""
from fabric.api import hosts
from fabric.api import local
from fabric.contrib.project import rsync_project
bootstrap_host = 'ubuntu@bmark'
bootstrap_server = '/var/www/bootstrap.py'
bootstrap_local = 'scripts/bootstrap/bootstrap.py'
def gen_bootstrap():
"""Run the generator that builds a custom virtualenv bootstrap file"""
local('python scripts/bootstrap/gen_bootstrap.py > scripts/bootstrap/bootstrap.py', capture=False)
@hosts(bootstrap_host)
def push_bootstrap():
"""Sync the bootstrap.py up to the server for download"""
rsync_project(bootstrap_server, bootstrap_local)
| """Fabric commands useful for working on developing Bookie are loaded here"""
import os
from fabric.api import hosts
from fabric.api import local
from fabric.contrib.project import rsync_project
bootstrap_host = 'ubuntu@bmark'
bootstrap_server = '/var/www/bootstrap.py'
bootstrap_local = 'scripts/bootstrap/bootstrap.py'
def gen_bootstrap():
"""Run the generator that builds a custom virtualenv bootstrap file"""
local('python scripts/bootstrap/gen_bootstrap.py > scripts/bootstrap/bootstrap.py', capture=False)
@hosts(bootstrap_host)
def push_bootstrap():
"""Sync the bootstrap.py up to the server for download"""
rsync_project(bootstrap_server, bootstrap_local)
def jstest():
"""Launch the JS tests we have in the system
Currently only the ones there are for extensions
"""
cwd = os.path.dirname(os.path.dirname(__file__))
local('google-chrome {0}/extensions/tests/index.html'.format(cwd))
| Add a fab command to run jstests | Add a fab command to run jstests
| Python | agpl-3.0 | adamlincoln/Bookie,skmezanul/Bookie,charany1/Bookie,charany1/Bookie,adamlincoln/Bookie,bookieio/Bookie,wangjun/Bookie,teodesson/Bookie,GreenLunar/Bookie,wangjun/Bookie,teodesson/Bookie,GreenLunar/Bookie,pombredanne/Bookie,pombredanne/Bookie,charany1/Bookie,GreenLunar/Bookie,skmezanul/Bookie,teodesson/Bookie,wangjun/Bookie,teodesson/Bookie,bookieio/Bookie,adamlincoln/Bookie,adamlincoln/Bookie,bookieio/Bookie,bookieio/Bookie,GreenLunar/Bookie,skmezanul/Bookie,pombredanne/Bookie,skmezanul/Bookie,wangjun/Bookie | """Fabric commands useful for working on developing Bookie are loaded here"""
from fabric.api import hosts
from fabric.api import local
from fabric.contrib.project import rsync_project
bootstrap_host = 'ubuntu@bmark'
bootstrap_server = '/var/www/bootstrap.py'
bootstrap_local = 'scripts/bootstrap/bootstrap.py'
def gen_bootstrap():
"""Run the generator that builds a custom virtualenv bootstrap file"""
local('python scripts/bootstrap/gen_bootstrap.py > scripts/bootstrap/bootstrap.py', capture=False)
@hosts(bootstrap_host)
def push_bootstrap():
"""Sync the bootstrap.py up to the server for download"""
rsync_project(bootstrap_server, bootstrap_local)
Add a fab command to run jstests | """Fabric commands useful for working on developing Bookie are loaded here"""
import os
from fabric.api import hosts
from fabric.api import local
from fabric.contrib.project import rsync_project
bootstrap_host = 'ubuntu@bmark'
bootstrap_server = '/var/www/bootstrap.py'
bootstrap_local = 'scripts/bootstrap/bootstrap.py'
def gen_bootstrap():
"""Run the generator that builds a custom virtualenv bootstrap file"""
local('python scripts/bootstrap/gen_bootstrap.py > scripts/bootstrap/bootstrap.py', capture=False)
@hosts(bootstrap_host)
def push_bootstrap():
"""Sync the bootstrap.py up to the server for download"""
rsync_project(bootstrap_server, bootstrap_local)
def jstest():
"""Launch the JS tests we have in the system
Currently only the ones there are for extensions
"""
cwd = os.path.dirname(os.path.dirname(__file__))
local('google-chrome {0}/extensions/tests/index.html'.format(cwd))
| <commit_before>"""Fabric commands useful for working on developing Bookie are loaded here"""
from fabric.api import hosts
from fabric.api import local
from fabric.contrib.project import rsync_project
bootstrap_host = 'ubuntu@bmark'
bootstrap_server = '/var/www/bootstrap.py'
bootstrap_local = 'scripts/bootstrap/bootstrap.py'
def gen_bootstrap():
"""Run the generator that builds a custom virtualenv bootstrap file"""
local('python scripts/bootstrap/gen_bootstrap.py > scripts/bootstrap/bootstrap.py', capture=False)
@hosts(bootstrap_host)
def push_bootstrap():
"""Sync the bootstrap.py up to the server for download"""
rsync_project(bootstrap_server, bootstrap_local)
<commit_msg>Add a fab command to run jstests<commit_after> | """Fabric commands useful for working on developing Bookie are loaded here"""
import os
from fabric.api import hosts
from fabric.api import local
from fabric.contrib.project import rsync_project
bootstrap_host = 'ubuntu@bmark'
bootstrap_server = '/var/www/bootstrap.py'
bootstrap_local = 'scripts/bootstrap/bootstrap.py'
def gen_bootstrap():
"""Run the generator that builds a custom virtualenv bootstrap file"""
local('python scripts/bootstrap/gen_bootstrap.py > scripts/bootstrap/bootstrap.py', capture=False)
@hosts(bootstrap_host)
def push_bootstrap():
"""Sync the bootstrap.py up to the server for download"""
rsync_project(bootstrap_server, bootstrap_local)
def jstest():
"""Launch the JS tests we have in the system
Currently only the ones there are for extensions
"""
cwd = os.path.dirname(os.path.dirname(__file__))
local('google-chrome {0}/extensions/tests/index.html'.format(cwd))
| """Fabric commands useful for working on developing Bookie are loaded here"""
from fabric.api import hosts
from fabric.api import local
from fabric.contrib.project import rsync_project
bootstrap_host = 'ubuntu@bmark'
bootstrap_server = '/var/www/bootstrap.py'
bootstrap_local = 'scripts/bootstrap/bootstrap.py'
def gen_bootstrap():
"""Run the generator that builds a custom virtualenv bootstrap file"""
local('python scripts/bootstrap/gen_bootstrap.py > scripts/bootstrap/bootstrap.py', capture=False)
@hosts(bootstrap_host)
def push_bootstrap():
"""Sync the bootstrap.py up to the server for download"""
rsync_project(bootstrap_server, bootstrap_local)
Add a fab command to run jstests"""Fabric commands useful for working on developing Bookie are loaded here"""
import os
from fabric.api import hosts
from fabric.api import local
from fabric.contrib.project import rsync_project
bootstrap_host = 'ubuntu@bmark'
bootstrap_server = '/var/www/bootstrap.py'
bootstrap_local = 'scripts/bootstrap/bootstrap.py'
def gen_bootstrap():
"""Run the generator that builds a custom virtualenv bootstrap file"""
local('python scripts/bootstrap/gen_bootstrap.py > scripts/bootstrap/bootstrap.py', capture=False)
@hosts(bootstrap_host)
def push_bootstrap():
"""Sync the bootstrap.py up to the server for download"""
rsync_project(bootstrap_server, bootstrap_local)
def jstest():
"""Launch the JS tests we have in the system
Currently only the ones there are for extensions
"""
cwd = os.path.dirname(os.path.dirname(__file__))
local('google-chrome {0}/extensions/tests/index.html'.format(cwd))
| <commit_before>"""Fabric commands useful for working on developing Bookie are loaded here"""
from fabric.api import hosts
from fabric.api import local
from fabric.contrib.project import rsync_project
bootstrap_host = 'ubuntu@bmark'
bootstrap_server = '/var/www/bootstrap.py'
bootstrap_local = 'scripts/bootstrap/bootstrap.py'
def gen_bootstrap():
"""Run the generator that builds a custom virtualenv bootstrap file"""
local('python scripts/bootstrap/gen_bootstrap.py > scripts/bootstrap/bootstrap.py', capture=False)
@hosts(bootstrap_host)
def push_bootstrap():
"""Sync the bootstrap.py up to the server for download"""
rsync_project(bootstrap_server, bootstrap_local)
<commit_msg>Add a fab command to run jstests<commit_after>"""Fabric commands useful for working on developing Bookie are loaded here"""
import os
from fabric.api import hosts
from fabric.api import local
from fabric.contrib.project import rsync_project
bootstrap_host = 'ubuntu@bmark'
bootstrap_server = '/var/www/bootstrap.py'
bootstrap_local = 'scripts/bootstrap/bootstrap.py'
def gen_bootstrap():
"""Run the generator that builds a custom virtualenv bootstrap file"""
local('python scripts/bootstrap/gen_bootstrap.py > scripts/bootstrap/bootstrap.py', capture=False)
@hosts(bootstrap_host)
def push_bootstrap():
"""Sync the bootstrap.py up to the server for download"""
rsync_project(bootstrap_server, bootstrap_local)
def jstest():
"""Launch the JS tests we have in the system
Currently only the ones there are for extensions
"""
cwd = os.path.dirname(os.path.dirname(__file__))
local('google-chrome {0}/extensions/tests/index.html'.format(cwd))
|
7197f1578335b38eb2037e8d82f15a27d786d5c1 | var/spack/repos/builtin/packages/py-setuptools/package.py | var/spack/repos/builtin/packages/py-setuptools/package.py | from spack import *
class PySetuptools(Package):
"""Easily download, build, install, upgrade, and uninstall Python packages."""
homepage = "https://pypi.python.org/pypi/setuptools"
url = "https://pypi.python.org/packages/source/s/setuptools/setuptools-11.3.tar.gz"
version('11.3.1', '01f69212e019a2420c1693fb43593930')
version('16.0', '0ace0b96233516fc5f7c857d086aa3ad')
version('18.1', 'f72e87f34fbf07f299f6cb46256a0b06')
version('19.2', '78353b1f80375ca5e088f4b4627ffe03')
version('20.5', 'fadc1e1123ddbe31006e5e43e927362b')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
| from spack import *
class PySetuptools(Package):
"""Easily download, build, install, upgrade, and uninstall Python packages."""
homepage = "https://pypi.python.org/pypi/setuptools"
url = "https://pypi.python.org/packages/source/s/setuptools/setuptools-11.3.tar.gz"
version('11.3.1', '01f69212e019a2420c1693fb43593930')
version('16.0', '0ace0b96233516fc5f7c857d086aa3ad')
version('18.1', 'f72e87f34fbf07f299f6cb46256a0b06')
version('19.2', '78353b1f80375ca5e088f4b4627ffe03')
version('20.5', 'fadc1e1123ddbe31006e5e43e927362b')
version('20.6.7', '45d6110f3ec14924e44c33411db64fe6')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
| Add version 2.6.7 of py-setuptools | Add version 2.6.7 of py-setuptools
| Python | lgpl-2.1 | skosukhin/spack,mfherbst/spack,tmerrick1/spack,lgarren/spack,skosukhin/spack,krafczyk/spack,iulian787/spack,skosukhin/spack,tmerrick1/spack,skosukhin/spack,matthiasdiener/spack,matthiasdiener/spack,TheTimmy/spack,LLNL/spack,LLNL/spack,krafczyk/spack,mfherbst/spack,mfherbst/spack,matthiasdiener/spack,mfherbst/spack,EmreAtes/spack,iulian787/spack,krafczyk/spack,matthiasdiener/spack,tmerrick1/spack,EmreAtes/spack,skosukhin/spack,matthiasdiener/spack,iulian787/spack,iulian787/spack,lgarren/spack,EmreAtes/spack,TheTimmy/spack,tmerrick1/spack,EmreAtes/spack,iulian787/spack,LLNL/spack,lgarren/spack,LLNL/spack,EmreAtes/spack,LLNL/spack,TheTimmy/spack,krafczyk/spack,mfherbst/spack,lgarren/spack,tmerrick1/spack,lgarren/spack,TheTimmy/spack,krafczyk/spack,TheTimmy/spack | from spack import *
class PySetuptools(Package):
"""Easily download, build, install, upgrade, and uninstall Python packages."""
homepage = "https://pypi.python.org/pypi/setuptools"
url = "https://pypi.python.org/packages/source/s/setuptools/setuptools-11.3.tar.gz"
version('11.3.1', '01f69212e019a2420c1693fb43593930')
version('16.0', '0ace0b96233516fc5f7c857d086aa3ad')
version('18.1', 'f72e87f34fbf07f299f6cb46256a0b06')
version('19.2', '78353b1f80375ca5e088f4b4627ffe03')
version('20.5', 'fadc1e1123ddbe31006e5e43e927362b')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
Add version 2.6.7 of py-setuptools | from spack import *
class PySetuptools(Package):
"""Easily download, build, install, upgrade, and uninstall Python packages."""
homepage = "https://pypi.python.org/pypi/setuptools"
url = "https://pypi.python.org/packages/source/s/setuptools/setuptools-11.3.tar.gz"
version('11.3.1', '01f69212e019a2420c1693fb43593930')
version('16.0', '0ace0b96233516fc5f7c857d086aa3ad')
version('18.1', 'f72e87f34fbf07f299f6cb46256a0b06')
version('19.2', '78353b1f80375ca5e088f4b4627ffe03')
version('20.5', 'fadc1e1123ddbe31006e5e43e927362b')
version('20.6.7', '45d6110f3ec14924e44c33411db64fe6')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
| <commit_before>from spack import *
class PySetuptools(Package):
"""Easily download, build, install, upgrade, and uninstall Python packages."""
homepage = "https://pypi.python.org/pypi/setuptools"
url = "https://pypi.python.org/packages/source/s/setuptools/setuptools-11.3.tar.gz"
version('11.3.1', '01f69212e019a2420c1693fb43593930')
version('16.0', '0ace0b96233516fc5f7c857d086aa3ad')
version('18.1', 'f72e87f34fbf07f299f6cb46256a0b06')
version('19.2', '78353b1f80375ca5e088f4b4627ffe03')
version('20.5', 'fadc1e1123ddbe31006e5e43e927362b')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
<commit_msg>Add version 2.6.7 of py-setuptools<commit_after> | from spack import *
class PySetuptools(Package):
"""Easily download, build, install, upgrade, and uninstall Python packages."""
homepage = "https://pypi.python.org/pypi/setuptools"
url = "https://pypi.python.org/packages/source/s/setuptools/setuptools-11.3.tar.gz"
version('11.3.1', '01f69212e019a2420c1693fb43593930')
version('16.0', '0ace0b96233516fc5f7c857d086aa3ad')
version('18.1', 'f72e87f34fbf07f299f6cb46256a0b06')
version('19.2', '78353b1f80375ca5e088f4b4627ffe03')
version('20.5', 'fadc1e1123ddbe31006e5e43e927362b')
version('20.6.7', '45d6110f3ec14924e44c33411db64fe6')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
| from spack import *
class PySetuptools(Package):
"""Easily download, build, install, upgrade, and uninstall Python packages."""
homepage = "https://pypi.python.org/pypi/setuptools"
url = "https://pypi.python.org/packages/source/s/setuptools/setuptools-11.3.tar.gz"
version('11.3.1', '01f69212e019a2420c1693fb43593930')
version('16.0', '0ace0b96233516fc5f7c857d086aa3ad')
version('18.1', 'f72e87f34fbf07f299f6cb46256a0b06')
version('19.2', '78353b1f80375ca5e088f4b4627ffe03')
version('20.5', 'fadc1e1123ddbe31006e5e43e927362b')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
Add version 2.6.7 of py-setuptoolsfrom spack import *
class PySetuptools(Package):
"""Easily download, build, install, upgrade, and uninstall Python packages."""
homepage = "https://pypi.python.org/pypi/setuptools"
url = "https://pypi.python.org/packages/source/s/setuptools/setuptools-11.3.tar.gz"
version('11.3.1', '01f69212e019a2420c1693fb43593930')
version('16.0', '0ace0b96233516fc5f7c857d086aa3ad')
version('18.1', 'f72e87f34fbf07f299f6cb46256a0b06')
version('19.2', '78353b1f80375ca5e088f4b4627ffe03')
version('20.5', 'fadc1e1123ddbe31006e5e43e927362b')
version('20.6.7', '45d6110f3ec14924e44c33411db64fe6')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
| <commit_before>from spack import *
class PySetuptools(Package):
"""Easily download, build, install, upgrade, and uninstall Python packages."""
homepage = "https://pypi.python.org/pypi/setuptools"
url = "https://pypi.python.org/packages/source/s/setuptools/setuptools-11.3.tar.gz"
version('11.3.1', '01f69212e019a2420c1693fb43593930')
version('16.0', '0ace0b96233516fc5f7c857d086aa3ad')
version('18.1', 'f72e87f34fbf07f299f6cb46256a0b06')
version('19.2', '78353b1f80375ca5e088f4b4627ffe03')
version('20.5', 'fadc1e1123ddbe31006e5e43e927362b')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
<commit_msg>Add version 2.6.7 of py-setuptools<commit_after>from spack import *
class PySetuptools(Package):
"""Easily download, build, install, upgrade, and uninstall Python packages."""
homepage = "https://pypi.python.org/pypi/setuptools"
url = "https://pypi.python.org/packages/source/s/setuptools/setuptools-11.3.tar.gz"
version('11.3.1', '01f69212e019a2420c1693fb43593930')
version('16.0', '0ace0b96233516fc5f7c857d086aa3ad')
version('18.1', 'f72e87f34fbf07f299f6cb46256a0b06')
version('19.2', '78353b1f80375ca5e088f4b4627ffe03')
version('20.5', 'fadc1e1123ddbe31006e5e43e927362b')
version('20.6.7', '45d6110f3ec14924e44c33411db64fe6')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
|
90e7bc2c8313de2a5054d5290441c527f5f2c253 | gameButton.py | gameButton.py | # Game Button class for menu
# Marshall Ehlinger
import pygame
class gameButton:
GRAY = [131, 131, 131]
PINK = [255, 55, 135]
def __init__(self, label, buttonWidth, buttonHeight, importedGameFunction):
self.label = label
self.height = buttonHeight
self.width = buttonWidth
self.importedGameFunction = importedGameFunction
def renderButton(self, surface, isSelected, origin_x, origin_y):
if isSelected:
# pygame.draw.rect(surface, self.PINK, [origin_x, origin_y, self.width, self.height])
surface.fill(self.PINK,[origin_x, origin_y, self.width, self.height])
else:
# pygame.draw.rect(surface, self.GRAY, [origin_x, origin_y, self.width, self.height])
surface.fill(self.GRAY,[origin_x, origin_y, self.width, self.height])
def runGame(self):
self.importedGameFunction()
| # Game Button class for menu
# Marshall Ehlinger
import pygame
class gameButton:
GRAY = [131, 131, 131]
PINK = [255, 55, 135]
WHITE = [255, 255, 255]
BLACK = [0, 0, 0]
def __init__(self, label, buttonWidth, buttonHeight, importedGameFunction):
self.label = label
self.height = buttonHeight
self.width = buttonWidth
self.importedGameFunction = importedGameFunction
self.font = pygame.font.SysFont("monospace", 15)
def renderButton(self, surface, isSelected, origin_x, origin_y):
label = self.font.render(self.label, True, self.BLACK)
if isSelected:
# pygame.draw.rect(surface, self.PINK, [origin_x, origin_y, self.width, self.height])
surface.fill(self.PINK,[origin_x, origin_y, self.width, self.height])
else:
# pygame.draw.rect(surface, self.GRAY, [origin_x, origin_y, self.width, self.height])
surface.fill(self.GRAY,[origin_x, origin_y, self.width, self.height])
surface.blit(label,[origin_x + 5, origin_y + (.3 * self.height)])
def runGame(self):
self.importedGameFunction()
| Add labels to menu buttons | Add labels to menu buttons
| Python | mit | MEhlinger/rpi_pushbutton_games | # Game Button class for menu
# Marshall Ehlinger
import pygame
class gameButton:
GRAY = [131, 131, 131]
PINK = [255, 55, 135]
def __init__(self, label, buttonWidth, buttonHeight, importedGameFunction):
self.label = label
self.height = buttonHeight
self.width = buttonWidth
self.importedGameFunction = importedGameFunction
def renderButton(self, surface, isSelected, origin_x, origin_y):
if isSelected:
# pygame.draw.rect(surface, self.PINK, [origin_x, origin_y, self.width, self.height])
surface.fill(self.PINK,[origin_x, origin_y, self.width, self.height])
else:
# pygame.draw.rect(surface, self.GRAY, [origin_x, origin_y, self.width, self.height])
surface.fill(self.GRAY,[origin_x, origin_y, self.width, self.height])
def runGame(self):
self.importedGameFunction()
Add labels to menu buttons | # Game Button class for menu
# Marshall Ehlinger
import pygame
class gameButton:
GRAY = [131, 131, 131]
PINK = [255, 55, 135]
WHITE = [255, 255, 255]
BLACK = [0, 0, 0]
def __init__(self, label, buttonWidth, buttonHeight, importedGameFunction):
self.label = label
self.height = buttonHeight
self.width = buttonWidth
self.importedGameFunction = importedGameFunction
self.font = pygame.font.SysFont("monospace", 15)
def renderButton(self, surface, isSelected, origin_x, origin_y):
label = self.font.render(self.label, True, self.BLACK)
if isSelected:
# pygame.draw.rect(surface, self.PINK, [origin_x, origin_y, self.width, self.height])
surface.fill(self.PINK,[origin_x, origin_y, self.width, self.height])
else:
# pygame.draw.rect(surface, self.GRAY, [origin_x, origin_y, self.width, self.height])
surface.fill(self.GRAY,[origin_x, origin_y, self.width, self.height])
surface.blit(label,[origin_x + 5, origin_y + (.3 * self.height)])
def runGame(self):
self.importedGameFunction()
| <commit_before># Game Button class for menu
# Marshall Ehlinger
import pygame
class gameButton:
GRAY = [131, 131, 131]
PINK = [255, 55, 135]
def __init__(self, label, buttonWidth, buttonHeight, importedGameFunction):
self.label = label
self.height = buttonHeight
self.width = buttonWidth
self.importedGameFunction = importedGameFunction
def renderButton(self, surface, isSelected, origin_x, origin_y):
if isSelected:
# pygame.draw.rect(surface, self.PINK, [origin_x, origin_y, self.width, self.height])
surface.fill(self.PINK,[origin_x, origin_y, self.width, self.height])
else:
# pygame.draw.rect(surface, self.GRAY, [origin_x, origin_y, self.width, self.height])
surface.fill(self.GRAY,[origin_x, origin_y, self.width, self.height])
def runGame(self):
self.importedGameFunction()
<commit_msg>Add labels to menu buttons<commit_after> | # Game Button class for menu
# Marshall Ehlinger
import pygame
class gameButton:
GRAY = [131, 131, 131]
PINK = [255, 55, 135]
WHITE = [255, 255, 255]
BLACK = [0, 0, 0]
def __init__(self, label, buttonWidth, buttonHeight, importedGameFunction):
self.label = label
self.height = buttonHeight
self.width = buttonWidth
self.importedGameFunction = importedGameFunction
self.font = pygame.font.SysFont("monospace", 15)
def renderButton(self, surface, isSelected, origin_x, origin_y):
label = self.font.render(self.label, True, self.BLACK)
if isSelected:
# pygame.draw.rect(surface, self.PINK, [origin_x, origin_y, self.width, self.height])
surface.fill(self.PINK,[origin_x, origin_y, self.width, self.height])
else:
# pygame.draw.rect(surface, self.GRAY, [origin_x, origin_y, self.width, self.height])
surface.fill(self.GRAY,[origin_x, origin_y, self.width, self.height])
surface.blit(label,[origin_x + 5, origin_y + (.3 * self.height)])
def runGame(self):
self.importedGameFunction()
| # Game Button class for menu
# Marshall Ehlinger
import pygame
class gameButton:
GRAY = [131, 131, 131]
PINK = [255, 55, 135]
def __init__(self, label, buttonWidth, buttonHeight, importedGameFunction):
self.label = label
self.height = buttonHeight
self.width = buttonWidth
self.importedGameFunction = importedGameFunction
def renderButton(self, surface, isSelected, origin_x, origin_y):
if isSelected:
# pygame.draw.rect(surface, self.PINK, [origin_x, origin_y, self.width, self.height])
surface.fill(self.PINK,[origin_x, origin_y, self.width, self.height])
else:
# pygame.draw.rect(surface, self.GRAY, [origin_x, origin_y, self.width, self.height])
surface.fill(self.GRAY,[origin_x, origin_y, self.width, self.height])
def runGame(self):
self.importedGameFunction()
Add labels to menu buttons# Game Button class for menu
# Marshall Ehlinger
import pygame
class gameButton:
GRAY = [131, 131, 131]
PINK = [255, 55, 135]
WHITE = [255, 255, 255]
BLACK = [0, 0, 0]
def __init__(self, label, buttonWidth, buttonHeight, importedGameFunction):
self.label = label
self.height = buttonHeight
self.width = buttonWidth
self.importedGameFunction = importedGameFunction
self.font = pygame.font.SysFont("monospace", 15)
def renderButton(self, surface, isSelected, origin_x, origin_y):
label = self.font.render(self.label, True, self.BLACK)
if isSelected:
# pygame.draw.rect(surface, self.PINK, [origin_x, origin_y, self.width, self.height])
surface.fill(self.PINK,[origin_x, origin_y, self.width, self.height])
else:
# pygame.draw.rect(surface, self.GRAY, [origin_x, origin_y, self.width, self.height])
surface.fill(self.GRAY,[origin_x, origin_y, self.width, self.height])
surface.blit(label,[origin_x + 5, origin_y + (.3 * self.height)])
def runGame(self):
self.importedGameFunction()
| <commit_before># Game Button class for menu
# Marshall Ehlinger
import pygame
class gameButton:
GRAY = [131, 131, 131]
PINK = [255, 55, 135]
def __init__(self, label, buttonWidth, buttonHeight, importedGameFunction):
self.label = label
self.height = buttonHeight
self.width = buttonWidth
self.importedGameFunction = importedGameFunction
def renderButton(self, surface, isSelected, origin_x, origin_y):
if isSelected:
# pygame.draw.rect(surface, self.PINK, [origin_x, origin_y, self.width, self.height])
surface.fill(self.PINK,[origin_x, origin_y, self.width, self.height])
else:
# pygame.draw.rect(surface, self.GRAY, [origin_x, origin_y, self.width, self.height])
surface.fill(self.GRAY,[origin_x, origin_y, self.width, self.height])
def runGame(self):
self.importedGameFunction()
<commit_msg>Add labels to menu buttons<commit_after># Game Button class for menu
# Marshall Ehlinger
import pygame
class gameButton:
GRAY = [131, 131, 131]
PINK = [255, 55, 135]
WHITE = [255, 255, 255]
BLACK = [0, 0, 0]
def __init__(self, label, buttonWidth, buttonHeight, importedGameFunction):
self.label = label
self.height = buttonHeight
self.width = buttonWidth
self.importedGameFunction = importedGameFunction
self.font = pygame.font.SysFont("monospace", 15)
def renderButton(self, surface, isSelected, origin_x, origin_y):
label = self.font.render(self.label, True, self.BLACK)
if isSelected:
# pygame.draw.rect(surface, self.PINK, [origin_x, origin_y, self.width, self.height])
surface.fill(self.PINK,[origin_x, origin_y, self.width, self.height])
else:
# pygame.draw.rect(surface, self.GRAY, [origin_x, origin_y, self.width, self.height])
surface.fill(self.GRAY,[origin_x, origin_y, self.width, self.height])
surface.blit(label,[origin_x + 5, origin_y + (.3 * self.height)])
def runGame(self):
self.importedGameFunction()
|
761367713658e2a436e1d600af026b375a7a332b | pymks/bases/real_ffts.py | pymks/bases/real_ffts.py | from .abstract import _AbstractMicrostructureBasis
import numpy as np
class _RealFFTBasis(_AbstractMicrostructureBasis):
def __init__(self, *args, **kwargs):
super(_RealFFTBasis, self).__init__(*args, **kwargs)
def _fftn(self, X, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.rfftn(np.ascontiguousarray(X),
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
overwrite_input=True,
avoid_copy=avoid_copy)()
else:
return self._fftmodule.rfftn(X, axes=self._axes)
def _ifftn(self, X, s=None, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.irfftn(np.ascontiguousarray(X), s=s,
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
avoid_copy=avoid_copy)().real
else:
return self._fftmodule.irfftn(X, axes=self._axes).real
def discretize(self, X):
raise NotImplementedError
| from .abstract import _AbstractMicrostructureBasis
import numpy as np
class _RealFFTBasis(_AbstractMicrostructureBasis):
def __init__(self, *args, **kwargs):
super(_RealFFTBasis, self).__init__(*args, **kwargs)
def _fftn(self, X, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.rfftn(np.ascontiguousarray(X),
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
overwrite_input=True,
avoid_copy=avoid_copy)()
else:
return self._fftmodule.rfftn(X, axes=self._axes)
def _ifftn(self, X, s=None, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.irfftn(np.ascontiguousarray(X), s=s,
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
avoid_copy=avoid_copy)().real
else:
return self._fftmodule.irfftn(X, axes=self._axes, s=s).real
def discretize(self, X):
raise NotImplementedError
| Fix bug for numpy's irfftn | Fix bug for numpy's irfftn
address #232
Fix bug for numpy's irfftn. The size of the returned array must be
passed because the returned size is potentially not unique. Without
this change only the return kernel would possibly have the wrong shape.
| Python | mit | davidbrough1/pymks,davidbrough1/pymks | from .abstract import _AbstractMicrostructureBasis
import numpy as np
class _RealFFTBasis(_AbstractMicrostructureBasis):
def __init__(self, *args, **kwargs):
super(_RealFFTBasis, self).__init__(*args, **kwargs)
def _fftn(self, X, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.rfftn(np.ascontiguousarray(X),
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
overwrite_input=True,
avoid_copy=avoid_copy)()
else:
return self._fftmodule.rfftn(X, axes=self._axes)
def _ifftn(self, X, s=None, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.irfftn(np.ascontiguousarray(X), s=s,
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
avoid_copy=avoid_copy)().real
else:
return self._fftmodule.irfftn(X, axes=self._axes).real
def discretize(self, X):
raise NotImplementedError
Fix bug for numpy's irfftn
address #232
Fix bug for numpy's irfftn. The size of the returned array must be
passed because the returned size is potentially not unique. Without
this change only the return kernel would possibly have the wrong shape. | from .abstract import _AbstractMicrostructureBasis
import numpy as np
class _RealFFTBasis(_AbstractMicrostructureBasis):
def __init__(self, *args, **kwargs):
super(_RealFFTBasis, self).__init__(*args, **kwargs)
def _fftn(self, X, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.rfftn(np.ascontiguousarray(X),
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
overwrite_input=True,
avoid_copy=avoid_copy)()
else:
return self._fftmodule.rfftn(X, axes=self._axes)
def _ifftn(self, X, s=None, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.irfftn(np.ascontiguousarray(X), s=s,
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
avoid_copy=avoid_copy)().real
else:
return self._fftmodule.irfftn(X, axes=self._axes, s=s).real
def discretize(self, X):
raise NotImplementedError
| <commit_before>from .abstract import _AbstractMicrostructureBasis
import numpy as np
class _RealFFTBasis(_AbstractMicrostructureBasis):
def __init__(self, *args, **kwargs):
super(_RealFFTBasis, self).__init__(*args, **kwargs)
def _fftn(self, X, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.rfftn(np.ascontiguousarray(X),
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
overwrite_input=True,
avoid_copy=avoid_copy)()
else:
return self._fftmodule.rfftn(X, axes=self._axes)
def _ifftn(self, X, s=None, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.irfftn(np.ascontiguousarray(X), s=s,
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
avoid_copy=avoid_copy)().real
else:
return self._fftmodule.irfftn(X, axes=self._axes).real
def discretize(self, X):
raise NotImplementedError
<commit_msg>Fix bug for numpy's irfftn
address #232
Fix bug for numpy's irfftn. The size of the returned array must be
passed because the returned size is potentially not unique. Without
this change only the return kernel would possibly have the wrong shape.<commit_after> | from .abstract import _AbstractMicrostructureBasis
import numpy as np
class _RealFFTBasis(_AbstractMicrostructureBasis):
def __init__(self, *args, **kwargs):
super(_RealFFTBasis, self).__init__(*args, **kwargs)
def _fftn(self, X, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.rfftn(np.ascontiguousarray(X),
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
overwrite_input=True,
avoid_copy=avoid_copy)()
else:
return self._fftmodule.rfftn(X, axes=self._axes)
def _ifftn(self, X, s=None, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.irfftn(np.ascontiguousarray(X), s=s,
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
avoid_copy=avoid_copy)().real
else:
return self._fftmodule.irfftn(X, axes=self._axes, s=s).real
def discretize(self, X):
raise NotImplementedError
| from .abstract import _AbstractMicrostructureBasis
import numpy as np
class _RealFFTBasis(_AbstractMicrostructureBasis):
def __init__(self, *args, **kwargs):
super(_RealFFTBasis, self).__init__(*args, **kwargs)
def _fftn(self, X, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.rfftn(np.ascontiguousarray(X),
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
overwrite_input=True,
avoid_copy=avoid_copy)()
else:
return self._fftmodule.rfftn(X, axes=self._axes)
def _ifftn(self, X, s=None, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.irfftn(np.ascontiguousarray(X), s=s,
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
avoid_copy=avoid_copy)().real
else:
return self._fftmodule.irfftn(X, axes=self._axes).real
def discretize(self, X):
raise NotImplementedError
Fix bug for numpy's irfftn
address #232
Fix bug for numpy's irfftn. The size of the returned array must be
passed because the returned size is potentially not unique. Without
this change only the return kernel would possibly have the wrong shape.from .abstract import _AbstractMicrostructureBasis
import numpy as np
class _RealFFTBasis(_AbstractMicrostructureBasis):
def __init__(self, *args, **kwargs):
super(_RealFFTBasis, self).__init__(*args, **kwargs)
def _fftn(self, X, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.rfftn(np.ascontiguousarray(X),
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
overwrite_input=True,
avoid_copy=avoid_copy)()
else:
return self._fftmodule.rfftn(X, axes=self._axes)
def _ifftn(self, X, s=None, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.irfftn(np.ascontiguousarray(X), s=s,
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
avoid_copy=avoid_copy)().real
else:
return self._fftmodule.irfftn(X, axes=self._axes, s=s).real
def discretize(self, X):
raise NotImplementedError
| <commit_before>from .abstract import _AbstractMicrostructureBasis
import numpy as np
class _RealFFTBasis(_AbstractMicrostructureBasis):
def __init__(self, *args, **kwargs):
super(_RealFFTBasis, self).__init__(*args, **kwargs)
def _fftn(self, X, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.rfftn(np.ascontiguousarray(X),
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
overwrite_input=True,
avoid_copy=avoid_copy)()
else:
return self._fftmodule.rfftn(X, axes=self._axes)
def _ifftn(self, X, s=None, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.irfftn(np.ascontiguousarray(X), s=s,
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
avoid_copy=avoid_copy)().real
else:
return self._fftmodule.irfftn(X, axes=self._axes).real
def discretize(self, X):
raise NotImplementedError
<commit_msg>Fix bug for numpy's irfftn
address #232
Fix bug for numpy's irfftn. The size of the returned array must be
passed because the returned size is potentially not unique. Without
this change only the return kernel would possibly have the wrong shape.<commit_after>from .abstract import _AbstractMicrostructureBasis
import numpy as np
class _RealFFTBasis(_AbstractMicrostructureBasis):
def __init__(self, *args, **kwargs):
super(_RealFFTBasis, self).__init__(*args, **kwargs)
def _fftn(self, X, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.rfftn(np.ascontiguousarray(X),
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
overwrite_input=True,
avoid_copy=avoid_copy)()
else:
return self._fftmodule.rfftn(X, axes=self._axes)
def _ifftn(self, X, s=None, threads=1, avoid_copy=True):
if self._pyfftw:
return self._fftmodule.irfftn(np.ascontiguousarray(X), s=s,
axes=self._axes,
threads=threads,
planner_effort='FFTW_ESTIMATE',
avoid_copy=avoid_copy)().real
else:
return self._fftmodule.irfftn(X, axes=self._axes, s=s).real
def discretize(self, X):
raise NotImplementedError
|
3c6c242fd42bd9acf9866f458fa70536d56f3ccd | tests/test_tabulate.py | tests/test_tabulate.py | from pgcli.packages.tabulate import tabulate
from textwrap import dedent
def test_dont_strip_leading_whitespace():
data = [[' abc']]
headers = ['xyz']
tbl, _ = tabulate(data, headers, tablefmt='psql')
assert tbl == dedent('''
+---------+
| xyz |
|---------|
| abc |
+---------+ ''').strip()
| from mycli.packages.tabulate import tabulate
from textwrap import dedent
def test_dont_strip_leading_whitespace():
data = [[' abc']]
headers = ['xyz']
tbl, _ = tabulate(data, headers, tablefmt='psql')
assert tbl == dedent('''
+---------+
| xyz |
|---------|
| abc |
+---------+ ''').strip()
| Change the pgcli import to mycli. | Change the pgcli import to mycli.
| Python | bsd-3-clause | mdsrosa/mycli,mdsrosa/mycli | from pgcli.packages.tabulate import tabulate
from textwrap import dedent
def test_dont_strip_leading_whitespace():
data = [[' abc']]
headers = ['xyz']
tbl, _ = tabulate(data, headers, tablefmt='psql')
assert tbl == dedent('''
+---------+
| xyz |
|---------|
| abc |
+---------+ ''').strip()
Change the pgcli import to mycli. | from mycli.packages.tabulate import tabulate
from textwrap import dedent
def test_dont_strip_leading_whitespace():
data = [[' abc']]
headers = ['xyz']
tbl, _ = tabulate(data, headers, tablefmt='psql')
assert tbl == dedent('''
+---------+
| xyz |
|---------|
| abc |
+---------+ ''').strip()
| <commit_before>from pgcli.packages.tabulate import tabulate
from textwrap import dedent
def test_dont_strip_leading_whitespace():
data = [[' abc']]
headers = ['xyz']
tbl, _ = tabulate(data, headers, tablefmt='psql')
assert tbl == dedent('''
+---------+
| xyz |
|---------|
| abc |
+---------+ ''').strip()
<commit_msg>Change the pgcli import to mycli.<commit_after> | from mycli.packages.tabulate import tabulate
from textwrap import dedent
def test_dont_strip_leading_whitespace():
data = [[' abc']]
headers = ['xyz']
tbl, _ = tabulate(data, headers, tablefmt='psql')
assert tbl == dedent('''
+---------+
| xyz |
|---------|
| abc |
+---------+ ''').strip()
| from pgcli.packages.tabulate import tabulate
from textwrap import dedent
def test_dont_strip_leading_whitespace():
data = [[' abc']]
headers = ['xyz']
tbl, _ = tabulate(data, headers, tablefmt='psql')
assert tbl == dedent('''
+---------+
| xyz |
|---------|
| abc |
+---------+ ''').strip()
Change the pgcli import to mycli.from mycli.packages.tabulate import tabulate
from textwrap import dedent
def test_dont_strip_leading_whitespace():
data = [[' abc']]
headers = ['xyz']
tbl, _ = tabulate(data, headers, tablefmt='psql')
assert tbl == dedent('''
+---------+
| xyz |
|---------|
| abc |
+---------+ ''').strip()
| <commit_before>from pgcli.packages.tabulate import tabulate
from textwrap import dedent
def test_dont_strip_leading_whitespace():
data = [[' abc']]
headers = ['xyz']
tbl, _ = tabulate(data, headers, tablefmt='psql')
assert tbl == dedent('''
+---------+
| xyz |
|---------|
| abc |
+---------+ ''').strip()
<commit_msg>Change the pgcli import to mycli.<commit_after>from mycli.packages.tabulate import tabulate
from textwrap import dedent
def test_dont_strip_leading_whitespace():
data = [[' abc']]
headers = ['xyz']
tbl, _ = tabulate(data, headers, tablefmt='psql')
assert tbl == dedent('''
+---------+
| xyz |
|---------|
| abc |
+---------+ ''').strip()
|
633b23fd862f152c3f7d9e88fbeb660635386c3f | qmtp_package/__init__.py | qmtp_package/__init__.py | import os
if not os.environ.get("RMG_workingDirectory"):
import os.path
message = "Please set your RMG_workingDirectory environment variable.\n" +\
"(eg. export RMG_workingDirectory=%s )" % \
os.path.abspath(os.path.join(os.path.dirname(__file__),'..'))
raise Exception(message)
| Check for RMG_workingDirectory environment variable in qmtp_package | Check for RMG_workingDirectory environment variable in qmtp_package
I dislike the way this is needed, but for now this commit will at least help people
discover their mistake if they forget. | Python | mit | nickvandewiele/RMG-Py,pierrelb/RMG-Py,KEHANG/RMG-Py,chatelak/RMG-Py,enochd/RMG-Py,comocheng/RMG-Py,nickvandewiele/RMG-Py,nyee/RMG-Py,nyee/RMG-Py,comocheng/RMG-Py,enochd/RMG-Py,pierrelb/RMG-Py,chatelak/RMG-Py,faribas/RMG-Py,KEHANG/RMG-Py,faribas/RMG-Py | Check for RMG_workingDirectory environment variable in qmtp_package
I dislike the way this is needed, but for now this commit will at least help people
discover their mistake if they forget. | import os
if not os.environ.get("RMG_workingDirectory"):
import os.path
message = "Please set your RMG_workingDirectory environment variable.\n" +\
"(eg. export RMG_workingDirectory=%s )" % \
os.path.abspath(os.path.join(os.path.dirname(__file__),'..'))
raise Exception(message)
| <commit_before><commit_msg>Check for RMG_workingDirectory environment variable in qmtp_package
I dislike the way this is needed, but for now this commit will at least help people
discover their mistake if they forget.<commit_after> | import os
if not os.environ.get("RMG_workingDirectory"):
import os.path
message = "Please set your RMG_workingDirectory environment variable.\n" +\
"(eg. export RMG_workingDirectory=%s )" % \
os.path.abspath(os.path.join(os.path.dirname(__file__),'..'))
raise Exception(message)
| Check for RMG_workingDirectory environment variable in qmtp_package
I dislike the way this is needed, but for now this commit will at least help people
discover their mistake if they forget.import os
if not os.environ.get("RMG_workingDirectory"):
import os.path
message = "Please set your RMG_workingDirectory environment variable.\n" +\
"(eg. export RMG_workingDirectory=%s )" % \
os.path.abspath(os.path.join(os.path.dirname(__file__),'..'))
raise Exception(message)
| <commit_before><commit_msg>Check for RMG_workingDirectory environment variable in qmtp_package
I dislike the way this is needed, but for now this commit will at least help people
discover their mistake if they forget.<commit_after>import os
if not os.environ.get("RMG_workingDirectory"):
import os.path
message = "Please set your RMG_workingDirectory environment variable.\n" +\
"(eg. export RMG_workingDirectory=%s )" % \
os.path.abspath(os.path.join(os.path.dirname(__file__),'..'))
raise Exception(message)
| |
f791354a098c32617f02f05dbbb53861b7a94139 | rapt/cmds/ingredients.py | rapt/cmds/ingredients.py | import click
from rapt.connection import get_vr
from rapt.models import query
from rapt.util import dump_yaml, load_yaml, edit_yaml
@click.command()
@click.option('--name', '-n')
def ingredients(name, verbose):
"""List builds.
"""
vr = get_vr()
q = {}
if name:
q['name'] = name
# add filters if we need to be...
for i, ingredient in enumerate(query('Ingredient', vr, q)):
click.echo(ingredient.name)
@click.command()
@click.argument('name')
def ingredient(name):
"""View a complete ingredient config."""
vr = get_vr()
q = {'name': name}
ingredient = query('Ingredient', vr, q).next()
doc = {
'config': load_yaml(ingredient.config_yaml),
'env': load_yaml(ingredient.env_yaml),
}
config = edit_yaml(dump_yaml(doc))
print(config)
| import click
from rapt.connection import get_vr
from rapt.models import query
from rapt.util import dump_yaml, load_yaml, edit_yaml
@click.command()
@click.option('--name', '-n')
def ingredients(name, verbose):
"""List builds.
"""
vr = get_vr()
q = {}
if name:
q['name'] = name
# add filters if we need to be...
for i, ingredient in enumerate(query('Ingredient', vr, q)):
click.echo(ingredient.name)
@click.command()
@click.argument('name')
def ingredient(name):
"""View a complete ingredient config."""
vr = get_vr()
q = {'name': name}
ingredient = query('Ingredient', vr, q).next()
doc = {
'config': load_yaml(ingredient.config_yaml),
'env': load_yaml(ingredient.env_yaml),
}
config = edit_yaml(dump_yaml(doc))
if not config:
click.echo('No changes')
return
ingredient.config_yaml = dump_yaml(config['config'])
ingredient.env_yaml = dump_yaml(config['env'])
ingredient.save()
| Update the ingredient or noop if there are no changes | Update the ingredient or noop if there are no changes
| Python | bsd-3-clause | yougov/rapt,yougov/rapt | import click
from rapt.connection import get_vr
from rapt.models import query
from rapt.util import dump_yaml, load_yaml, edit_yaml
@click.command()
@click.option('--name', '-n')
def ingredients(name, verbose):
"""List builds.
"""
vr = get_vr()
q = {}
if name:
q['name'] = name
# add filters if we need to be...
for i, ingredient in enumerate(query('Ingredient', vr, q)):
click.echo(ingredient.name)
@click.command()
@click.argument('name')
def ingredient(name):
"""View a complete ingredient config."""
vr = get_vr()
q = {'name': name}
ingredient = query('Ingredient', vr, q).next()
doc = {
'config': load_yaml(ingredient.config_yaml),
'env': load_yaml(ingredient.env_yaml),
}
config = edit_yaml(dump_yaml(doc))
print(config)
Update the ingredient or noop if there are no changes | import click
from rapt.connection import get_vr
from rapt.models import query
from rapt.util import dump_yaml, load_yaml, edit_yaml
@click.command()
@click.option('--name', '-n')
def ingredients(name, verbose):
"""List builds.
"""
vr = get_vr()
q = {}
if name:
q['name'] = name
# add filters if we need to be...
for i, ingredient in enumerate(query('Ingredient', vr, q)):
click.echo(ingredient.name)
@click.command()
@click.argument('name')
def ingredient(name):
"""View a complete ingredient config."""
vr = get_vr()
q = {'name': name}
ingredient = query('Ingredient', vr, q).next()
doc = {
'config': load_yaml(ingredient.config_yaml),
'env': load_yaml(ingredient.env_yaml),
}
config = edit_yaml(dump_yaml(doc))
if not config:
click.echo('No changes')
return
ingredient.config_yaml = dump_yaml(config['config'])
ingredient.env_yaml = dump_yaml(config['env'])
ingredient.save()
| <commit_before>import click
from rapt.connection import get_vr
from rapt.models import query
from rapt.util import dump_yaml, load_yaml, edit_yaml
@click.command()
@click.option('--name', '-n')
def ingredients(name, verbose):
"""List builds.
"""
vr = get_vr()
q = {}
if name:
q['name'] = name
# add filters if we need to be...
for i, ingredient in enumerate(query('Ingredient', vr, q)):
click.echo(ingredient.name)
@click.command()
@click.argument('name')
def ingredient(name):
"""View a complete ingredient config."""
vr = get_vr()
q = {'name': name}
ingredient = query('Ingredient', vr, q).next()
doc = {
'config': load_yaml(ingredient.config_yaml),
'env': load_yaml(ingredient.env_yaml),
}
config = edit_yaml(dump_yaml(doc))
print(config)
<commit_msg>Update the ingredient or noop if there are no changes<commit_after> | import click
from rapt.connection import get_vr
from rapt.models import query
from rapt.util import dump_yaml, load_yaml, edit_yaml
@click.command()
@click.option('--name', '-n')
def ingredients(name, verbose):
"""List builds.
"""
vr = get_vr()
q = {}
if name:
q['name'] = name
# add filters if we need to be...
for i, ingredient in enumerate(query('Ingredient', vr, q)):
click.echo(ingredient.name)
@click.command()
@click.argument('name')
def ingredient(name):
"""View a complete ingredient config."""
vr = get_vr()
q = {'name': name}
ingredient = query('Ingredient', vr, q).next()
doc = {
'config': load_yaml(ingredient.config_yaml),
'env': load_yaml(ingredient.env_yaml),
}
config = edit_yaml(dump_yaml(doc))
if not config:
click.echo('No changes')
return
ingredient.config_yaml = dump_yaml(config['config'])
ingredient.env_yaml = dump_yaml(config['env'])
ingredient.save()
| import click
from rapt.connection import get_vr
from rapt.models import query
from rapt.util import dump_yaml, load_yaml, edit_yaml
@click.command()
@click.option('--name', '-n')
def ingredients(name, verbose):
"""List builds.
"""
vr = get_vr()
q = {}
if name:
q['name'] = name
# add filters if we need to be...
for i, ingredient in enumerate(query('Ingredient', vr, q)):
click.echo(ingredient.name)
@click.command()
@click.argument('name')
def ingredient(name):
"""View a complete ingredient config."""
vr = get_vr()
q = {'name': name}
ingredient = query('Ingredient', vr, q).next()
doc = {
'config': load_yaml(ingredient.config_yaml),
'env': load_yaml(ingredient.env_yaml),
}
config = edit_yaml(dump_yaml(doc))
print(config)
Update the ingredient or noop if there are no changesimport click
from rapt.connection import get_vr
from rapt.models import query
from rapt.util import dump_yaml, load_yaml, edit_yaml
@click.command()
@click.option('--name', '-n')
def ingredients(name, verbose):
"""List builds.
"""
vr = get_vr()
q = {}
if name:
q['name'] = name
# add filters if we need to be...
for i, ingredient in enumerate(query('Ingredient', vr, q)):
click.echo(ingredient.name)
@click.command()
@click.argument('name')
def ingredient(name):
"""View a complete ingredient config."""
vr = get_vr()
q = {'name': name}
ingredient = query('Ingredient', vr, q).next()
doc = {
'config': load_yaml(ingredient.config_yaml),
'env': load_yaml(ingredient.env_yaml),
}
config = edit_yaml(dump_yaml(doc))
if not config:
click.echo('No changes')
return
ingredient.config_yaml = dump_yaml(config['config'])
ingredient.env_yaml = dump_yaml(config['env'])
ingredient.save()
| <commit_before>import click
from rapt.connection import get_vr
from rapt.models import query
from rapt.util import dump_yaml, load_yaml, edit_yaml
@click.command()
@click.option('--name', '-n')
def ingredients(name, verbose):
"""List builds.
"""
vr = get_vr()
q = {}
if name:
q['name'] = name
# add filters if we need to be...
for i, ingredient in enumerate(query('Ingredient', vr, q)):
click.echo(ingredient.name)
@click.command()
@click.argument('name')
def ingredient(name):
"""View a complete ingredient config."""
vr = get_vr()
q = {'name': name}
ingredient = query('Ingredient', vr, q).next()
doc = {
'config': load_yaml(ingredient.config_yaml),
'env': load_yaml(ingredient.env_yaml),
}
config = edit_yaml(dump_yaml(doc))
print(config)
<commit_msg>Update the ingredient or noop if there are no changes<commit_after>import click
from rapt.connection import get_vr
from rapt.models import query
from rapt.util import dump_yaml, load_yaml, edit_yaml
@click.command()
@click.option('--name', '-n')
def ingredients(name, verbose):
"""List builds.
"""
vr = get_vr()
q = {}
if name:
q['name'] = name
# add filters if we need to be...
for i, ingredient in enumerate(query('Ingredient', vr, q)):
click.echo(ingredient.name)
@click.command()
@click.argument('name')
def ingredient(name):
"""View a complete ingredient config."""
vr = get_vr()
q = {'name': name}
ingredient = query('Ingredient', vr, q).next()
doc = {
'config': load_yaml(ingredient.config_yaml),
'env': load_yaml(ingredient.env_yaml),
}
config = edit_yaml(dump_yaml(doc))
if not config:
click.echo('No changes')
return
ingredient.config_yaml = dump_yaml(config['config'])
ingredient.env_yaml = dump_yaml(config['env'])
ingredient.save()
|
903458640ec8db1c39c822b229e466bc717efe40 | registration/__init__.py | registration/__init__.py | from django.utils.version import get_version as django_get_version
VERSION = (0, 9, 0, 'beta', 1)
def get_version():
return django_get_version(VERSION) # pragma: no cover
| VERSION = (0, 9, 0, 'beta', 1)
def get_version():
from django.utils.version import get_version as django_get_version
return django_get_version(VERSION) # pragma: no cover
| Move import of Django's get_version into django-registration's get_version, to avoid dependency-order problems. | Move import of Django's get_version into django-registration's get_version, to avoid dependency-order problems.
| Python | bsd-3-clause | christang/django-registration-1.5,AndrewLvov/django-registration,AndrewLvov/django-registration,fedenko/django-registration,fedenko/django-registration,christang/django-registration-1.5 | from django.utils.version import get_version as django_get_version
VERSION = (0, 9, 0, 'beta', 1)
def get_version():
return django_get_version(VERSION) # pragma: no cover
Move import of Django's get_version into django-registration's get_version, to avoid dependency-order problems. | VERSION = (0, 9, 0, 'beta', 1)
def get_version():
from django.utils.version import get_version as django_get_version
return django_get_version(VERSION) # pragma: no cover
| <commit_before>from django.utils.version import get_version as django_get_version
VERSION = (0, 9, 0, 'beta', 1)
def get_version():
return django_get_version(VERSION) # pragma: no cover
<commit_msg>Move import of Django's get_version into django-registration's get_version, to avoid dependency-order problems.<commit_after> | VERSION = (0, 9, 0, 'beta', 1)
def get_version():
from django.utils.version import get_version as django_get_version
return django_get_version(VERSION) # pragma: no cover
| from django.utils.version import get_version as django_get_version
VERSION = (0, 9, 0, 'beta', 1)
def get_version():
return django_get_version(VERSION) # pragma: no cover
Move import of Django's get_version into django-registration's get_version, to avoid dependency-order problems.VERSION = (0, 9, 0, 'beta', 1)
def get_version():
from django.utils.version import get_version as django_get_version
return django_get_version(VERSION) # pragma: no cover
| <commit_before>from django.utils.version import get_version as django_get_version
VERSION = (0, 9, 0, 'beta', 1)
def get_version():
return django_get_version(VERSION) # pragma: no cover
<commit_msg>Move import of Django's get_version into django-registration's get_version, to avoid dependency-order problems.<commit_after>VERSION = (0, 9, 0, 'beta', 1)
def get_version():
from django.utils.version import get_version as django_get_version
return django_get_version(VERSION) # pragma: no cover
|
c763d42e48d501461ab6c8c875d691f52045ada8 | intelmq/bots/outputs/mongodb/output.py | intelmq/bots/outputs/mongodb/output.py | # -*- coding: utf-8 -*-
"""
pymongo library automatically tries to reconnect if connection has been lost
"""
from intelmq.lib.bot import Bot
try:
import pymongo
except ImportError:
pymongo = None
class MongoDBOutputBot(Bot):
def init(self):
if pymongo is None:
self.logger.error('Could not import pymongo. Please install it.')
self.stop()
self.connect()
def connect(self):
self.logger.debug('Connecting to mongodb server.')
try:
self.client = pymongo.MongoClient(self.parameters.host,
int(self.parameters.port))
except pymongo.errors.ConnectionFailure:
raise ValueError('Connection to mongodb server failed.')
else:
db = self.client[self.parameters.database]
self.collection = db[self.parameters.collection]
self.logger.info('Successfully connected to mongodb server.')
def process(self):
event = self.receive_message()
try:
self.collection.insert(event.to_dict(hierarchical=self.parameters.hierarchical_output))
except pymongo.errors.AutoReconnect:
self.logger.error('Connection Lost. Connecting again.')
self.connect()
else:
self.acknowledge_message()
def shutdown(self):
self.client.close()
BOT = MongoDBOutputBot
| # -*- coding: utf-8 -*-
"""
pymongo library automatically tries to reconnect if connection has been lost
"""
from intelmq.lib.bot import Bot
try:
import pymongo
except ImportError:
pymongo = None
class MongoDBOutputBot(Bot):
def init(self):
if pymongo is None:
self.logger.error('Could not import pymongo. Please install it.')
self.stop()
self.connect()
def connect(self):
self.logger.debug('Connecting to mongodb server.')
try:
self.client = pymongo.MongoClient(self.parameters.host,
int(self.parameters.port))
except pymongo.errors.ConnectionFailure:
raise ValueError('Connection to mongodb server failed.')
else:
db = self.client[self.parameters.database]
if self.parameters.db_user and self.parameters.db_pass:
try:
db.authenticate(name=self.parameters.db_user,
password=self.parameters.db_pass)
except pymongo.errors.OperationFailure:
raise ValueError('Authentication to {} failed'.format(self.parameters.database))
self.collection = db[self.parameters.collection]
self.logger.info('Successfully connected to mongodb server.')
def process(self):
event = self.receive_message()
try:
self.collection.insert(event.to_dict(hierarchical=self.parameters.hierarchical_output))
except pymongo.errors.AutoReconnect:
self.logger.error('Connection Lost. Connecting again.')
self.connect()
else:
self.acknowledge_message()
def shutdown(self):
self.client.close()
BOT = MongoDBOutputBot
| Add authentication otpion to mongodb | Add authentication otpion to mongodb
| Python | agpl-3.0 | certtools/intelmq,certtools/intelmq,certtools/intelmq,aaronkaplan/intelmq,aaronkaplan/intelmq,aaronkaplan/intelmq | # -*- coding: utf-8 -*-
"""
pymongo library automatically tries to reconnect if connection has been lost
"""
from intelmq.lib.bot import Bot
try:
import pymongo
except ImportError:
pymongo = None
class MongoDBOutputBot(Bot):
def init(self):
if pymongo is None:
self.logger.error('Could not import pymongo. Please install it.')
self.stop()
self.connect()
def connect(self):
self.logger.debug('Connecting to mongodb server.')
try:
self.client = pymongo.MongoClient(self.parameters.host,
int(self.parameters.port))
except pymongo.errors.ConnectionFailure:
raise ValueError('Connection to mongodb server failed.')
else:
db = self.client[self.parameters.database]
self.collection = db[self.parameters.collection]
self.logger.info('Successfully connected to mongodb server.')
def process(self):
event = self.receive_message()
try:
self.collection.insert(event.to_dict(hierarchical=self.parameters.hierarchical_output))
except pymongo.errors.AutoReconnect:
self.logger.error('Connection Lost. Connecting again.')
self.connect()
else:
self.acknowledge_message()
def shutdown(self):
self.client.close()
BOT = MongoDBOutputBot
Add authentication otpion to mongodb | # -*- coding: utf-8 -*-
"""
pymongo library automatically tries to reconnect if connection has been lost
"""
from intelmq.lib.bot import Bot
try:
import pymongo
except ImportError:
pymongo = None
class MongoDBOutputBot(Bot):
def init(self):
if pymongo is None:
self.logger.error('Could not import pymongo. Please install it.')
self.stop()
self.connect()
def connect(self):
self.logger.debug('Connecting to mongodb server.')
try:
self.client = pymongo.MongoClient(self.parameters.host,
int(self.parameters.port))
except pymongo.errors.ConnectionFailure:
raise ValueError('Connection to mongodb server failed.')
else:
db = self.client[self.parameters.database]
if self.parameters.db_user and self.parameters.db_pass:
try:
db.authenticate(name=self.parameters.db_user,
password=self.parameters.db_pass)
except pymongo.errors.OperationFailure:
raise ValueError('Authentication to {} failed'.format(self.parameters.database))
self.collection = db[self.parameters.collection]
self.logger.info('Successfully connected to mongodb server.')
def process(self):
event = self.receive_message()
try:
self.collection.insert(event.to_dict(hierarchical=self.parameters.hierarchical_output))
except pymongo.errors.AutoReconnect:
self.logger.error('Connection Lost. Connecting again.')
self.connect()
else:
self.acknowledge_message()
def shutdown(self):
self.client.close()
BOT = MongoDBOutputBot
| <commit_before># -*- coding: utf-8 -*-
"""
pymongo library automatically tries to reconnect if connection has been lost
"""
from intelmq.lib.bot import Bot
try:
import pymongo
except ImportError:
pymongo = None
class MongoDBOutputBot(Bot):
def init(self):
if pymongo is None:
self.logger.error('Could not import pymongo. Please install it.')
self.stop()
self.connect()
def connect(self):
self.logger.debug('Connecting to mongodb server.')
try:
self.client = pymongo.MongoClient(self.parameters.host,
int(self.parameters.port))
except pymongo.errors.ConnectionFailure:
raise ValueError('Connection to mongodb server failed.')
else:
db = self.client[self.parameters.database]
self.collection = db[self.parameters.collection]
self.logger.info('Successfully connected to mongodb server.')
def process(self):
event = self.receive_message()
try:
self.collection.insert(event.to_dict(hierarchical=self.parameters.hierarchical_output))
except pymongo.errors.AutoReconnect:
self.logger.error('Connection Lost. Connecting again.')
self.connect()
else:
self.acknowledge_message()
def shutdown(self):
self.client.close()
BOT = MongoDBOutputBot
<commit_msg>Add authentication otpion to mongodb<commit_after> | # -*- coding: utf-8 -*-
"""
pymongo library automatically tries to reconnect if connection has been lost
"""
from intelmq.lib.bot import Bot
try:
import pymongo
except ImportError:
pymongo = None
class MongoDBOutputBot(Bot):
def init(self):
if pymongo is None:
self.logger.error('Could not import pymongo. Please install it.')
self.stop()
self.connect()
def connect(self):
self.logger.debug('Connecting to mongodb server.')
try:
self.client = pymongo.MongoClient(self.parameters.host,
int(self.parameters.port))
except pymongo.errors.ConnectionFailure:
raise ValueError('Connection to mongodb server failed.')
else:
db = self.client[self.parameters.database]
if self.parameters.db_user and self.parameters.db_pass:
try:
db.authenticate(name=self.parameters.db_user,
password=self.parameters.db_pass)
except pymongo.errors.OperationFailure:
raise ValueError('Authentication to {} failed'.format(self.parameters.database))
self.collection = db[self.parameters.collection]
self.logger.info('Successfully connected to mongodb server.')
def process(self):
event = self.receive_message()
try:
self.collection.insert(event.to_dict(hierarchical=self.parameters.hierarchical_output))
except pymongo.errors.AutoReconnect:
self.logger.error('Connection Lost. Connecting again.')
self.connect()
else:
self.acknowledge_message()
def shutdown(self):
self.client.close()
BOT = MongoDBOutputBot
| # -*- coding: utf-8 -*-
"""
pymongo library automatically tries to reconnect if connection has been lost
"""
from intelmq.lib.bot import Bot
try:
import pymongo
except ImportError:
pymongo = None
class MongoDBOutputBot(Bot):
def init(self):
if pymongo is None:
self.logger.error('Could not import pymongo. Please install it.')
self.stop()
self.connect()
def connect(self):
self.logger.debug('Connecting to mongodb server.')
try:
self.client = pymongo.MongoClient(self.parameters.host,
int(self.parameters.port))
except pymongo.errors.ConnectionFailure:
raise ValueError('Connection to mongodb server failed.')
else:
db = self.client[self.parameters.database]
self.collection = db[self.parameters.collection]
self.logger.info('Successfully connected to mongodb server.')
def process(self):
event = self.receive_message()
try:
self.collection.insert(event.to_dict(hierarchical=self.parameters.hierarchical_output))
except pymongo.errors.AutoReconnect:
self.logger.error('Connection Lost. Connecting again.')
self.connect()
else:
self.acknowledge_message()
def shutdown(self):
self.client.close()
BOT = MongoDBOutputBot
Add authentication otpion to mongodb# -*- coding: utf-8 -*-
"""
pymongo library automatically tries to reconnect if connection has been lost
"""
from intelmq.lib.bot import Bot
try:
import pymongo
except ImportError:
pymongo = None
class MongoDBOutputBot(Bot):
def init(self):
if pymongo is None:
self.logger.error('Could not import pymongo. Please install it.')
self.stop()
self.connect()
def connect(self):
self.logger.debug('Connecting to mongodb server.')
try:
self.client = pymongo.MongoClient(self.parameters.host,
int(self.parameters.port))
except pymongo.errors.ConnectionFailure:
raise ValueError('Connection to mongodb server failed.')
else:
db = self.client[self.parameters.database]
if self.parameters.db_user and self.parameters.db_pass:
try:
db.authenticate(name=self.parameters.db_user,
password=self.parameters.db_pass)
except pymongo.errors.OperationFailure:
raise ValueError('Authentication to {} failed'.format(self.parameters.database))
self.collection = db[self.parameters.collection]
self.logger.info('Successfully connected to mongodb server.')
def process(self):
event = self.receive_message()
try:
self.collection.insert(event.to_dict(hierarchical=self.parameters.hierarchical_output))
except pymongo.errors.AutoReconnect:
self.logger.error('Connection Lost. Connecting again.')
self.connect()
else:
self.acknowledge_message()
def shutdown(self):
self.client.close()
BOT = MongoDBOutputBot
| <commit_before># -*- coding: utf-8 -*-
"""
pymongo library automatically tries to reconnect if connection has been lost
"""
from intelmq.lib.bot import Bot
try:
import pymongo
except ImportError:
pymongo = None
class MongoDBOutputBot(Bot):
def init(self):
if pymongo is None:
self.logger.error('Could not import pymongo. Please install it.')
self.stop()
self.connect()
def connect(self):
self.logger.debug('Connecting to mongodb server.')
try:
self.client = pymongo.MongoClient(self.parameters.host,
int(self.parameters.port))
except pymongo.errors.ConnectionFailure:
raise ValueError('Connection to mongodb server failed.')
else:
db = self.client[self.parameters.database]
self.collection = db[self.parameters.collection]
self.logger.info('Successfully connected to mongodb server.')
def process(self):
event = self.receive_message()
try:
self.collection.insert(event.to_dict(hierarchical=self.parameters.hierarchical_output))
except pymongo.errors.AutoReconnect:
self.logger.error('Connection Lost. Connecting again.')
self.connect()
else:
self.acknowledge_message()
def shutdown(self):
self.client.close()
BOT = MongoDBOutputBot
<commit_msg>Add authentication otpion to mongodb<commit_after># -*- coding: utf-8 -*-
"""
pymongo library automatically tries to reconnect if connection has been lost
"""
from intelmq.lib.bot import Bot
try:
import pymongo
except ImportError:
pymongo = None
class MongoDBOutputBot(Bot):
def init(self):
if pymongo is None:
self.logger.error('Could not import pymongo. Please install it.')
self.stop()
self.connect()
def connect(self):
self.logger.debug('Connecting to mongodb server.')
try:
self.client = pymongo.MongoClient(self.parameters.host,
int(self.parameters.port))
except pymongo.errors.ConnectionFailure:
raise ValueError('Connection to mongodb server failed.')
else:
db = self.client[self.parameters.database]
if self.parameters.db_user and self.parameters.db_pass:
try:
db.authenticate(name=self.parameters.db_user,
password=self.parameters.db_pass)
except pymongo.errors.OperationFailure:
raise ValueError('Authentication to {} failed'.format(self.parameters.database))
self.collection = db[self.parameters.collection]
self.logger.info('Successfully connected to mongodb server.')
def process(self):
event = self.receive_message()
try:
self.collection.insert(event.to_dict(hierarchical=self.parameters.hierarchical_output))
except pymongo.errors.AutoReconnect:
self.logger.error('Connection Lost. Connecting again.')
self.connect()
else:
self.acknowledge_message()
def shutdown(self):
self.client.close()
BOT = MongoDBOutputBot
|
afe90ba2a9720ffd80780e7696353510501362c7 | studygroups/management/commands/generate_reminders.py | studygroups/management/commands/generate_reminders.py | from django.core.management.base import BaseCommand, CommandError
from studygroups.tasks import gen_reminders
class Command(BaseCommand):
help = 'Generate reminders for all study groups happening in 3 days from now'
def handle(self, *args, **options):
gen_reminders()
| from django.core.management.base import BaseCommand, CommandError
from django.utils import timezone
from studygroups.models import Meeting
from studygroups.models.learningcircle import generate_meeting_reminder
class Command(BaseCommand):
help = 'Transitional command to generate reminders for all meetings in the future.'
def handle(self, *args, **options):
today = timezone.now().date()
meetings = Meeting.objects.active().filter(study_group__deleted_at__isnull=True).filter(meeting_date__gte=today)
for meeting in meetings:
print(f'Generating meeting reminder for meeting happening {meeting.meeting_date}')
generate_meeting_reminder(meeting)
| Update task to generate reminders for all future meetings | Update task to generate reminders for all future meetings
| Python | mit | p2pu/learning-circles,p2pu/learning-circles,p2pu/learning-circles,p2pu/learning-circles | from django.core.management.base import BaseCommand, CommandError
from studygroups.tasks import gen_reminders
class Command(BaseCommand):
help = 'Generate reminders for all study groups happening in 3 days from now'
def handle(self, *args, **options):
gen_reminders()
Update task to generate reminders for all future meetings | from django.core.management.base import BaseCommand, CommandError
from django.utils import timezone
from studygroups.models import Meeting
from studygroups.models.learningcircle import generate_meeting_reminder
class Command(BaseCommand):
help = 'Transitional command to generate reminders for all meetings in the future.'
def handle(self, *args, **options):
today = timezone.now().date()
meetings = Meeting.objects.active().filter(study_group__deleted_at__isnull=True).filter(meeting_date__gte=today)
for meeting in meetings:
print(f'Generating meeting reminder for meeting happening {meeting.meeting_date}')
generate_meeting_reminder(meeting)
| <commit_before>from django.core.management.base import BaseCommand, CommandError
from studygroups.tasks import gen_reminders
class Command(BaseCommand):
help = 'Generate reminders for all study groups happening in 3 days from now'
def handle(self, *args, **options):
gen_reminders()
<commit_msg>Update task to generate reminders for all future meetings<commit_after> | from django.core.management.base import BaseCommand, CommandError
from django.utils import timezone
from studygroups.models import Meeting
from studygroups.models.learningcircle import generate_meeting_reminder
class Command(BaseCommand):
help = 'Transitional command to generate reminders for all meetings in the future.'
def handle(self, *args, **options):
today = timezone.now().date()
meetings = Meeting.objects.active().filter(study_group__deleted_at__isnull=True).filter(meeting_date__gte=today)
for meeting in meetings:
print(f'Generating meeting reminder for meeting happening {meeting.meeting_date}')
generate_meeting_reminder(meeting)
| from django.core.management.base import BaseCommand, CommandError
from studygroups.tasks import gen_reminders
class Command(BaseCommand):
help = 'Generate reminders for all study groups happening in 3 days from now'
def handle(self, *args, **options):
gen_reminders()
Update task to generate reminders for all future meetingsfrom django.core.management.base import BaseCommand, CommandError
from django.utils import timezone
from studygroups.models import Meeting
from studygroups.models.learningcircle import generate_meeting_reminder
class Command(BaseCommand):
help = 'Transitional command to generate reminders for all meetings in the future.'
def handle(self, *args, **options):
today = timezone.now().date()
meetings = Meeting.objects.active().filter(study_group__deleted_at__isnull=True).filter(meeting_date__gte=today)
for meeting in meetings:
print(f'Generating meeting reminder for meeting happening {meeting.meeting_date}')
generate_meeting_reminder(meeting)
| <commit_before>from django.core.management.base import BaseCommand, CommandError
from studygroups.tasks import gen_reminders
class Command(BaseCommand):
help = 'Generate reminders for all study groups happening in 3 days from now'
def handle(self, *args, **options):
gen_reminders()
<commit_msg>Update task to generate reminders for all future meetings<commit_after>from django.core.management.base import BaseCommand, CommandError
from django.utils import timezone
from studygroups.models import Meeting
from studygroups.models.learningcircle import generate_meeting_reminder
class Command(BaseCommand):
help = 'Transitional command to generate reminders for all meetings in the future.'
def handle(self, *args, **options):
today = timezone.now().date()
meetings = Meeting.objects.active().filter(study_group__deleted_at__isnull=True).filter(meeting_date__gte=today)
for meeting in meetings:
print(f'Generating meeting reminder for meeting happening {meeting.meeting_date}')
generate_meeting_reminder(meeting)
|
52a9e0b5f3f0df4d2a9a092ecf6935def7a3e5cf | lib/ansiblelint/formatters/__init__.py | lib/ansiblelint/formatters/__init__.py | class Formatter(object):
def format(self, match):
formatstr = u"[{0}] {1}\n{2}:{3}\n{4}\n"
return formatstr.format(match.rule.id,
match.message,
match.filename,
match.linenumber,
match.line)
class QuietFormatter(object):
def format(self, match):
formatstr = u"[{0}] {1}:{2}"
return formatstr.format(match.rule.id, match.filename,
match.linenumber)
class ParseableFormatter(object):
def format(self, match):
formatstr = u"{0}:{1}: [{2}] {3}"
return formatstr.format(match.filename,
match.linenumber,
match.rule.id,
match.message,
)
| class Formatter(object):
def format(self, match):
formatstr = u"[{0}] {1}\n{2}:{3}\n{4}\n"
return formatstr.format(match.rule.id,
match.message,
match.filename,
match.linenumber,
match.line)
class QuietFormatter(object):
def format(self, match):
formatstr = u"[{0}] {1}:{2}"
return formatstr.format(match.rule.id, match.filename,
match.linenumber)
class ParseableFormatter(object):
def format(self, match):
formatstr = u"{0}:{1}: [{2}] {3}"
return formatstr.format(match.filename,
match.linenumber,
"E" + match.rule.id,
match.message,
)
| Improve ParseableFormatter to be more like pylint | Improve ParseableFormatter to be more like pylint
Add an E in front of the rule ID so that pylint detects
it as an error.
Fixes #154
| Python | mit | willthames/ansible-lint,dataxu/ansible-lint,MatrixCrawler/ansible-lint | class Formatter(object):
def format(self, match):
formatstr = u"[{0}] {1}\n{2}:{3}\n{4}\n"
return formatstr.format(match.rule.id,
match.message,
match.filename,
match.linenumber,
match.line)
class QuietFormatter(object):
def format(self, match):
formatstr = u"[{0}] {1}:{2}"
return formatstr.format(match.rule.id, match.filename,
match.linenumber)
class ParseableFormatter(object):
def format(self, match):
formatstr = u"{0}:{1}: [{2}] {3}"
return formatstr.format(match.filename,
match.linenumber,
match.rule.id,
match.message,
)
Improve ParseableFormatter to be more like pylint
Add an E in front of the rule ID so that pylint detects
it as an error.
Fixes #154 | class Formatter(object):
def format(self, match):
formatstr = u"[{0}] {1}\n{2}:{3}\n{4}\n"
return formatstr.format(match.rule.id,
match.message,
match.filename,
match.linenumber,
match.line)
class QuietFormatter(object):
def format(self, match):
formatstr = u"[{0}] {1}:{2}"
return formatstr.format(match.rule.id, match.filename,
match.linenumber)
class ParseableFormatter(object):
def format(self, match):
formatstr = u"{0}:{1}: [{2}] {3}"
return formatstr.format(match.filename,
match.linenumber,
"E" + match.rule.id,
match.message,
)
| <commit_before>class Formatter(object):
def format(self, match):
formatstr = u"[{0}] {1}\n{2}:{3}\n{4}\n"
return formatstr.format(match.rule.id,
match.message,
match.filename,
match.linenumber,
match.line)
class QuietFormatter(object):
def format(self, match):
formatstr = u"[{0}] {1}:{2}"
return formatstr.format(match.rule.id, match.filename,
match.linenumber)
class ParseableFormatter(object):
def format(self, match):
formatstr = u"{0}:{1}: [{2}] {3}"
return formatstr.format(match.filename,
match.linenumber,
match.rule.id,
match.message,
)
<commit_msg>Improve ParseableFormatter to be more like pylint
Add an E in front of the rule ID so that pylint detects
it as an error.
Fixes #154<commit_after> | class Formatter(object):
def format(self, match):
formatstr = u"[{0}] {1}\n{2}:{3}\n{4}\n"
return formatstr.format(match.rule.id,
match.message,
match.filename,
match.linenumber,
match.line)
class QuietFormatter(object):
def format(self, match):
formatstr = u"[{0}] {1}:{2}"
return formatstr.format(match.rule.id, match.filename,
match.linenumber)
class ParseableFormatter(object):
def format(self, match):
formatstr = u"{0}:{1}: [{2}] {3}"
return formatstr.format(match.filename,
match.linenumber,
"E" + match.rule.id,
match.message,
)
| class Formatter(object):
def format(self, match):
formatstr = u"[{0}] {1}\n{2}:{3}\n{4}\n"
return formatstr.format(match.rule.id,
match.message,
match.filename,
match.linenumber,
match.line)
class QuietFormatter(object):
def format(self, match):
formatstr = u"[{0}] {1}:{2}"
return formatstr.format(match.rule.id, match.filename,
match.linenumber)
class ParseableFormatter(object):
def format(self, match):
formatstr = u"{0}:{1}: [{2}] {3}"
return formatstr.format(match.filename,
match.linenumber,
match.rule.id,
match.message,
)
Improve ParseableFormatter to be more like pylint
Add an E in front of the rule ID so that pylint detects
it as an error.
Fixes #154class Formatter(object):
def format(self, match):
formatstr = u"[{0}] {1}\n{2}:{3}\n{4}\n"
return formatstr.format(match.rule.id,
match.message,
match.filename,
match.linenumber,
match.line)
class QuietFormatter(object):
def format(self, match):
formatstr = u"[{0}] {1}:{2}"
return formatstr.format(match.rule.id, match.filename,
match.linenumber)
class ParseableFormatter(object):
def format(self, match):
formatstr = u"{0}:{1}: [{2}] {3}"
return formatstr.format(match.filename,
match.linenumber,
"E" + match.rule.id,
match.message,
)
| <commit_before>class Formatter(object):
def format(self, match):
formatstr = u"[{0}] {1}\n{2}:{3}\n{4}\n"
return formatstr.format(match.rule.id,
match.message,
match.filename,
match.linenumber,
match.line)
class QuietFormatter(object):
def format(self, match):
formatstr = u"[{0}] {1}:{2}"
return formatstr.format(match.rule.id, match.filename,
match.linenumber)
class ParseableFormatter(object):
def format(self, match):
formatstr = u"{0}:{1}: [{2}] {3}"
return formatstr.format(match.filename,
match.linenumber,
match.rule.id,
match.message,
)
<commit_msg>Improve ParseableFormatter to be more like pylint
Add an E in front of the rule ID so that pylint detects
it as an error.
Fixes #154<commit_after>class Formatter(object):
def format(self, match):
formatstr = u"[{0}] {1}\n{2}:{3}\n{4}\n"
return formatstr.format(match.rule.id,
match.message,
match.filename,
match.linenumber,
match.line)
class QuietFormatter(object):
def format(self, match):
formatstr = u"[{0}] {1}:{2}"
return formatstr.format(match.rule.id, match.filename,
match.linenumber)
class ParseableFormatter(object):
def format(self, match):
formatstr = u"{0}:{1}: [{2}] {3}"
return formatstr.format(match.filename,
match.linenumber,
"E" + match.rule.id,
match.message,
)
|
d1d55450db13766f51f264c9bfef1bcea74ef7b1 | convert.py | convert.py | #!/usr/bin/env python
import os, sys
import pexpect
import geom
from sfepy.fem.mesh import Mesh
try:
from site_cfg import tetgen_path
except ImportError:
tetgen_path = '/usr/bin/tetgen'
def mesh():
if len( sys.argv ) == 3:
geomFileName = sys.argv[1]
vtkFileName = sys.argv[2]
if len( sys.argv ) == 2:
geomFileName = sys.argv[1]
vtkFileName = "tmp/t.1.vtk"
else:
geomFileName = "database/box.geo"
vtkFileName = "tmp/t.1.vtk"
pexpect.run( "gmsh -0 %s -o tmp/x.geo" % geomFileName )
g=geom.read_gmsh("tmp/x.geo")
g.printinfo()
geom.write_tetgen(g,"tmp/t.poly")
geom.runtetgen("tmp/t.poly",a=0.0003,Q=1.0,quadratic=False,
tetgenpath = tetgen_path)
m = Mesh.fromFile("tmp/t.1.node")
m.write( vtkFileName, io = "auto" )
try:
os.makedirs( "tmp" )
except OSError, e:
if e.errno != 17: # [Errno 17] File exists
raise
mesh()
| #!/usr/bin/env python
import os, sys
import geom
from sfepy.fem.mesh import Mesh
try:
from site_cfg import tetgen_path
except ImportError:
tetgen_path = '/usr/bin/tetgen'
def mesh():
if len( sys.argv ) == 3:
geomFileName = sys.argv[1]
vtkFileName = sys.argv[2]
if len( sys.argv ) == 2:
geomFileName = sys.argv[1]
vtkFileName = "tmp/t.1.vtk"
else:
geomFileName = "database/box.geo"
vtkFileName = "tmp/t.1.vtk"
os.system( "gmsh -0 %s -o tmp/x.geo" % geomFileName )
g=geom.read_gmsh("tmp/x.geo")
g.printinfo()
geom.write_tetgen(g,"tmp/t.poly")
geom.runtetgen("tmp/t.poly",a=0.03,Q=1.0,quadratic=False,
tetgenpath = tetgen_path)
m = Mesh.fromFile("tmp/t.1.node")
m.write( vtkFileName, io = "auto" )
try:
os.makedirs( "tmp" )
except OSError, e:
if e.errno != 17: # [Errno 17] File exists
raise
mesh()
| Use os.system() instead of pexpect.run(). | Use os.system() instead of pexpect.run().
Also do not generate too dense mesh by default, so that it's faster.
| Python | bsd-3-clause | BubuLK/sfepy,sfepy/sfepy,BubuLK/sfepy,RexFuzzle/sfepy,vlukes/sfepy,sfepy/sfepy,rc/sfepy,lokik/sfepy,RexFuzzle/sfepy,BubuLK/sfepy,olivierverdier/sfepy,sfepy/sfepy,olivierverdier/sfepy,vlukes/sfepy,lokik/sfepy,vlukes/sfepy,RexFuzzle/sfepy,RexFuzzle/sfepy,olivierverdier/sfepy,lokik/sfepy,lokik/sfepy,rc/sfepy,rc/sfepy | #!/usr/bin/env python
import os, sys
import pexpect
import geom
from sfepy.fem.mesh import Mesh
try:
from site_cfg import tetgen_path
except ImportError:
tetgen_path = '/usr/bin/tetgen'
def mesh():
if len( sys.argv ) == 3:
geomFileName = sys.argv[1]
vtkFileName = sys.argv[2]
if len( sys.argv ) == 2:
geomFileName = sys.argv[1]
vtkFileName = "tmp/t.1.vtk"
else:
geomFileName = "database/box.geo"
vtkFileName = "tmp/t.1.vtk"
pexpect.run( "gmsh -0 %s -o tmp/x.geo" % geomFileName )
g=geom.read_gmsh("tmp/x.geo")
g.printinfo()
geom.write_tetgen(g,"tmp/t.poly")
geom.runtetgen("tmp/t.poly",a=0.0003,Q=1.0,quadratic=False,
tetgenpath = tetgen_path)
m = Mesh.fromFile("tmp/t.1.node")
m.write( vtkFileName, io = "auto" )
try:
os.makedirs( "tmp" )
except OSError, e:
if e.errno != 17: # [Errno 17] File exists
raise
mesh()
Use os.system() instead of pexpect.run().
Also do not generate too dense mesh by default, so that it's faster. | #!/usr/bin/env python
import os, sys
import geom
from sfepy.fem.mesh import Mesh
try:
from site_cfg import tetgen_path
except ImportError:
tetgen_path = '/usr/bin/tetgen'
def mesh():
if len( sys.argv ) == 3:
geomFileName = sys.argv[1]
vtkFileName = sys.argv[2]
if len( sys.argv ) == 2:
geomFileName = sys.argv[1]
vtkFileName = "tmp/t.1.vtk"
else:
geomFileName = "database/box.geo"
vtkFileName = "tmp/t.1.vtk"
os.system( "gmsh -0 %s -o tmp/x.geo" % geomFileName )
g=geom.read_gmsh("tmp/x.geo")
g.printinfo()
geom.write_tetgen(g,"tmp/t.poly")
geom.runtetgen("tmp/t.poly",a=0.03,Q=1.0,quadratic=False,
tetgenpath = tetgen_path)
m = Mesh.fromFile("tmp/t.1.node")
m.write( vtkFileName, io = "auto" )
try:
os.makedirs( "tmp" )
except OSError, e:
if e.errno != 17: # [Errno 17] File exists
raise
mesh()
| <commit_before>#!/usr/bin/env python
import os, sys
import pexpect
import geom
from sfepy.fem.mesh import Mesh
try:
from site_cfg import tetgen_path
except ImportError:
tetgen_path = '/usr/bin/tetgen'
def mesh():
if len( sys.argv ) == 3:
geomFileName = sys.argv[1]
vtkFileName = sys.argv[2]
if len( sys.argv ) == 2:
geomFileName = sys.argv[1]
vtkFileName = "tmp/t.1.vtk"
else:
geomFileName = "database/box.geo"
vtkFileName = "tmp/t.1.vtk"
pexpect.run( "gmsh -0 %s -o tmp/x.geo" % geomFileName )
g=geom.read_gmsh("tmp/x.geo")
g.printinfo()
geom.write_tetgen(g,"tmp/t.poly")
geom.runtetgen("tmp/t.poly",a=0.0003,Q=1.0,quadratic=False,
tetgenpath = tetgen_path)
m = Mesh.fromFile("tmp/t.1.node")
m.write( vtkFileName, io = "auto" )
try:
os.makedirs( "tmp" )
except OSError, e:
if e.errno != 17: # [Errno 17] File exists
raise
mesh()
<commit_msg>Use os.system() instead of pexpect.run().
Also do not generate too dense mesh by default, so that it's faster.<commit_after> | #!/usr/bin/env python
import os, sys
import geom
from sfepy.fem.mesh import Mesh
try:
from site_cfg import tetgen_path
except ImportError:
tetgen_path = '/usr/bin/tetgen'
def mesh():
if len( sys.argv ) == 3:
geomFileName = sys.argv[1]
vtkFileName = sys.argv[2]
if len( sys.argv ) == 2:
geomFileName = sys.argv[1]
vtkFileName = "tmp/t.1.vtk"
else:
geomFileName = "database/box.geo"
vtkFileName = "tmp/t.1.vtk"
os.system( "gmsh -0 %s -o tmp/x.geo" % geomFileName )
g=geom.read_gmsh("tmp/x.geo")
g.printinfo()
geom.write_tetgen(g,"tmp/t.poly")
geom.runtetgen("tmp/t.poly",a=0.03,Q=1.0,quadratic=False,
tetgenpath = tetgen_path)
m = Mesh.fromFile("tmp/t.1.node")
m.write( vtkFileName, io = "auto" )
try:
os.makedirs( "tmp" )
except OSError, e:
if e.errno != 17: # [Errno 17] File exists
raise
mesh()
| #!/usr/bin/env python
import os, sys
import pexpect
import geom
from sfepy.fem.mesh import Mesh
try:
from site_cfg import tetgen_path
except ImportError:
tetgen_path = '/usr/bin/tetgen'
def mesh():
if len( sys.argv ) == 3:
geomFileName = sys.argv[1]
vtkFileName = sys.argv[2]
if len( sys.argv ) == 2:
geomFileName = sys.argv[1]
vtkFileName = "tmp/t.1.vtk"
else:
geomFileName = "database/box.geo"
vtkFileName = "tmp/t.1.vtk"
pexpect.run( "gmsh -0 %s -o tmp/x.geo" % geomFileName )
g=geom.read_gmsh("tmp/x.geo")
g.printinfo()
geom.write_tetgen(g,"tmp/t.poly")
geom.runtetgen("tmp/t.poly",a=0.0003,Q=1.0,quadratic=False,
tetgenpath = tetgen_path)
m = Mesh.fromFile("tmp/t.1.node")
m.write( vtkFileName, io = "auto" )
try:
os.makedirs( "tmp" )
except OSError, e:
if e.errno != 17: # [Errno 17] File exists
raise
mesh()
Use os.system() instead of pexpect.run().
Also do not generate too dense mesh by default, so that it's faster.#!/usr/bin/env python
import os, sys
import geom
from sfepy.fem.mesh import Mesh
try:
from site_cfg import tetgen_path
except ImportError:
tetgen_path = '/usr/bin/tetgen'
def mesh():
if len( sys.argv ) == 3:
geomFileName = sys.argv[1]
vtkFileName = sys.argv[2]
if len( sys.argv ) == 2:
geomFileName = sys.argv[1]
vtkFileName = "tmp/t.1.vtk"
else:
geomFileName = "database/box.geo"
vtkFileName = "tmp/t.1.vtk"
os.system( "gmsh -0 %s -o tmp/x.geo" % geomFileName )
g=geom.read_gmsh("tmp/x.geo")
g.printinfo()
geom.write_tetgen(g,"tmp/t.poly")
geom.runtetgen("tmp/t.poly",a=0.03,Q=1.0,quadratic=False,
tetgenpath = tetgen_path)
m = Mesh.fromFile("tmp/t.1.node")
m.write( vtkFileName, io = "auto" )
try:
os.makedirs( "tmp" )
except OSError, e:
if e.errno != 17: # [Errno 17] File exists
raise
mesh()
| <commit_before>#!/usr/bin/env python
import os, sys
import pexpect
import geom
from sfepy.fem.mesh import Mesh
try:
from site_cfg import tetgen_path
except ImportError:
tetgen_path = '/usr/bin/tetgen'
def mesh():
if len( sys.argv ) == 3:
geomFileName = sys.argv[1]
vtkFileName = sys.argv[2]
if len( sys.argv ) == 2:
geomFileName = sys.argv[1]
vtkFileName = "tmp/t.1.vtk"
else:
geomFileName = "database/box.geo"
vtkFileName = "tmp/t.1.vtk"
pexpect.run( "gmsh -0 %s -o tmp/x.geo" % geomFileName )
g=geom.read_gmsh("tmp/x.geo")
g.printinfo()
geom.write_tetgen(g,"tmp/t.poly")
geom.runtetgen("tmp/t.poly",a=0.0003,Q=1.0,quadratic=False,
tetgenpath = tetgen_path)
m = Mesh.fromFile("tmp/t.1.node")
m.write( vtkFileName, io = "auto" )
try:
os.makedirs( "tmp" )
except OSError, e:
if e.errno != 17: # [Errno 17] File exists
raise
mesh()
<commit_msg>Use os.system() instead of pexpect.run().
Also do not generate too dense mesh by default, so that it's faster.<commit_after>#!/usr/bin/env python
import os, sys
import geom
from sfepy.fem.mesh import Mesh
try:
from site_cfg import tetgen_path
except ImportError:
tetgen_path = '/usr/bin/tetgen'
def mesh():
if len( sys.argv ) == 3:
geomFileName = sys.argv[1]
vtkFileName = sys.argv[2]
if len( sys.argv ) == 2:
geomFileName = sys.argv[1]
vtkFileName = "tmp/t.1.vtk"
else:
geomFileName = "database/box.geo"
vtkFileName = "tmp/t.1.vtk"
os.system( "gmsh -0 %s -o tmp/x.geo" % geomFileName )
g=geom.read_gmsh("tmp/x.geo")
g.printinfo()
geom.write_tetgen(g,"tmp/t.poly")
geom.runtetgen("tmp/t.poly",a=0.03,Q=1.0,quadratic=False,
tetgenpath = tetgen_path)
m = Mesh.fromFile("tmp/t.1.node")
m.write( vtkFileName, io = "auto" )
try:
os.makedirs( "tmp" )
except OSError, e:
if e.errno != 17: # [Errno 17] File exists
raise
mesh()
|
d178fb001b8b6869038ed6ec288acf5fb427205c | rssmailer/tasks/mail.py | rssmailer/tasks/mail.py | from celery.decorators import task
from django.core.mail import send_mail
from ..models import Email
@task(ignore_result=True, name="rssmailer.tasks.mail.send")
def send(entry, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending entry: %s" % entry.title)
emails_all = Email.objects.all()
step = 3 # how many recipients in one e-mail
for i in range(0, len(emails_all), step):
recipients = map(lambda e: e.email, emails_all[i:i+step])
send_entry_to.delay(entry.title, entry.summary, recipients)
@task(ignore_result=True, name="rssmailer.tasks.mail.send_entry_to")
def send_entry_to(title, body, recipients, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending to: %s" % ','.join(recipients))
send_mail(title, body, "rssmailer@praus.net", recipients)
| from celery.decorators import task
from django.core.mail import send_mail
from ..models import Email
@task(ignore_result=True, name="rssmailer.tasks.send")
def send(entry, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending entry: %s" % entry.title)
emails_all = Email.objects.all()
step = 3 # how many recipients in one e-mail
for i in range(0, len(emails_all), step):
recipients = map(lambda e: e.email, emails_all[i:i+step])
send_entry_to.delay(entry.title, entry.summary, recipients)
@task(ignore_result=True, name="rssmailer.tasks.send_entry_to")
def send_entry_to(title, body, recipients, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending to: %s" % ','.join(recipients))
send_mail(title, body, "rssmailer@praus.net", recipients) | Fix naming issues with tasks | Fix naming issues with tasks
| Python | bsd-3-clause | praus/django-rssmailer | from celery.decorators import task
from django.core.mail import send_mail
from ..models import Email
@task(ignore_result=True, name="rssmailer.tasks.mail.send")
def send(entry, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending entry: %s" % entry.title)
emails_all = Email.objects.all()
step = 3 # how many recipients in one e-mail
for i in range(0, len(emails_all), step):
recipients = map(lambda e: e.email, emails_all[i:i+step])
send_entry_to.delay(entry.title, entry.summary, recipients)
@task(ignore_result=True, name="rssmailer.tasks.mail.send_entry_to")
def send_entry_to(title, body, recipients, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending to: %s" % ','.join(recipients))
send_mail(title, body, "rssmailer@praus.net", recipients)
Fix naming issues with tasks | from celery.decorators import task
from django.core.mail import send_mail
from ..models import Email
@task(ignore_result=True, name="rssmailer.tasks.send")
def send(entry, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending entry: %s" % entry.title)
emails_all = Email.objects.all()
step = 3 # how many recipients in one e-mail
for i in range(0, len(emails_all), step):
recipients = map(lambda e: e.email, emails_all[i:i+step])
send_entry_to.delay(entry.title, entry.summary, recipients)
@task(ignore_result=True, name="rssmailer.tasks.send_entry_to")
def send_entry_to(title, body, recipients, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending to: %s" % ','.join(recipients))
send_mail(title, body, "rssmailer@praus.net", recipients) | <commit_before>from celery.decorators import task
from django.core.mail import send_mail
from ..models import Email
@task(ignore_result=True, name="rssmailer.tasks.mail.send")
def send(entry, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending entry: %s" % entry.title)
emails_all = Email.objects.all()
step = 3 # how many recipients in one e-mail
for i in range(0, len(emails_all), step):
recipients = map(lambda e: e.email, emails_all[i:i+step])
send_entry_to.delay(entry.title, entry.summary, recipients)
@task(ignore_result=True, name="rssmailer.tasks.mail.send_entry_to")
def send_entry_to(title, body, recipients, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending to: %s" % ','.join(recipients))
send_mail(title, body, "rssmailer@praus.net", recipients)
<commit_msg>Fix naming issues with tasks<commit_after> | from celery.decorators import task
from django.core.mail import send_mail
from ..models import Email
@task(ignore_result=True, name="rssmailer.tasks.send")
def send(entry, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending entry: %s" % entry.title)
emails_all = Email.objects.all()
step = 3 # how many recipients in one e-mail
for i in range(0, len(emails_all), step):
recipients = map(lambda e: e.email, emails_all[i:i+step])
send_entry_to.delay(entry.title, entry.summary, recipients)
@task(ignore_result=True, name="rssmailer.tasks.send_entry_to")
def send_entry_to(title, body, recipients, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending to: %s" % ','.join(recipients))
send_mail(title, body, "rssmailer@praus.net", recipients) | from celery.decorators import task
from django.core.mail import send_mail
from ..models import Email
@task(ignore_result=True, name="rssmailer.tasks.mail.send")
def send(entry, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending entry: %s" % entry.title)
emails_all = Email.objects.all()
step = 3 # how many recipients in one e-mail
for i in range(0, len(emails_all), step):
recipients = map(lambda e: e.email, emails_all[i:i+step])
send_entry_to.delay(entry.title, entry.summary, recipients)
@task(ignore_result=True, name="rssmailer.tasks.mail.send_entry_to")
def send_entry_to(title, body, recipients, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending to: %s" % ','.join(recipients))
send_mail(title, body, "rssmailer@praus.net", recipients)
Fix naming issues with tasksfrom celery.decorators import task
from django.core.mail import send_mail
from ..models import Email
@task(ignore_result=True, name="rssmailer.tasks.send")
def send(entry, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending entry: %s" % entry.title)
emails_all = Email.objects.all()
step = 3 # how many recipients in one e-mail
for i in range(0, len(emails_all), step):
recipients = map(lambda e: e.email, emails_all[i:i+step])
send_entry_to.delay(entry.title, entry.summary, recipients)
@task(ignore_result=True, name="rssmailer.tasks.send_entry_to")
def send_entry_to(title, body, recipients, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending to: %s" % ','.join(recipients))
send_mail(title, body, "rssmailer@praus.net", recipients) | <commit_before>from celery.decorators import task
from django.core.mail import send_mail
from ..models import Email
@task(ignore_result=True, name="rssmailer.tasks.mail.send")
def send(entry, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending entry: %s" % entry.title)
emails_all = Email.objects.all()
step = 3 # how many recipients in one e-mail
for i in range(0, len(emails_all), step):
recipients = map(lambda e: e.email, emails_all[i:i+step])
send_entry_to.delay(entry.title, entry.summary, recipients)
@task(ignore_result=True, name="rssmailer.tasks.mail.send_entry_to")
def send_entry_to(title, body, recipients, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending to: %s" % ','.join(recipients))
send_mail(title, body, "rssmailer@praus.net", recipients)
<commit_msg>Fix naming issues with tasks<commit_after>from celery.decorators import task
from django.core.mail import send_mail
from ..models import Email
@task(ignore_result=True, name="rssmailer.tasks.send")
def send(entry, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending entry: %s" % entry.title)
emails_all = Email.objects.all()
step = 3 # how many recipients in one e-mail
for i in range(0, len(emails_all), step):
recipients = map(lambda e: e.email, emails_all[i:i+step])
send_entry_to.delay(entry.title, entry.summary, recipients)
@task(ignore_result=True, name="rssmailer.tasks.send_entry_to")
def send_entry_to(title, body, recipients, **kwargs):
logger = send.get_logger(**kwargs)
logger.info("Sending to: %s" % ','.join(recipients))
send_mail(title, body, "rssmailer@praus.net", recipients) |
615627cf6ea4725bed7886e822bc01c12d9fdead | nodewatcher/web/sanitize-dump.py | nodewatcher/web/sanitize-dump.py | #!/usr/bin/python
# Setup import paths, since we are using Django models
import sys, os
sys.path.append('/var/www/django')
os.environ['DJANGO_SETTINGS_MODULE'] = 'wlanlj.settings_production'
# Imports
from django.core import serializers
if len(sys.argv) != 4:
print "Usage: %s format input-file output-file" % sys.argv[0]
exit(1)
if sys.argv[1] not in ('json', 'xml'):
print "Invalid format '%s'! Valid formats are: json xml" % sys.argv[1]
exit(1)
def object_transformator():
# Read all objects one by one
for holder in serializers.deserialize(sys.argv[1], open(sys.argv[2], 'r')):
name = holder.object.__class__.__name__
object = holder.object
# Some objects need to be sanitized
if name == 'Node':
object.notes = ''
elif name == 'UserAccount':
object.vpn_password = 'XXX'
object.phone = '5551234'
elif name == 'User':
object.password = 'XXX'
elif name == 'Profile':
object.root_pass = 'XXX'
yield holder.object
# Write transformed objects out
out = open(sys.argv[3], 'w')
serializers.serialize(sys.argv[1], object_transformator(), stream = out)
out.close()
| #!/usr/bin/python
# Setup import paths, since we are using Django models
import sys, os
sys.path.append('/var/www/django')
os.environ['DJANGO_SETTINGS_MODULE'] = 'wlanlj.settings_production'
# Imports
from django.core import serializers
if len(sys.argv) != 4:
print "Usage: %s format input-file output-file" % sys.argv[0]
exit(1)
if sys.argv[1] not in ('json', 'xml'):
print "Invalid format '%s'! Valid formats are: json xml" % sys.argv[1]
exit(1)
def object_transformator():
# Read all objects one by one
for holder in serializers.deserialize(sys.argv[1], open(sys.argv[2], 'r')):
name = holder.object.__class__.__name__
object = holder.object
# Some objects need to be sanitized
if name == 'Node':
object.notes = ''
elif name == 'UserAccount':
object.vpn_password = 'XXX'
object.phone = '5551234'
elif name == 'User':
object.password = 'XXX'
elif name == 'Profile':
object.root_pass = 'XXX'
elif name == 'StatsSolar':
continue
yield holder.object
# Write transformed objects out
out = open(sys.argv[3], 'w')
serializers.serialize(sys.argv[1], object_transformator(), stream = out)
out.close()
| Remove solar statistics data from dumps. | Remove solar statistics data from dumps.
| Python | agpl-3.0 | galaxor/Nodewatcher,galaxor/Nodewatcher,galaxor/Nodewatcher,galaxor/Nodewatcher | #!/usr/bin/python
# Setup import paths, since we are using Django models
import sys, os
sys.path.append('/var/www/django')
os.environ['DJANGO_SETTINGS_MODULE'] = 'wlanlj.settings_production'
# Imports
from django.core import serializers
if len(sys.argv) != 4:
print "Usage: %s format input-file output-file" % sys.argv[0]
exit(1)
if sys.argv[1] not in ('json', 'xml'):
print "Invalid format '%s'! Valid formats are: json xml" % sys.argv[1]
exit(1)
def object_transformator():
# Read all objects one by one
for holder in serializers.deserialize(sys.argv[1], open(sys.argv[2], 'r')):
name = holder.object.__class__.__name__
object = holder.object
# Some objects need to be sanitized
if name == 'Node':
object.notes = ''
elif name == 'UserAccount':
object.vpn_password = 'XXX'
object.phone = '5551234'
elif name == 'User':
object.password = 'XXX'
elif name == 'Profile':
object.root_pass = 'XXX'
yield holder.object
# Write transformed objects out
out = open(sys.argv[3], 'w')
serializers.serialize(sys.argv[1], object_transformator(), stream = out)
out.close()
Remove solar statistics data from dumps. | #!/usr/bin/python
# Setup import paths, since we are using Django models
import sys, os
sys.path.append('/var/www/django')
os.environ['DJANGO_SETTINGS_MODULE'] = 'wlanlj.settings_production'
# Imports
from django.core import serializers
if len(sys.argv) != 4:
print "Usage: %s format input-file output-file" % sys.argv[0]
exit(1)
if sys.argv[1] not in ('json', 'xml'):
print "Invalid format '%s'! Valid formats are: json xml" % sys.argv[1]
exit(1)
def object_transformator():
# Read all objects one by one
for holder in serializers.deserialize(sys.argv[1], open(sys.argv[2], 'r')):
name = holder.object.__class__.__name__
object = holder.object
# Some objects need to be sanitized
if name == 'Node':
object.notes = ''
elif name == 'UserAccount':
object.vpn_password = 'XXX'
object.phone = '5551234'
elif name == 'User':
object.password = 'XXX'
elif name == 'Profile':
object.root_pass = 'XXX'
elif name == 'StatsSolar':
continue
yield holder.object
# Write transformed objects out
out = open(sys.argv[3], 'w')
serializers.serialize(sys.argv[1], object_transformator(), stream = out)
out.close()
| <commit_before>#!/usr/bin/python
# Setup import paths, since we are using Django models
import sys, os
sys.path.append('/var/www/django')
os.environ['DJANGO_SETTINGS_MODULE'] = 'wlanlj.settings_production'
# Imports
from django.core import serializers
if len(sys.argv) != 4:
print "Usage: %s format input-file output-file" % sys.argv[0]
exit(1)
if sys.argv[1] not in ('json', 'xml'):
print "Invalid format '%s'! Valid formats are: json xml" % sys.argv[1]
exit(1)
def object_transformator():
# Read all objects one by one
for holder in serializers.deserialize(sys.argv[1], open(sys.argv[2], 'r')):
name = holder.object.__class__.__name__
object = holder.object
# Some objects need to be sanitized
if name == 'Node':
object.notes = ''
elif name == 'UserAccount':
object.vpn_password = 'XXX'
object.phone = '5551234'
elif name == 'User':
object.password = 'XXX'
elif name == 'Profile':
object.root_pass = 'XXX'
yield holder.object
# Write transformed objects out
out = open(sys.argv[3], 'w')
serializers.serialize(sys.argv[1], object_transformator(), stream = out)
out.close()
<commit_msg>Remove solar statistics data from dumps.<commit_after> | #!/usr/bin/python
# Setup import paths, since we are using Django models
import sys, os
sys.path.append('/var/www/django')
os.environ['DJANGO_SETTINGS_MODULE'] = 'wlanlj.settings_production'
# Imports
from django.core import serializers
if len(sys.argv) != 4:
print "Usage: %s format input-file output-file" % sys.argv[0]
exit(1)
if sys.argv[1] not in ('json', 'xml'):
print "Invalid format '%s'! Valid formats are: json xml" % sys.argv[1]
exit(1)
def object_transformator():
# Read all objects one by one
for holder in serializers.deserialize(sys.argv[1], open(sys.argv[2], 'r')):
name = holder.object.__class__.__name__
object = holder.object
# Some objects need to be sanitized
if name == 'Node':
object.notes = ''
elif name == 'UserAccount':
object.vpn_password = 'XXX'
object.phone = '5551234'
elif name == 'User':
object.password = 'XXX'
elif name == 'Profile':
object.root_pass = 'XXX'
elif name == 'StatsSolar':
continue
yield holder.object
# Write transformed objects out
out = open(sys.argv[3], 'w')
serializers.serialize(sys.argv[1], object_transformator(), stream = out)
out.close()
| #!/usr/bin/python
# Setup import paths, since we are using Django models
import sys, os
sys.path.append('/var/www/django')
os.environ['DJANGO_SETTINGS_MODULE'] = 'wlanlj.settings_production'
# Imports
from django.core import serializers
if len(sys.argv) != 4:
print "Usage: %s format input-file output-file" % sys.argv[0]
exit(1)
if sys.argv[1] not in ('json', 'xml'):
print "Invalid format '%s'! Valid formats are: json xml" % sys.argv[1]
exit(1)
def object_transformator():
# Read all objects one by one
for holder in serializers.deserialize(sys.argv[1], open(sys.argv[2], 'r')):
name = holder.object.__class__.__name__
object = holder.object
# Some objects need to be sanitized
if name == 'Node':
object.notes = ''
elif name == 'UserAccount':
object.vpn_password = 'XXX'
object.phone = '5551234'
elif name == 'User':
object.password = 'XXX'
elif name == 'Profile':
object.root_pass = 'XXX'
yield holder.object
# Write transformed objects out
out = open(sys.argv[3], 'w')
serializers.serialize(sys.argv[1], object_transformator(), stream = out)
out.close()
Remove solar statistics data from dumps.#!/usr/bin/python
# Setup import paths, since we are using Django models
import sys, os
sys.path.append('/var/www/django')
os.environ['DJANGO_SETTINGS_MODULE'] = 'wlanlj.settings_production'
# Imports
from django.core import serializers
if len(sys.argv) != 4:
print "Usage: %s format input-file output-file" % sys.argv[0]
exit(1)
if sys.argv[1] not in ('json', 'xml'):
print "Invalid format '%s'! Valid formats are: json xml" % sys.argv[1]
exit(1)
def object_transformator():
# Read all objects one by one
for holder in serializers.deserialize(sys.argv[1], open(sys.argv[2], 'r')):
name = holder.object.__class__.__name__
object = holder.object
# Some objects need to be sanitized
if name == 'Node':
object.notes = ''
elif name == 'UserAccount':
object.vpn_password = 'XXX'
object.phone = '5551234'
elif name == 'User':
object.password = 'XXX'
elif name == 'Profile':
object.root_pass = 'XXX'
elif name == 'StatsSolar':
continue
yield holder.object
# Write transformed objects out
out = open(sys.argv[3], 'w')
serializers.serialize(sys.argv[1], object_transformator(), stream = out)
out.close()
| <commit_before>#!/usr/bin/python
# Setup import paths, since we are using Django models
import sys, os
sys.path.append('/var/www/django')
os.environ['DJANGO_SETTINGS_MODULE'] = 'wlanlj.settings_production'
# Imports
from django.core import serializers
if len(sys.argv) != 4:
print "Usage: %s format input-file output-file" % sys.argv[0]
exit(1)
if sys.argv[1] not in ('json', 'xml'):
print "Invalid format '%s'! Valid formats are: json xml" % sys.argv[1]
exit(1)
def object_transformator():
# Read all objects one by one
for holder in serializers.deserialize(sys.argv[1], open(sys.argv[2], 'r')):
name = holder.object.__class__.__name__
object = holder.object
# Some objects need to be sanitized
if name == 'Node':
object.notes = ''
elif name == 'UserAccount':
object.vpn_password = 'XXX'
object.phone = '5551234'
elif name == 'User':
object.password = 'XXX'
elif name == 'Profile':
object.root_pass = 'XXX'
yield holder.object
# Write transformed objects out
out = open(sys.argv[3], 'w')
serializers.serialize(sys.argv[1], object_transformator(), stream = out)
out.close()
<commit_msg>Remove solar statistics data from dumps.<commit_after>#!/usr/bin/python
# Setup import paths, since we are using Django models
import sys, os
sys.path.append('/var/www/django')
os.environ['DJANGO_SETTINGS_MODULE'] = 'wlanlj.settings_production'
# Imports
from django.core import serializers
if len(sys.argv) != 4:
print "Usage: %s format input-file output-file" % sys.argv[0]
exit(1)
if sys.argv[1] not in ('json', 'xml'):
print "Invalid format '%s'! Valid formats are: json xml" % sys.argv[1]
exit(1)
def object_transformator():
# Read all objects one by one
for holder in serializers.deserialize(sys.argv[1], open(sys.argv[2], 'r')):
name = holder.object.__class__.__name__
object = holder.object
# Some objects need to be sanitized
if name == 'Node':
object.notes = ''
elif name == 'UserAccount':
object.vpn_password = 'XXX'
object.phone = '5551234'
elif name == 'User':
object.password = 'XXX'
elif name == 'Profile':
object.root_pass = 'XXX'
elif name == 'StatsSolar':
continue
yield holder.object
# Write transformed objects out
out = open(sys.argv[3], 'w')
serializers.serialize(sys.argv[1], object_transformator(), stream = out)
out.close()
|
8ca20bec63b8f8aaff55a7012c69a2644e292095 | mltsp/science_features/lomb_scargle_fast.py | mltsp/science_features/lomb_scargle_fast.py | import numpy as np
import gatspy
def lomb_scargle_fast_period(t, m, e):
"""Fits a simple sinuosidal model
y(t) = A sin(2*pi*w*t) + B cos(2*pi*w*t) + c
and returns the estimated period 1/w. Much faster than fitting the
full multi-frequency model used by `science_features.lomb_scargle`.
"""
opt_args = {'period_range': (2*t.max() / len(t), t.max()), 'quiet': True}
model = gatspy.periodic.LombScargleFast(fit_period=True, optimizer_kwds=opt_args)
model.fit(t, m, e)
return model.best_period
| import numpy as np
import gatspy
def lomb_scargle_fast_period(t, m, e):
"""Fits a simple sinuosidal model
y(t) = A sin(2*pi*w*t + phi) + c
and returns the estimated period 1/w. Much faster than fitting the
full multi-frequency model used by `science_features.lomb_scargle`.
"""
opt_args = {'period_range': (2*t.max() / len(t), t.max()), 'quiet': True}
model = gatspy.periodic.LombScargleFast(fit_period=True, optimizer_kwds=opt_args)
model.fit(t, m, e)
return model.best_period
| Change docstring for `period_fast` feature | Change docstring for `period_fast` feature
| Python | bsd-3-clause | bnaul/mltsp,mltsp/mltsp,mltsp/mltsp,bnaul/mltsp,acrellin/mltsp,bnaul/mltsp,acrellin/mltsp,mltsp/mltsp,mltsp/mltsp,bnaul/mltsp,acrellin/mltsp,mltsp/mltsp,mltsp/mltsp,acrellin/mltsp,bnaul/mltsp,acrellin/mltsp,bnaul/mltsp,acrellin/mltsp | import numpy as np
import gatspy
def lomb_scargle_fast_period(t, m, e):
"""Fits a simple sinuosidal model
y(t) = A sin(2*pi*w*t) + B cos(2*pi*w*t) + c
and returns the estimated period 1/w. Much faster than fitting the
full multi-frequency model used by `science_features.lomb_scargle`.
"""
opt_args = {'period_range': (2*t.max() / len(t), t.max()), 'quiet': True}
model = gatspy.periodic.LombScargleFast(fit_period=True, optimizer_kwds=opt_args)
model.fit(t, m, e)
return model.best_period
Change docstring for `period_fast` feature | import numpy as np
import gatspy
def lomb_scargle_fast_period(t, m, e):
"""Fits a simple sinuosidal model
y(t) = A sin(2*pi*w*t + phi) + c
and returns the estimated period 1/w. Much faster than fitting the
full multi-frequency model used by `science_features.lomb_scargle`.
"""
opt_args = {'period_range': (2*t.max() / len(t), t.max()), 'quiet': True}
model = gatspy.periodic.LombScargleFast(fit_period=True, optimizer_kwds=opt_args)
model.fit(t, m, e)
return model.best_period
| <commit_before>import numpy as np
import gatspy
def lomb_scargle_fast_period(t, m, e):
"""Fits a simple sinuosidal model
y(t) = A sin(2*pi*w*t) + B cos(2*pi*w*t) + c
and returns the estimated period 1/w. Much faster than fitting the
full multi-frequency model used by `science_features.lomb_scargle`.
"""
opt_args = {'period_range': (2*t.max() / len(t), t.max()), 'quiet': True}
model = gatspy.periodic.LombScargleFast(fit_period=True, optimizer_kwds=opt_args)
model.fit(t, m, e)
return model.best_period
<commit_msg>Change docstring for `period_fast` feature<commit_after> | import numpy as np
import gatspy
def lomb_scargle_fast_period(t, m, e):
"""Fits a simple sinuosidal model
y(t) = A sin(2*pi*w*t + phi) + c
and returns the estimated period 1/w. Much faster than fitting the
full multi-frequency model used by `science_features.lomb_scargle`.
"""
opt_args = {'period_range': (2*t.max() / len(t), t.max()), 'quiet': True}
model = gatspy.periodic.LombScargleFast(fit_period=True, optimizer_kwds=opt_args)
model.fit(t, m, e)
return model.best_period
| import numpy as np
import gatspy
def lomb_scargle_fast_period(t, m, e):
"""Fits a simple sinuosidal model
y(t) = A sin(2*pi*w*t) + B cos(2*pi*w*t) + c
and returns the estimated period 1/w. Much faster than fitting the
full multi-frequency model used by `science_features.lomb_scargle`.
"""
opt_args = {'period_range': (2*t.max() / len(t), t.max()), 'quiet': True}
model = gatspy.periodic.LombScargleFast(fit_period=True, optimizer_kwds=opt_args)
model.fit(t, m, e)
return model.best_period
Change docstring for `period_fast` featureimport numpy as np
import gatspy
def lomb_scargle_fast_period(t, m, e):
"""Fits a simple sinuosidal model
y(t) = A sin(2*pi*w*t + phi) + c
and returns the estimated period 1/w. Much faster than fitting the
full multi-frequency model used by `science_features.lomb_scargle`.
"""
opt_args = {'period_range': (2*t.max() / len(t), t.max()), 'quiet': True}
model = gatspy.periodic.LombScargleFast(fit_period=True, optimizer_kwds=opt_args)
model.fit(t, m, e)
return model.best_period
| <commit_before>import numpy as np
import gatspy
def lomb_scargle_fast_period(t, m, e):
"""Fits a simple sinuosidal model
y(t) = A sin(2*pi*w*t) + B cos(2*pi*w*t) + c
and returns the estimated period 1/w. Much faster than fitting the
full multi-frequency model used by `science_features.lomb_scargle`.
"""
opt_args = {'period_range': (2*t.max() / len(t), t.max()), 'quiet': True}
model = gatspy.periodic.LombScargleFast(fit_period=True, optimizer_kwds=opt_args)
model.fit(t, m, e)
return model.best_period
<commit_msg>Change docstring for `period_fast` feature<commit_after>import numpy as np
import gatspy
def lomb_scargle_fast_period(t, m, e):
"""Fits a simple sinuosidal model
y(t) = A sin(2*pi*w*t + phi) + c
and returns the estimated period 1/w. Much faster than fitting the
full multi-frequency model used by `science_features.lomb_scargle`.
"""
opt_args = {'period_range': (2*t.max() / len(t), t.max()), 'quiet': True}
model = gatspy.periodic.LombScargleFast(fit_period=True, optimizer_kwds=opt_args)
model.fit(t, m, e)
return model.best_period
|
4d5c8ec9c2006b78a42461af43944de8ab7bc9ea | us_ignite/common/sanitizer.py | us_ignite/common/sanitizer.py | import bleach
ALLOWED_TAGS = [
'a',
'abbr',
'acronym',
'b',
'blockquote',
'code',
'em',
'i',
'li',
'ol',
'strong',
'ul',
'p',
'br',
]
ALLOWED_ATTRIBUTES = {
'a': ['href', 'title'],
'abbr': ['title'],
'acronym': ['title'],
}
ALLOWED_STYLES = []
def sanitize(text):
"""Cleans the HTML received."""
cleaned_text = bleach.clean(
text, tags=ALLOWED_TAGS, attributes=ALLOWED_ATTRIBUTES,
styles=ALLOWED_STYLES, strip=True)
return cleaned_text
| import bleach
ALLOWED_TAGS = [
'a',
'abbr',
'acronym',
'b',
'blockquote',
'code',
'em',
'i',
'li',
'ol',
'strong',
'ul',
'p',
'br',
'h3',
'h4',
'h5',
'h6',
]
ALLOWED_ATTRIBUTES = {
'a': ['href', 'title'],
'abbr': ['title'],
'acronym': ['title'],
}
ALLOWED_STYLES = []
def sanitize(text):
"""Cleans the HTML received."""
cleaned_text = bleach.clean(
text, tags=ALLOWED_TAGS, attributes=ALLOWED_ATTRIBUTES,
styles=ALLOWED_STYLES, strip=True)
return cleaned_text
| Allow low level titles when sanitising. | Allow low level titles when sanitising.
| Python | bsd-3-clause | us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite | import bleach
ALLOWED_TAGS = [
'a',
'abbr',
'acronym',
'b',
'blockquote',
'code',
'em',
'i',
'li',
'ol',
'strong',
'ul',
'p',
'br',
]
ALLOWED_ATTRIBUTES = {
'a': ['href', 'title'],
'abbr': ['title'],
'acronym': ['title'],
}
ALLOWED_STYLES = []
def sanitize(text):
"""Cleans the HTML received."""
cleaned_text = bleach.clean(
text, tags=ALLOWED_TAGS, attributes=ALLOWED_ATTRIBUTES,
styles=ALLOWED_STYLES, strip=True)
return cleaned_text
Allow low level titles when sanitising. | import bleach
ALLOWED_TAGS = [
'a',
'abbr',
'acronym',
'b',
'blockquote',
'code',
'em',
'i',
'li',
'ol',
'strong',
'ul',
'p',
'br',
'h3',
'h4',
'h5',
'h6',
]
ALLOWED_ATTRIBUTES = {
'a': ['href', 'title'],
'abbr': ['title'],
'acronym': ['title'],
}
ALLOWED_STYLES = []
def sanitize(text):
"""Cleans the HTML received."""
cleaned_text = bleach.clean(
text, tags=ALLOWED_TAGS, attributes=ALLOWED_ATTRIBUTES,
styles=ALLOWED_STYLES, strip=True)
return cleaned_text
| <commit_before>import bleach
ALLOWED_TAGS = [
'a',
'abbr',
'acronym',
'b',
'blockquote',
'code',
'em',
'i',
'li',
'ol',
'strong',
'ul',
'p',
'br',
]
ALLOWED_ATTRIBUTES = {
'a': ['href', 'title'],
'abbr': ['title'],
'acronym': ['title'],
}
ALLOWED_STYLES = []
def sanitize(text):
"""Cleans the HTML received."""
cleaned_text = bleach.clean(
text, tags=ALLOWED_TAGS, attributes=ALLOWED_ATTRIBUTES,
styles=ALLOWED_STYLES, strip=True)
return cleaned_text
<commit_msg>Allow low level titles when sanitising.<commit_after> | import bleach
ALLOWED_TAGS = [
'a',
'abbr',
'acronym',
'b',
'blockquote',
'code',
'em',
'i',
'li',
'ol',
'strong',
'ul',
'p',
'br',
'h3',
'h4',
'h5',
'h6',
]
ALLOWED_ATTRIBUTES = {
'a': ['href', 'title'],
'abbr': ['title'],
'acronym': ['title'],
}
ALLOWED_STYLES = []
def sanitize(text):
"""Cleans the HTML received."""
cleaned_text = bleach.clean(
text, tags=ALLOWED_TAGS, attributes=ALLOWED_ATTRIBUTES,
styles=ALLOWED_STYLES, strip=True)
return cleaned_text
| import bleach
ALLOWED_TAGS = [
'a',
'abbr',
'acronym',
'b',
'blockquote',
'code',
'em',
'i',
'li',
'ol',
'strong',
'ul',
'p',
'br',
]
ALLOWED_ATTRIBUTES = {
'a': ['href', 'title'],
'abbr': ['title'],
'acronym': ['title'],
}
ALLOWED_STYLES = []
def sanitize(text):
"""Cleans the HTML received."""
cleaned_text = bleach.clean(
text, tags=ALLOWED_TAGS, attributes=ALLOWED_ATTRIBUTES,
styles=ALLOWED_STYLES, strip=True)
return cleaned_text
Allow low level titles when sanitising.import bleach
ALLOWED_TAGS = [
'a',
'abbr',
'acronym',
'b',
'blockquote',
'code',
'em',
'i',
'li',
'ol',
'strong',
'ul',
'p',
'br',
'h3',
'h4',
'h5',
'h6',
]
ALLOWED_ATTRIBUTES = {
'a': ['href', 'title'],
'abbr': ['title'],
'acronym': ['title'],
}
ALLOWED_STYLES = []
def sanitize(text):
"""Cleans the HTML received."""
cleaned_text = bleach.clean(
text, tags=ALLOWED_TAGS, attributes=ALLOWED_ATTRIBUTES,
styles=ALLOWED_STYLES, strip=True)
return cleaned_text
| <commit_before>import bleach
ALLOWED_TAGS = [
'a',
'abbr',
'acronym',
'b',
'blockquote',
'code',
'em',
'i',
'li',
'ol',
'strong',
'ul',
'p',
'br',
]
ALLOWED_ATTRIBUTES = {
'a': ['href', 'title'],
'abbr': ['title'],
'acronym': ['title'],
}
ALLOWED_STYLES = []
def sanitize(text):
"""Cleans the HTML received."""
cleaned_text = bleach.clean(
text, tags=ALLOWED_TAGS, attributes=ALLOWED_ATTRIBUTES,
styles=ALLOWED_STYLES, strip=True)
return cleaned_text
<commit_msg>Allow low level titles when sanitising.<commit_after>import bleach
ALLOWED_TAGS = [
'a',
'abbr',
'acronym',
'b',
'blockquote',
'code',
'em',
'i',
'li',
'ol',
'strong',
'ul',
'p',
'br',
'h3',
'h4',
'h5',
'h6',
]
ALLOWED_ATTRIBUTES = {
'a': ['href', 'title'],
'abbr': ['title'],
'acronym': ['title'],
}
ALLOWED_STYLES = []
def sanitize(text):
"""Cleans the HTML received."""
cleaned_text = bleach.clean(
text, tags=ALLOWED_TAGS, attributes=ALLOWED_ATTRIBUTES,
styles=ALLOWED_STYLES, strip=True)
return cleaned_text
|
858bc6f152a87298f9bd3568712aed49b6e02e42 | suave/suave.py | suave/suave.py | #!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Redraw the screen only when it changes.
if screen.is_wintouched():
screen.clear()
screen.refresh()
# Give every box an opportunity to redraw if it has changed.
[box.redraw_if_changed() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
| #!/usr/bin/env python
import curses
import os
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Redraw the screen only when it changes.
if screen.is_wintouched():
screen.clear()
screen.refresh()
# Give every box an opportunity to redraw if it has changed.
[box.redraw_if_changed() for box in boxes]
# Wait before redrawing again.
curses.napms(1000)
curses.wrapper(main)
| Use napms method from curses rather than sleep method from time | Use napms method from curses rather than sleep method from time
| Python | mit | countermeasure/suave | #!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Redraw the screen only when it changes.
if screen.is_wintouched():
screen.clear()
screen.refresh()
# Give every box an opportunity to redraw if it has changed.
[box.redraw_if_changed() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
Use napms method from curses rather than sleep method from time | #!/usr/bin/env python
import curses
import os
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Redraw the screen only when it changes.
if screen.is_wintouched():
screen.clear()
screen.refresh()
# Give every box an opportunity to redraw if it has changed.
[box.redraw_if_changed() for box in boxes]
# Wait before redrawing again.
curses.napms(1000)
curses.wrapper(main)
| <commit_before>#!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Redraw the screen only when it changes.
if screen.is_wintouched():
screen.clear()
screen.refresh()
# Give every box an opportunity to redraw if it has changed.
[box.redraw_if_changed() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
<commit_msg>Use napms method from curses rather than sleep method from time<commit_after> | #!/usr/bin/env python
import curses
import os
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Redraw the screen only when it changes.
if screen.is_wintouched():
screen.clear()
screen.refresh()
# Give every box an opportunity to redraw if it has changed.
[box.redraw_if_changed() for box in boxes]
# Wait before redrawing again.
curses.napms(1000)
curses.wrapper(main)
| #!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Redraw the screen only when it changes.
if screen.is_wintouched():
screen.clear()
screen.refresh()
# Give every box an opportunity to redraw if it has changed.
[box.redraw_if_changed() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
Use napms method from curses rather than sleep method from time#!/usr/bin/env python
import curses
import os
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Redraw the screen only when it changes.
if screen.is_wintouched():
screen.clear()
screen.refresh()
# Give every box an opportunity to redraw if it has changed.
[box.redraw_if_changed() for box in boxes]
# Wait before redrawing again.
curses.napms(1000)
curses.wrapper(main)
| <commit_before>#!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Redraw the screen only when it changes.
if screen.is_wintouched():
screen.clear()
screen.refresh()
# Give every box an opportunity to redraw if it has changed.
[box.redraw_if_changed() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
<commit_msg>Use napms method from curses rather than sleep method from time<commit_after>#!/usr/bin/env python
import curses
import os
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Redraw the screen only when it changes.
if screen.is_wintouched():
screen.clear()
screen.refresh()
# Give every box an opportunity to redraw if it has changed.
[box.redraw_if_changed() for box in boxes]
# Wait before redrawing again.
curses.napms(1000)
curses.wrapper(main)
|
59627d96975b2735fabd0d44e34e018ca97dec2b | tweepy/error.py | tweepy/error.py | # Tweepy
# Copyright 2009-2010 Joshua Roesslein
# See LICENSE for details.
from __future__ import print_function
import six
class TweepError(Exception):
"""Tweepy exception"""
def __init__(self, reason, response=None, api_code=None):
self.reason = six.text_type(reason)
self.response = response
self.api_code = api_code
super(TweepError, self).__init__(self, reason)
def __str__(self):
return self.reason
def is_rate_limit_error_message(message):
"""Check if the supplied error message belongs to a rate limit error."""
return isinstance(message, list) \
and len(message) > 0 \
and 'code' in message[0] \
and message[0]['code'] == 88
class RateLimitError(TweepError):
"""Exception for Tweepy hitting the rate limit."""
# RateLimitError has the exact same properties and inner workings
# as TweepError for backwards compatibility reasons.
pass
| # Tweepy
# Copyright 2009-2010 Joshua Roesslein
# See LICENSE for details.
from __future__ import print_function
import six
class TweepError(Exception):
"""Tweepy exception"""
def __init__(self, reason, response=None, api_code=None):
self.reason = six.text_type(reason)
self.response = response
self.api_code = api_code
super(TweepError, self).__init__(reason)
def __str__(self):
return self.reason
def is_rate_limit_error_message(message):
"""Check if the supplied error message belongs to a rate limit error."""
return isinstance(message, list) \
and len(message) > 0 \
and 'code' in message[0] \
and message[0]['code'] == 88
class RateLimitError(TweepError):
"""Exception for Tweepy hitting the rate limit."""
# RateLimitError has the exact same properties and inner workings
# as TweepError for backwards compatibility reasons.
pass
| Fix super usage in TweepError initialization | Fix super usage in TweepError initialization
| Python | mit | svven/tweepy,tweepy/tweepy | # Tweepy
# Copyright 2009-2010 Joshua Roesslein
# See LICENSE for details.
from __future__ import print_function
import six
class TweepError(Exception):
"""Tweepy exception"""
def __init__(self, reason, response=None, api_code=None):
self.reason = six.text_type(reason)
self.response = response
self.api_code = api_code
super(TweepError, self).__init__(self, reason)
def __str__(self):
return self.reason
def is_rate_limit_error_message(message):
"""Check if the supplied error message belongs to a rate limit error."""
return isinstance(message, list) \
and len(message) > 0 \
and 'code' in message[0] \
and message[0]['code'] == 88
class RateLimitError(TweepError):
"""Exception for Tweepy hitting the rate limit."""
# RateLimitError has the exact same properties and inner workings
# as TweepError for backwards compatibility reasons.
pass
Fix super usage in TweepError initialization | # Tweepy
# Copyright 2009-2010 Joshua Roesslein
# See LICENSE for details.
from __future__ import print_function
import six
class TweepError(Exception):
"""Tweepy exception"""
def __init__(self, reason, response=None, api_code=None):
self.reason = six.text_type(reason)
self.response = response
self.api_code = api_code
super(TweepError, self).__init__(reason)
def __str__(self):
return self.reason
def is_rate_limit_error_message(message):
"""Check if the supplied error message belongs to a rate limit error."""
return isinstance(message, list) \
and len(message) > 0 \
and 'code' in message[0] \
and message[0]['code'] == 88
class RateLimitError(TweepError):
"""Exception for Tweepy hitting the rate limit."""
# RateLimitError has the exact same properties and inner workings
# as TweepError for backwards compatibility reasons.
pass
| <commit_before># Tweepy
# Copyright 2009-2010 Joshua Roesslein
# See LICENSE for details.
from __future__ import print_function
import six
class TweepError(Exception):
"""Tweepy exception"""
def __init__(self, reason, response=None, api_code=None):
self.reason = six.text_type(reason)
self.response = response
self.api_code = api_code
super(TweepError, self).__init__(self, reason)
def __str__(self):
return self.reason
def is_rate_limit_error_message(message):
"""Check if the supplied error message belongs to a rate limit error."""
return isinstance(message, list) \
and len(message) > 0 \
and 'code' in message[0] \
and message[0]['code'] == 88
class RateLimitError(TweepError):
"""Exception for Tweepy hitting the rate limit."""
# RateLimitError has the exact same properties and inner workings
# as TweepError for backwards compatibility reasons.
pass
<commit_msg>Fix super usage in TweepError initialization<commit_after> | # Tweepy
# Copyright 2009-2010 Joshua Roesslein
# See LICENSE for details.
from __future__ import print_function
import six
class TweepError(Exception):
"""Tweepy exception"""
def __init__(self, reason, response=None, api_code=None):
self.reason = six.text_type(reason)
self.response = response
self.api_code = api_code
super(TweepError, self).__init__(reason)
def __str__(self):
return self.reason
def is_rate_limit_error_message(message):
"""Check if the supplied error message belongs to a rate limit error."""
return isinstance(message, list) \
and len(message) > 0 \
and 'code' in message[0] \
and message[0]['code'] == 88
class RateLimitError(TweepError):
"""Exception for Tweepy hitting the rate limit."""
# RateLimitError has the exact same properties and inner workings
# as TweepError for backwards compatibility reasons.
pass
| # Tweepy
# Copyright 2009-2010 Joshua Roesslein
# See LICENSE for details.
from __future__ import print_function
import six
class TweepError(Exception):
"""Tweepy exception"""
def __init__(self, reason, response=None, api_code=None):
self.reason = six.text_type(reason)
self.response = response
self.api_code = api_code
super(TweepError, self).__init__(self, reason)
def __str__(self):
return self.reason
def is_rate_limit_error_message(message):
"""Check if the supplied error message belongs to a rate limit error."""
return isinstance(message, list) \
and len(message) > 0 \
and 'code' in message[0] \
and message[0]['code'] == 88
class RateLimitError(TweepError):
"""Exception for Tweepy hitting the rate limit."""
# RateLimitError has the exact same properties and inner workings
# as TweepError for backwards compatibility reasons.
pass
Fix super usage in TweepError initialization# Tweepy
# Copyright 2009-2010 Joshua Roesslein
# See LICENSE for details.
from __future__ import print_function
import six
class TweepError(Exception):
"""Tweepy exception"""
def __init__(self, reason, response=None, api_code=None):
self.reason = six.text_type(reason)
self.response = response
self.api_code = api_code
super(TweepError, self).__init__(reason)
def __str__(self):
return self.reason
def is_rate_limit_error_message(message):
"""Check if the supplied error message belongs to a rate limit error."""
return isinstance(message, list) \
and len(message) > 0 \
and 'code' in message[0] \
and message[0]['code'] == 88
class RateLimitError(TweepError):
"""Exception for Tweepy hitting the rate limit."""
# RateLimitError has the exact same properties and inner workings
# as TweepError for backwards compatibility reasons.
pass
| <commit_before># Tweepy
# Copyright 2009-2010 Joshua Roesslein
# See LICENSE for details.
from __future__ import print_function
import six
class TweepError(Exception):
"""Tweepy exception"""
def __init__(self, reason, response=None, api_code=None):
self.reason = six.text_type(reason)
self.response = response
self.api_code = api_code
super(TweepError, self).__init__(self, reason)
def __str__(self):
return self.reason
def is_rate_limit_error_message(message):
"""Check if the supplied error message belongs to a rate limit error."""
return isinstance(message, list) \
and len(message) > 0 \
and 'code' in message[0] \
and message[0]['code'] == 88
class RateLimitError(TweepError):
"""Exception for Tweepy hitting the rate limit."""
# RateLimitError has the exact same properties and inner workings
# as TweepError for backwards compatibility reasons.
pass
<commit_msg>Fix super usage in TweepError initialization<commit_after># Tweepy
# Copyright 2009-2010 Joshua Roesslein
# See LICENSE for details.
from __future__ import print_function
import six
class TweepError(Exception):
"""Tweepy exception"""
def __init__(self, reason, response=None, api_code=None):
self.reason = six.text_type(reason)
self.response = response
self.api_code = api_code
super(TweepError, self).__init__(reason)
def __str__(self):
return self.reason
def is_rate_limit_error_message(message):
"""Check if the supplied error message belongs to a rate limit error."""
return isinstance(message, list) \
and len(message) > 0 \
and 'code' in message[0] \
and message[0]['code'] == 88
class RateLimitError(TweepError):
"""Exception for Tweepy hitting the rate limit."""
# RateLimitError has the exact same properties and inner workings
# as TweepError for backwards compatibility reasons.
pass
|
be9ce58461e56873b0d8f60c85c0af96e48ce3fb | fabfile.py | fabfile.py | import logging
import yaml
from fabric.api import lcd, env, task
from fabric.contrib.project import rsync_project
logging.basicConfig(level=logging.DEBUG)
log = logging.getLogger()
try:
conf = yaml.load(open('deploy.yaml', 'rb').read())
except:
log.exception('error: unable to read deply.yaml config file:')
env.user = conf['user']
env.hosts = ['{}@{}:22'.format(env.user, host) for host in conf['hosts']]
def deploy_project(local_dir, remote_dir, exclusions=[]):
"""Deploy the entire project at local_dir to remote_dir, excluding the given paths."""
with lcd(local_dir):
rsync_project(remote_dir=remote_dir, local_dir='.', exclude=exclusions)
rsync_project(remote_dir=remote_dir, local_dir='resources', exclude=exclusions, delete=True)
@task
def deploy():
"""Deploys web and script to remote server."""
deploy_project('web', conf['web_remote_dir'],
['.git', 'fabfile.py', 'cache', 'config', 'template'])
deploy_project('script', conf['script_remote_dir'],
['.git', 'fabfile.py', 'cache', 'js', 'image'])
| import logging
import os
import yaml
from fabric.api import lcd, env, task, local
from fabric.contrib.project import rsync_project
logging.basicConfig(level=logging.DEBUG)
log = logging.getLogger()
repo_root = local('git rev-parse --show-toplevel', capture=True)
try:
conf = yaml.load(open(os.path.join(repo_root, 'deploy.yaml'), 'rb').read())
except:
log.exception('error: unable to read deply.yaml config file:')
env.user = conf['user']
env.hosts = ['{}@{}:22'.format(env.user, host) for host in conf['hosts']]
def deploy_project(local_dir, remote_dir, exclusions=[]):
"""Deploy the entire project at local_dir to remote_dir, excluding the given paths."""
with lcd(repo_root):
with lcd(local_dir):
rsync_project(remote_dir=remote_dir, local_dir='.', exclude=exclusions)
rsync_project(remote_dir=remote_dir, local_dir='resources', exclude=exclusions, delete=True)
@task
def deploy():
"""Deploys web and script to remote server."""
deploy_project('web', conf['web_remote_dir'],
['.git', 'fabfile.py', 'cache', 'config', 'template'])
deploy_project('script', conf['script_remote_dir'],
['.git', 'fabfile.py', 'cache', 'js', 'image'])
| Make deployment script work from anywhere. | Make deployment script work from anywhere.
| Python | mit | lexicalunit/pancake-master,lexicalunit/pancake-master,lexicalunit/pancake-master | import logging
import yaml
from fabric.api import lcd, env, task
from fabric.contrib.project import rsync_project
logging.basicConfig(level=logging.DEBUG)
log = logging.getLogger()
try:
conf = yaml.load(open('deploy.yaml', 'rb').read())
except:
log.exception('error: unable to read deply.yaml config file:')
env.user = conf['user']
env.hosts = ['{}@{}:22'.format(env.user, host) for host in conf['hosts']]
def deploy_project(local_dir, remote_dir, exclusions=[]):
"""Deploy the entire project at local_dir to remote_dir, excluding the given paths."""
with lcd(local_dir):
rsync_project(remote_dir=remote_dir, local_dir='.', exclude=exclusions)
rsync_project(remote_dir=remote_dir, local_dir='resources', exclude=exclusions, delete=True)
@task
def deploy():
"""Deploys web and script to remote server."""
deploy_project('web', conf['web_remote_dir'],
['.git', 'fabfile.py', 'cache', 'config', 'template'])
deploy_project('script', conf['script_remote_dir'],
['.git', 'fabfile.py', 'cache', 'js', 'image'])
Make deployment script work from anywhere. | import logging
import os
import yaml
from fabric.api import lcd, env, task, local
from fabric.contrib.project import rsync_project
logging.basicConfig(level=logging.DEBUG)
log = logging.getLogger()
repo_root = local('git rev-parse --show-toplevel', capture=True)
try:
conf = yaml.load(open(os.path.join(repo_root, 'deploy.yaml'), 'rb').read())
except:
log.exception('error: unable to read deply.yaml config file:')
env.user = conf['user']
env.hosts = ['{}@{}:22'.format(env.user, host) for host in conf['hosts']]
def deploy_project(local_dir, remote_dir, exclusions=[]):
"""Deploy the entire project at local_dir to remote_dir, excluding the given paths."""
with lcd(repo_root):
with lcd(local_dir):
rsync_project(remote_dir=remote_dir, local_dir='.', exclude=exclusions)
rsync_project(remote_dir=remote_dir, local_dir='resources', exclude=exclusions, delete=True)
@task
def deploy():
"""Deploys web and script to remote server."""
deploy_project('web', conf['web_remote_dir'],
['.git', 'fabfile.py', 'cache', 'config', 'template'])
deploy_project('script', conf['script_remote_dir'],
['.git', 'fabfile.py', 'cache', 'js', 'image'])
| <commit_before>import logging
import yaml
from fabric.api import lcd, env, task
from fabric.contrib.project import rsync_project
logging.basicConfig(level=logging.DEBUG)
log = logging.getLogger()
try:
conf = yaml.load(open('deploy.yaml', 'rb').read())
except:
log.exception('error: unable to read deply.yaml config file:')
env.user = conf['user']
env.hosts = ['{}@{}:22'.format(env.user, host) for host in conf['hosts']]
def deploy_project(local_dir, remote_dir, exclusions=[]):
"""Deploy the entire project at local_dir to remote_dir, excluding the given paths."""
with lcd(local_dir):
rsync_project(remote_dir=remote_dir, local_dir='.', exclude=exclusions)
rsync_project(remote_dir=remote_dir, local_dir='resources', exclude=exclusions, delete=True)
@task
def deploy():
"""Deploys web and script to remote server."""
deploy_project('web', conf['web_remote_dir'],
['.git', 'fabfile.py', 'cache', 'config', 'template'])
deploy_project('script', conf['script_remote_dir'],
['.git', 'fabfile.py', 'cache', 'js', 'image'])
<commit_msg>Make deployment script work from anywhere.<commit_after> | import logging
import os
import yaml
from fabric.api import lcd, env, task, local
from fabric.contrib.project import rsync_project
logging.basicConfig(level=logging.DEBUG)
log = logging.getLogger()
repo_root = local('git rev-parse --show-toplevel', capture=True)
try:
conf = yaml.load(open(os.path.join(repo_root, 'deploy.yaml'), 'rb').read())
except:
log.exception('error: unable to read deply.yaml config file:')
env.user = conf['user']
env.hosts = ['{}@{}:22'.format(env.user, host) for host in conf['hosts']]
def deploy_project(local_dir, remote_dir, exclusions=[]):
"""Deploy the entire project at local_dir to remote_dir, excluding the given paths."""
with lcd(repo_root):
with lcd(local_dir):
rsync_project(remote_dir=remote_dir, local_dir='.', exclude=exclusions)
rsync_project(remote_dir=remote_dir, local_dir='resources', exclude=exclusions, delete=True)
@task
def deploy():
"""Deploys web and script to remote server."""
deploy_project('web', conf['web_remote_dir'],
['.git', 'fabfile.py', 'cache', 'config', 'template'])
deploy_project('script', conf['script_remote_dir'],
['.git', 'fabfile.py', 'cache', 'js', 'image'])
| import logging
import yaml
from fabric.api import lcd, env, task
from fabric.contrib.project import rsync_project
logging.basicConfig(level=logging.DEBUG)
log = logging.getLogger()
try:
conf = yaml.load(open('deploy.yaml', 'rb').read())
except:
log.exception('error: unable to read deply.yaml config file:')
env.user = conf['user']
env.hosts = ['{}@{}:22'.format(env.user, host) for host in conf['hosts']]
def deploy_project(local_dir, remote_dir, exclusions=[]):
"""Deploy the entire project at local_dir to remote_dir, excluding the given paths."""
with lcd(local_dir):
rsync_project(remote_dir=remote_dir, local_dir='.', exclude=exclusions)
rsync_project(remote_dir=remote_dir, local_dir='resources', exclude=exclusions, delete=True)
@task
def deploy():
"""Deploys web and script to remote server."""
deploy_project('web', conf['web_remote_dir'],
['.git', 'fabfile.py', 'cache', 'config', 'template'])
deploy_project('script', conf['script_remote_dir'],
['.git', 'fabfile.py', 'cache', 'js', 'image'])
Make deployment script work from anywhere.import logging
import os
import yaml
from fabric.api import lcd, env, task, local
from fabric.contrib.project import rsync_project
logging.basicConfig(level=logging.DEBUG)
log = logging.getLogger()
repo_root = local('git rev-parse --show-toplevel', capture=True)
try:
conf = yaml.load(open(os.path.join(repo_root, 'deploy.yaml'), 'rb').read())
except:
log.exception('error: unable to read deply.yaml config file:')
env.user = conf['user']
env.hosts = ['{}@{}:22'.format(env.user, host) for host in conf['hosts']]
def deploy_project(local_dir, remote_dir, exclusions=[]):
"""Deploy the entire project at local_dir to remote_dir, excluding the given paths."""
with lcd(repo_root):
with lcd(local_dir):
rsync_project(remote_dir=remote_dir, local_dir='.', exclude=exclusions)
rsync_project(remote_dir=remote_dir, local_dir='resources', exclude=exclusions, delete=True)
@task
def deploy():
"""Deploys web and script to remote server."""
deploy_project('web', conf['web_remote_dir'],
['.git', 'fabfile.py', 'cache', 'config', 'template'])
deploy_project('script', conf['script_remote_dir'],
['.git', 'fabfile.py', 'cache', 'js', 'image'])
| <commit_before>import logging
import yaml
from fabric.api import lcd, env, task
from fabric.contrib.project import rsync_project
logging.basicConfig(level=logging.DEBUG)
log = logging.getLogger()
try:
conf = yaml.load(open('deploy.yaml', 'rb').read())
except:
log.exception('error: unable to read deply.yaml config file:')
env.user = conf['user']
env.hosts = ['{}@{}:22'.format(env.user, host) for host in conf['hosts']]
def deploy_project(local_dir, remote_dir, exclusions=[]):
"""Deploy the entire project at local_dir to remote_dir, excluding the given paths."""
with lcd(local_dir):
rsync_project(remote_dir=remote_dir, local_dir='.', exclude=exclusions)
rsync_project(remote_dir=remote_dir, local_dir='resources', exclude=exclusions, delete=True)
@task
def deploy():
"""Deploys web and script to remote server."""
deploy_project('web', conf['web_remote_dir'],
['.git', 'fabfile.py', 'cache', 'config', 'template'])
deploy_project('script', conf['script_remote_dir'],
['.git', 'fabfile.py', 'cache', 'js', 'image'])
<commit_msg>Make deployment script work from anywhere.<commit_after>import logging
import os
import yaml
from fabric.api import lcd, env, task, local
from fabric.contrib.project import rsync_project
logging.basicConfig(level=logging.DEBUG)
log = logging.getLogger()
repo_root = local('git rev-parse --show-toplevel', capture=True)
try:
conf = yaml.load(open(os.path.join(repo_root, 'deploy.yaml'), 'rb').read())
except:
log.exception('error: unable to read deply.yaml config file:')
env.user = conf['user']
env.hosts = ['{}@{}:22'.format(env.user, host) for host in conf['hosts']]
def deploy_project(local_dir, remote_dir, exclusions=[]):
"""Deploy the entire project at local_dir to remote_dir, excluding the given paths."""
with lcd(repo_root):
with lcd(local_dir):
rsync_project(remote_dir=remote_dir, local_dir='.', exclude=exclusions)
rsync_project(remote_dir=remote_dir, local_dir='resources', exclude=exclusions, delete=True)
@task
def deploy():
"""Deploys web and script to remote server."""
deploy_project('web', conf['web_remote_dir'],
['.git', 'fabfile.py', 'cache', 'config', 'template'])
deploy_project('script', conf['script_remote_dir'],
['.git', 'fabfile.py', 'cache', 'js', 'image'])
|
d648598d669144d589ffbbb03bf56edad4050aff | connector/__manifest__.py | connector/__manifest__.py | # -*- coding: utf-8 -*-
# Copyright 2013-2017 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html)
{'name': 'Connector',
'version': '10.0.1.0.0',
'author': 'Camptocamp,Openerp Connector Core Editors,'
'Odoo Community Association (OCA)',
'website': 'http://odoo-connector.com',
'license': 'AGPL-3',
'category': 'Generic Modules',
'depends': ['mail',
'queue_job',
],
'data': ['security/connector_security.xml',
'security/ir.model.access.csv',
'checkpoint/checkpoint_view.xml',
'connector_menu.xml',
'setting_view.xml',
'res_partner_view.xml',
],
'installable': True,
'application': True,
}
| # -*- coding: utf-8 -*-
# Copyright 2013-2017 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html)
{'name': 'Connector',
'version': '10.0.1.0.0',
'author': 'Camptocamp,Openerp Connector Core Editors,'
'Odoo Community Association (OCA)',
'website': 'http://odoo-connector.com',
'license': 'AGPL-3',
'category': 'Generic Modules',
'depends': ['mail',
'queue_job',
],
'data': ['security/connector_security.xml',
'security/ir.model.access.csv',
'checkpoint/checkpoint_view.xml',
'connector_menu.xml',
'setting_view.xml',
'res_partner_view.xml',
],
'installable': True,
}
| Remove application flag, not an application | Remove application flag, not an application
| Python | agpl-3.0 | js-landoo/connector,js-landoo/connector | # -*- coding: utf-8 -*-
# Copyright 2013-2017 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html)
{'name': 'Connector',
'version': '10.0.1.0.0',
'author': 'Camptocamp,Openerp Connector Core Editors,'
'Odoo Community Association (OCA)',
'website': 'http://odoo-connector.com',
'license': 'AGPL-3',
'category': 'Generic Modules',
'depends': ['mail',
'queue_job',
],
'data': ['security/connector_security.xml',
'security/ir.model.access.csv',
'checkpoint/checkpoint_view.xml',
'connector_menu.xml',
'setting_view.xml',
'res_partner_view.xml',
],
'installable': True,
'application': True,
}
Remove application flag, not an application | # -*- coding: utf-8 -*-
# Copyright 2013-2017 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html)
{'name': 'Connector',
'version': '10.0.1.0.0',
'author': 'Camptocamp,Openerp Connector Core Editors,'
'Odoo Community Association (OCA)',
'website': 'http://odoo-connector.com',
'license': 'AGPL-3',
'category': 'Generic Modules',
'depends': ['mail',
'queue_job',
],
'data': ['security/connector_security.xml',
'security/ir.model.access.csv',
'checkpoint/checkpoint_view.xml',
'connector_menu.xml',
'setting_view.xml',
'res_partner_view.xml',
],
'installable': True,
}
| <commit_before># -*- coding: utf-8 -*-
# Copyright 2013-2017 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html)
{'name': 'Connector',
'version': '10.0.1.0.0',
'author': 'Camptocamp,Openerp Connector Core Editors,'
'Odoo Community Association (OCA)',
'website': 'http://odoo-connector.com',
'license': 'AGPL-3',
'category': 'Generic Modules',
'depends': ['mail',
'queue_job',
],
'data': ['security/connector_security.xml',
'security/ir.model.access.csv',
'checkpoint/checkpoint_view.xml',
'connector_menu.xml',
'setting_view.xml',
'res_partner_view.xml',
],
'installable': True,
'application': True,
}
<commit_msg>Remove application flag, not an application<commit_after> | # -*- coding: utf-8 -*-
# Copyright 2013-2017 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html)
{'name': 'Connector',
'version': '10.0.1.0.0',
'author': 'Camptocamp,Openerp Connector Core Editors,'
'Odoo Community Association (OCA)',
'website': 'http://odoo-connector.com',
'license': 'AGPL-3',
'category': 'Generic Modules',
'depends': ['mail',
'queue_job',
],
'data': ['security/connector_security.xml',
'security/ir.model.access.csv',
'checkpoint/checkpoint_view.xml',
'connector_menu.xml',
'setting_view.xml',
'res_partner_view.xml',
],
'installable': True,
}
| # -*- coding: utf-8 -*-
# Copyright 2013-2017 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html)
{'name': 'Connector',
'version': '10.0.1.0.0',
'author': 'Camptocamp,Openerp Connector Core Editors,'
'Odoo Community Association (OCA)',
'website': 'http://odoo-connector.com',
'license': 'AGPL-3',
'category': 'Generic Modules',
'depends': ['mail',
'queue_job',
],
'data': ['security/connector_security.xml',
'security/ir.model.access.csv',
'checkpoint/checkpoint_view.xml',
'connector_menu.xml',
'setting_view.xml',
'res_partner_view.xml',
],
'installable': True,
'application': True,
}
Remove application flag, not an application# -*- coding: utf-8 -*-
# Copyright 2013-2017 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html)
{'name': 'Connector',
'version': '10.0.1.0.0',
'author': 'Camptocamp,Openerp Connector Core Editors,'
'Odoo Community Association (OCA)',
'website': 'http://odoo-connector.com',
'license': 'AGPL-3',
'category': 'Generic Modules',
'depends': ['mail',
'queue_job',
],
'data': ['security/connector_security.xml',
'security/ir.model.access.csv',
'checkpoint/checkpoint_view.xml',
'connector_menu.xml',
'setting_view.xml',
'res_partner_view.xml',
],
'installable': True,
}
| <commit_before># -*- coding: utf-8 -*-
# Copyright 2013-2017 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html)
{'name': 'Connector',
'version': '10.0.1.0.0',
'author': 'Camptocamp,Openerp Connector Core Editors,'
'Odoo Community Association (OCA)',
'website': 'http://odoo-connector.com',
'license': 'AGPL-3',
'category': 'Generic Modules',
'depends': ['mail',
'queue_job',
],
'data': ['security/connector_security.xml',
'security/ir.model.access.csv',
'checkpoint/checkpoint_view.xml',
'connector_menu.xml',
'setting_view.xml',
'res_partner_view.xml',
],
'installable': True,
'application': True,
}
<commit_msg>Remove application flag, not an application<commit_after># -*- coding: utf-8 -*-
# Copyright 2013-2017 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html)
{'name': 'Connector',
'version': '10.0.1.0.0',
'author': 'Camptocamp,Openerp Connector Core Editors,'
'Odoo Community Association (OCA)',
'website': 'http://odoo-connector.com',
'license': 'AGPL-3',
'category': 'Generic Modules',
'depends': ['mail',
'queue_job',
],
'data': ['security/connector_security.xml',
'security/ir.model.access.csv',
'checkpoint/checkpoint_view.xml',
'connector_menu.xml',
'setting_view.xml',
'res_partner_view.xml',
],
'installable': True,
}
|
15cb279724a646368066591e81467e1b26d61938 | examples/charts/file/steps.py | examples/charts/file/steps.py | from bokeh.charts import Step, show, output_file
# build a dataset where multiple columns measure the same thing
data = dict(python=[2, 3, 7, 5, 26, 221, 44, 233, 254, 265, 266, 267, 120, 111],
pypy=[12, 33, 47, 15, 126, 121, 144, 233, 254, 225, 226, 267, 110, 130],
jython=[22, 43, 10, 25, 26, 101, 114, 203, 194, 215, 201, 227, 139, 160],
test=['foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar',
'foo', 'bar', 'foo', 'bar']
)
# create a line chart where each column of measures receives a unique color and dash style
line = Step(data, y=['python', 'pypy', 'jython'],
dash=['python', 'pypy', 'jython'],
color=['python', 'pypy', 'jython'],
title="Interpreter Sample Data", ylabel='Duration', legend=True)
output_file("steps.html")
show(line)
| """ This example uses the U.S. postage rate per ounce for stamps and
postcards.
Source: https://en.wikipedia.org/wiki/History_of_United_States_postage_rates
"""
from bokeh.charts import Step, show, output_file
# build a dataset where multiple columns measure the same thing
data = dict(stamp=[
.33, .33, .34, .37, .37, .37, .37, .39, .41, .42,
.44, .44, .44, .45, .46, .49, .49],
postcard=[
.20, .20, .21, .23, .23, .23, .23, .24, .26, .27,
.28, .28, .29, .32, .33, .34, .35],
)
# create a line chart where each column of measures receives a unique color and dash style
line = Step(data, y=['stamp', 'postcard'],
dash=['stamp', 'postcard'],
color=['stamp', 'postcard'],
title="U.S. Postage Rates (1999-2015)", ylabel='Rate per ounce', legend=True)
output_file("steps.html")
show(line)
| Change step example to plot US postage rates | Change step example to plot US postage rates
| Python | bsd-3-clause | ptitjano/bokeh,timsnyder/bokeh,draperjames/bokeh,percyfal/bokeh,justacec/bokeh,clairetang6/bokeh,philippjfr/bokeh,ericmjl/bokeh,rs2/bokeh,azjps/bokeh,DuCorey/bokeh,clairetang6/bokeh,draperjames/bokeh,clairetang6/bokeh,DuCorey/bokeh,aavanian/bokeh,KasperPRasmussen/bokeh,justacec/bokeh,bokeh/bokeh,aiguofer/bokeh,rs2/bokeh,aavanian/bokeh,stonebig/bokeh,schoolie/bokeh,msarahan/bokeh,msarahan/bokeh,jakirkham/bokeh,schoolie/bokeh,Karel-van-de-Plassche/bokeh,Karel-van-de-Plassche/bokeh,phobson/bokeh,stonebig/bokeh,azjps/bokeh,aavanian/bokeh,jakirkham/bokeh,ptitjano/bokeh,bokeh/bokeh,azjps/bokeh,ptitjano/bokeh,msarahan/bokeh,Karel-van-de-Plassche/bokeh,schoolie/bokeh,KasperPRasmussen/bokeh,draperjames/bokeh,phobson/bokeh,schoolie/bokeh,timsnyder/bokeh,azjps/bokeh,quasiben/bokeh,percyfal/bokeh,ericmjl/bokeh,jakirkham/bokeh,timsnyder/bokeh,phobson/bokeh,justacec/bokeh,mindriot101/bokeh,azjps/bokeh,Karel-van-de-Plassche/bokeh,ericmjl/bokeh,quasiben/bokeh,dennisobrien/bokeh,philippjfr/bokeh,clairetang6/bokeh,ptitjano/bokeh,ericmjl/bokeh,quasiben/bokeh,aiguofer/bokeh,timsnyder/bokeh,Karel-van-de-Plassche/bokeh,draperjames/bokeh,dennisobrien/bokeh,phobson/bokeh,dennisobrien/bokeh,aavanian/bokeh,stonebig/bokeh,percyfal/bokeh,aiguofer/bokeh,aiguofer/bokeh,jakirkham/bokeh,rs2/bokeh,draperjames/bokeh,ericmjl/bokeh,bokeh/bokeh,dennisobrien/bokeh,schoolie/bokeh,bokeh/bokeh,KasperPRasmussen/bokeh,percyfal/bokeh,aavanian/bokeh,phobson/bokeh,justacec/bokeh,mindriot101/bokeh,DuCorey/bokeh,stonebig/bokeh,bokeh/bokeh,philippjfr/bokeh,philippjfr/bokeh,DuCorey/bokeh,rs2/bokeh,mindriot101/bokeh,timsnyder/bokeh,DuCorey/bokeh,dennisobrien/bokeh,ptitjano/bokeh,percyfal/bokeh,msarahan/bokeh,rs2/bokeh,KasperPRasmussen/bokeh,jakirkham/bokeh,KasperPRasmussen/bokeh,mindriot101/bokeh,aiguofer/bokeh,philippjfr/bokeh | from bokeh.charts import Step, show, output_file
# build a dataset where multiple columns measure the same thing
data = dict(python=[2, 3, 7, 5, 26, 221, 44, 233, 254, 265, 266, 267, 120, 111],
pypy=[12, 33, 47, 15, 126, 121, 144, 233, 254, 225, 226, 267, 110, 130],
jython=[22, 43, 10, 25, 26, 101, 114, 203, 194, 215, 201, 227, 139, 160],
test=['foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar',
'foo', 'bar', 'foo', 'bar']
)
# create a line chart where each column of measures receives a unique color and dash style
line = Step(data, y=['python', 'pypy', 'jython'],
dash=['python', 'pypy', 'jython'],
color=['python', 'pypy', 'jython'],
title="Interpreter Sample Data", ylabel='Duration', legend=True)
output_file("steps.html")
show(line)
Change step example to plot US postage rates | """ This example uses the U.S. postage rate per ounce for stamps and
postcards.
Source: https://en.wikipedia.org/wiki/History_of_United_States_postage_rates
"""
from bokeh.charts import Step, show, output_file
# build a dataset where multiple columns measure the same thing
data = dict(stamp=[
.33, .33, .34, .37, .37, .37, .37, .39, .41, .42,
.44, .44, .44, .45, .46, .49, .49],
postcard=[
.20, .20, .21, .23, .23, .23, .23, .24, .26, .27,
.28, .28, .29, .32, .33, .34, .35],
)
# create a line chart where each column of measures receives a unique color and dash style
line = Step(data, y=['stamp', 'postcard'],
dash=['stamp', 'postcard'],
color=['stamp', 'postcard'],
title="U.S. Postage Rates (1999-2015)", ylabel='Rate per ounce', legend=True)
output_file("steps.html")
show(line)
| <commit_before>from bokeh.charts import Step, show, output_file
# build a dataset where multiple columns measure the same thing
data = dict(python=[2, 3, 7, 5, 26, 221, 44, 233, 254, 265, 266, 267, 120, 111],
pypy=[12, 33, 47, 15, 126, 121, 144, 233, 254, 225, 226, 267, 110, 130],
jython=[22, 43, 10, 25, 26, 101, 114, 203, 194, 215, 201, 227, 139, 160],
test=['foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar',
'foo', 'bar', 'foo', 'bar']
)
# create a line chart where each column of measures receives a unique color and dash style
line = Step(data, y=['python', 'pypy', 'jython'],
dash=['python', 'pypy', 'jython'],
color=['python', 'pypy', 'jython'],
title="Interpreter Sample Data", ylabel='Duration', legend=True)
output_file("steps.html")
show(line)
<commit_msg>Change step example to plot US postage rates<commit_after> | """ This example uses the U.S. postage rate per ounce for stamps and
postcards.
Source: https://en.wikipedia.org/wiki/History_of_United_States_postage_rates
"""
from bokeh.charts import Step, show, output_file
# build a dataset where multiple columns measure the same thing
data = dict(stamp=[
.33, .33, .34, .37, .37, .37, .37, .39, .41, .42,
.44, .44, .44, .45, .46, .49, .49],
postcard=[
.20, .20, .21, .23, .23, .23, .23, .24, .26, .27,
.28, .28, .29, .32, .33, .34, .35],
)
# create a line chart where each column of measures receives a unique color and dash style
line = Step(data, y=['stamp', 'postcard'],
dash=['stamp', 'postcard'],
color=['stamp', 'postcard'],
title="U.S. Postage Rates (1999-2015)", ylabel='Rate per ounce', legend=True)
output_file("steps.html")
show(line)
| from bokeh.charts import Step, show, output_file
# build a dataset where multiple columns measure the same thing
data = dict(python=[2, 3, 7, 5, 26, 221, 44, 233, 254, 265, 266, 267, 120, 111],
pypy=[12, 33, 47, 15, 126, 121, 144, 233, 254, 225, 226, 267, 110, 130],
jython=[22, 43, 10, 25, 26, 101, 114, 203, 194, 215, 201, 227, 139, 160],
test=['foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar',
'foo', 'bar', 'foo', 'bar']
)
# create a line chart where each column of measures receives a unique color and dash style
line = Step(data, y=['python', 'pypy', 'jython'],
dash=['python', 'pypy', 'jython'],
color=['python', 'pypy', 'jython'],
title="Interpreter Sample Data", ylabel='Duration', legend=True)
output_file("steps.html")
show(line)
Change step example to plot US postage rates""" This example uses the U.S. postage rate per ounce for stamps and
postcards.
Source: https://en.wikipedia.org/wiki/History_of_United_States_postage_rates
"""
from bokeh.charts import Step, show, output_file
# build a dataset where multiple columns measure the same thing
data = dict(stamp=[
.33, .33, .34, .37, .37, .37, .37, .39, .41, .42,
.44, .44, .44, .45, .46, .49, .49],
postcard=[
.20, .20, .21, .23, .23, .23, .23, .24, .26, .27,
.28, .28, .29, .32, .33, .34, .35],
)
# create a line chart where each column of measures receives a unique color and dash style
line = Step(data, y=['stamp', 'postcard'],
dash=['stamp', 'postcard'],
color=['stamp', 'postcard'],
title="U.S. Postage Rates (1999-2015)", ylabel='Rate per ounce', legend=True)
output_file("steps.html")
show(line)
| <commit_before>from bokeh.charts import Step, show, output_file
# build a dataset where multiple columns measure the same thing
data = dict(python=[2, 3, 7, 5, 26, 221, 44, 233, 254, 265, 266, 267, 120, 111],
pypy=[12, 33, 47, 15, 126, 121, 144, 233, 254, 225, 226, 267, 110, 130],
jython=[22, 43, 10, 25, 26, 101, 114, 203, 194, 215, 201, 227, 139, 160],
test=['foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar',
'foo', 'bar', 'foo', 'bar']
)
# create a line chart where each column of measures receives a unique color and dash style
line = Step(data, y=['python', 'pypy', 'jython'],
dash=['python', 'pypy', 'jython'],
color=['python', 'pypy', 'jython'],
title="Interpreter Sample Data", ylabel='Duration', legend=True)
output_file("steps.html")
show(line)
<commit_msg>Change step example to plot US postage rates<commit_after>""" This example uses the U.S. postage rate per ounce for stamps and
postcards.
Source: https://en.wikipedia.org/wiki/History_of_United_States_postage_rates
"""
from bokeh.charts import Step, show, output_file
# build a dataset where multiple columns measure the same thing
data = dict(stamp=[
.33, .33, .34, .37, .37, .37, .37, .39, .41, .42,
.44, .44, .44, .45, .46, .49, .49],
postcard=[
.20, .20, .21, .23, .23, .23, .23, .24, .26, .27,
.28, .28, .29, .32, .33, .34, .35],
)
# create a line chart where each column of measures receives a unique color and dash style
line = Step(data, y=['stamp', 'postcard'],
dash=['stamp', 'postcard'],
color=['stamp', 'postcard'],
title="U.S. Postage Rates (1999-2015)", ylabel='Rate per ounce', legend=True)
output_file("steps.html")
show(line)
|
dd03a3d323594c7836525a1be733b689731d98c4 | core/settings/__init__.py | core/settings/__init__.py | """Settings package initialization."""
import dotenv
dotenv.load()
# Ensure development settings are not used in testing and production:
if dotenv.get('ENVIRONMENT') == 'HEROKU':
from .production import *
elif dotenv.get('ENVIRONMENT') == 'TRAVIS':
from .testing import *
else:
from .local import *
| """Settings package initialization."""
import dotenv
dotenv.load()
# Ensure development settings are not used in testing and production:
if dotenv.get('ENVIRONMENT') == 'PRODUCTION':
from .production import *
elif dotenv.get('ENVIRONMENT') == 'TRAVIS':
from .testing import *
else:
from .local import *
| Change ENVIRONMENT variable from HEROKU to PRODUCTION | Change ENVIRONMENT variable from HEROKU to PRODUCTION
| Python | mit | teamtaverna/core | """Settings package initialization."""
import dotenv
dotenv.load()
# Ensure development settings are not used in testing and production:
if dotenv.get('ENVIRONMENT') == 'HEROKU':
from .production import *
elif dotenv.get('ENVIRONMENT') == 'TRAVIS':
from .testing import *
else:
from .local import *
Change ENVIRONMENT variable from HEROKU to PRODUCTION | """Settings package initialization."""
import dotenv
dotenv.load()
# Ensure development settings are not used in testing and production:
if dotenv.get('ENVIRONMENT') == 'PRODUCTION':
from .production import *
elif dotenv.get('ENVIRONMENT') == 'TRAVIS':
from .testing import *
else:
from .local import *
| <commit_before>"""Settings package initialization."""
import dotenv
dotenv.load()
# Ensure development settings are not used in testing and production:
if dotenv.get('ENVIRONMENT') == 'HEROKU':
from .production import *
elif dotenv.get('ENVIRONMENT') == 'TRAVIS':
from .testing import *
else:
from .local import *
<commit_msg>Change ENVIRONMENT variable from HEROKU to PRODUCTION<commit_after> | """Settings package initialization."""
import dotenv
dotenv.load()
# Ensure development settings are not used in testing and production:
if dotenv.get('ENVIRONMENT') == 'PRODUCTION':
from .production import *
elif dotenv.get('ENVIRONMENT') == 'TRAVIS':
from .testing import *
else:
from .local import *
| """Settings package initialization."""
import dotenv
dotenv.load()
# Ensure development settings are not used in testing and production:
if dotenv.get('ENVIRONMENT') == 'HEROKU':
from .production import *
elif dotenv.get('ENVIRONMENT') == 'TRAVIS':
from .testing import *
else:
from .local import *
Change ENVIRONMENT variable from HEROKU to PRODUCTION"""Settings package initialization."""
import dotenv
dotenv.load()
# Ensure development settings are not used in testing and production:
if dotenv.get('ENVIRONMENT') == 'PRODUCTION':
from .production import *
elif dotenv.get('ENVIRONMENT') == 'TRAVIS':
from .testing import *
else:
from .local import *
| <commit_before>"""Settings package initialization."""
import dotenv
dotenv.load()
# Ensure development settings are not used in testing and production:
if dotenv.get('ENVIRONMENT') == 'HEROKU':
from .production import *
elif dotenv.get('ENVIRONMENT') == 'TRAVIS':
from .testing import *
else:
from .local import *
<commit_msg>Change ENVIRONMENT variable from HEROKU to PRODUCTION<commit_after>"""Settings package initialization."""
import dotenv
dotenv.load()
# Ensure development settings are not used in testing and production:
if dotenv.get('ENVIRONMENT') == 'PRODUCTION':
from .production import *
elif dotenv.get('ENVIRONMENT') == 'TRAVIS':
from .testing import *
else:
from .local import *
|
a55f0fa9f80042c3fa673f263b259e70dd52f7d6 | streak-podium/read.py | streak-podium/read.py | import requests
def input_file(filename):
"""
Read a file and return list of usernames.
Assumes one username per line and ignores blank lines.
"""
with open(filename, 'r') as f:
return list(line.strip() for line in f if line.strip())
def org_members(org_name):
"""
Query Github API and return list of members from a Github organization.
"""
# TODO: Return github org members, not a placeholder
return ['supermitch', 'Jollyra']
def svg_data(username):
"""
Returns the contribution streak SVG file contents from Github
for a specific username.
"""
url = 'https://github.com/users/{}/contributions'.format(username)
r = requests.get(url)
return r.text
| import requests
def input_file(filename):
"""
Read a file and return list of usernames.
Assumes one username per line and ignores blank lines.
"""
with open(filename, 'r') as f:
return list(line.strip() for line in f if line.strip())
def org_members(org_name):
"""
Query Github API and return list of members from a Github organization.
"""
# TODO: Return github org members, not a placeholder
return ['supermitch', 'Jollyra']
def svg_data(username):
"""
Returns the contribution streak SVG file contents from Github
for a specific username.
"""
url = 'https://github.com/users/{}/contributions'.format(username)
try:
r = requests.get(url)
except requests.exceptions.ConnectionError:
logging.warn('Connection error trying to get url: [{}]'.format(url))
return None
return r.text
| Handle exception where connection fails | Handle exception where connection fails
| Python | mit | jollyra/hubot-streak-podium,jollyra/hubot-commit-streak,supermitch/streak-podium,jollyra/hubot-streak-podium,jollyra/hubot-commit-streak,supermitch/streak-podium | import requests
def input_file(filename):
"""
Read a file and return list of usernames.
Assumes one username per line and ignores blank lines.
"""
with open(filename, 'r') as f:
return list(line.strip() for line in f if line.strip())
def org_members(org_name):
"""
Query Github API and return list of members from a Github organization.
"""
# TODO: Return github org members, not a placeholder
return ['supermitch', 'Jollyra']
def svg_data(username):
"""
Returns the contribution streak SVG file contents from Github
for a specific username.
"""
url = 'https://github.com/users/{}/contributions'.format(username)
r = requests.get(url)
return r.text
Handle exception where connection fails | import requests
def input_file(filename):
"""
Read a file and return list of usernames.
Assumes one username per line and ignores blank lines.
"""
with open(filename, 'r') as f:
return list(line.strip() for line in f if line.strip())
def org_members(org_name):
"""
Query Github API and return list of members from a Github organization.
"""
# TODO: Return github org members, not a placeholder
return ['supermitch', 'Jollyra']
def svg_data(username):
"""
Returns the contribution streak SVG file contents from Github
for a specific username.
"""
url = 'https://github.com/users/{}/contributions'.format(username)
try:
r = requests.get(url)
except requests.exceptions.ConnectionError:
logging.warn('Connection error trying to get url: [{}]'.format(url))
return None
return r.text
| <commit_before>import requests
def input_file(filename):
"""
Read a file and return list of usernames.
Assumes one username per line and ignores blank lines.
"""
with open(filename, 'r') as f:
return list(line.strip() for line in f if line.strip())
def org_members(org_name):
"""
Query Github API and return list of members from a Github organization.
"""
# TODO: Return github org members, not a placeholder
return ['supermitch', 'Jollyra']
def svg_data(username):
"""
Returns the contribution streak SVG file contents from Github
for a specific username.
"""
url = 'https://github.com/users/{}/contributions'.format(username)
r = requests.get(url)
return r.text
<commit_msg>Handle exception where connection fails<commit_after> | import requests
def input_file(filename):
"""
Read a file and return list of usernames.
Assumes one username per line and ignores blank lines.
"""
with open(filename, 'r') as f:
return list(line.strip() for line in f if line.strip())
def org_members(org_name):
"""
Query Github API and return list of members from a Github organization.
"""
# TODO: Return github org members, not a placeholder
return ['supermitch', 'Jollyra']
def svg_data(username):
"""
Returns the contribution streak SVG file contents from Github
for a specific username.
"""
url = 'https://github.com/users/{}/contributions'.format(username)
try:
r = requests.get(url)
except requests.exceptions.ConnectionError:
logging.warn('Connection error trying to get url: [{}]'.format(url))
return None
return r.text
| import requests
def input_file(filename):
"""
Read a file and return list of usernames.
Assumes one username per line and ignores blank lines.
"""
with open(filename, 'r') as f:
return list(line.strip() for line in f if line.strip())
def org_members(org_name):
"""
Query Github API and return list of members from a Github organization.
"""
# TODO: Return github org members, not a placeholder
return ['supermitch', 'Jollyra']
def svg_data(username):
"""
Returns the contribution streak SVG file contents from Github
for a specific username.
"""
url = 'https://github.com/users/{}/contributions'.format(username)
r = requests.get(url)
return r.text
Handle exception where connection failsimport requests
def input_file(filename):
"""
Read a file and return list of usernames.
Assumes one username per line and ignores blank lines.
"""
with open(filename, 'r') as f:
return list(line.strip() for line in f if line.strip())
def org_members(org_name):
"""
Query Github API and return list of members from a Github organization.
"""
# TODO: Return github org members, not a placeholder
return ['supermitch', 'Jollyra']
def svg_data(username):
"""
Returns the contribution streak SVG file contents from Github
for a specific username.
"""
url = 'https://github.com/users/{}/contributions'.format(username)
try:
r = requests.get(url)
except requests.exceptions.ConnectionError:
logging.warn('Connection error trying to get url: [{}]'.format(url))
return None
return r.text
| <commit_before>import requests
def input_file(filename):
"""
Read a file and return list of usernames.
Assumes one username per line and ignores blank lines.
"""
with open(filename, 'r') as f:
return list(line.strip() for line in f if line.strip())
def org_members(org_name):
"""
Query Github API and return list of members from a Github organization.
"""
# TODO: Return github org members, not a placeholder
return ['supermitch', 'Jollyra']
def svg_data(username):
"""
Returns the contribution streak SVG file contents from Github
for a specific username.
"""
url = 'https://github.com/users/{}/contributions'.format(username)
r = requests.get(url)
return r.text
<commit_msg>Handle exception where connection fails<commit_after>import requests
def input_file(filename):
"""
Read a file and return list of usernames.
Assumes one username per line and ignores blank lines.
"""
with open(filename, 'r') as f:
return list(line.strip() for line in f if line.strip())
def org_members(org_name):
"""
Query Github API and return list of members from a Github organization.
"""
# TODO: Return github org members, not a placeholder
return ['supermitch', 'Jollyra']
def svg_data(username):
"""
Returns the contribution streak SVG file contents from Github
for a specific username.
"""
url = 'https://github.com/users/{}/contributions'.format(username)
try:
r = requests.get(url)
except requests.exceptions.ConnectionError:
logging.warn('Connection error trying to get url: [{}]'.format(url))
return None
return r.text
|
764256427ea7c0dbf73accf63ed05e8372f58a75 | test/pyfrontend/util.py | test/pyfrontend/util.py | import contextlib
import tempfile
import shutil
@contextlib.contextmanager
def temporary_directory():
"""Simple context manager to make a temporary directory"""
tmpdir = tempfile.mkdtemp()
yield tmpdir
shutil.rmtree(tmpdir, ignore_errors=True)
| import contextlib
import tempfile
import shutil
import saliweb.test # for python 2.6 support
@contextlib.contextmanager
def temporary_directory():
"""Simple context manager to make a temporary directory"""
tmpdir = tempfile.mkdtemp()
yield tmpdir
shutil.rmtree(tmpdir, ignore_errors=True)
| Add unittest methods to Python 2.6 | Add unittest methods to Python 2.6
| Python | lgpl-2.1 | salilab/saliweb,salilab/saliweb,salilab/saliweb,salilab/saliweb,salilab/saliweb | import contextlib
import tempfile
import shutil
@contextlib.contextmanager
def temporary_directory():
"""Simple context manager to make a temporary directory"""
tmpdir = tempfile.mkdtemp()
yield tmpdir
shutil.rmtree(tmpdir, ignore_errors=True)
Add unittest methods to Python 2.6 | import contextlib
import tempfile
import shutil
import saliweb.test # for python 2.6 support
@contextlib.contextmanager
def temporary_directory():
"""Simple context manager to make a temporary directory"""
tmpdir = tempfile.mkdtemp()
yield tmpdir
shutil.rmtree(tmpdir, ignore_errors=True)
| <commit_before>import contextlib
import tempfile
import shutil
@contextlib.contextmanager
def temporary_directory():
"""Simple context manager to make a temporary directory"""
tmpdir = tempfile.mkdtemp()
yield tmpdir
shutil.rmtree(tmpdir, ignore_errors=True)
<commit_msg>Add unittest methods to Python 2.6<commit_after> | import contextlib
import tempfile
import shutil
import saliweb.test # for python 2.6 support
@contextlib.contextmanager
def temporary_directory():
"""Simple context manager to make a temporary directory"""
tmpdir = tempfile.mkdtemp()
yield tmpdir
shutil.rmtree(tmpdir, ignore_errors=True)
| import contextlib
import tempfile
import shutil
@contextlib.contextmanager
def temporary_directory():
"""Simple context manager to make a temporary directory"""
tmpdir = tempfile.mkdtemp()
yield tmpdir
shutil.rmtree(tmpdir, ignore_errors=True)
Add unittest methods to Python 2.6import contextlib
import tempfile
import shutil
import saliweb.test # for python 2.6 support
@contextlib.contextmanager
def temporary_directory():
"""Simple context manager to make a temporary directory"""
tmpdir = tempfile.mkdtemp()
yield tmpdir
shutil.rmtree(tmpdir, ignore_errors=True)
| <commit_before>import contextlib
import tempfile
import shutil
@contextlib.contextmanager
def temporary_directory():
"""Simple context manager to make a temporary directory"""
tmpdir = tempfile.mkdtemp()
yield tmpdir
shutil.rmtree(tmpdir, ignore_errors=True)
<commit_msg>Add unittest methods to Python 2.6<commit_after>import contextlib
import tempfile
import shutil
import saliweb.test # for python 2.6 support
@contextlib.contextmanager
def temporary_directory():
"""Simple context manager to make a temporary directory"""
tmpdir = tempfile.mkdtemp()
yield tmpdir
shutil.rmtree(tmpdir, ignore_errors=True)
|
46b60c5886ede34db8998d7cfd5ae36f9211a0e8 | ovp_users/tests/test_views/__init__.py | ovp_users/tests/test_views/__init__.py | from ovp_users.tests.test_views.user import UserResourceViewSetTestCase
from ovp_users.tests.test_views.auth import JWTAuthTestCase
from ovp_users.tests.test_views.password_recovery import RecoveryTokenViewSetTestCase
from ovp_users.tests.test_views.password_recovery import RecoverPasswordViewSetTestCase
| from ovp_users.tests.test_views.user import UserResourceViewSetTestCase
from ovp_users.tests.test_views.auth import JWTAuthTestCase
from ovp_users.tests.test_views.profile import ProfileTestCase
from ovp_users.tests.test_views.password_recovery import RecoveryTokenViewSetTestCase
from ovp_users.tests.test_views.password_recovery import RecoverPasswordViewSetTestCase
| Add profile views tests to suite | Add profile views tests to suite
| Python | agpl-3.0 | OpenVolunteeringPlatform/django-ovp-users,OpenVolunteeringPlatform/django-ovp-users | from ovp_users.tests.test_views.user import UserResourceViewSetTestCase
from ovp_users.tests.test_views.auth import JWTAuthTestCase
from ovp_users.tests.test_views.password_recovery import RecoveryTokenViewSetTestCase
from ovp_users.tests.test_views.password_recovery import RecoverPasswordViewSetTestCase
Add profile views tests to suite | from ovp_users.tests.test_views.user import UserResourceViewSetTestCase
from ovp_users.tests.test_views.auth import JWTAuthTestCase
from ovp_users.tests.test_views.profile import ProfileTestCase
from ovp_users.tests.test_views.password_recovery import RecoveryTokenViewSetTestCase
from ovp_users.tests.test_views.password_recovery import RecoverPasswordViewSetTestCase
| <commit_before>from ovp_users.tests.test_views.user import UserResourceViewSetTestCase
from ovp_users.tests.test_views.auth import JWTAuthTestCase
from ovp_users.tests.test_views.password_recovery import RecoveryTokenViewSetTestCase
from ovp_users.tests.test_views.password_recovery import RecoverPasswordViewSetTestCase
<commit_msg>Add profile views tests to suite<commit_after> | from ovp_users.tests.test_views.user import UserResourceViewSetTestCase
from ovp_users.tests.test_views.auth import JWTAuthTestCase
from ovp_users.tests.test_views.profile import ProfileTestCase
from ovp_users.tests.test_views.password_recovery import RecoveryTokenViewSetTestCase
from ovp_users.tests.test_views.password_recovery import RecoverPasswordViewSetTestCase
| from ovp_users.tests.test_views.user import UserResourceViewSetTestCase
from ovp_users.tests.test_views.auth import JWTAuthTestCase
from ovp_users.tests.test_views.password_recovery import RecoveryTokenViewSetTestCase
from ovp_users.tests.test_views.password_recovery import RecoverPasswordViewSetTestCase
Add profile views tests to suitefrom ovp_users.tests.test_views.user import UserResourceViewSetTestCase
from ovp_users.tests.test_views.auth import JWTAuthTestCase
from ovp_users.tests.test_views.profile import ProfileTestCase
from ovp_users.tests.test_views.password_recovery import RecoveryTokenViewSetTestCase
from ovp_users.tests.test_views.password_recovery import RecoverPasswordViewSetTestCase
| <commit_before>from ovp_users.tests.test_views.user import UserResourceViewSetTestCase
from ovp_users.tests.test_views.auth import JWTAuthTestCase
from ovp_users.tests.test_views.password_recovery import RecoveryTokenViewSetTestCase
from ovp_users.tests.test_views.password_recovery import RecoverPasswordViewSetTestCase
<commit_msg>Add profile views tests to suite<commit_after>from ovp_users.tests.test_views.user import UserResourceViewSetTestCase
from ovp_users.tests.test_views.auth import JWTAuthTestCase
from ovp_users.tests.test_views.profile import ProfileTestCase
from ovp_users.tests.test_views.password_recovery import RecoveryTokenViewSetTestCase
from ovp_users.tests.test_views.password_recovery import RecoverPasswordViewSetTestCase
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.