commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
6547d653491adb6ab46e4a3a5f8251129719d3f7
login/middleware.py
login/middleware.py
from django.conf import settings from django.http import HttpResponseRedirect DETACH_PATH = '/user/detach' ACTIVATE_PATH = '/user/activate' class DetachMiddleware(object): def process_request(self, request): if not request.path == '/login/' \ and not request.path.startswith('/api'): if not request.user.is_anonymous \ and not request.user.is_native \ and not request.path == DETACH_PATH: return HttpResponseRedirect(DETACH_PATH) if not request.user.is_anonymous \ and not request.user.is_mail_verified \ and not request.path.startswith(ACTIVATE_PATH): return HttpResponseRedirect(ACTIVATE_PATH)
from django.conf import settings from django.http import HttpResponseRedirect DETACH_PATH = '/user/detach' ACTIVATE_PATH = '/user/activate' class DetachMiddleware(object): def process_request(self, request): if not request.path == '/login/' \ and not request.path.startswith('/api') \ and not request.user.is_anonymous: if not request.user.is_native: if not request.path == DETACH_PATH: return HttpResponseRedirect(DETACH_PATH) elif not request.user.is_mail_verified \ and not request.path.startswith(ACTIVATE_PATH): return HttpResponseRedirect(ACTIVATE_PATH)
Remove infinite loop if user is neither native nor verified
Remove infinite loop if user is neither native nor verified
Python
agpl-3.0
openego/oeplatform,openego/oeplatform,openego/oeplatform,openego/oeplatform
from django.conf import settings from django.http import HttpResponseRedirect DETACH_PATH = '/user/detach' ACTIVATE_PATH = '/user/activate' class DetachMiddleware(object): def process_request(self, request): if not request.path == '/login/' \ and not request.path.startswith('/api'): if not request.user.is_anonymous \ and not request.user.is_native \ and not request.path == DETACH_PATH: return HttpResponseRedirect(DETACH_PATH) if not request.user.is_anonymous \ and not request.user.is_mail_verified \ and not request.path.startswith(ACTIVATE_PATH): return HttpResponseRedirect(ACTIVATE_PATH)Remove infinite loop if user is neither native nor verified
from django.conf import settings from django.http import HttpResponseRedirect DETACH_PATH = '/user/detach' ACTIVATE_PATH = '/user/activate' class DetachMiddleware(object): def process_request(self, request): if not request.path == '/login/' \ and not request.path.startswith('/api') \ and not request.user.is_anonymous: if not request.user.is_native: if not request.path == DETACH_PATH: return HttpResponseRedirect(DETACH_PATH) elif not request.user.is_mail_verified \ and not request.path.startswith(ACTIVATE_PATH): return HttpResponseRedirect(ACTIVATE_PATH)
<commit_before>from django.conf import settings from django.http import HttpResponseRedirect DETACH_PATH = '/user/detach' ACTIVATE_PATH = '/user/activate' class DetachMiddleware(object): def process_request(self, request): if not request.path == '/login/' \ and not request.path.startswith('/api'): if not request.user.is_anonymous \ and not request.user.is_native \ and not request.path == DETACH_PATH: return HttpResponseRedirect(DETACH_PATH) if not request.user.is_anonymous \ and not request.user.is_mail_verified \ and not request.path.startswith(ACTIVATE_PATH): return HttpResponseRedirect(ACTIVATE_PATH)<commit_msg>Remove infinite loop if user is neither native nor verified<commit_after>
from django.conf import settings from django.http import HttpResponseRedirect DETACH_PATH = '/user/detach' ACTIVATE_PATH = '/user/activate' class DetachMiddleware(object): def process_request(self, request): if not request.path == '/login/' \ and not request.path.startswith('/api') \ and not request.user.is_anonymous: if not request.user.is_native: if not request.path == DETACH_PATH: return HttpResponseRedirect(DETACH_PATH) elif not request.user.is_mail_verified \ and not request.path.startswith(ACTIVATE_PATH): return HttpResponseRedirect(ACTIVATE_PATH)
from django.conf import settings from django.http import HttpResponseRedirect DETACH_PATH = '/user/detach' ACTIVATE_PATH = '/user/activate' class DetachMiddleware(object): def process_request(self, request): if not request.path == '/login/' \ and not request.path.startswith('/api'): if not request.user.is_anonymous \ and not request.user.is_native \ and not request.path == DETACH_PATH: return HttpResponseRedirect(DETACH_PATH) if not request.user.is_anonymous \ and not request.user.is_mail_verified \ and not request.path.startswith(ACTIVATE_PATH): return HttpResponseRedirect(ACTIVATE_PATH)Remove infinite loop if user is neither native nor verifiedfrom django.conf import settings from django.http import HttpResponseRedirect DETACH_PATH = '/user/detach' ACTIVATE_PATH = '/user/activate' class DetachMiddleware(object): def process_request(self, request): if not request.path == '/login/' \ and not request.path.startswith('/api') \ and not request.user.is_anonymous: if not request.user.is_native: if not request.path == DETACH_PATH: return HttpResponseRedirect(DETACH_PATH) elif not request.user.is_mail_verified \ and not request.path.startswith(ACTIVATE_PATH): return HttpResponseRedirect(ACTIVATE_PATH)
<commit_before>from django.conf import settings from django.http import HttpResponseRedirect DETACH_PATH = '/user/detach' ACTIVATE_PATH = '/user/activate' class DetachMiddleware(object): def process_request(self, request): if not request.path == '/login/' \ and not request.path.startswith('/api'): if not request.user.is_anonymous \ and not request.user.is_native \ and not request.path == DETACH_PATH: return HttpResponseRedirect(DETACH_PATH) if not request.user.is_anonymous \ and not request.user.is_mail_verified \ and not request.path.startswith(ACTIVATE_PATH): return HttpResponseRedirect(ACTIVATE_PATH)<commit_msg>Remove infinite loop if user is neither native nor verified<commit_after>from django.conf import settings from django.http import HttpResponseRedirect DETACH_PATH = '/user/detach' ACTIVATE_PATH = '/user/activate' class DetachMiddleware(object): def process_request(self, request): if not request.path == '/login/' \ and not request.path.startswith('/api') \ and not request.user.is_anonymous: if not request.user.is_native: if not request.path == DETACH_PATH: return HttpResponseRedirect(DETACH_PATH) elif not request.user.is_mail_verified \ and not request.path.startswith(ACTIVATE_PATH): return HttpResponseRedirect(ACTIVATE_PATH)
de6de7c43932224f2845cc540170eeaf46650491
mkerefuse/refuse.py
mkerefuse/refuse.py
from .util import XPathObject class RefusePickup(XPathObject): """Defines attribute to XPath specification matching""" input_properties = { 'success_msg': '//*[@id="nConf"]/h1', 'route_garbage': '//*[@id="nConf"]/strong[1]', 'next_pickup_garbage': '//*[@id="nConf"]/strong[2]', 'route_recyle': '//*[@id="nConf"]/strong[3]', 'next_pickup_recycle_after': '//*[@id="nConf"]/strong[4]', 'next_pickup_recycle_before': '//*[@id="nConf"]/strong[5]', } """Maps the key to an attr name & value to an XPath lookup"""
Add RefusePickup class (for parsing)
Add RefusePickup class (for parsing)
Python
unlicense
tomislacker/python-mke-trash-pickup,tomislacker/python-mke-trash-pickup
Add RefusePickup class (for parsing)
from .util import XPathObject class RefusePickup(XPathObject): """Defines attribute to XPath specification matching""" input_properties = { 'success_msg': '//*[@id="nConf"]/h1', 'route_garbage': '//*[@id="nConf"]/strong[1]', 'next_pickup_garbage': '//*[@id="nConf"]/strong[2]', 'route_recyle': '//*[@id="nConf"]/strong[3]', 'next_pickup_recycle_after': '//*[@id="nConf"]/strong[4]', 'next_pickup_recycle_before': '//*[@id="nConf"]/strong[5]', } """Maps the key to an attr name & value to an XPath lookup"""
<commit_before> <commit_msg>Add RefusePickup class (for parsing)<commit_after>
from .util import XPathObject class RefusePickup(XPathObject): """Defines attribute to XPath specification matching""" input_properties = { 'success_msg': '//*[@id="nConf"]/h1', 'route_garbage': '//*[@id="nConf"]/strong[1]', 'next_pickup_garbage': '//*[@id="nConf"]/strong[2]', 'route_recyle': '//*[@id="nConf"]/strong[3]', 'next_pickup_recycle_after': '//*[@id="nConf"]/strong[4]', 'next_pickup_recycle_before': '//*[@id="nConf"]/strong[5]', } """Maps the key to an attr name & value to an XPath lookup"""
Add RefusePickup class (for parsing)from .util import XPathObject class RefusePickup(XPathObject): """Defines attribute to XPath specification matching""" input_properties = { 'success_msg': '//*[@id="nConf"]/h1', 'route_garbage': '//*[@id="nConf"]/strong[1]', 'next_pickup_garbage': '//*[@id="nConf"]/strong[2]', 'route_recyle': '//*[@id="nConf"]/strong[3]', 'next_pickup_recycle_after': '//*[@id="nConf"]/strong[4]', 'next_pickup_recycle_before': '//*[@id="nConf"]/strong[5]', } """Maps the key to an attr name & value to an XPath lookup"""
<commit_before> <commit_msg>Add RefusePickup class (for parsing)<commit_after>from .util import XPathObject class RefusePickup(XPathObject): """Defines attribute to XPath specification matching""" input_properties = { 'success_msg': '//*[@id="nConf"]/h1', 'route_garbage': '//*[@id="nConf"]/strong[1]', 'next_pickup_garbage': '//*[@id="nConf"]/strong[2]', 'route_recyle': '//*[@id="nConf"]/strong[3]', 'next_pickup_recycle_after': '//*[@id="nConf"]/strong[4]', 'next_pickup_recycle_before': '//*[@id="nConf"]/strong[5]', } """Maps the key to an attr name & value to an XPath lookup"""
016f14304b6b86634c4608927d3345f993178682
config.py
config.py
### # Copyright (c) 2012, spline # All rights reserved. # # ### import supybot.conf as conf import supybot.registry as registry from supybot.i18n import PluginInternationalization, internationalizeDocstring _ = PluginInternationalization('Scores') def configure(advanced): # This will be called by supybot to configure this module. advanced is # a bool that specifies whether the user identified himself as an advanced # user or not. You should effect your configuration by manipulating the # registry as appropriate. from supybot.questions import expect, anything, something, yn conf.registerPlugin('Scores', True) Scores = conf.registerPlugin('Scores') # This is where your configuration variables (if any) should go. For example: # conf.registerGlobalValue(Scores, 'someConfigVariableName', # registry.Boolean(False, _("""Help for someConfigVariableName."""))) # vim:set shiftwidth=4 tabstop=4 expandtab textwidth=79:
### # Copyright (c) 2012, spline # All rights reserved. # # ### import supybot.conf as conf import supybot.registry as registry from supybot.i18n import PluginInternationalization, internationalizeDocstring _ = PluginInternationalization('Scores') def configure(advanced): # This will be called by supybot to configure this module. advanced is # a bool that specifies whether the user identified himself as an advanced # user or not. You should effect your configuration by manipulating the # registry as appropriate. from supybot.questions import expect, anything, something, yn conf.registerPlugin('Scores', True) Scores = conf.registerPlugin('Scores') conf.registerChannelValue(Scores, 'disableANSI', registry.Boolean(False, """Do not display any ANSI (color/bold) in output.""")) # vim:set shiftwidth=4 tabstop=4 expandtab textwidth=250:
Add in channel value for disabling ansi
Add in channel value for disabling ansi
Python
mit
reticulatingspline/Scores,cottongin/Scores
### # Copyright (c) 2012, spline # All rights reserved. # # ### import supybot.conf as conf import supybot.registry as registry from supybot.i18n import PluginInternationalization, internationalizeDocstring _ = PluginInternationalization('Scores') def configure(advanced): # This will be called by supybot to configure this module. advanced is # a bool that specifies whether the user identified himself as an advanced # user or not. You should effect your configuration by manipulating the # registry as appropriate. from supybot.questions import expect, anything, something, yn conf.registerPlugin('Scores', True) Scores = conf.registerPlugin('Scores') # This is where your configuration variables (if any) should go. For example: # conf.registerGlobalValue(Scores, 'someConfigVariableName', # registry.Boolean(False, _("""Help for someConfigVariableName."""))) # vim:set shiftwidth=4 tabstop=4 expandtab textwidth=79: Add in channel value for disabling ansi
### # Copyright (c) 2012, spline # All rights reserved. # # ### import supybot.conf as conf import supybot.registry as registry from supybot.i18n import PluginInternationalization, internationalizeDocstring _ = PluginInternationalization('Scores') def configure(advanced): # This will be called by supybot to configure this module. advanced is # a bool that specifies whether the user identified himself as an advanced # user or not. You should effect your configuration by manipulating the # registry as appropriate. from supybot.questions import expect, anything, something, yn conf.registerPlugin('Scores', True) Scores = conf.registerPlugin('Scores') conf.registerChannelValue(Scores, 'disableANSI', registry.Boolean(False, """Do not display any ANSI (color/bold) in output.""")) # vim:set shiftwidth=4 tabstop=4 expandtab textwidth=250:
<commit_before>### # Copyright (c) 2012, spline # All rights reserved. # # ### import supybot.conf as conf import supybot.registry as registry from supybot.i18n import PluginInternationalization, internationalizeDocstring _ = PluginInternationalization('Scores') def configure(advanced): # This will be called by supybot to configure this module. advanced is # a bool that specifies whether the user identified himself as an advanced # user or not. You should effect your configuration by manipulating the # registry as appropriate. from supybot.questions import expect, anything, something, yn conf.registerPlugin('Scores', True) Scores = conf.registerPlugin('Scores') # This is where your configuration variables (if any) should go. For example: # conf.registerGlobalValue(Scores, 'someConfigVariableName', # registry.Boolean(False, _("""Help for someConfigVariableName."""))) # vim:set shiftwidth=4 tabstop=4 expandtab textwidth=79: <commit_msg>Add in channel value for disabling ansi<commit_after>
### # Copyright (c) 2012, spline # All rights reserved. # # ### import supybot.conf as conf import supybot.registry as registry from supybot.i18n import PluginInternationalization, internationalizeDocstring _ = PluginInternationalization('Scores') def configure(advanced): # This will be called by supybot to configure this module. advanced is # a bool that specifies whether the user identified himself as an advanced # user or not. You should effect your configuration by manipulating the # registry as appropriate. from supybot.questions import expect, anything, something, yn conf.registerPlugin('Scores', True) Scores = conf.registerPlugin('Scores') conf.registerChannelValue(Scores, 'disableANSI', registry.Boolean(False, """Do not display any ANSI (color/bold) in output.""")) # vim:set shiftwidth=4 tabstop=4 expandtab textwidth=250:
### # Copyright (c) 2012, spline # All rights reserved. # # ### import supybot.conf as conf import supybot.registry as registry from supybot.i18n import PluginInternationalization, internationalizeDocstring _ = PluginInternationalization('Scores') def configure(advanced): # This will be called by supybot to configure this module. advanced is # a bool that specifies whether the user identified himself as an advanced # user or not. You should effect your configuration by manipulating the # registry as appropriate. from supybot.questions import expect, anything, something, yn conf.registerPlugin('Scores', True) Scores = conf.registerPlugin('Scores') # This is where your configuration variables (if any) should go. For example: # conf.registerGlobalValue(Scores, 'someConfigVariableName', # registry.Boolean(False, _("""Help for someConfigVariableName."""))) # vim:set shiftwidth=4 tabstop=4 expandtab textwidth=79: Add in channel value for disabling ansi### # Copyright (c) 2012, spline # All rights reserved. # # ### import supybot.conf as conf import supybot.registry as registry from supybot.i18n import PluginInternationalization, internationalizeDocstring _ = PluginInternationalization('Scores') def configure(advanced): # This will be called by supybot to configure this module. advanced is # a bool that specifies whether the user identified himself as an advanced # user or not. You should effect your configuration by manipulating the # registry as appropriate. from supybot.questions import expect, anything, something, yn conf.registerPlugin('Scores', True) Scores = conf.registerPlugin('Scores') conf.registerChannelValue(Scores, 'disableANSI', registry.Boolean(False, """Do not display any ANSI (color/bold) in output.""")) # vim:set shiftwidth=4 tabstop=4 expandtab textwidth=250:
<commit_before>### # Copyright (c) 2012, spline # All rights reserved. # # ### import supybot.conf as conf import supybot.registry as registry from supybot.i18n import PluginInternationalization, internationalizeDocstring _ = PluginInternationalization('Scores') def configure(advanced): # This will be called by supybot to configure this module. advanced is # a bool that specifies whether the user identified himself as an advanced # user or not. You should effect your configuration by manipulating the # registry as appropriate. from supybot.questions import expect, anything, something, yn conf.registerPlugin('Scores', True) Scores = conf.registerPlugin('Scores') # This is where your configuration variables (if any) should go. For example: # conf.registerGlobalValue(Scores, 'someConfigVariableName', # registry.Boolean(False, _("""Help for someConfigVariableName."""))) # vim:set shiftwidth=4 tabstop=4 expandtab textwidth=79: <commit_msg>Add in channel value for disabling ansi<commit_after>### # Copyright (c) 2012, spline # All rights reserved. # # ### import supybot.conf as conf import supybot.registry as registry from supybot.i18n import PluginInternationalization, internationalizeDocstring _ = PluginInternationalization('Scores') def configure(advanced): # This will be called by supybot to configure this module. advanced is # a bool that specifies whether the user identified himself as an advanced # user or not. You should effect your configuration by manipulating the # registry as appropriate. from supybot.questions import expect, anything, something, yn conf.registerPlugin('Scores', True) Scores = conf.registerPlugin('Scores') conf.registerChannelValue(Scores, 'disableANSI', registry.Boolean(False, """Do not display any ANSI (color/bold) in output.""")) # vim:set shiftwidth=4 tabstop=4 expandtab textwidth=250:
61fe55efba2c491da6a93421fa702f123615bc32
spacy/lang/en/__init__.py
spacy/lang/en/__init__.py
# coding: utf8 from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .tag_map import TAG_MAP from .stop_words import STOP_WORDS from .lex_attrs import LEX_ATTRS from .morph_rules import MORPH_RULES from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC from .syntax_iterators import SYNTAX_ITERATORS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class English(Language): lang = 'en' class Defaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'en' lex_attr_getters.update(LEX_ATTRS) tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) tag_map = dict(TAG_MAP) stop_words = set(STOP_WORDS) morph_rules = dict(MORPH_RULES) lemma_rules = dict(LEMMA_RULES) lemma_index = dict(LEMMA_INDEX) lemma_exc = dict(LEMMA_EXC) sytax_iterators = dict(SYNTAX_ITERATORS) __all__ = ['English']
# coding: utf8 from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .tag_map import TAG_MAP from .stop_words import STOP_WORDS from .lex_attrs import LEX_ATTRS from .morph_rules import MORPH_RULES from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC from .syntax_iterators import SYNTAX_ITERATORS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class EnglishDefaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'en' lex_attr_getters.update(LEX_ATTRS) tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) tag_map = dict(TAG_MAP) stop_words = set(STOP_WORDS) morph_rules = dict(MORPH_RULES) lemma_rules = dict(LEMMA_RULES) lemma_index = dict(LEMMA_INDEX) lemma_exc = dict(LEMMA_EXC) sytax_iterators = dict(SYNTAX_ITERATORS) class English(Language): lang = 'en' Defaults = EnglishDefaults __all__ = ['English', 'EnglishDefaults']
Move EnglishDefaults class out of English
Move EnglishDefaults class out of English
Python
mit
honnibal/spaCy,aikramer2/spaCy,recognai/spaCy,recognai/spaCy,honnibal/spaCy,aikramer2/spaCy,explosion/spaCy,aikramer2/spaCy,recognai/spaCy,aikramer2/spaCy,spacy-io/spaCy,recognai/spaCy,spacy-io/spaCy,explosion/spaCy,explosion/spaCy,explosion/spaCy,honnibal/spaCy,spacy-io/spaCy,recognai/spaCy,aikramer2/spaCy,recognai/spaCy,spacy-io/spaCy,honnibal/spaCy,explosion/spaCy,aikramer2/spaCy,explosion/spaCy,spacy-io/spaCy,spacy-io/spaCy
# coding: utf8 from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .tag_map import TAG_MAP from .stop_words import STOP_WORDS from .lex_attrs import LEX_ATTRS from .morph_rules import MORPH_RULES from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC from .syntax_iterators import SYNTAX_ITERATORS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class English(Language): lang = 'en' class Defaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'en' lex_attr_getters.update(LEX_ATTRS) tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) tag_map = dict(TAG_MAP) stop_words = set(STOP_WORDS) morph_rules = dict(MORPH_RULES) lemma_rules = dict(LEMMA_RULES) lemma_index = dict(LEMMA_INDEX) lemma_exc = dict(LEMMA_EXC) sytax_iterators = dict(SYNTAX_ITERATORS) __all__ = ['English'] Move EnglishDefaults class out of English
# coding: utf8 from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .tag_map import TAG_MAP from .stop_words import STOP_WORDS from .lex_attrs import LEX_ATTRS from .morph_rules import MORPH_RULES from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC from .syntax_iterators import SYNTAX_ITERATORS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class EnglishDefaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'en' lex_attr_getters.update(LEX_ATTRS) tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) tag_map = dict(TAG_MAP) stop_words = set(STOP_WORDS) morph_rules = dict(MORPH_RULES) lemma_rules = dict(LEMMA_RULES) lemma_index = dict(LEMMA_INDEX) lemma_exc = dict(LEMMA_EXC) sytax_iterators = dict(SYNTAX_ITERATORS) class English(Language): lang = 'en' Defaults = EnglishDefaults __all__ = ['English', 'EnglishDefaults']
<commit_before># coding: utf8 from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .tag_map import TAG_MAP from .stop_words import STOP_WORDS from .lex_attrs import LEX_ATTRS from .morph_rules import MORPH_RULES from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC from .syntax_iterators import SYNTAX_ITERATORS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class English(Language): lang = 'en' class Defaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'en' lex_attr_getters.update(LEX_ATTRS) tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) tag_map = dict(TAG_MAP) stop_words = set(STOP_WORDS) morph_rules = dict(MORPH_RULES) lemma_rules = dict(LEMMA_RULES) lemma_index = dict(LEMMA_INDEX) lemma_exc = dict(LEMMA_EXC) sytax_iterators = dict(SYNTAX_ITERATORS) __all__ = ['English'] <commit_msg>Move EnglishDefaults class out of English<commit_after>
# coding: utf8 from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .tag_map import TAG_MAP from .stop_words import STOP_WORDS from .lex_attrs import LEX_ATTRS from .morph_rules import MORPH_RULES from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC from .syntax_iterators import SYNTAX_ITERATORS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class EnglishDefaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'en' lex_attr_getters.update(LEX_ATTRS) tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) tag_map = dict(TAG_MAP) stop_words = set(STOP_WORDS) morph_rules = dict(MORPH_RULES) lemma_rules = dict(LEMMA_RULES) lemma_index = dict(LEMMA_INDEX) lemma_exc = dict(LEMMA_EXC) sytax_iterators = dict(SYNTAX_ITERATORS) class English(Language): lang = 'en' Defaults = EnglishDefaults __all__ = ['English', 'EnglishDefaults']
# coding: utf8 from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .tag_map import TAG_MAP from .stop_words import STOP_WORDS from .lex_attrs import LEX_ATTRS from .morph_rules import MORPH_RULES from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC from .syntax_iterators import SYNTAX_ITERATORS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class English(Language): lang = 'en' class Defaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'en' lex_attr_getters.update(LEX_ATTRS) tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) tag_map = dict(TAG_MAP) stop_words = set(STOP_WORDS) morph_rules = dict(MORPH_RULES) lemma_rules = dict(LEMMA_RULES) lemma_index = dict(LEMMA_INDEX) lemma_exc = dict(LEMMA_EXC) sytax_iterators = dict(SYNTAX_ITERATORS) __all__ = ['English'] Move EnglishDefaults class out of English# coding: utf8 from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .tag_map import TAG_MAP from .stop_words import STOP_WORDS from .lex_attrs import LEX_ATTRS from .morph_rules import MORPH_RULES from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC from .syntax_iterators import SYNTAX_ITERATORS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class EnglishDefaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'en' lex_attr_getters.update(LEX_ATTRS) tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) tag_map = dict(TAG_MAP) stop_words = set(STOP_WORDS) morph_rules = dict(MORPH_RULES) lemma_rules = dict(LEMMA_RULES) lemma_index = dict(LEMMA_INDEX) lemma_exc = dict(LEMMA_EXC) sytax_iterators = dict(SYNTAX_ITERATORS) class English(Language): lang = 'en' Defaults = EnglishDefaults __all__ = ['English', 'EnglishDefaults']
<commit_before># coding: utf8 from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .tag_map import TAG_MAP from .stop_words import STOP_WORDS from .lex_attrs import LEX_ATTRS from .morph_rules import MORPH_RULES from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC from .syntax_iterators import SYNTAX_ITERATORS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class English(Language): lang = 'en' class Defaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'en' lex_attr_getters.update(LEX_ATTRS) tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) tag_map = dict(TAG_MAP) stop_words = set(STOP_WORDS) morph_rules = dict(MORPH_RULES) lemma_rules = dict(LEMMA_RULES) lemma_index = dict(LEMMA_INDEX) lemma_exc = dict(LEMMA_EXC) sytax_iterators = dict(SYNTAX_ITERATORS) __all__ = ['English'] <commit_msg>Move EnglishDefaults class out of English<commit_after># coding: utf8 from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .tag_map import TAG_MAP from .stop_words import STOP_WORDS from .lex_attrs import LEX_ATTRS from .morph_rules import MORPH_RULES from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC from .syntax_iterators import SYNTAX_ITERATORS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class EnglishDefaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'en' lex_attr_getters.update(LEX_ATTRS) tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS) tag_map = dict(TAG_MAP) stop_words = set(STOP_WORDS) morph_rules = dict(MORPH_RULES) lemma_rules = dict(LEMMA_RULES) lemma_index = dict(LEMMA_INDEX) lemma_exc = dict(LEMMA_EXC) sytax_iterators = dict(SYNTAX_ITERATORS) class English(Language): lang = 'en' Defaults = EnglishDefaults __all__ = ['English', 'EnglishDefaults']
d6cd7c16e32f64c4fd3627953d751e8c8bc26f1c
premis_event_service/forms.py
premis_event_service/forms.py
from django import forms import settings OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES class EventSearchForm(forms.Form): event_outcome = forms.ChoiceField( widget=forms.Select( attrs={ 'id': 'prependedInput', 'class': 'input-small', } ), choices=OUTCOME_CHOICES, required=False) event_type = forms.ChoiceField( widget=forms.Select( attrs={ 'id': 'prependedInput', 'class': 'input-medium', } ), choices=EVENT_TYPE_CHOICES, required=False) start_date = forms.DateField( widget=forms.DateInput( attrs={ 'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small', } ), required=False) end_date = forms.DateField( widget=forms.DateInput( attrs={ 'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small', } ), required=False) linked_object_id = forms.CharField( widget=forms.TextInput( attrs={ 'placeholder': 'Linked Object ID', 'class': 'input-medium', } ), max_length=20, required=False, )
from django import forms import settings OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES class EventSearchForm(forms.Form): event_outcome = forms.ChoiceField( widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-small'}), choices=OUTCOME_CHOICES, required=False) event_type = forms.ChoiceField( widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-medium'}), choices=EVENT_TYPE_CHOICES, required=False) start_date = forms.DateField( widget=forms.DateInput(attrs={'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small'}), # noqa required=False) end_date = forms.DateField( widget=forms.DateInput(attrs={'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small'}), # noqa required=False) linked_object_id = forms.CharField( widget=forms.TextInput(attrs={'placeholder': 'Linked Object ID', 'class': 'input-medium'}), max_length=20, required=False)
Fix formatting for the EventSearchForm class.
Fix formatting for the EventSearchForm class.
Python
bsd-3-clause
unt-libraries/django-premis-event-service,unt-libraries/django-premis-event-service,unt-libraries/django-premis-event-service
from django import forms import settings OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES class EventSearchForm(forms.Form): event_outcome = forms.ChoiceField( widget=forms.Select( attrs={ 'id': 'prependedInput', 'class': 'input-small', } ), choices=OUTCOME_CHOICES, required=False) event_type = forms.ChoiceField( widget=forms.Select( attrs={ 'id': 'prependedInput', 'class': 'input-medium', } ), choices=EVENT_TYPE_CHOICES, required=False) start_date = forms.DateField( widget=forms.DateInput( attrs={ 'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small', } ), required=False) end_date = forms.DateField( widget=forms.DateInput( attrs={ 'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small', } ), required=False) linked_object_id = forms.CharField( widget=forms.TextInput( attrs={ 'placeholder': 'Linked Object ID', 'class': 'input-medium', } ), max_length=20, required=False, ) Fix formatting for the EventSearchForm class.
from django import forms import settings OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES class EventSearchForm(forms.Form): event_outcome = forms.ChoiceField( widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-small'}), choices=OUTCOME_CHOICES, required=False) event_type = forms.ChoiceField( widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-medium'}), choices=EVENT_TYPE_CHOICES, required=False) start_date = forms.DateField( widget=forms.DateInput(attrs={'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small'}), # noqa required=False) end_date = forms.DateField( widget=forms.DateInput(attrs={'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small'}), # noqa required=False) linked_object_id = forms.CharField( widget=forms.TextInput(attrs={'placeholder': 'Linked Object ID', 'class': 'input-medium'}), max_length=20, required=False)
<commit_before>from django import forms import settings OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES class EventSearchForm(forms.Form): event_outcome = forms.ChoiceField( widget=forms.Select( attrs={ 'id': 'prependedInput', 'class': 'input-small', } ), choices=OUTCOME_CHOICES, required=False) event_type = forms.ChoiceField( widget=forms.Select( attrs={ 'id': 'prependedInput', 'class': 'input-medium', } ), choices=EVENT_TYPE_CHOICES, required=False) start_date = forms.DateField( widget=forms.DateInput( attrs={ 'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small', } ), required=False) end_date = forms.DateField( widget=forms.DateInput( attrs={ 'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small', } ), required=False) linked_object_id = forms.CharField( widget=forms.TextInput( attrs={ 'placeholder': 'Linked Object ID', 'class': 'input-medium', } ), max_length=20, required=False, ) <commit_msg>Fix formatting for the EventSearchForm class.<commit_after>
from django import forms import settings OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES class EventSearchForm(forms.Form): event_outcome = forms.ChoiceField( widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-small'}), choices=OUTCOME_CHOICES, required=False) event_type = forms.ChoiceField( widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-medium'}), choices=EVENT_TYPE_CHOICES, required=False) start_date = forms.DateField( widget=forms.DateInput(attrs={'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small'}), # noqa required=False) end_date = forms.DateField( widget=forms.DateInput(attrs={'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small'}), # noqa required=False) linked_object_id = forms.CharField( widget=forms.TextInput(attrs={'placeholder': 'Linked Object ID', 'class': 'input-medium'}), max_length=20, required=False)
from django import forms import settings OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES class EventSearchForm(forms.Form): event_outcome = forms.ChoiceField( widget=forms.Select( attrs={ 'id': 'prependedInput', 'class': 'input-small', } ), choices=OUTCOME_CHOICES, required=False) event_type = forms.ChoiceField( widget=forms.Select( attrs={ 'id': 'prependedInput', 'class': 'input-medium', } ), choices=EVENT_TYPE_CHOICES, required=False) start_date = forms.DateField( widget=forms.DateInput( attrs={ 'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small', } ), required=False) end_date = forms.DateField( widget=forms.DateInput( attrs={ 'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small', } ), required=False) linked_object_id = forms.CharField( widget=forms.TextInput( attrs={ 'placeholder': 'Linked Object ID', 'class': 'input-medium', } ), max_length=20, required=False, ) Fix formatting for the EventSearchForm class.from django import forms import settings OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES class EventSearchForm(forms.Form): event_outcome = forms.ChoiceField( widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-small'}), choices=OUTCOME_CHOICES, required=False) event_type = forms.ChoiceField( widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-medium'}), choices=EVENT_TYPE_CHOICES, required=False) start_date = forms.DateField( widget=forms.DateInput(attrs={'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small'}), # noqa required=False) end_date = forms.DateField( widget=forms.DateInput(attrs={'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small'}), # noqa required=False) linked_object_id = forms.CharField( widget=forms.TextInput(attrs={'placeholder': 'Linked Object ID', 'class': 'input-medium'}), max_length=20, required=False)
<commit_before>from django import forms import settings OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES class EventSearchForm(forms.Form): event_outcome = forms.ChoiceField( widget=forms.Select( attrs={ 'id': 'prependedInput', 'class': 'input-small', } ), choices=OUTCOME_CHOICES, required=False) event_type = forms.ChoiceField( widget=forms.Select( attrs={ 'id': 'prependedInput', 'class': 'input-medium', } ), choices=EVENT_TYPE_CHOICES, required=False) start_date = forms.DateField( widget=forms.DateInput( attrs={ 'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small', } ), required=False) end_date = forms.DateField( widget=forms.DateInput( attrs={ 'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small', } ), required=False) linked_object_id = forms.CharField( widget=forms.TextInput( attrs={ 'placeholder': 'Linked Object ID', 'class': 'input-medium', } ), max_length=20, required=False, ) <commit_msg>Fix formatting for the EventSearchForm class.<commit_after>from django import forms import settings OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES class EventSearchForm(forms.Form): event_outcome = forms.ChoiceField( widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-small'}), choices=OUTCOME_CHOICES, required=False) event_type = forms.ChoiceField( widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-medium'}), choices=EVENT_TYPE_CHOICES, required=False) start_date = forms.DateField( widget=forms.DateInput(attrs={'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small'}), # noqa required=False) end_date = forms.DateField( widget=forms.DateInput(attrs={'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small'}), # noqa required=False) linked_object_id = forms.CharField( widget=forms.TextInput(attrs={'placeholder': 'Linked Object ID', 'class': 'input-medium'}), max_length=20, required=False)
585fe3142860a294a6b30edb9879f979b6ac9c49
apex/lib/db.py
apex/lib/db.py
from pyramid.httpexceptions import HTTPNotFound from pyramid.i18n import TranslationString as _ def get_or_create(session, model, **kw): """ Django's get_or_create function http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create """ obj = session.query(model).filter_by(**kw).first() if obj: return obj else: obj = model(**kw) session.add(obj) session.flush() return obj def get_object_or_404(session, model, **kw): """ Django's get_object_or_404 function """ obj = session.query(model).filter_by(**kw).first() if obj is None: raise HTTPNotFound(detail=_('No %s matches the given query.') % model.__name__) return obj def merge_session_with_post(session, post): """ Basic function to merge data into an sql object. This function doesn't work with relations. """ for key, value in post: setattr(session, key, value) return session
from pyramid.httpexceptions import HTTPNotFound from pyramid.i18n import TranslationString as _ def get_or_create(session, model, **kw): """ Django's get_or_create function http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create """ obj = session.query(model).filter_by(**kw).first() if obj: return obj else: obj = model(**kw) session.add(obj) session.flush() return obj def get_object_or_404(session, model, **kw): """ Django's get_object_or_404 function """ obj = session.query(model).filter_by(**kw).first() if obj is None: raise HTTPNotFound(detail='No %s matches the given query.' % model.__name__) return obj def merge_session_with_post(session, post): """ Basic function to merge data into an sql object. This function doesn't work with relations. """ for key, value in post: setattr(session, key, value) return session
Fix issue with wsgi and translations.
Fix issue with wsgi and translations.
Python
mit
Qwait/apex,Qwait/apex
from pyramid.httpexceptions import HTTPNotFound from pyramid.i18n import TranslationString as _ def get_or_create(session, model, **kw): """ Django's get_or_create function http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create """ obj = session.query(model).filter_by(**kw).first() if obj: return obj else: obj = model(**kw) session.add(obj) session.flush() return obj def get_object_or_404(session, model, **kw): """ Django's get_object_or_404 function """ obj = session.query(model).filter_by(**kw).first() if obj is None: raise HTTPNotFound(detail=_('No %s matches the given query.') % model.__name__) return obj def merge_session_with_post(session, post): """ Basic function to merge data into an sql object. This function doesn't work with relations. """ for key, value in post: setattr(session, key, value) return session Fix issue with wsgi and translations.
from pyramid.httpexceptions import HTTPNotFound from pyramid.i18n import TranslationString as _ def get_or_create(session, model, **kw): """ Django's get_or_create function http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create """ obj = session.query(model).filter_by(**kw).first() if obj: return obj else: obj = model(**kw) session.add(obj) session.flush() return obj def get_object_or_404(session, model, **kw): """ Django's get_object_or_404 function """ obj = session.query(model).filter_by(**kw).first() if obj is None: raise HTTPNotFound(detail='No %s matches the given query.' % model.__name__) return obj def merge_session_with_post(session, post): """ Basic function to merge data into an sql object. This function doesn't work with relations. """ for key, value in post: setattr(session, key, value) return session
<commit_before>from pyramid.httpexceptions import HTTPNotFound from pyramid.i18n import TranslationString as _ def get_or_create(session, model, **kw): """ Django's get_or_create function http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create """ obj = session.query(model).filter_by(**kw).first() if obj: return obj else: obj = model(**kw) session.add(obj) session.flush() return obj def get_object_or_404(session, model, **kw): """ Django's get_object_or_404 function """ obj = session.query(model).filter_by(**kw).first() if obj is None: raise HTTPNotFound(detail=_('No %s matches the given query.') % model.__name__) return obj def merge_session_with_post(session, post): """ Basic function to merge data into an sql object. This function doesn't work with relations. """ for key, value in post: setattr(session, key, value) return session <commit_msg>Fix issue with wsgi and translations.<commit_after>
from pyramid.httpexceptions import HTTPNotFound from pyramid.i18n import TranslationString as _ def get_or_create(session, model, **kw): """ Django's get_or_create function http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create """ obj = session.query(model).filter_by(**kw).first() if obj: return obj else: obj = model(**kw) session.add(obj) session.flush() return obj def get_object_or_404(session, model, **kw): """ Django's get_object_or_404 function """ obj = session.query(model).filter_by(**kw).first() if obj is None: raise HTTPNotFound(detail='No %s matches the given query.' % model.__name__) return obj def merge_session_with_post(session, post): """ Basic function to merge data into an sql object. This function doesn't work with relations. """ for key, value in post: setattr(session, key, value) return session
from pyramid.httpexceptions import HTTPNotFound from pyramid.i18n import TranslationString as _ def get_or_create(session, model, **kw): """ Django's get_or_create function http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create """ obj = session.query(model).filter_by(**kw).first() if obj: return obj else: obj = model(**kw) session.add(obj) session.flush() return obj def get_object_or_404(session, model, **kw): """ Django's get_object_or_404 function """ obj = session.query(model).filter_by(**kw).first() if obj is None: raise HTTPNotFound(detail=_('No %s matches the given query.') % model.__name__) return obj def merge_session_with_post(session, post): """ Basic function to merge data into an sql object. This function doesn't work with relations. """ for key, value in post: setattr(session, key, value) return session Fix issue with wsgi and translations.from pyramid.httpexceptions import HTTPNotFound from pyramid.i18n import TranslationString as _ def get_or_create(session, model, **kw): """ Django's get_or_create function http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create """ obj = session.query(model).filter_by(**kw).first() if obj: return obj else: obj = model(**kw) session.add(obj) session.flush() return obj def get_object_or_404(session, model, **kw): """ Django's get_object_or_404 function """ obj = session.query(model).filter_by(**kw).first() if obj is None: raise HTTPNotFound(detail='No %s matches the given query.' % model.__name__) return obj def merge_session_with_post(session, post): """ Basic function to merge data into an sql object. This function doesn't work with relations. """ for key, value in post: setattr(session, key, value) return session
<commit_before>from pyramid.httpexceptions import HTTPNotFound from pyramid.i18n import TranslationString as _ def get_or_create(session, model, **kw): """ Django's get_or_create function http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create """ obj = session.query(model).filter_by(**kw).first() if obj: return obj else: obj = model(**kw) session.add(obj) session.flush() return obj def get_object_or_404(session, model, **kw): """ Django's get_object_or_404 function """ obj = session.query(model).filter_by(**kw).first() if obj is None: raise HTTPNotFound(detail=_('No %s matches the given query.') % model.__name__) return obj def merge_session_with_post(session, post): """ Basic function to merge data into an sql object. This function doesn't work with relations. """ for key, value in post: setattr(session, key, value) return session <commit_msg>Fix issue with wsgi and translations.<commit_after>from pyramid.httpexceptions import HTTPNotFound from pyramid.i18n import TranslationString as _ def get_or_create(session, model, **kw): """ Django's get_or_create function http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create """ obj = session.query(model).filter_by(**kw).first() if obj: return obj else: obj = model(**kw) session.add(obj) session.flush() return obj def get_object_or_404(session, model, **kw): """ Django's get_object_or_404 function """ obj = session.query(model).filter_by(**kw).first() if obj is None: raise HTTPNotFound(detail='No %s matches the given query.' % model.__name__) return obj def merge_session_with_post(session, post): """ Basic function to merge data into an sql object. This function doesn't work with relations. """ for key, value in post: setattr(session, key, value) return session
2937af0fe2f28ed9381b6b43c337c4cca14e4e78
apps/polls/admin.py
apps/polls/admin.py
from django.contrib import admin from apps.polls.models import Poll, Choice class ChoiceInline(admin.TabularInline): model = Choice extra = 3 class PollAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['question']}), ('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}), ] inlines = [ChoiceInline] list_display = ('question', 'pub_date', 'was_published_recently') list_filter = ['pub_date'] admin.site.register(Poll, PollAdmin) admin.site.register(Choice)
from django.contrib import admin from apps.polls.models import Poll, Choice class ChoiceInline(admin.TabularInline): model = Choice extra = 3 class PollAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['question']}), ('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}), ] inlines = [ChoiceInline] list_display = ('question', 'pub_date', 'was_published_recently') list_filter = ['pub_date'] search_fields = ['question'] admin.site.register(Poll, PollAdmin) admin.site.register(Choice)
Add search_fields = ['question'] to PollAdmin
Add search_fields = ['question'] to PollAdmin
Python
bsd-3-clause
datphan/teracy-tutorial,teracyhq/django-tutorial
from django.contrib import admin from apps.polls.models import Poll, Choice class ChoiceInline(admin.TabularInline): model = Choice extra = 3 class PollAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['question']}), ('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}), ] inlines = [ChoiceInline] list_display = ('question', 'pub_date', 'was_published_recently') list_filter = ['pub_date'] admin.site.register(Poll, PollAdmin) admin.site.register(Choice)Add search_fields = ['question'] to PollAdmin
from django.contrib import admin from apps.polls.models import Poll, Choice class ChoiceInline(admin.TabularInline): model = Choice extra = 3 class PollAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['question']}), ('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}), ] inlines = [ChoiceInline] list_display = ('question', 'pub_date', 'was_published_recently') list_filter = ['pub_date'] search_fields = ['question'] admin.site.register(Poll, PollAdmin) admin.site.register(Choice)
<commit_before>from django.contrib import admin from apps.polls.models import Poll, Choice class ChoiceInline(admin.TabularInline): model = Choice extra = 3 class PollAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['question']}), ('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}), ] inlines = [ChoiceInline] list_display = ('question', 'pub_date', 'was_published_recently') list_filter = ['pub_date'] admin.site.register(Poll, PollAdmin) admin.site.register(Choice)<commit_msg>Add search_fields = ['question'] to PollAdmin<commit_after>
from django.contrib import admin from apps.polls.models import Poll, Choice class ChoiceInline(admin.TabularInline): model = Choice extra = 3 class PollAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['question']}), ('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}), ] inlines = [ChoiceInline] list_display = ('question', 'pub_date', 'was_published_recently') list_filter = ['pub_date'] search_fields = ['question'] admin.site.register(Poll, PollAdmin) admin.site.register(Choice)
from django.contrib import admin from apps.polls.models import Poll, Choice class ChoiceInline(admin.TabularInline): model = Choice extra = 3 class PollAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['question']}), ('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}), ] inlines = [ChoiceInline] list_display = ('question', 'pub_date', 'was_published_recently') list_filter = ['pub_date'] admin.site.register(Poll, PollAdmin) admin.site.register(Choice)Add search_fields = ['question'] to PollAdminfrom django.contrib import admin from apps.polls.models import Poll, Choice class ChoiceInline(admin.TabularInline): model = Choice extra = 3 class PollAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['question']}), ('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}), ] inlines = [ChoiceInline] list_display = ('question', 'pub_date', 'was_published_recently') list_filter = ['pub_date'] search_fields = ['question'] admin.site.register(Poll, PollAdmin) admin.site.register(Choice)
<commit_before>from django.contrib import admin from apps.polls.models import Poll, Choice class ChoiceInline(admin.TabularInline): model = Choice extra = 3 class PollAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['question']}), ('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}), ] inlines = [ChoiceInline] list_display = ('question', 'pub_date', 'was_published_recently') list_filter = ['pub_date'] admin.site.register(Poll, PollAdmin) admin.site.register(Choice)<commit_msg>Add search_fields = ['question'] to PollAdmin<commit_after>from django.contrib import admin from apps.polls.models import Poll, Choice class ChoiceInline(admin.TabularInline): model = Choice extra = 3 class PollAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['question']}), ('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}), ] inlines = [ChoiceInline] list_display = ('question', 'pub_date', 'was_published_recently') list_filter = ['pub_date'] search_fields = ['question'] admin.site.register(Poll, PollAdmin) admin.site.register(Choice)
f50589ec9b61efbd2bd56cca802ffc542f5b3336
pyrene/constants.py
pyrene/constants.py
class REPO: '''Repo attributes''' TYPE = 'type' DIRECTORY = 'directory' VOLATILE = 'volatile' SERVE_INTERFACE = 'serve_interface' SERVE_PORT = 'serve_port' SERVE_USERNAME = 'serve_username' SERVE_PASSWORD = 'serve_password' USERNAME = 'username' PASSWORD = 'password' DOWNLOAD_URL = 'download_url' UPLOAD_URL = 'upload_url' class REPOTYPE: '''Values for REPO.TYPE''' DIRECTORY = 'directory' HTTP = 'http'
class REPO: '''Repo attributes''' TYPE = 'type' DIRECTORY = 'directory' VOLATILE = 'volatile' SERVE_INTERFACE = 'interface' SERVE_PORT = 'port' SERVE_USERNAME = 'username' SERVE_PASSWORD = 'password' USERNAME = 'username' PASSWORD = 'password' DOWNLOAD_URL = 'download_url' UPLOAD_URL = 'upload_url' class REPOTYPE: '''Values for REPO.TYPE''' DIRECTORY = 'directory' HTTP = 'http'
Revert "make REPO.SERVE_* attributes distinct from other attributes (username, password)"
Revert "make REPO.SERVE_* attributes distinct from other attributes (username, password)" This reverts commit 1553f4bae5f315666fac5ad9f6600ba8b076a84b.
Python
mit
krisztianfekete/pyrene
class REPO: '''Repo attributes''' TYPE = 'type' DIRECTORY = 'directory' VOLATILE = 'volatile' SERVE_INTERFACE = 'serve_interface' SERVE_PORT = 'serve_port' SERVE_USERNAME = 'serve_username' SERVE_PASSWORD = 'serve_password' USERNAME = 'username' PASSWORD = 'password' DOWNLOAD_URL = 'download_url' UPLOAD_URL = 'upload_url' class REPOTYPE: '''Values for REPO.TYPE''' DIRECTORY = 'directory' HTTP = 'http' Revert "make REPO.SERVE_* attributes distinct from other attributes (username, password)" This reverts commit 1553f4bae5f315666fac5ad9f6600ba8b076a84b.
class REPO: '''Repo attributes''' TYPE = 'type' DIRECTORY = 'directory' VOLATILE = 'volatile' SERVE_INTERFACE = 'interface' SERVE_PORT = 'port' SERVE_USERNAME = 'username' SERVE_PASSWORD = 'password' USERNAME = 'username' PASSWORD = 'password' DOWNLOAD_URL = 'download_url' UPLOAD_URL = 'upload_url' class REPOTYPE: '''Values for REPO.TYPE''' DIRECTORY = 'directory' HTTP = 'http'
<commit_before>class REPO: '''Repo attributes''' TYPE = 'type' DIRECTORY = 'directory' VOLATILE = 'volatile' SERVE_INTERFACE = 'serve_interface' SERVE_PORT = 'serve_port' SERVE_USERNAME = 'serve_username' SERVE_PASSWORD = 'serve_password' USERNAME = 'username' PASSWORD = 'password' DOWNLOAD_URL = 'download_url' UPLOAD_URL = 'upload_url' class REPOTYPE: '''Values for REPO.TYPE''' DIRECTORY = 'directory' HTTP = 'http' <commit_msg>Revert "make REPO.SERVE_* attributes distinct from other attributes (username, password)" This reverts commit 1553f4bae5f315666fac5ad9f6600ba8b076a84b.<commit_after>
class REPO: '''Repo attributes''' TYPE = 'type' DIRECTORY = 'directory' VOLATILE = 'volatile' SERVE_INTERFACE = 'interface' SERVE_PORT = 'port' SERVE_USERNAME = 'username' SERVE_PASSWORD = 'password' USERNAME = 'username' PASSWORD = 'password' DOWNLOAD_URL = 'download_url' UPLOAD_URL = 'upload_url' class REPOTYPE: '''Values for REPO.TYPE''' DIRECTORY = 'directory' HTTP = 'http'
class REPO: '''Repo attributes''' TYPE = 'type' DIRECTORY = 'directory' VOLATILE = 'volatile' SERVE_INTERFACE = 'serve_interface' SERVE_PORT = 'serve_port' SERVE_USERNAME = 'serve_username' SERVE_PASSWORD = 'serve_password' USERNAME = 'username' PASSWORD = 'password' DOWNLOAD_URL = 'download_url' UPLOAD_URL = 'upload_url' class REPOTYPE: '''Values for REPO.TYPE''' DIRECTORY = 'directory' HTTP = 'http' Revert "make REPO.SERVE_* attributes distinct from other attributes (username, password)" This reverts commit 1553f4bae5f315666fac5ad9f6600ba8b076a84b.class REPO: '''Repo attributes''' TYPE = 'type' DIRECTORY = 'directory' VOLATILE = 'volatile' SERVE_INTERFACE = 'interface' SERVE_PORT = 'port' SERVE_USERNAME = 'username' SERVE_PASSWORD = 'password' USERNAME = 'username' PASSWORD = 'password' DOWNLOAD_URL = 'download_url' UPLOAD_URL = 'upload_url' class REPOTYPE: '''Values for REPO.TYPE''' DIRECTORY = 'directory' HTTP = 'http'
<commit_before>class REPO: '''Repo attributes''' TYPE = 'type' DIRECTORY = 'directory' VOLATILE = 'volatile' SERVE_INTERFACE = 'serve_interface' SERVE_PORT = 'serve_port' SERVE_USERNAME = 'serve_username' SERVE_PASSWORD = 'serve_password' USERNAME = 'username' PASSWORD = 'password' DOWNLOAD_URL = 'download_url' UPLOAD_URL = 'upload_url' class REPOTYPE: '''Values for REPO.TYPE''' DIRECTORY = 'directory' HTTP = 'http' <commit_msg>Revert "make REPO.SERVE_* attributes distinct from other attributes (username, password)" This reverts commit 1553f4bae5f315666fac5ad9f6600ba8b076a84b.<commit_after>class REPO: '''Repo attributes''' TYPE = 'type' DIRECTORY = 'directory' VOLATILE = 'volatile' SERVE_INTERFACE = 'interface' SERVE_PORT = 'port' SERVE_USERNAME = 'username' SERVE_PASSWORD = 'password' USERNAME = 'username' PASSWORD = 'password' DOWNLOAD_URL = 'download_url' UPLOAD_URL = 'upload_url' class REPOTYPE: '''Values for REPO.TYPE''' DIRECTORY = 'directory' HTTP = 'http'
2973b664e8c9cf551d5d7277ab4995125be5fad0
python/reference.py
python/reference.py
import os # Current directory # If you call this from the current directory without abspath, # then it will not work since __file__ is a relative path os.path.dirname(os.path.abspath(__file__))
import os # Current directory # If you call this from the current directory without abspath, # then it will not work since __file__ is a relative path os.path.dirname(os.path.abspath(__file__)) # Get all files in a directory # Never use os.walk again def all_sub_files(root): for path, subdirs, files in os.walk(root): for name in files: yield os.path.join(path, name)
Add util for not using os.walk
Add util for not using os.walk
Python
mit
brycepg/how-to
import os # Current directory # If you call this from the current directory without abspath, # then it will not work since __file__ is a relative path os.path.dirname(os.path.abspath(__file__)) Add util for not using os.walk
import os # Current directory # If you call this from the current directory without abspath, # then it will not work since __file__ is a relative path os.path.dirname(os.path.abspath(__file__)) # Get all files in a directory # Never use os.walk again def all_sub_files(root): for path, subdirs, files in os.walk(root): for name in files: yield os.path.join(path, name)
<commit_before>import os # Current directory # If you call this from the current directory without abspath, # then it will not work since __file__ is a relative path os.path.dirname(os.path.abspath(__file__)) <commit_msg>Add util for not using os.walk<commit_after>
import os # Current directory # If you call this from the current directory without abspath, # then it will not work since __file__ is a relative path os.path.dirname(os.path.abspath(__file__)) # Get all files in a directory # Never use os.walk again def all_sub_files(root): for path, subdirs, files in os.walk(root): for name in files: yield os.path.join(path, name)
import os # Current directory # If you call this from the current directory without abspath, # then it will not work since __file__ is a relative path os.path.dirname(os.path.abspath(__file__)) Add util for not using os.walkimport os # Current directory # If you call this from the current directory without abspath, # then it will not work since __file__ is a relative path os.path.dirname(os.path.abspath(__file__)) # Get all files in a directory # Never use os.walk again def all_sub_files(root): for path, subdirs, files in os.walk(root): for name in files: yield os.path.join(path, name)
<commit_before>import os # Current directory # If you call this from the current directory without abspath, # then it will not work since __file__ is a relative path os.path.dirname(os.path.abspath(__file__)) <commit_msg>Add util for not using os.walk<commit_after>import os # Current directory # If you call this from the current directory without abspath, # then it will not work since __file__ is a relative path os.path.dirname(os.path.abspath(__file__)) # Get all files in a directory # Never use os.walk again def all_sub_files(root): for path, subdirs, files in os.walk(root): for name in files: yield os.path.join(path, name)
38dfe0948b063b93f5375e065129838335d5ca95
raiden/constants.py
raiden/constants.py
# -*- coding: utf-8 -*- UINT64_MAX = 2 ** 64 - 1 UINT64_MIN = 0 INT64_MAX = 2 ** 63 - 1 INT64_MIN = -(2 ** 63) UINT256_MAX = 2 ** 256 - 1 # Deployed to Ropsten revival on 2017-06-19 from commit 2677298a69c1b1f35b9ab26beafe457acfdcc0ee ROPSTEN_REGISTRY_ADDRESS = 'aff1f958c69a6820b08a02549ff9041629ae8257' ROPSTEN_DISCOVERY_ADDRESS = 'cf56165f4f6e8ec38bb463854c1fe28a5d320f4f' DISCOVERY_REGISTRATION_GAS = 500000 MINUTE_SEC = 60 MINUTE_MS = 60 * 1000 NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6 # TODO: add this as an attribute of the transport class UDP_MAX_MESSAGE_SIZE = 1200
# -*- coding: utf-8 -*- UINT64_MAX = 2 ** 64 - 1 UINT64_MIN = 0 INT64_MAX = 2 ** 63 - 1 INT64_MIN = -(2 ** 63) UINT256_MAX = 2 ** 256 - 1 # Deployed to Ropsten revival on 2017-08-03 from commit 17aa7671159779ceef22fe90001970bed0685c4d ROPSTEN_REGISTRY_ADDRESS = '25926b6d29f56ba8466601d7ce7dd29985af1f14' ROPSTEN_DISCOVERY_ADDRESS = 'a647336e0b4eaddcee11fe6c13e9163ad2700d6d' DISCOVERY_REGISTRATION_GAS = 500000 MINUTE_SEC = 60 MINUTE_MS = 60 * 1000 NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6 # TODO: add this as an attribute of the transport class UDP_MAX_MESSAGE_SIZE = 1200
Update default smart contract addresses
Update default smart contract addresses
Python
mit
tomashaber/raiden,tomashaber/raiden,hackaugusto/raiden,tomashaber/raiden,hackaugusto/raiden,tomashaber/raiden,tomashaber/raiden
# -*- coding: utf-8 -*- UINT64_MAX = 2 ** 64 - 1 UINT64_MIN = 0 INT64_MAX = 2 ** 63 - 1 INT64_MIN = -(2 ** 63) UINT256_MAX = 2 ** 256 - 1 # Deployed to Ropsten revival on 2017-06-19 from commit 2677298a69c1b1f35b9ab26beafe457acfdcc0ee ROPSTEN_REGISTRY_ADDRESS = 'aff1f958c69a6820b08a02549ff9041629ae8257' ROPSTEN_DISCOVERY_ADDRESS = 'cf56165f4f6e8ec38bb463854c1fe28a5d320f4f' DISCOVERY_REGISTRATION_GAS = 500000 MINUTE_SEC = 60 MINUTE_MS = 60 * 1000 NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6 # TODO: add this as an attribute of the transport class UDP_MAX_MESSAGE_SIZE = 1200 Update default smart contract addresses
# -*- coding: utf-8 -*- UINT64_MAX = 2 ** 64 - 1 UINT64_MIN = 0 INT64_MAX = 2 ** 63 - 1 INT64_MIN = -(2 ** 63) UINT256_MAX = 2 ** 256 - 1 # Deployed to Ropsten revival on 2017-08-03 from commit 17aa7671159779ceef22fe90001970bed0685c4d ROPSTEN_REGISTRY_ADDRESS = '25926b6d29f56ba8466601d7ce7dd29985af1f14' ROPSTEN_DISCOVERY_ADDRESS = 'a647336e0b4eaddcee11fe6c13e9163ad2700d6d' DISCOVERY_REGISTRATION_GAS = 500000 MINUTE_SEC = 60 MINUTE_MS = 60 * 1000 NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6 # TODO: add this as an attribute of the transport class UDP_MAX_MESSAGE_SIZE = 1200
<commit_before># -*- coding: utf-8 -*- UINT64_MAX = 2 ** 64 - 1 UINT64_MIN = 0 INT64_MAX = 2 ** 63 - 1 INT64_MIN = -(2 ** 63) UINT256_MAX = 2 ** 256 - 1 # Deployed to Ropsten revival on 2017-06-19 from commit 2677298a69c1b1f35b9ab26beafe457acfdcc0ee ROPSTEN_REGISTRY_ADDRESS = 'aff1f958c69a6820b08a02549ff9041629ae8257' ROPSTEN_DISCOVERY_ADDRESS = 'cf56165f4f6e8ec38bb463854c1fe28a5d320f4f' DISCOVERY_REGISTRATION_GAS = 500000 MINUTE_SEC = 60 MINUTE_MS = 60 * 1000 NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6 # TODO: add this as an attribute of the transport class UDP_MAX_MESSAGE_SIZE = 1200 <commit_msg>Update default smart contract addresses<commit_after>
# -*- coding: utf-8 -*- UINT64_MAX = 2 ** 64 - 1 UINT64_MIN = 0 INT64_MAX = 2 ** 63 - 1 INT64_MIN = -(2 ** 63) UINT256_MAX = 2 ** 256 - 1 # Deployed to Ropsten revival on 2017-08-03 from commit 17aa7671159779ceef22fe90001970bed0685c4d ROPSTEN_REGISTRY_ADDRESS = '25926b6d29f56ba8466601d7ce7dd29985af1f14' ROPSTEN_DISCOVERY_ADDRESS = 'a647336e0b4eaddcee11fe6c13e9163ad2700d6d' DISCOVERY_REGISTRATION_GAS = 500000 MINUTE_SEC = 60 MINUTE_MS = 60 * 1000 NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6 # TODO: add this as an attribute of the transport class UDP_MAX_MESSAGE_SIZE = 1200
# -*- coding: utf-8 -*- UINT64_MAX = 2 ** 64 - 1 UINT64_MIN = 0 INT64_MAX = 2 ** 63 - 1 INT64_MIN = -(2 ** 63) UINT256_MAX = 2 ** 256 - 1 # Deployed to Ropsten revival on 2017-06-19 from commit 2677298a69c1b1f35b9ab26beafe457acfdcc0ee ROPSTEN_REGISTRY_ADDRESS = 'aff1f958c69a6820b08a02549ff9041629ae8257' ROPSTEN_DISCOVERY_ADDRESS = 'cf56165f4f6e8ec38bb463854c1fe28a5d320f4f' DISCOVERY_REGISTRATION_GAS = 500000 MINUTE_SEC = 60 MINUTE_MS = 60 * 1000 NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6 # TODO: add this as an attribute of the transport class UDP_MAX_MESSAGE_SIZE = 1200 Update default smart contract addresses# -*- coding: utf-8 -*- UINT64_MAX = 2 ** 64 - 1 UINT64_MIN = 0 INT64_MAX = 2 ** 63 - 1 INT64_MIN = -(2 ** 63) UINT256_MAX = 2 ** 256 - 1 # Deployed to Ropsten revival on 2017-08-03 from commit 17aa7671159779ceef22fe90001970bed0685c4d ROPSTEN_REGISTRY_ADDRESS = '25926b6d29f56ba8466601d7ce7dd29985af1f14' ROPSTEN_DISCOVERY_ADDRESS = 'a647336e0b4eaddcee11fe6c13e9163ad2700d6d' DISCOVERY_REGISTRATION_GAS = 500000 MINUTE_SEC = 60 MINUTE_MS = 60 * 1000 NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6 # TODO: add this as an attribute of the transport class UDP_MAX_MESSAGE_SIZE = 1200
<commit_before># -*- coding: utf-8 -*- UINT64_MAX = 2 ** 64 - 1 UINT64_MIN = 0 INT64_MAX = 2 ** 63 - 1 INT64_MIN = -(2 ** 63) UINT256_MAX = 2 ** 256 - 1 # Deployed to Ropsten revival on 2017-06-19 from commit 2677298a69c1b1f35b9ab26beafe457acfdcc0ee ROPSTEN_REGISTRY_ADDRESS = 'aff1f958c69a6820b08a02549ff9041629ae8257' ROPSTEN_DISCOVERY_ADDRESS = 'cf56165f4f6e8ec38bb463854c1fe28a5d320f4f' DISCOVERY_REGISTRATION_GAS = 500000 MINUTE_SEC = 60 MINUTE_MS = 60 * 1000 NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6 # TODO: add this as an attribute of the transport class UDP_MAX_MESSAGE_SIZE = 1200 <commit_msg>Update default smart contract addresses<commit_after># -*- coding: utf-8 -*- UINT64_MAX = 2 ** 64 - 1 UINT64_MIN = 0 INT64_MAX = 2 ** 63 - 1 INT64_MIN = -(2 ** 63) UINT256_MAX = 2 ** 256 - 1 # Deployed to Ropsten revival on 2017-08-03 from commit 17aa7671159779ceef22fe90001970bed0685c4d ROPSTEN_REGISTRY_ADDRESS = '25926b6d29f56ba8466601d7ce7dd29985af1f14' ROPSTEN_DISCOVERY_ADDRESS = 'a647336e0b4eaddcee11fe6c13e9163ad2700d6d' DISCOVERY_REGISTRATION_GAS = 500000 MINUTE_SEC = 60 MINUTE_MS = 60 * 1000 NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6 # TODO: add this as an attribute of the transport class UDP_MAX_MESSAGE_SIZE = 1200
ab1893301736e2cf5cd87cc30671898884377661
test/test_events.py
test/test_events.py
import os import sys import unittest sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from smartbot import events class TestEvents(unittest.TestCase): def test_empty(self): event = events.Event() self.assertEqual(len(event.trigger()), 0) def test_with_handlers(self): event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger()), 1) def test_custom_comparator(self): comparator = lambda *args, **kwargs: False event = events.Event(default_comparator=comparator) event.register(lambda: None) self.assertEqual(len(event.trigger()), 0) event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger(comparator=comparator)), 0) def test_default_comparator(self): event = events.Event() event.register(lambda *args, **kwargs: None, a=10) self.assertEqual(len(event.trigger()), 0) self.assertEqual(len(event.trigger(a=10)), 1) def test_decorator(self): event = events.Event() event()(lambda: None) self.assertEqual(len(event.trigger()), 1)
import os import sys import unittest sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from smartbot import events class TestEvents(unittest.TestCase): def test_empty(self): event = events.Event() self.assertEqual(len(event.trigger()), 0) def test_with_handlers(self): event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger()), 1) def test_custom_comparator(self): comparator = lambda *args, **kwargs: False event = events.Event(default_comparator=comparator) event.register(lambda: None) self.assertEqual(len(event.trigger()), 0) event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger(comparator=comparator)), 0) def test_default_comparator(self): event = events.Event() event.register(lambda *args, **kwargs: None, a=10) self.assertEqual(len(event.trigger()), 1) self.assertEqual(len(event.trigger(a=10)), 1) def test_decorator(self): event = events.Event() event()(lambda: None) self.assertEqual(len(event.trigger()), 1)
Update unit tests to match correct behaviour
Update unit tests to match correct behaviour
Python
mit
Cyanogenoid/smartbot,tomleese/smartbot,Muzer/smartbot,thomasleese/smartbot-old
import os import sys import unittest sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from smartbot import events class TestEvents(unittest.TestCase): def test_empty(self): event = events.Event() self.assertEqual(len(event.trigger()), 0) def test_with_handlers(self): event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger()), 1) def test_custom_comparator(self): comparator = lambda *args, **kwargs: False event = events.Event(default_comparator=comparator) event.register(lambda: None) self.assertEqual(len(event.trigger()), 0) event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger(comparator=comparator)), 0) def test_default_comparator(self): event = events.Event() event.register(lambda *args, **kwargs: None, a=10) self.assertEqual(len(event.trigger()), 0) self.assertEqual(len(event.trigger(a=10)), 1) def test_decorator(self): event = events.Event() event()(lambda: None) self.assertEqual(len(event.trigger()), 1) Update unit tests to match correct behaviour
import os import sys import unittest sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from smartbot import events class TestEvents(unittest.TestCase): def test_empty(self): event = events.Event() self.assertEqual(len(event.trigger()), 0) def test_with_handlers(self): event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger()), 1) def test_custom_comparator(self): comparator = lambda *args, **kwargs: False event = events.Event(default_comparator=comparator) event.register(lambda: None) self.assertEqual(len(event.trigger()), 0) event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger(comparator=comparator)), 0) def test_default_comparator(self): event = events.Event() event.register(lambda *args, **kwargs: None, a=10) self.assertEqual(len(event.trigger()), 1) self.assertEqual(len(event.trigger(a=10)), 1) def test_decorator(self): event = events.Event() event()(lambda: None) self.assertEqual(len(event.trigger()), 1)
<commit_before>import os import sys import unittest sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from smartbot import events class TestEvents(unittest.TestCase): def test_empty(self): event = events.Event() self.assertEqual(len(event.trigger()), 0) def test_with_handlers(self): event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger()), 1) def test_custom_comparator(self): comparator = lambda *args, **kwargs: False event = events.Event(default_comparator=comparator) event.register(lambda: None) self.assertEqual(len(event.trigger()), 0) event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger(comparator=comparator)), 0) def test_default_comparator(self): event = events.Event() event.register(lambda *args, **kwargs: None, a=10) self.assertEqual(len(event.trigger()), 0) self.assertEqual(len(event.trigger(a=10)), 1) def test_decorator(self): event = events.Event() event()(lambda: None) self.assertEqual(len(event.trigger()), 1) <commit_msg>Update unit tests to match correct behaviour<commit_after>
import os import sys import unittest sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from smartbot import events class TestEvents(unittest.TestCase): def test_empty(self): event = events.Event() self.assertEqual(len(event.trigger()), 0) def test_with_handlers(self): event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger()), 1) def test_custom_comparator(self): comparator = lambda *args, **kwargs: False event = events.Event(default_comparator=comparator) event.register(lambda: None) self.assertEqual(len(event.trigger()), 0) event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger(comparator=comparator)), 0) def test_default_comparator(self): event = events.Event() event.register(lambda *args, **kwargs: None, a=10) self.assertEqual(len(event.trigger()), 1) self.assertEqual(len(event.trigger(a=10)), 1) def test_decorator(self): event = events.Event() event()(lambda: None) self.assertEqual(len(event.trigger()), 1)
import os import sys import unittest sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from smartbot import events class TestEvents(unittest.TestCase): def test_empty(self): event = events.Event() self.assertEqual(len(event.trigger()), 0) def test_with_handlers(self): event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger()), 1) def test_custom_comparator(self): comparator = lambda *args, **kwargs: False event = events.Event(default_comparator=comparator) event.register(lambda: None) self.assertEqual(len(event.trigger()), 0) event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger(comparator=comparator)), 0) def test_default_comparator(self): event = events.Event() event.register(lambda *args, **kwargs: None, a=10) self.assertEqual(len(event.trigger()), 0) self.assertEqual(len(event.trigger(a=10)), 1) def test_decorator(self): event = events.Event() event()(lambda: None) self.assertEqual(len(event.trigger()), 1) Update unit tests to match correct behaviourimport os import sys import unittest sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from smartbot import events class TestEvents(unittest.TestCase): def test_empty(self): event = events.Event() self.assertEqual(len(event.trigger()), 0) def test_with_handlers(self): event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger()), 1) def test_custom_comparator(self): comparator = lambda *args, **kwargs: False event = events.Event(default_comparator=comparator) event.register(lambda: None) self.assertEqual(len(event.trigger()), 0) event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger(comparator=comparator)), 0) def test_default_comparator(self): event = events.Event() event.register(lambda *args, **kwargs: None, a=10) self.assertEqual(len(event.trigger()), 1) self.assertEqual(len(event.trigger(a=10)), 1) def test_decorator(self): event = events.Event() event()(lambda: None) self.assertEqual(len(event.trigger()), 1)
<commit_before>import os import sys import unittest sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from smartbot import events class TestEvents(unittest.TestCase): def test_empty(self): event = events.Event() self.assertEqual(len(event.trigger()), 0) def test_with_handlers(self): event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger()), 1) def test_custom_comparator(self): comparator = lambda *args, **kwargs: False event = events.Event(default_comparator=comparator) event.register(lambda: None) self.assertEqual(len(event.trigger()), 0) event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger(comparator=comparator)), 0) def test_default_comparator(self): event = events.Event() event.register(lambda *args, **kwargs: None, a=10) self.assertEqual(len(event.trigger()), 0) self.assertEqual(len(event.trigger(a=10)), 1) def test_decorator(self): event = events.Event() event()(lambda: None) self.assertEqual(len(event.trigger()), 1) <commit_msg>Update unit tests to match correct behaviour<commit_after>import os import sys import unittest sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from smartbot import events class TestEvents(unittest.TestCase): def test_empty(self): event = events.Event() self.assertEqual(len(event.trigger()), 0) def test_with_handlers(self): event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger()), 1) def test_custom_comparator(self): comparator = lambda *args, **kwargs: False event = events.Event(default_comparator=comparator) event.register(lambda: None) self.assertEqual(len(event.trigger()), 0) event = events.Event() event.register(lambda: None) self.assertEqual(len(event.trigger(comparator=comparator)), 0) def test_default_comparator(self): event = events.Event() event.register(lambda *args, **kwargs: None, a=10) self.assertEqual(len(event.trigger()), 1) self.assertEqual(len(event.trigger(a=10)), 1) def test_decorator(self): event = events.Event() event()(lambda: None) self.assertEqual(len(event.trigger()), 1)
0540b78a5c83cf307c4d629bb814c8359edd8709
comrade/core/context_processors.py
comrade/core/context_processors.py
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def context_processor(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context
Rename the only context processor.
Rename the only context processor.
Python
mit
bueda/django-comrade
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def context_processor(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context Rename the only context processor.
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context
<commit_before>from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def context_processor(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context <commit_msg>Rename the only context processor.<commit_after>
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def context_processor(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context Rename the only context processor.from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context
<commit_before>from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def context_processor(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context <commit_msg>Rename the only context processor.<commit_after>from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context
b7939c13622d1134364e874da1d1903bcea6cffe
tests/test_graph.py
tests/test_graph.py
import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="graph", warningiserror=True) def test_graph(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") pretty_print_xml(tree.getroot()) # Verify that 1 graphviz node is found. assert len(tree.findall(".//graphviz")) == 1
import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="graph", warningiserror=True) def test_graph(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") # Verify that 1 graphviz node is found. assert len(tree.findall(".//graphviz")) == 1
Remove debug printing of doctree during a test
Remove debug printing of doctree during a test
Python
apache-2.0
t4ngo/sphinxcontrib-traceables
import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="graph", warningiserror=True) def test_graph(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") pretty_print_xml(tree.getroot()) # Verify that 1 graphviz node is found. assert len(tree.findall(".//graphviz")) == 1 Remove debug printing of doctree during a test
import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="graph", warningiserror=True) def test_graph(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") # Verify that 1 graphviz node is found. assert len(tree.findall(".//graphviz")) == 1
<commit_before> import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="graph", warningiserror=True) def test_graph(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") pretty_print_xml(tree.getroot()) # Verify that 1 graphviz node is found. assert len(tree.findall(".//graphviz")) == 1 <commit_msg>Remove debug printing of doctree during a test<commit_after>
import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="graph", warningiserror=True) def test_graph(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") # Verify that 1 graphviz node is found. assert len(tree.findall(".//graphviz")) == 1
import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="graph", warningiserror=True) def test_graph(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") pretty_print_xml(tree.getroot()) # Verify that 1 graphviz node is found. assert len(tree.findall(".//graphviz")) == 1 Remove debug printing of doctree during a test import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="graph", warningiserror=True) def test_graph(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") # Verify that 1 graphviz node is found. assert len(tree.findall(".//graphviz")) == 1
<commit_before> import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="graph", warningiserror=True) def test_graph(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") pretty_print_xml(tree.getroot()) # Verify that 1 graphviz node is found. assert len(tree.findall(".//graphviz")) == 1 <commit_msg>Remove debug printing of doctree during a test<commit_after> import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="graph", warningiserror=True) def test_graph(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") # Verify that 1 graphviz node is found. assert len(tree.findall(".//graphviz")) == 1
942bd1abf9d4e74b0f1dbeb8fc5acaaaeed1a2a6
tests/test_nexus.py
tests/test_nexus.py
import os, sys sys.path.append(os.path.abspath('..')) import unittest from mock import patch from pynexus import api_client class NexusTest(unittest.TestCase): def test_constructor_appends_base(self): n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') self.assertEquals(n.uri, 'http://test.com/nexus/service/local/') @patch.object(api_client.requests, 'get') def test_get_users_return_list_with_just_anonymous_user(self, mock_get): mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \ 'service/local/users/anonymous","userId":"anonymous",' \ '"firstName":"Nexus","lastName":"Anonymous User",' \ '"status":"active","email":"changeme2@yourcompany.com"' \ ',"roles":["anonymous","repository-any-read"]}' mock_get.return_value = mock_output n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') result = n.get_users() self.assertEqual(result, mock_output) def main(): unittest.main() if __name__ == '__main__': main()
import os, sys sys.path.append(os.path.abspath('..')) import unittest from mock import patch from pynexus import api_client class NexusTest(unittest.TestCase): def test_constructor_appends_base(self): n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') self.assertEquals(n.uri, 'http://test.com/nexus/service/local') @patch.object(api_client.requests, 'get') def test_get_users_return_list_with_just_anonymous_user(self, mock_get): mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \ 'service/local/users/anonymous","userId":"anonymous",' \ '"firstName":"Nexus","lastName":"Anonymous User",' \ '"status":"active","email":"changeme2@yourcompany.com"' \ ',"roles":["anonymous","repository-any-read"]}' mock_get.return_value = mock_output n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') result = n.get_users() self.assertEqual(result, mock_output) def main(): unittest.main() if __name__ == '__main__': main()
Fix constructor test, since constructor does not append / at the end
Fix constructor test, since constructor does not append / at the end
Python
apache-2.0
rcarrillocruz/pynexus
import os, sys sys.path.append(os.path.abspath('..')) import unittest from mock import patch from pynexus import api_client class NexusTest(unittest.TestCase): def test_constructor_appends_base(self): n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') self.assertEquals(n.uri, 'http://test.com/nexus/service/local/') @patch.object(api_client.requests, 'get') def test_get_users_return_list_with_just_anonymous_user(self, mock_get): mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \ 'service/local/users/anonymous","userId":"anonymous",' \ '"firstName":"Nexus","lastName":"Anonymous User",' \ '"status":"active","email":"changeme2@yourcompany.com"' \ ',"roles":["anonymous","repository-any-read"]}' mock_get.return_value = mock_output n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') result = n.get_users() self.assertEqual(result, mock_output) def main(): unittest.main() if __name__ == '__main__': main() Fix constructor test, since constructor does not append / at the end
import os, sys sys.path.append(os.path.abspath('..')) import unittest from mock import patch from pynexus import api_client class NexusTest(unittest.TestCase): def test_constructor_appends_base(self): n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') self.assertEquals(n.uri, 'http://test.com/nexus/service/local') @patch.object(api_client.requests, 'get') def test_get_users_return_list_with_just_anonymous_user(self, mock_get): mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \ 'service/local/users/anonymous","userId":"anonymous",' \ '"firstName":"Nexus","lastName":"Anonymous User",' \ '"status":"active","email":"changeme2@yourcompany.com"' \ ',"roles":["anonymous","repository-any-read"]}' mock_get.return_value = mock_output n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') result = n.get_users() self.assertEqual(result, mock_output) def main(): unittest.main() if __name__ == '__main__': main()
<commit_before>import os, sys sys.path.append(os.path.abspath('..')) import unittest from mock import patch from pynexus import api_client class NexusTest(unittest.TestCase): def test_constructor_appends_base(self): n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') self.assertEquals(n.uri, 'http://test.com/nexus/service/local/') @patch.object(api_client.requests, 'get') def test_get_users_return_list_with_just_anonymous_user(self, mock_get): mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \ 'service/local/users/anonymous","userId":"anonymous",' \ '"firstName":"Nexus","lastName":"Anonymous User",' \ '"status":"active","email":"changeme2@yourcompany.com"' \ ',"roles":["anonymous","repository-any-read"]}' mock_get.return_value = mock_output n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') result = n.get_users() self.assertEqual(result, mock_output) def main(): unittest.main() if __name__ == '__main__': main() <commit_msg>Fix constructor test, since constructor does not append / at the end<commit_after>
import os, sys sys.path.append(os.path.abspath('..')) import unittest from mock import patch from pynexus import api_client class NexusTest(unittest.TestCase): def test_constructor_appends_base(self): n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') self.assertEquals(n.uri, 'http://test.com/nexus/service/local') @patch.object(api_client.requests, 'get') def test_get_users_return_list_with_just_anonymous_user(self, mock_get): mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \ 'service/local/users/anonymous","userId":"anonymous",' \ '"firstName":"Nexus","lastName":"Anonymous User",' \ '"status":"active","email":"changeme2@yourcompany.com"' \ ',"roles":["anonymous","repository-any-read"]}' mock_get.return_value = mock_output n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') result = n.get_users() self.assertEqual(result, mock_output) def main(): unittest.main() if __name__ == '__main__': main()
import os, sys sys.path.append(os.path.abspath('..')) import unittest from mock import patch from pynexus import api_client class NexusTest(unittest.TestCase): def test_constructor_appends_base(self): n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') self.assertEquals(n.uri, 'http://test.com/nexus/service/local/') @patch.object(api_client.requests, 'get') def test_get_users_return_list_with_just_anonymous_user(self, mock_get): mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \ 'service/local/users/anonymous","userId":"anonymous",' \ '"firstName":"Nexus","lastName":"Anonymous User",' \ '"status":"active","email":"changeme2@yourcompany.com"' \ ',"roles":["anonymous","repository-any-read"]}' mock_get.return_value = mock_output n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') result = n.get_users() self.assertEqual(result, mock_output) def main(): unittest.main() if __name__ == '__main__': main() Fix constructor test, since constructor does not append / at the endimport os, sys sys.path.append(os.path.abspath('..')) import unittest from mock import patch from pynexus import api_client class NexusTest(unittest.TestCase): def test_constructor_appends_base(self): n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') self.assertEquals(n.uri, 'http://test.com/nexus/service/local') @patch.object(api_client.requests, 'get') def test_get_users_return_list_with_just_anonymous_user(self, mock_get): mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \ 'service/local/users/anonymous","userId":"anonymous",' \ '"firstName":"Nexus","lastName":"Anonymous User",' \ '"status":"active","email":"changeme2@yourcompany.com"' \ ',"roles":["anonymous","repository-any-read"]}' mock_get.return_value = mock_output n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') result = n.get_users() self.assertEqual(result, mock_output) def main(): unittest.main() if __name__ == '__main__': main()
<commit_before>import os, sys sys.path.append(os.path.abspath('..')) import unittest from mock import patch from pynexus import api_client class NexusTest(unittest.TestCase): def test_constructor_appends_base(self): n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') self.assertEquals(n.uri, 'http://test.com/nexus/service/local/') @patch.object(api_client.requests, 'get') def test_get_users_return_list_with_just_anonymous_user(self, mock_get): mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \ 'service/local/users/anonymous","userId":"anonymous",' \ '"firstName":"Nexus","lastName":"Anonymous User",' \ '"status":"active","email":"changeme2@yourcompany.com"' \ ',"roles":["anonymous","repository-any-read"]}' mock_get.return_value = mock_output n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') result = n.get_users() self.assertEqual(result, mock_output) def main(): unittest.main() if __name__ == '__main__': main() <commit_msg>Fix constructor test, since constructor does not append / at the end<commit_after>import os, sys sys.path.append(os.path.abspath('..')) import unittest from mock import patch from pynexus import api_client class NexusTest(unittest.TestCase): def test_constructor_appends_base(self): n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') self.assertEquals(n.uri, 'http://test.com/nexus/service/local') @patch.object(api_client.requests, 'get') def test_get_users_return_list_with_just_anonymous_user(self, mock_get): mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \ 'service/local/users/anonymous","userId":"anonymous",' \ '"firstName":"Nexus","lastName":"Anonymous User",' \ '"status":"active","email":"changeme2@yourcompany.com"' \ ',"roles":["anonymous","repository-any-read"]}' mock_get.return_value = mock_output n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd') result = n.get_users() self.assertEqual(result, mock_output) def main(): unittest.main() if __name__ == '__main__': main()
c76e3accff36eb993ea44f4e38adad9466af1f54
tests/wsgi_tests.py
tests/wsgi_tests.py
import mock import unittest from resto.wsgi import Middleware class AppTestCase(unittest.TestCase): def setUp(self): self.environ = { 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': True, } self.mock_start = mock.Mock() def test_application_init(self): result = Middleware(self.environ, self.mock_start) content = '' for data in result: content += data self.assertGreater(content, '') if __name__ == '__main__': unittest.main()
import mock import unittest from resto.wsgi import Middleware class AppTestCase(unittest.TestCase): def setUp(self): self.environ = { 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': True, } self.mock_start = mock.Mock() def test_application_init(self): result = Middleware(self.environ, self.mock_start) content = '' for data in result: content += data self.assertGreater(content, '') self.assertTrue(self.mock_start.called) if __name__ == '__main__': unittest.main()
Test WSGI start_response is called
Test WSGI start_response is called
Python
mit
rafaelpivato/resto
import mock import unittest from resto.wsgi import Middleware class AppTestCase(unittest.TestCase): def setUp(self): self.environ = { 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': True, } self.mock_start = mock.Mock() def test_application_init(self): result = Middleware(self.environ, self.mock_start) content = '' for data in result: content += data self.assertGreater(content, '') if __name__ == '__main__': unittest.main() Test WSGI start_response is called
import mock import unittest from resto.wsgi import Middleware class AppTestCase(unittest.TestCase): def setUp(self): self.environ = { 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': True, } self.mock_start = mock.Mock() def test_application_init(self): result = Middleware(self.environ, self.mock_start) content = '' for data in result: content += data self.assertGreater(content, '') self.assertTrue(self.mock_start.called) if __name__ == '__main__': unittest.main()
<commit_before>import mock import unittest from resto.wsgi import Middleware class AppTestCase(unittest.TestCase): def setUp(self): self.environ = { 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': True, } self.mock_start = mock.Mock() def test_application_init(self): result = Middleware(self.environ, self.mock_start) content = '' for data in result: content += data self.assertGreater(content, '') if __name__ == '__main__': unittest.main() <commit_msg>Test WSGI start_response is called<commit_after>
import mock import unittest from resto.wsgi import Middleware class AppTestCase(unittest.TestCase): def setUp(self): self.environ = { 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': True, } self.mock_start = mock.Mock() def test_application_init(self): result = Middleware(self.environ, self.mock_start) content = '' for data in result: content += data self.assertGreater(content, '') self.assertTrue(self.mock_start.called) if __name__ == '__main__': unittest.main()
import mock import unittest from resto.wsgi import Middleware class AppTestCase(unittest.TestCase): def setUp(self): self.environ = { 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': True, } self.mock_start = mock.Mock() def test_application_init(self): result = Middleware(self.environ, self.mock_start) content = '' for data in result: content += data self.assertGreater(content, '') if __name__ == '__main__': unittest.main() Test WSGI start_response is calledimport mock import unittest from resto.wsgi import Middleware class AppTestCase(unittest.TestCase): def setUp(self): self.environ = { 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': True, } self.mock_start = mock.Mock() def test_application_init(self): result = Middleware(self.environ, self.mock_start) content = '' for data in result: content += data self.assertGreater(content, '') self.assertTrue(self.mock_start.called) if __name__ == '__main__': unittest.main()
<commit_before>import mock import unittest from resto.wsgi import Middleware class AppTestCase(unittest.TestCase): def setUp(self): self.environ = { 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': True, } self.mock_start = mock.Mock() def test_application_init(self): result = Middleware(self.environ, self.mock_start) content = '' for data in result: content += data self.assertGreater(content, '') if __name__ == '__main__': unittest.main() <commit_msg>Test WSGI start_response is called<commit_after>import mock import unittest from resto.wsgi import Middleware class AppTestCase(unittest.TestCase): def setUp(self): self.environ = { 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': True, } self.mock_start = mock.Mock() def test_application_init(self): result = Middleware(self.environ, self.mock_start) content = '' for data in result: content += data self.assertGreater(content, '') self.assertTrue(self.mock_start.called) if __name__ == '__main__': unittest.main()
ff662e931cf4d57016c97698d490bc75368ecb67
pmxbot/testing/fixtures.py
pmxbot/testing/fixtures.py
import pytest @pytest.fixture(scope='session', autouse=True) def init_config(): __import__('pmxbot').config = {} @pytest.fixture(params=['mongodb', 'sqlite']) def db_uri(request): if request.param == 'mongodb': return request.getfuncargvalue('mongodb_uri') return 'sqlite:pmxbot.sqlite'
import pytest @pytest.fixture(scope='session', autouse=True) def init_config(): __import__('pmxbot').config = {} @pytest.fixture(params=['mongodb', 'sqlite']) def db_uri(request): if request.param == 'mongodb': return request.getfixturevalue('mongodb_uri') return 'sqlite:pmxbot.sqlite'
Fix deprecation warning in fixture retrieval.
Fix deprecation warning in fixture retrieval.
Python
mit
yougov/pmxbot,yougov/pmxbot,yougov/pmxbot
import pytest @pytest.fixture(scope='session', autouse=True) def init_config(): __import__('pmxbot').config = {} @pytest.fixture(params=['mongodb', 'sqlite']) def db_uri(request): if request.param == 'mongodb': return request.getfuncargvalue('mongodb_uri') return 'sqlite:pmxbot.sqlite' Fix deprecation warning in fixture retrieval.
import pytest @pytest.fixture(scope='session', autouse=True) def init_config(): __import__('pmxbot').config = {} @pytest.fixture(params=['mongodb', 'sqlite']) def db_uri(request): if request.param == 'mongodb': return request.getfixturevalue('mongodb_uri') return 'sqlite:pmxbot.sqlite'
<commit_before>import pytest @pytest.fixture(scope='session', autouse=True) def init_config(): __import__('pmxbot').config = {} @pytest.fixture(params=['mongodb', 'sqlite']) def db_uri(request): if request.param == 'mongodb': return request.getfuncargvalue('mongodb_uri') return 'sqlite:pmxbot.sqlite' <commit_msg>Fix deprecation warning in fixture retrieval.<commit_after>
import pytest @pytest.fixture(scope='session', autouse=True) def init_config(): __import__('pmxbot').config = {} @pytest.fixture(params=['mongodb', 'sqlite']) def db_uri(request): if request.param == 'mongodb': return request.getfixturevalue('mongodb_uri') return 'sqlite:pmxbot.sqlite'
import pytest @pytest.fixture(scope='session', autouse=True) def init_config(): __import__('pmxbot').config = {} @pytest.fixture(params=['mongodb', 'sqlite']) def db_uri(request): if request.param == 'mongodb': return request.getfuncargvalue('mongodb_uri') return 'sqlite:pmxbot.sqlite' Fix deprecation warning in fixture retrieval.import pytest @pytest.fixture(scope='session', autouse=True) def init_config(): __import__('pmxbot').config = {} @pytest.fixture(params=['mongodb', 'sqlite']) def db_uri(request): if request.param == 'mongodb': return request.getfixturevalue('mongodb_uri') return 'sqlite:pmxbot.sqlite'
<commit_before>import pytest @pytest.fixture(scope='session', autouse=True) def init_config(): __import__('pmxbot').config = {} @pytest.fixture(params=['mongodb', 'sqlite']) def db_uri(request): if request.param == 'mongodb': return request.getfuncargvalue('mongodb_uri') return 'sqlite:pmxbot.sqlite' <commit_msg>Fix deprecation warning in fixture retrieval.<commit_after>import pytest @pytest.fixture(scope='session', autouse=True) def init_config(): __import__('pmxbot').config = {} @pytest.fixture(params=['mongodb', 'sqlite']) def db_uri(request): if request.param == 'mongodb': return request.getfixturevalue('mongodb_uri') return 'sqlite:pmxbot.sqlite'
32f7b016d0e06fa85bef95119223636dc74918ad
reclass/defaults.py
reclass/defaults.py
# # -*- coding: utf-8 -*- # # This file is part of reclass (http://github.com/madduck/reclass) # # Copyright © 2007–14 martin f. krafft <madduck@madduck.net> # Released under the terms of the Artistic Licence 2.0 # import os, sys from version import RECLASS_NAME # defaults for the command-line options OPT_STORAGE_TYPE = 'yaml_fs' OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME) OPT_NODES_URI = 'nodes' OPT_CLASSES_URI = 'classes' OPT_PRETTY_PRINT = True OPT_OUTPUT = 'yaml' CONFIG_FILE_SEARCH_PATH = [os.getcwd(), os.path.expanduser('~'), OPT_INVENTORY_BASE_URI, os.path.dirname(sys.argv[0]) ] CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml' PARAMETER_INTERPOLATION_SENTINELS = ('${', '}') PARAMETER_INTERPOLATION_DELIMITER = ':'
# # -*- coding: utf-8 -*- # # This file is part of reclass (http://github.com/madduck/reclass) # # Copyright © 2007–14 martin f. krafft <madduck@madduck.net> # Released under the terms of the Artistic Licence 2.0 # import os, sys from version import RECLASS_NAME # defaults for the command-line options OPT_STORAGE_TYPE = 'yaml_fs' OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME) OPT_NODES_URI = 'nodes' OPT_CLASSES_URI = 'classes' OPT_PRETTY_PRINT = True OPT_OUTPUT = 'yaml' CONFIG_FILE_SEARCH_PATH = [os.getcwd(), os.path.expanduser('~'), OPT_INVENTORY_BASE_URI, os.path.dirname(sys.argv[0]) ] CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml' PARAMETER_INTERPOLATION_SENTINELS = ('{{', '}}') PARAMETER_INTERPOLATION_DELIMITER = ':'
Change interpolation sentinels to prevent collissions with pillar data
Change interpolation sentinels to prevent collissions with pillar data
Python
artistic-2.0
jeroen92/reclass,jeroen92/reclass
# # -*- coding: utf-8 -*- # # This file is part of reclass (http://github.com/madduck/reclass) # # Copyright © 2007–14 martin f. krafft <madduck@madduck.net> # Released under the terms of the Artistic Licence 2.0 # import os, sys from version import RECLASS_NAME # defaults for the command-line options OPT_STORAGE_TYPE = 'yaml_fs' OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME) OPT_NODES_URI = 'nodes' OPT_CLASSES_URI = 'classes' OPT_PRETTY_PRINT = True OPT_OUTPUT = 'yaml' CONFIG_FILE_SEARCH_PATH = [os.getcwd(), os.path.expanduser('~'), OPT_INVENTORY_BASE_URI, os.path.dirname(sys.argv[0]) ] CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml' PARAMETER_INTERPOLATION_SENTINELS = ('${', '}') PARAMETER_INTERPOLATION_DELIMITER = ':' Change interpolation sentinels to prevent collissions with pillar data
# # -*- coding: utf-8 -*- # # This file is part of reclass (http://github.com/madduck/reclass) # # Copyright © 2007–14 martin f. krafft <madduck@madduck.net> # Released under the terms of the Artistic Licence 2.0 # import os, sys from version import RECLASS_NAME # defaults for the command-line options OPT_STORAGE_TYPE = 'yaml_fs' OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME) OPT_NODES_URI = 'nodes' OPT_CLASSES_URI = 'classes' OPT_PRETTY_PRINT = True OPT_OUTPUT = 'yaml' CONFIG_FILE_SEARCH_PATH = [os.getcwd(), os.path.expanduser('~'), OPT_INVENTORY_BASE_URI, os.path.dirname(sys.argv[0]) ] CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml' PARAMETER_INTERPOLATION_SENTINELS = ('{{', '}}') PARAMETER_INTERPOLATION_DELIMITER = ':'
<commit_before># # -*- coding: utf-8 -*- # # This file is part of reclass (http://github.com/madduck/reclass) # # Copyright © 2007–14 martin f. krafft <madduck@madduck.net> # Released under the terms of the Artistic Licence 2.0 # import os, sys from version import RECLASS_NAME # defaults for the command-line options OPT_STORAGE_TYPE = 'yaml_fs' OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME) OPT_NODES_URI = 'nodes' OPT_CLASSES_URI = 'classes' OPT_PRETTY_PRINT = True OPT_OUTPUT = 'yaml' CONFIG_FILE_SEARCH_PATH = [os.getcwd(), os.path.expanduser('~'), OPT_INVENTORY_BASE_URI, os.path.dirname(sys.argv[0]) ] CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml' PARAMETER_INTERPOLATION_SENTINELS = ('${', '}') PARAMETER_INTERPOLATION_DELIMITER = ':' <commit_msg>Change interpolation sentinels to prevent collissions with pillar data<commit_after>
# # -*- coding: utf-8 -*- # # This file is part of reclass (http://github.com/madduck/reclass) # # Copyright © 2007–14 martin f. krafft <madduck@madduck.net> # Released under the terms of the Artistic Licence 2.0 # import os, sys from version import RECLASS_NAME # defaults for the command-line options OPT_STORAGE_TYPE = 'yaml_fs' OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME) OPT_NODES_URI = 'nodes' OPT_CLASSES_URI = 'classes' OPT_PRETTY_PRINT = True OPT_OUTPUT = 'yaml' CONFIG_FILE_SEARCH_PATH = [os.getcwd(), os.path.expanduser('~'), OPT_INVENTORY_BASE_URI, os.path.dirname(sys.argv[0]) ] CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml' PARAMETER_INTERPOLATION_SENTINELS = ('{{', '}}') PARAMETER_INTERPOLATION_DELIMITER = ':'
# # -*- coding: utf-8 -*- # # This file is part of reclass (http://github.com/madduck/reclass) # # Copyright © 2007–14 martin f. krafft <madduck@madduck.net> # Released under the terms of the Artistic Licence 2.0 # import os, sys from version import RECLASS_NAME # defaults for the command-line options OPT_STORAGE_TYPE = 'yaml_fs' OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME) OPT_NODES_URI = 'nodes' OPT_CLASSES_URI = 'classes' OPT_PRETTY_PRINT = True OPT_OUTPUT = 'yaml' CONFIG_FILE_SEARCH_PATH = [os.getcwd(), os.path.expanduser('~'), OPT_INVENTORY_BASE_URI, os.path.dirname(sys.argv[0]) ] CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml' PARAMETER_INTERPOLATION_SENTINELS = ('${', '}') PARAMETER_INTERPOLATION_DELIMITER = ':' Change interpolation sentinels to prevent collissions with pillar data# # -*- coding: utf-8 -*- # # This file is part of reclass (http://github.com/madduck/reclass) # # Copyright © 2007–14 martin f. krafft <madduck@madduck.net> # Released under the terms of the Artistic Licence 2.0 # import os, sys from version import RECLASS_NAME # defaults for the command-line options OPT_STORAGE_TYPE = 'yaml_fs' OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME) OPT_NODES_URI = 'nodes' OPT_CLASSES_URI = 'classes' OPT_PRETTY_PRINT = True OPT_OUTPUT = 'yaml' CONFIG_FILE_SEARCH_PATH = [os.getcwd(), os.path.expanduser('~'), OPT_INVENTORY_BASE_URI, os.path.dirname(sys.argv[0]) ] CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml' PARAMETER_INTERPOLATION_SENTINELS = ('{{', '}}') PARAMETER_INTERPOLATION_DELIMITER = ':'
<commit_before># # -*- coding: utf-8 -*- # # This file is part of reclass (http://github.com/madduck/reclass) # # Copyright © 2007–14 martin f. krafft <madduck@madduck.net> # Released under the terms of the Artistic Licence 2.0 # import os, sys from version import RECLASS_NAME # defaults for the command-line options OPT_STORAGE_TYPE = 'yaml_fs' OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME) OPT_NODES_URI = 'nodes' OPT_CLASSES_URI = 'classes' OPT_PRETTY_PRINT = True OPT_OUTPUT = 'yaml' CONFIG_FILE_SEARCH_PATH = [os.getcwd(), os.path.expanduser('~'), OPT_INVENTORY_BASE_URI, os.path.dirname(sys.argv[0]) ] CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml' PARAMETER_INTERPOLATION_SENTINELS = ('${', '}') PARAMETER_INTERPOLATION_DELIMITER = ':' <commit_msg>Change interpolation sentinels to prevent collissions with pillar data<commit_after># # -*- coding: utf-8 -*- # # This file is part of reclass (http://github.com/madduck/reclass) # # Copyright © 2007–14 martin f. krafft <madduck@madduck.net> # Released under the terms of the Artistic Licence 2.0 # import os, sys from version import RECLASS_NAME # defaults for the command-line options OPT_STORAGE_TYPE = 'yaml_fs' OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME) OPT_NODES_URI = 'nodes' OPT_CLASSES_URI = 'classes' OPT_PRETTY_PRINT = True OPT_OUTPUT = 'yaml' CONFIG_FILE_SEARCH_PATH = [os.getcwd(), os.path.expanduser('~'), OPT_INVENTORY_BASE_URI, os.path.dirname(sys.argv[0]) ] CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml' PARAMETER_INTERPOLATION_SENTINELS = ('{{', '}}') PARAMETER_INTERPOLATION_DELIMITER = ':'
019aa0d78fbf54dda405cf8df3aab92dfdaba188
tests/grammar_atomic_tests.py
tests/grammar_atomic_tests.py
from unittest import TestCase from pyparsing import ParseException from regparser.grammar.atomic import * class GrammarAtomicTests(TestCase): def test_em_digit_p(self): result = em_digit_p.parseString('(<E T="03">2</E>)') self.assertEqual('2', result.p5) def test_double_alpha(self): # Match (aa), (bb), etc. result = lower_p.parseString('(a)') self.assertEqual('a', result.p1) result = lower_p.parseString('(aa)') self.assertEqual('aa', result.p1) result = lower_p.parseString('(i)') self.assertEqual('i', result.p1) # Except for roman numerals with self.assertRaises(ParseException): result = lower_p.parseString('(ii)') with self.assertRaises(ParseException): result = lower_p.parseString('(iv)')
from unittest import TestCase from pyparsing import ParseException from regparser.grammar.atomic import * class GrammarAtomicTests(TestCase): def test_em_digit_p(self): result = em_digit_p.parseString('(<E T="03">2</E>)') self.assertEqual('2', result.p5) def test_double_alpha(self): for text, p1 in [('(a)', 'a'), ('(aa)', 'aa'), ('(i)','i')]: result = lower_p.parseString(text) self.assertEqual(p1, result.p1) for text in ['(ii)', '(iv)', '(vi)']: try: result = lower_p.parseString(text) except ParseException: pass except e: self.fail("Unexpected error:", e) else: self.fail("Didn't raise ParseException")
Refactor test to eliminate assertRaises() error with Python 2.6
Refactor test to eliminate assertRaises() error with Python 2.6
Python
cc0-1.0
adderall/regulations-parser,willbarton/regulations-parser,grapesmoker/regulations-parser
from unittest import TestCase from pyparsing import ParseException from regparser.grammar.atomic import * class GrammarAtomicTests(TestCase): def test_em_digit_p(self): result = em_digit_p.parseString('(<E T="03">2</E>)') self.assertEqual('2', result.p5) def test_double_alpha(self): # Match (aa), (bb), etc. result = lower_p.parseString('(a)') self.assertEqual('a', result.p1) result = lower_p.parseString('(aa)') self.assertEqual('aa', result.p1) result = lower_p.parseString('(i)') self.assertEqual('i', result.p1) # Except for roman numerals with self.assertRaises(ParseException): result = lower_p.parseString('(ii)') with self.assertRaises(ParseException): result = lower_p.parseString('(iv)') Refactor test to eliminate assertRaises() error with Python 2.6
from unittest import TestCase from pyparsing import ParseException from regparser.grammar.atomic import * class GrammarAtomicTests(TestCase): def test_em_digit_p(self): result = em_digit_p.parseString('(<E T="03">2</E>)') self.assertEqual('2', result.p5) def test_double_alpha(self): for text, p1 in [('(a)', 'a'), ('(aa)', 'aa'), ('(i)','i')]: result = lower_p.parseString(text) self.assertEqual(p1, result.p1) for text in ['(ii)', '(iv)', '(vi)']: try: result = lower_p.parseString(text) except ParseException: pass except e: self.fail("Unexpected error:", e) else: self.fail("Didn't raise ParseException")
<commit_before>from unittest import TestCase from pyparsing import ParseException from regparser.grammar.atomic import * class GrammarAtomicTests(TestCase): def test_em_digit_p(self): result = em_digit_p.parseString('(<E T="03">2</E>)') self.assertEqual('2', result.p5) def test_double_alpha(self): # Match (aa), (bb), etc. result = lower_p.parseString('(a)') self.assertEqual('a', result.p1) result = lower_p.parseString('(aa)') self.assertEqual('aa', result.p1) result = lower_p.parseString('(i)') self.assertEqual('i', result.p1) # Except for roman numerals with self.assertRaises(ParseException): result = lower_p.parseString('(ii)') with self.assertRaises(ParseException): result = lower_p.parseString('(iv)') <commit_msg>Refactor test to eliminate assertRaises() error with Python 2.6<commit_after>
from unittest import TestCase from pyparsing import ParseException from regparser.grammar.atomic import * class GrammarAtomicTests(TestCase): def test_em_digit_p(self): result = em_digit_p.parseString('(<E T="03">2</E>)') self.assertEqual('2', result.p5) def test_double_alpha(self): for text, p1 in [('(a)', 'a'), ('(aa)', 'aa'), ('(i)','i')]: result = lower_p.parseString(text) self.assertEqual(p1, result.p1) for text in ['(ii)', '(iv)', '(vi)']: try: result = lower_p.parseString(text) except ParseException: pass except e: self.fail("Unexpected error:", e) else: self.fail("Didn't raise ParseException")
from unittest import TestCase from pyparsing import ParseException from regparser.grammar.atomic import * class GrammarAtomicTests(TestCase): def test_em_digit_p(self): result = em_digit_p.parseString('(<E T="03">2</E>)') self.assertEqual('2', result.p5) def test_double_alpha(self): # Match (aa), (bb), etc. result = lower_p.parseString('(a)') self.assertEqual('a', result.p1) result = lower_p.parseString('(aa)') self.assertEqual('aa', result.p1) result = lower_p.parseString('(i)') self.assertEqual('i', result.p1) # Except for roman numerals with self.assertRaises(ParseException): result = lower_p.parseString('(ii)') with self.assertRaises(ParseException): result = lower_p.parseString('(iv)') Refactor test to eliminate assertRaises() error with Python 2.6from unittest import TestCase from pyparsing import ParseException from regparser.grammar.atomic import * class GrammarAtomicTests(TestCase): def test_em_digit_p(self): result = em_digit_p.parseString('(<E T="03">2</E>)') self.assertEqual('2', result.p5) def test_double_alpha(self): for text, p1 in [('(a)', 'a'), ('(aa)', 'aa'), ('(i)','i')]: result = lower_p.parseString(text) self.assertEqual(p1, result.p1) for text in ['(ii)', '(iv)', '(vi)']: try: result = lower_p.parseString(text) except ParseException: pass except e: self.fail("Unexpected error:", e) else: self.fail("Didn't raise ParseException")
<commit_before>from unittest import TestCase from pyparsing import ParseException from regparser.grammar.atomic import * class GrammarAtomicTests(TestCase): def test_em_digit_p(self): result = em_digit_p.parseString('(<E T="03">2</E>)') self.assertEqual('2', result.p5) def test_double_alpha(self): # Match (aa), (bb), etc. result = lower_p.parseString('(a)') self.assertEqual('a', result.p1) result = lower_p.parseString('(aa)') self.assertEqual('aa', result.p1) result = lower_p.parseString('(i)') self.assertEqual('i', result.p1) # Except for roman numerals with self.assertRaises(ParseException): result = lower_p.parseString('(ii)') with self.assertRaises(ParseException): result = lower_p.parseString('(iv)') <commit_msg>Refactor test to eliminate assertRaises() error with Python 2.6<commit_after>from unittest import TestCase from pyparsing import ParseException from regparser.grammar.atomic import * class GrammarAtomicTests(TestCase): def test_em_digit_p(self): result = em_digit_p.parseString('(<E T="03">2</E>)') self.assertEqual('2', result.p5) def test_double_alpha(self): for text, p1 in [('(a)', 'a'), ('(aa)', 'aa'), ('(i)','i')]: result = lower_p.parseString(text) self.assertEqual(p1, result.p1) for text in ['(ii)', '(iv)', '(vi)']: try: result = lower_p.parseString(text) except ParseException: pass except e: self.fail("Unexpected error:", e) else: self.fail("Didn't raise ParseException")
c34817c2740e860493692b630a11fdb7acab76aa
tests/test_simple_features.py
tests/test_simple_features.py
from wordgraph.points import Point import wordgraph EPOCH_START = 1407109280 def time_values(values, start=EPOCH_START, increment=1): datapoints = [] for index, value in enumerate(values): datapoints.append(Point(x=value, y=start + (increment * index))) return datapoints def test_monotonic_up_per_second(): datapoints = time_values(float(i) for i in range(POINTS)) features = wordgraph.describe(datapoints) assert "" in features
from wordgraph.points import Point import wordgraph EPOCH_START = 1407109280 def time_values(values, start=EPOCH_START, increment=1): datapoints = [] for index, value in enumerate(values): datapoints.append(Point(x=value, y=start + (increment * index))) return datapoints def test_monotonic_up_per_second(): datapoints = time_values(float(i) for i in range(10)) features = wordgraph.describe(datapoints) assert "" in features def test_monotonic_down_per_second(): datapoints = time_values(10.0 - i for i in range(10)) features = wordgraph.describe(datapoints) assert "" in features
Test case for monotonically decreasing graphs
Test case for monotonically decreasing graphs Generate time series data for values that decrease monotonically over time.
Python
apache-2.0
tleeuwenburg/wordgraph,tleeuwenburg/wordgraph
from wordgraph.points import Point import wordgraph EPOCH_START = 1407109280 def time_values(values, start=EPOCH_START, increment=1): datapoints = [] for index, value in enumerate(values): datapoints.append(Point(x=value, y=start + (increment * index))) return datapoints def test_monotonic_up_per_second(): datapoints = time_values(float(i) for i in range(POINTS)) features = wordgraph.describe(datapoints) assert "" in features Test case for monotonically decreasing graphs Generate time series data for values that decrease monotonically over time.
from wordgraph.points import Point import wordgraph EPOCH_START = 1407109280 def time_values(values, start=EPOCH_START, increment=1): datapoints = [] for index, value in enumerate(values): datapoints.append(Point(x=value, y=start + (increment * index))) return datapoints def test_monotonic_up_per_second(): datapoints = time_values(float(i) for i in range(10)) features = wordgraph.describe(datapoints) assert "" in features def test_monotonic_down_per_second(): datapoints = time_values(10.0 - i for i in range(10)) features = wordgraph.describe(datapoints) assert "" in features
<commit_before>from wordgraph.points import Point import wordgraph EPOCH_START = 1407109280 def time_values(values, start=EPOCH_START, increment=1): datapoints = [] for index, value in enumerate(values): datapoints.append(Point(x=value, y=start + (increment * index))) return datapoints def test_monotonic_up_per_second(): datapoints = time_values(float(i) for i in range(POINTS)) features = wordgraph.describe(datapoints) assert "" in features <commit_msg>Test case for monotonically decreasing graphs Generate time series data for values that decrease monotonically over time.<commit_after>
from wordgraph.points import Point import wordgraph EPOCH_START = 1407109280 def time_values(values, start=EPOCH_START, increment=1): datapoints = [] for index, value in enumerate(values): datapoints.append(Point(x=value, y=start + (increment * index))) return datapoints def test_monotonic_up_per_second(): datapoints = time_values(float(i) for i in range(10)) features = wordgraph.describe(datapoints) assert "" in features def test_monotonic_down_per_second(): datapoints = time_values(10.0 - i for i in range(10)) features = wordgraph.describe(datapoints) assert "" in features
from wordgraph.points import Point import wordgraph EPOCH_START = 1407109280 def time_values(values, start=EPOCH_START, increment=1): datapoints = [] for index, value in enumerate(values): datapoints.append(Point(x=value, y=start + (increment * index))) return datapoints def test_monotonic_up_per_second(): datapoints = time_values(float(i) for i in range(POINTS)) features = wordgraph.describe(datapoints) assert "" in features Test case for monotonically decreasing graphs Generate time series data for values that decrease monotonically over time.from wordgraph.points import Point import wordgraph EPOCH_START = 1407109280 def time_values(values, start=EPOCH_START, increment=1): datapoints = [] for index, value in enumerate(values): datapoints.append(Point(x=value, y=start + (increment * index))) return datapoints def test_monotonic_up_per_second(): datapoints = time_values(float(i) for i in range(10)) features = wordgraph.describe(datapoints) assert "" in features def test_monotonic_down_per_second(): datapoints = time_values(10.0 - i for i in range(10)) features = wordgraph.describe(datapoints) assert "" in features
<commit_before>from wordgraph.points import Point import wordgraph EPOCH_START = 1407109280 def time_values(values, start=EPOCH_START, increment=1): datapoints = [] for index, value in enumerate(values): datapoints.append(Point(x=value, y=start + (increment * index))) return datapoints def test_monotonic_up_per_second(): datapoints = time_values(float(i) for i in range(POINTS)) features = wordgraph.describe(datapoints) assert "" in features <commit_msg>Test case for monotonically decreasing graphs Generate time series data for values that decrease monotonically over time.<commit_after>from wordgraph.points import Point import wordgraph EPOCH_START = 1407109280 def time_values(values, start=EPOCH_START, increment=1): datapoints = [] for index, value in enumerate(values): datapoints.append(Point(x=value, y=start + (increment * index))) return datapoints def test_monotonic_up_per_second(): datapoints = time_values(float(i) for i in range(10)) features = wordgraph.describe(datapoints) assert "" in features def test_monotonic_down_per_second(): datapoints = time_values(10.0 - i for i in range(10)) features = wordgraph.describe(datapoints) assert "" in features
e2991360bb35060e570e59bbea6a1e6d23a54aa7
transport_tester.py
transport_tester.py
from gevent import monkey monkey.patch_all() # noqa import sys import time from raiden.network.transport import UDPTransport from raiden.network.sockfactory import socket_factory class DummyProtocol(object): def __init__(self): self.raiden = None def receive(self, data): print data if __name__ == "__main__": with socket_factory('0.0.0.0', 8885) as mapped_socket: print mapped_socket t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) while True: time.sleep(1) if len(sys.argv) > 1: t.send(None, (sys.argv[1], 8885), b'hello')
from gevent import monkey monkey.patch_all() # noqa import sys import time from ethereum import slogging from raiden.network.transport import UDPTransport from raiden.network.sockfactory import socket_factory class DummyProtocol(object): def __init__(self): self.raiden = None def receive(self, data): print data if __name__ == "__main__": slogging.configure(':DEBUG') with socket_factory('0.0.0.0', 8885) as mapped_socket: print mapped_socket t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) while True: time.sleep(1) if len(sys.argv) > 1: t.send(None, (sys.argv[1], 8885), b'hello')
Add logging to script test
Add logging to script test
Python
mit
hackaugusto/raiden,tomashaber/raiden,tomashaber/raiden,tomashaber/raiden,hackaugusto/raiden,tomashaber/raiden,tomashaber/raiden,charles-cooper/raiden,charles-cooper/raiden
from gevent import monkey monkey.patch_all() # noqa import sys import time from raiden.network.transport import UDPTransport from raiden.network.sockfactory import socket_factory class DummyProtocol(object): def __init__(self): self.raiden = None def receive(self, data): print data if __name__ == "__main__": with socket_factory('0.0.0.0', 8885) as mapped_socket: print mapped_socket t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) while True: time.sleep(1) if len(sys.argv) > 1: t.send(None, (sys.argv[1], 8885), b'hello') Add logging to script test
from gevent import monkey monkey.patch_all() # noqa import sys import time from ethereum import slogging from raiden.network.transport import UDPTransport from raiden.network.sockfactory import socket_factory class DummyProtocol(object): def __init__(self): self.raiden = None def receive(self, data): print data if __name__ == "__main__": slogging.configure(':DEBUG') with socket_factory('0.0.0.0', 8885) as mapped_socket: print mapped_socket t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) while True: time.sleep(1) if len(sys.argv) > 1: t.send(None, (sys.argv[1], 8885), b'hello')
<commit_before>from gevent import monkey monkey.patch_all() # noqa import sys import time from raiden.network.transport import UDPTransport from raiden.network.sockfactory import socket_factory class DummyProtocol(object): def __init__(self): self.raiden = None def receive(self, data): print data if __name__ == "__main__": with socket_factory('0.0.0.0', 8885) as mapped_socket: print mapped_socket t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) while True: time.sleep(1) if len(sys.argv) > 1: t.send(None, (sys.argv[1], 8885), b'hello') <commit_msg>Add logging to script test<commit_after>
from gevent import monkey monkey.patch_all() # noqa import sys import time from ethereum import slogging from raiden.network.transport import UDPTransport from raiden.network.sockfactory import socket_factory class DummyProtocol(object): def __init__(self): self.raiden = None def receive(self, data): print data if __name__ == "__main__": slogging.configure(':DEBUG') with socket_factory('0.0.0.0', 8885) as mapped_socket: print mapped_socket t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) while True: time.sleep(1) if len(sys.argv) > 1: t.send(None, (sys.argv[1], 8885), b'hello')
from gevent import monkey monkey.patch_all() # noqa import sys import time from raiden.network.transport import UDPTransport from raiden.network.sockfactory import socket_factory class DummyProtocol(object): def __init__(self): self.raiden = None def receive(self, data): print data if __name__ == "__main__": with socket_factory('0.0.0.0', 8885) as mapped_socket: print mapped_socket t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) while True: time.sleep(1) if len(sys.argv) > 1: t.send(None, (sys.argv[1], 8885), b'hello') Add logging to script testfrom gevent import monkey monkey.patch_all() # noqa import sys import time from ethereum import slogging from raiden.network.transport import UDPTransport from raiden.network.sockfactory import socket_factory class DummyProtocol(object): def __init__(self): self.raiden = None def receive(self, data): print data if __name__ == "__main__": slogging.configure(':DEBUG') with socket_factory('0.0.0.0', 8885) as mapped_socket: print mapped_socket t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) while True: time.sleep(1) if len(sys.argv) > 1: t.send(None, (sys.argv[1], 8885), b'hello')
<commit_before>from gevent import monkey monkey.patch_all() # noqa import sys import time from raiden.network.transport import UDPTransport from raiden.network.sockfactory import socket_factory class DummyProtocol(object): def __init__(self): self.raiden = None def receive(self, data): print data if __name__ == "__main__": with socket_factory('0.0.0.0', 8885) as mapped_socket: print mapped_socket t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) while True: time.sleep(1) if len(sys.argv) > 1: t.send(None, (sys.argv[1], 8885), b'hello') <commit_msg>Add logging to script test<commit_after>from gevent import monkey monkey.patch_all() # noqa import sys import time from ethereum import slogging from raiden.network.transport import UDPTransport from raiden.network.sockfactory import socket_factory class DummyProtocol(object): def __init__(self): self.raiden = None def receive(self, data): print data if __name__ == "__main__": slogging.configure(':DEBUG') with socket_factory('0.0.0.0', 8885) as mapped_socket: print mapped_socket t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) while True: time.sleep(1) if len(sys.argv) > 1: t.send(None, (sys.argv[1], 8885), b'hello')
156ebae630c3690db875b8925bfbdc5ded396fdd
src/Sensors/Factory.py
src/Sensors/Factory.py
from src.Sensors.BME280 import BME280 from src.Sensors.BME680 import BME680 from src.Sensors.DS18B20 import DS18B20 from src.Notification.Subscriber.LED.RGB import RGB class Factory: @staticmethod def create_sensor(device, address): if device == 'BME280': return BME280(address=address) elif device == 'BME680': return BME680(address=address) elif device == 'DS18B20': return DS18B20(address=address) @staticmethod def create_led(device, configuration, notification_manager): if device == 'rgb': return RGB(configuration=configuration, notification_manager=notification_manager)
from src.Sensors.BME280 import BME280 from src.Sensors.BME680 import BME680 from src.Sensors.DS18B20 import DS18B20 from src.Sensors.CCS811 import CCS811 from src.Notification.Subscriber.LED.RGB import RGB class Factory: @staticmethod def create_sensor(device, address): if device == 'BME280': return BME280(address=address) elif device == 'BME680': return BME680(address=address) elif device == 'DS18B20': return DS18B20(address=address) elif device == 'CCS811': return CCS811(address=address) @staticmethod def create_led(device, configuration, notification_manager): if device == 'rgb': return RGB(configuration=configuration, notification_manager=notification_manager)
Add CCS811 to device factory
Add CCS811 to device factory
Python
mit
dashford/sentinel
from src.Sensors.BME280 import BME280 from src.Sensors.BME680 import BME680 from src.Sensors.DS18B20 import DS18B20 from src.Notification.Subscriber.LED.RGB import RGB class Factory: @staticmethod def create_sensor(device, address): if device == 'BME280': return BME280(address=address) elif device == 'BME680': return BME680(address=address) elif device == 'DS18B20': return DS18B20(address=address) @staticmethod def create_led(device, configuration, notification_manager): if device == 'rgb': return RGB(configuration=configuration, notification_manager=notification_manager) Add CCS811 to device factory
from src.Sensors.BME280 import BME280 from src.Sensors.BME680 import BME680 from src.Sensors.DS18B20 import DS18B20 from src.Sensors.CCS811 import CCS811 from src.Notification.Subscriber.LED.RGB import RGB class Factory: @staticmethod def create_sensor(device, address): if device == 'BME280': return BME280(address=address) elif device == 'BME680': return BME680(address=address) elif device == 'DS18B20': return DS18B20(address=address) elif device == 'CCS811': return CCS811(address=address) @staticmethod def create_led(device, configuration, notification_manager): if device == 'rgb': return RGB(configuration=configuration, notification_manager=notification_manager)
<commit_before>from src.Sensors.BME280 import BME280 from src.Sensors.BME680 import BME680 from src.Sensors.DS18B20 import DS18B20 from src.Notification.Subscriber.LED.RGB import RGB class Factory: @staticmethod def create_sensor(device, address): if device == 'BME280': return BME280(address=address) elif device == 'BME680': return BME680(address=address) elif device == 'DS18B20': return DS18B20(address=address) @staticmethod def create_led(device, configuration, notification_manager): if device == 'rgb': return RGB(configuration=configuration, notification_manager=notification_manager) <commit_msg>Add CCS811 to device factory<commit_after>
from src.Sensors.BME280 import BME280 from src.Sensors.BME680 import BME680 from src.Sensors.DS18B20 import DS18B20 from src.Sensors.CCS811 import CCS811 from src.Notification.Subscriber.LED.RGB import RGB class Factory: @staticmethod def create_sensor(device, address): if device == 'BME280': return BME280(address=address) elif device == 'BME680': return BME680(address=address) elif device == 'DS18B20': return DS18B20(address=address) elif device == 'CCS811': return CCS811(address=address) @staticmethod def create_led(device, configuration, notification_manager): if device == 'rgb': return RGB(configuration=configuration, notification_manager=notification_manager)
from src.Sensors.BME280 import BME280 from src.Sensors.BME680 import BME680 from src.Sensors.DS18B20 import DS18B20 from src.Notification.Subscriber.LED.RGB import RGB class Factory: @staticmethod def create_sensor(device, address): if device == 'BME280': return BME280(address=address) elif device == 'BME680': return BME680(address=address) elif device == 'DS18B20': return DS18B20(address=address) @staticmethod def create_led(device, configuration, notification_manager): if device == 'rgb': return RGB(configuration=configuration, notification_manager=notification_manager) Add CCS811 to device factoryfrom src.Sensors.BME280 import BME280 from src.Sensors.BME680 import BME680 from src.Sensors.DS18B20 import DS18B20 from src.Sensors.CCS811 import CCS811 from src.Notification.Subscriber.LED.RGB import RGB class Factory: @staticmethod def create_sensor(device, address): if device == 'BME280': return BME280(address=address) elif device == 'BME680': return BME680(address=address) elif device == 'DS18B20': return DS18B20(address=address) elif device == 'CCS811': return CCS811(address=address) @staticmethod def create_led(device, configuration, notification_manager): if device == 'rgb': return RGB(configuration=configuration, notification_manager=notification_manager)
<commit_before>from src.Sensors.BME280 import BME280 from src.Sensors.BME680 import BME680 from src.Sensors.DS18B20 import DS18B20 from src.Notification.Subscriber.LED.RGB import RGB class Factory: @staticmethod def create_sensor(device, address): if device == 'BME280': return BME280(address=address) elif device == 'BME680': return BME680(address=address) elif device == 'DS18B20': return DS18B20(address=address) @staticmethod def create_led(device, configuration, notification_manager): if device == 'rgb': return RGB(configuration=configuration, notification_manager=notification_manager) <commit_msg>Add CCS811 to device factory<commit_after>from src.Sensors.BME280 import BME280 from src.Sensors.BME680 import BME680 from src.Sensors.DS18B20 import DS18B20 from src.Sensors.CCS811 import CCS811 from src.Notification.Subscriber.LED.RGB import RGB class Factory: @staticmethod def create_sensor(device, address): if device == 'BME280': return BME280(address=address) elif device == 'BME680': return BME680(address=address) elif device == 'DS18B20': return DS18B20(address=address) elif device == 'CCS811': return CCS811(address=address) @staticmethod def create_led(device, configuration, notification_manager): if device == 'rgb': return RGB(configuration=configuration, notification_manager=notification_manager)
b05ba019143cc39ba0d02d822824172313e78591
aubergine/celery.py
aubergine/celery.py
from __future__ import absolute_import from celery import Celery app = Celery('aubergine') app.config_from_object('aubergine.settings.celeryconfig') app.autodiscover_tasks(['aubergine'], related_name='tasks') app.setup_security() @app.task(bind=True) def debug_task(self): print('Request: {0!r}'.format(self.request)) @app.task() def tadd(x, y): return x + y @app.task() def tsum(numbers): return sum(numbers)
from __future__ import absolute_import from celery import Celery app = Celery('aubergine') app.config_from_object('aubergine.settings.celeryconfig') app.autodiscover_tasks(['aubergine'], related_name='tasks') # For message signing # app.setup_security() @app.task(bind=True) def debug_task(self): print('Request: {0!r}'.format(self.request)) @app.task() def tadd(x, y): return x + y @app.task() def tsum(numbers): return sum(numbers)
Disable message signing for now
Disable message signing for now
Python
bsd-3-clause
michaelwisely/aubergine
from __future__ import absolute_import from celery import Celery app = Celery('aubergine') app.config_from_object('aubergine.settings.celeryconfig') app.autodiscover_tasks(['aubergine'], related_name='tasks') app.setup_security() @app.task(bind=True) def debug_task(self): print('Request: {0!r}'.format(self.request)) @app.task() def tadd(x, y): return x + y @app.task() def tsum(numbers): return sum(numbers) Disable message signing for now
from __future__ import absolute_import from celery import Celery app = Celery('aubergine') app.config_from_object('aubergine.settings.celeryconfig') app.autodiscover_tasks(['aubergine'], related_name='tasks') # For message signing # app.setup_security() @app.task(bind=True) def debug_task(self): print('Request: {0!r}'.format(self.request)) @app.task() def tadd(x, y): return x + y @app.task() def tsum(numbers): return sum(numbers)
<commit_before>from __future__ import absolute_import from celery import Celery app = Celery('aubergine') app.config_from_object('aubergine.settings.celeryconfig') app.autodiscover_tasks(['aubergine'], related_name='tasks') app.setup_security() @app.task(bind=True) def debug_task(self): print('Request: {0!r}'.format(self.request)) @app.task() def tadd(x, y): return x + y @app.task() def tsum(numbers): return sum(numbers) <commit_msg>Disable message signing for now<commit_after>
from __future__ import absolute_import from celery import Celery app = Celery('aubergine') app.config_from_object('aubergine.settings.celeryconfig') app.autodiscover_tasks(['aubergine'], related_name='tasks') # For message signing # app.setup_security() @app.task(bind=True) def debug_task(self): print('Request: {0!r}'.format(self.request)) @app.task() def tadd(x, y): return x + y @app.task() def tsum(numbers): return sum(numbers)
from __future__ import absolute_import from celery import Celery app = Celery('aubergine') app.config_from_object('aubergine.settings.celeryconfig') app.autodiscover_tasks(['aubergine'], related_name='tasks') app.setup_security() @app.task(bind=True) def debug_task(self): print('Request: {0!r}'.format(self.request)) @app.task() def tadd(x, y): return x + y @app.task() def tsum(numbers): return sum(numbers) Disable message signing for nowfrom __future__ import absolute_import from celery import Celery app = Celery('aubergine') app.config_from_object('aubergine.settings.celeryconfig') app.autodiscover_tasks(['aubergine'], related_name='tasks') # For message signing # app.setup_security() @app.task(bind=True) def debug_task(self): print('Request: {0!r}'.format(self.request)) @app.task() def tadd(x, y): return x + y @app.task() def tsum(numbers): return sum(numbers)
<commit_before>from __future__ import absolute_import from celery import Celery app = Celery('aubergine') app.config_from_object('aubergine.settings.celeryconfig') app.autodiscover_tasks(['aubergine'], related_name='tasks') app.setup_security() @app.task(bind=True) def debug_task(self): print('Request: {0!r}'.format(self.request)) @app.task() def tadd(x, y): return x + y @app.task() def tsum(numbers): return sum(numbers) <commit_msg>Disable message signing for now<commit_after>from __future__ import absolute_import from celery import Celery app = Celery('aubergine') app.config_from_object('aubergine.settings.celeryconfig') app.autodiscover_tasks(['aubergine'], related_name='tasks') # For message signing # app.setup_security() @app.task(bind=True) def debug_task(self): print('Request: {0!r}'.format(self.request)) @app.task() def tadd(x, y): return x + y @app.task() def tsum(numbers): return sum(numbers)
f1af7dad41992b53e90a5f8dd20e1635f11a7ce1
pstats_print2list/__init__.py
pstats_print2list/__init__.py
# -*- coding: utf-8 -*- __author__ = 'Vauxoo' __email__ = 'info@vauxoo.com' __version__ = '0.1.0'
# -*- coding: utf-8 -*- from pstats_print2list import print_stats __author__ = 'Vauxoo' __email__ = 'info@vauxoo.com' __version__ = '0.1.0'
Add print_stats to init file
[REF] pstats_print2list: Add print_stats to init file
Python
isc
Vauxoo/pstats-print2list
# -*- coding: utf-8 -*- __author__ = 'Vauxoo' __email__ = 'info@vauxoo.com' __version__ = '0.1.0' [REF] pstats_print2list: Add print_stats to init file
# -*- coding: utf-8 -*- from pstats_print2list import print_stats __author__ = 'Vauxoo' __email__ = 'info@vauxoo.com' __version__ = '0.1.0'
<commit_before># -*- coding: utf-8 -*- __author__ = 'Vauxoo' __email__ = 'info@vauxoo.com' __version__ = '0.1.0' <commit_msg>[REF] pstats_print2list: Add print_stats to init file<commit_after>
# -*- coding: utf-8 -*- from pstats_print2list import print_stats __author__ = 'Vauxoo' __email__ = 'info@vauxoo.com' __version__ = '0.1.0'
# -*- coding: utf-8 -*- __author__ = 'Vauxoo' __email__ = 'info@vauxoo.com' __version__ = '0.1.0' [REF] pstats_print2list: Add print_stats to init file# -*- coding: utf-8 -*- from pstats_print2list import print_stats __author__ = 'Vauxoo' __email__ = 'info@vauxoo.com' __version__ = '0.1.0'
<commit_before># -*- coding: utf-8 -*- __author__ = 'Vauxoo' __email__ = 'info@vauxoo.com' __version__ = '0.1.0' <commit_msg>[REF] pstats_print2list: Add print_stats to init file<commit_after># -*- coding: utf-8 -*- from pstats_print2list import print_stats __author__ = 'Vauxoo' __email__ = 'info@vauxoo.com' __version__ = '0.1.0'
875e25231dec76760bde97107d52036560347914
salt/modules/win_shadow.py
salt/modules/win_shadow.py
''' Manage the shadow file ''' import salt.utils def __virtual__(): ''' Only works on Windows systems ''' if salt.utils.is_windows(): return 'shadow' return False def info(name): ''' Return information for the specified user This is just returns dummy data so that salt states can work. CLI Example:: salt '*' shadow.info root ''' ret = { 'name': name, 'pwd': '', 'lstchg': '', 'min': '', 'max': '', 'warn': '', 'inact': '', 'expire': ''} return ret def set_password(name, password): ''' Set the password for a named user. CLI Example:: salt '*' shadow.set_password root mysecretpassword ''' cmd = 'net user {0} {1}'.format(name, password) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode']
''' Manage the shadow file ''' import salt.utils def __virtual__(): ''' Only works on Windows systems ''' if salt.utils.is_windows(): return 'shadow' return False def info(name): ''' Return information for the specified user This is just returns dummy data so that salt states can work. CLI Example:: salt '*' shadow.info root ''' ret = { 'name': name, 'passwd': '', 'lstchg': '', 'min': '', 'max': '', 'warn': '', 'inact': '', 'expire': ''} return ret def set_password(name, password): ''' Set the password for a named user. CLI Example:: salt '*' shadow.set_password root mysecretpassword ''' cmd = 'net user {0} {1}'.format(name, password) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode']
Rename 'pwd' field in windows shadow.info output
Rename 'pwd' field in windows shadow.info output This makes the field name consistent with the other shadow modules. Note that the passwd field is not used at all in Windows user management, so this is merely a cosmetic change.
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
''' Manage the shadow file ''' import salt.utils def __virtual__(): ''' Only works on Windows systems ''' if salt.utils.is_windows(): return 'shadow' return False def info(name): ''' Return information for the specified user This is just returns dummy data so that salt states can work. CLI Example:: salt '*' shadow.info root ''' ret = { 'name': name, 'pwd': '', 'lstchg': '', 'min': '', 'max': '', 'warn': '', 'inact': '', 'expire': ''} return ret def set_password(name, password): ''' Set the password for a named user. CLI Example:: salt '*' shadow.set_password root mysecretpassword ''' cmd = 'net user {0} {1}'.format(name, password) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode'] Rename 'pwd' field in windows shadow.info output This makes the field name consistent with the other shadow modules. Note that the passwd field is not used at all in Windows user management, so this is merely a cosmetic change.
''' Manage the shadow file ''' import salt.utils def __virtual__(): ''' Only works on Windows systems ''' if salt.utils.is_windows(): return 'shadow' return False def info(name): ''' Return information for the specified user This is just returns dummy data so that salt states can work. CLI Example:: salt '*' shadow.info root ''' ret = { 'name': name, 'passwd': '', 'lstchg': '', 'min': '', 'max': '', 'warn': '', 'inact': '', 'expire': ''} return ret def set_password(name, password): ''' Set the password for a named user. CLI Example:: salt '*' shadow.set_password root mysecretpassword ''' cmd = 'net user {0} {1}'.format(name, password) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode']
<commit_before>''' Manage the shadow file ''' import salt.utils def __virtual__(): ''' Only works on Windows systems ''' if salt.utils.is_windows(): return 'shadow' return False def info(name): ''' Return information for the specified user This is just returns dummy data so that salt states can work. CLI Example:: salt '*' shadow.info root ''' ret = { 'name': name, 'pwd': '', 'lstchg': '', 'min': '', 'max': '', 'warn': '', 'inact': '', 'expire': ''} return ret def set_password(name, password): ''' Set the password for a named user. CLI Example:: salt '*' shadow.set_password root mysecretpassword ''' cmd = 'net user {0} {1}'.format(name, password) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode'] <commit_msg>Rename 'pwd' field in windows shadow.info output This makes the field name consistent with the other shadow modules. Note that the passwd field is not used at all in Windows user management, so this is merely a cosmetic change.<commit_after>
''' Manage the shadow file ''' import salt.utils def __virtual__(): ''' Only works on Windows systems ''' if salt.utils.is_windows(): return 'shadow' return False def info(name): ''' Return information for the specified user This is just returns dummy data so that salt states can work. CLI Example:: salt '*' shadow.info root ''' ret = { 'name': name, 'passwd': '', 'lstchg': '', 'min': '', 'max': '', 'warn': '', 'inact': '', 'expire': ''} return ret def set_password(name, password): ''' Set the password for a named user. CLI Example:: salt '*' shadow.set_password root mysecretpassword ''' cmd = 'net user {0} {1}'.format(name, password) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode']
''' Manage the shadow file ''' import salt.utils def __virtual__(): ''' Only works on Windows systems ''' if salt.utils.is_windows(): return 'shadow' return False def info(name): ''' Return information for the specified user This is just returns dummy data so that salt states can work. CLI Example:: salt '*' shadow.info root ''' ret = { 'name': name, 'pwd': '', 'lstchg': '', 'min': '', 'max': '', 'warn': '', 'inact': '', 'expire': ''} return ret def set_password(name, password): ''' Set the password for a named user. CLI Example:: salt '*' shadow.set_password root mysecretpassword ''' cmd = 'net user {0} {1}'.format(name, password) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode'] Rename 'pwd' field in windows shadow.info output This makes the field name consistent with the other shadow modules. Note that the passwd field is not used at all in Windows user management, so this is merely a cosmetic change.''' Manage the shadow file ''' import salt.utils def __virtual__(): ''' Only works on Windows systems ''' if salt.utils.is_windows(): return 'shadow' return False def info(name): ''' Return information for the specified user This is just returns dummy data so that salt states can work. CLI Example:: salt '*' shadow.info root ''' ret = { 'name': name, 'passwd': '', 'lstchg': '', 'min': '', 'max': '', 'warn': '', 'inact': '', 'expire': ''} return ret def set_password(name, password): ''' Set the password for a named user. CLI Example:: salt '*' shadow.set_password root mysecretpassword ''' cmd = 'net user {0} {1}'.format(name, password) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode']
<commit_before>''' Manage the shadow file ''' import salt.utils def __virtual__(): ''' Only works on Windows systems ''' if salt.utils.is_windows(): return 'shadow' return False def info(name): ''' Return information for the specified user This is just returns dummy data so that salt states can work. CLI Example:: salt '*' shadow.info root ''' ret = { 'name': name, 'pwd': '', 'lstchg': '', 'min': '', 'max': '', 'warn': '', 'inact': '', 'expire': ''} return ret def set_password(name, password): ''' Set the password for a named user. CLI Example:: salt '*' shadow.set_password root mysecretpassword ''' cmd = 'net user {0} {1}'.format(name, password) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode'] <commit_msg>Rename 'pwd' field in windows shadow.info output This makes the field name consistent with the other shadow modules. Note that the passwd field is not used at all in Windows user management, so this is merely a cosmetic change.<commit_after>''' Manage the shadow file ''' import salt.utils def __virtual__(): ''' Only works on Windows systems ''' if salt.utils.is_windows(): return 'shadow' return False def info(name): ''' Return information for the specified user This is just returns dummy data so that salt states can work. CLI Example:: salt '*' shadow.info root ''' ret = { 'name': name, 'passwd': '', 'lstchg': '', 'min': '', 'max': '', 'warn': '', 'inact': '', 'expire': ''} return ret def set_password(name, password): ''' Set the password for a named user. CLI Example:: salt '*' shadow.set_password root mysecretpassword ''' cmd = 'net user {0} {1}'.format(name, password) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode']
d279572c255a302dd5f191b0f047c46c9184ec2a
tests/test_write.py
tests/test_write.py
from __future__ import absolute_import from ofxparse import OfxParser as op, OfxPrinter from unittest import TestCase from os import close, remove from tempfile import mkstemp import sys sys.path.append('..') from .support import open_file class TestOfxWrite(TestCase): def test_write(self): test_file = open_file('fidelity.ofx') ofx_doc = op.parse(test_file) self.assertEqual(str(ofx_doc), "") def test_using_ofx_printer(self): test_file = open_file('checking.ofx') ofx_doc = op.parse(test_file) fd, name = mkstemp() close(fd) printer = OfxPrinter(ofx=ofx_doc, filename=name) printer.write(tabs=1) if __name__ == "__main__": import unittest unittest.main()
from __future__ import absolute_import from ofxparse import OfxParser, OfxPrinter from unittest import TestCase from os import close, remove from tempfile import mkstemp import sys sys.path.append('..') from .support import open_file class TestOfxWrite(TestCase): def test_write(self): with open_file('fidelity.ofx') as f: ofx = OfxParser.parse(f) self.assertEqual(str(ofx), "") def test_using_ofx_printer(self): with open_file('checking.ofx') as f: ofx = OfxParser.parse(f) fd, name = mkstemp() close(fd) printer = OfxPrinter(ofx=ofx, filename=name) printer.write(tabs=1) if __name__ == "__main__": import unittest unittest.main()
Fix warnings under Python 3
Fix warnings under Python 3
Python
mit
rdsteed/ofxparse,jaraco/ofxparse,jseutter/ofxparse,udibr/ofxparse
from __future__ import absolute_import from ofxparse import OfxParser as op, OfxPrinter from unittest import TestCase from os import close, remove from tempfile import mkstemp import sys sys.path.append('..') from .support import open_file class TestOfxWrite(TestCase): def test_write(self): test_file = open_file('fidelity.ofx') ofx_doc = op.parse(test_file) self.assertEqual(str(ofx_doc), "") def test_using_ofx_printer(self): test_file = open_file('checking.ofx') ofx_doc = op.parse(test_file) fd, name = mkstemp() close(fd) printer = OfxPrinter(ofx=ofx_doc, filename=name) printer.write(tabs=1) if __name__ == "__main__": import unittest unittest.main() Fix warnings under Python 3
from __future__ import absolute_import from ofxparse import OfxParser, OfxPrinter from unittest import TestCase from os import close, remove from tempfile import mkstemp import sys sys.path.append('..') from .support import open_file class TestOfxWrite(TestCase): def test_write(self): with open_file('fidelity.ofx') as f: ofx = OfxParser.parse(f) self.assertEqual(str(ofx), "") def test_using_ofx_printer(self): with open_file('checking.ofx') as f: ofx = OfxParser.parse(f) fd, name = mkstemp() close(fd) printer = OfxPrinter(ofx=ofx, filename=name) printer.write(tabs=1) if __name__ == "__main__": import unittest unittest.main()
<commit_before>from __future__ import absolute_import from ofxparse import OfxParser as op, OfxPrinter from unittest import TestCase from os import close, remove from tempfile import mkstemp import sys sys.path.append('..') from .support import open_file class TestOfxWrite(TestCase): def test_write(self): test_file = open_file('fidelity.ofx') ofx_doc = op.parse(test_file) self.assertEqual(str(ofx_doc), "") def test_using_ofx_printer(self): test_file = open_file('checking.ofx') ofx_doc = op.parse(test_file) fd, name = mkstemp() close(fd) printer = OfxPrinter(ofx=ofx_doc, filename=name) printer.write(tabs=1) if __name__ == "__main__": import unittest unittest.main() <commit_msg>Fix warnings under Python 3<commit_after>
from __future__ import absolute_import from ofxparse import OfxParser, OfxPrinter from unittest import TestCase from os import close, remove from tempfile import mkstemp import sys sys.path.append('..') from .support import open_file class TestOfxWrite(TestCase): def test_write(self): with open_file('fidelity.ofx') as f: ofx = OfxParser.parse(f) self.assertEqual(str(ofx), "") def test_using_ofx_printer(self): with open_file('checking.ofx') as f: ofx = OfxParser.parse(f) fd, name = mkstemp() close(fd) printer = OfxPrinter(ofx=ofx, filename=name) printer.write(tabs=1) if __name__ == "__main__": import unittest unittest.main()
from __future__ import absolute_import from ofxparse import OfxParser as op, OfxPrinter from unittest import TestCase from os import close, remove from tempfile import mkstemp import sys sys.path.append('..') from .support import open_file class TestOfxWrite(TestCase): def test_write(self): test_file = open_file('fidelity.ofx') ofx_doc = op.parse(test_file) self.assertEqual(str(ofx_doc), "") def test_using_ofx_printer(self): test_file = open_file('checking.ofx') ofx_doc = op.parse(test_file) fd, name = mkstemp() close(fd) printer = OfxPrinter(ofx=ofx_doc, filename=name) printer.write(tabs=1) if __name__ == "__main__": import unittest unittest.main() Fix warnings under Python 3from __future__ import absolute_import from ofxparse import OfxParser, OfxPrinter from unittest import TestCase from os import close, remove from tempfile import mkstemp import sys sys.path.append('..') from .support import open_file class TestOfxWrite(TestCase): def test_write(self): with open_file('fidelity.ofx') as f: ofx = OfxParser.parse(f) self.assertEqual(str(ofx), "") def test_using_ofx_printer(self): with open_file('checking.ofx') as f: ofx = OfxParser.parse(f) fd, name = mkstemp() close(fd) printer = OfxPrinter(ofx=ofx, filename=name) printer.write(tabs=1) if __name__ == "__main__": import unittest unittest.main()
<commit_before>from __future__ import absolute_import from ofxparse import OfxParser as op, OfxPrinter from unittest import TestCase from os import close, remove from tempfile import mkstemp import sys sys.path.append('..') from .support import open_file class TestOfxWrite(TestCase): def test_write(self): test_file = open_file('fidelity.ofx') ofx_doc = op.parse(test_file) self.assertEqual(str(ofx_doc), "") def test_using_ofx_printer(self): test_file = open_file('checking.ofx') ofx_doc = op.parse(test_file) fd, name = mkstemp() close(fd) printer = OfxPrinter(ofx=ofx_doc, filename=name) printer.write(tabs=1) if __name__ == "__main__": import unittest unittest.main() <commit_msg>Fix warnings under Python 3<commit_after>from __future__ import absolute_import from ofxparse import OfxParser, OfxPrinter from unittest import TestCase from os import close, remove from tempfile import mkstemp import sys sys.path.append('..') from .support import open_file class TestOfxWrite(TestCase): def test_write(self): with open_file('fidelity.ofx') as f: ofx = OfxParser.parse(f) self.assertEqual(str(ofx), "") def test_using_ofx_printer(self): with open_file('checking.ofx') as f: ofx = OfxParser.parse(f) fd, name = mkstemp() close(fd) printer = OfxPrinter(ofx=ofx, filename=name) printer.write(tabs=1) if __name__ == "__main__": import unittest unittest.main()
491613b34cb3c89e8d49670457a46b924a109529
pypinksign/__init__.py
pypinksign/__init__.py
""" Basic Template system for project pinksign, similar to the template part of PasteScript but without any dependencies. """ from .pypinksign import ( PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator, bit2string, separate_p12_into_npki )
""" Basic Template system for project pinksign, similar to the template part of PasteScript but without any dependencies. """ from .pypinksign import ( PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator, bit2string, separate_p12_into_npki, encrypt_decrypted_prikey )
Add new function in init
Add new function in init
Python
mit
bandoche/PyPinkSign
""" Basic Template system for project pinksign, similar to the template part of PasteScript but without any dependencies. """ from .pypinksign import ( PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator, bit2string, separate_p12_into_npki ) Add new function in init
""" Basic Template system for project pinksign, similar to the template part of PasteScript but without any dependencies. """ from .pypinksign import ( PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator, bit2string, separate_p12_into_npki, encrypt_decrypted_prikey )
<commit_before>""" Basic Template system for project pinksign, similar to the template part of PasteScript but without any dependencies. """ from .pypinksign import ( PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator, bit2string, separate_p12_into_npki ) <commit_msg>Add new function in init<commit_after>
""" Basic Template system for project pinksign, similar to the template part of PasteScript but without any dependencies. """ from .pypinksign import ( PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator, bit2string, separate_p12_into_npki, encrypt_decrypted_prikey )
""" Basic Template system for project pinksign, similar to the template part of PasteScript but without any dependencies. """ from .pypinksign import ( PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator, bit2string, separate_p12_into_npki ) Add new function in init""" Basic Template system for project pinksign, similar to the template part of PasteScript but without any dependencies. """ from .pypinksign import ( PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator, bit2string, separate_p12_into_npki, encrypt_decrypted_prikey )
<commit_before>""" Basic Template system for project pinksign, similar to the template part of PasteScript but without any dependencies. """ from .pypinksign import ( PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator, bit2string, separate_p12_into_npki ) <commit_msg>Add new function in init<commit_after>""" Basic Template system for project pinksign, similar to the template part of PasteScript but without any dependencies. """ from .pypinksign import ( PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator, bit2string, separate_p12_into_npki, encrypt_decrypted_prikey )
d7534dc3536ebe035abf063d83aa8d471cdadb16
python/pyqt_version.py
python/pyqt_version.py
import PySide2.QtCore # Prints PySide2 version # e.g. 5.11.1a1 print(PySide2.__version__) # Gets a tuple with each version component # e.g. (5, 11, 1, 'a', 1) print(PySide2.__version_info__) # Prints the Qt version used to compile PySide2 # e.g. "5.11.2" print(PySide2.QtCore.__version__) # Gets a tuple with each version components of Qt used to compile PySide2 # e.g. (5, 11, 2) print(PySide2.QtCore.__version_info__) print(PySide2.QtCore.qVersion())
#!/usr/bin/env python3 # coding: utf-8 ''' PySide2 ''' import sys try: import PySide2.QtCore except ImportError: print('cannot load module: PySide2.QtCore') sys.exit(1) # Prints PySide2 version # e.g. 5.11.1a1 print(PySide2.__version__) # Gets a tuple with each version component # e.g. (5, 11, 1, 'a', 1) print(PySide2.__version_info__) # Prints the Qt version used to compile PySide2 # e.g. "5.11.2" print(PySide2.QtCore.__version__) # Gets a tuple with each version components of Qt used to compile PySide2 # e.g. (5, 11, 2) print(PySide2.QtCore.__version_info__) print(PySide2.QtCore.qVersion())
ADD try-except to handle import error
ADD try-except to handle import error
Python
mit
ericosur/myqt,ericosur/myqt,ericosur/myqt,ericosur/myqt,ericosur/myqt,ericosur/myqt
import PySide2.QtCore # Prints PySide2 version # e.g. 5.11.1a1 print(PySide2.__version__) # Gets a tuple with each version component # e.g. (5, 11, 1, 'a', 1) print(PySide2.__version_info__) # Prints the Qt version used to compile PySide2 # e.g. "5.11.2" print(PySide2.QtCore.__version__) # Gets a tuple with each version components of Qt used to compile PySide2 # e.g. (5, 11, 2) print(PySide2.QtCore.__version_info__) print(PySide2.QtCore.qVersion()) ADD try-except to handle import error
#!/usr/bin/env python3 # coding: utf-8 ''' PySide2 ''' import sys try: import PySide2.QtCore except ImportError: print('cannot load module: PySide2.QtCore') sys.exit(1) # Prints PySide2 version # e.g. 5.11.1a1 print(PySide2.__version__) # Gets a tuple with each version component # e.g. (5, 11, 1, 'a', 1) print(PySide2.__version_info__) # Prints the Qt version used to compile PySide2 # e.g. "5.11.2" print(PySide2.QtCore.__version__) # Gets a tuple with each version components of Qt used to compile PySide2 # e.g. (5, 11, 2) print(PySide2.QtCore.__version_info__) print(PySide2.QtCore.qVersion())
<commit_before>import PySide2.QtCore # Prints PySide2 version # e.g. 5.11.1a1 print(PySide2.__version__) # Gets a tuple with each version component # e.g. (5, 11, 1, 'a', 1) print(PySide2.__version_info__) # Prints the Qt version used to compile PySide2 # e.g. "5.11.2" print(PySide2.QtCore.__version__) # Gets a tuple with each version components of Qt used to compile PySide2 # e.g. (5, 11, 2) print(PySide2.QtCore.__version_info__) print(PySide2.QtCore.qVersion()) <commit_msg>ADD try-except to handle import error<commit_after>
#!/usr/bin/env python3 # coding: utf-8 ''' PySide2 ''' import sys try: import PySide2.QtCore except ImportError: print('cannot load module: PySide2.QtCore') sys.exit(1) # Prints PySide2 version # e.g. 5.11.1a1 print(PySide2.__version__) # Gets a tuple with each version component # e.g. (5, 11, 1, 'a', 1) print(PySide2.__version_info__) # Prints the Qt version used to compile PySide2 # e.g. "5.11.2" print(PySide2.QtCore.__version__) # Gets a tuple with each version components of Qt used to compile PySide2 # e.g. (5, 11, 2) print(PySide2.QtCore.__version_info__) print(PySide2.QtCore.qVersion())
import PySide2.QtCore # Prints PySide2 version # e.g. 5.11.1a1 print(PySide2.__version__) # Gets a tuple with each version component # e.g. (5, 11, 1, 'a', 1) print(PySide2.__version_info__) # Prints the Qt version used to compile PySide2 # e.g. "5.11.2" print(PySide2.QtCore.__version__) # Gets a tuple with each version components of Qt used to compile PySide2 # e.g. (5, 11, 2) print(PySide2.QtCore.__version_info__) print(PySide2.QtCore.qVersion()) ADD try-except to handle import error#!/usr/bin/env python3 # coding: utf-8 ''' PySide2 ''' import sys try: import PySide2.QtCore except ImportError: print('cannot load module: PySide2.QtCore') sys.exit(1) # Prints PySide2 version # e.g. 5.11.1a1 print(PySide2.__version__) # Gets a tuple with each version component # e.g. (5, 11, 1, 'a', 1) print(PySide2.__version_info__) # Prints the Qt version used to compile PySide2 # e.g. "5.11.2" print(PySide2.QtCore.__version__) # Gets a tuple with each version components of Qt used to compile PySide2 # e.g. (5, 11, 2) print(PySide2.QtCore.__version_info__) print(PySide2.QtCore.qVersion())
<commit_before>import PySide2.QtCore # Prints PySide2 version # e.g. 5.11.1a1 print(PySide2.__version__) # Gets a tuple with each version component # e.g. (5, 11, 1, 'a', 1) print(PySide2.__version_info__) # Prints the Qt version used to compile PySide2 # e.g. "5.11.2" print(PySide2.QtCore.__version__) # Gets a tuple with each version components of Qt used to compile PySide2 # e.g. (5, 11, 2) print(PySide2.QtCore.__version_info__) print(PySide2.QtCore.qVersion()) <commit_msg>ADD try-except to handle import error<commit_after>#!/usr/bin/env python3 # coding: utf-8 ''' PySide2 ''' import sys try: import PySide2.QtCore except ImportError: print('cannot load module: PySide2.QtCore') sys.exit(1) # Prints PySide2 version # e.g. 5.11.1a1 print(PySide2.__version__) # Gets a tuple with each version component # e.g. (5, 11, 1, 'a', 1) print(PySide2.__version_info__) # Prints the Qt version used to compile PySide2 # e.g. "5.11.2" print(PySide2.QtCore.__version__) # Gets a tuple with each version components of Qt used to compile PySide2 # e.g. (5, 11, 2) print(PySide2.QtCore.__version_info__) print(PySide2.QtCore.qVersion())
da67ce3f25a708b99cb81f17703e74965dbea960
rtrss/filestorage/httputil.py
rtrss/filestorage/httputil.py
import logging import time import requests from googleapiclient.errors import HttpError # Number of retries in case of API errors NUM_RETRIES = 3 # Delay between retry attempts, seconds RETRY_DELAY = 1 _logger = logging.getLogger(__name__) def is_retryable(exc): retryable_codes = [500, 502, 503, 504] """Returns True if exception is "retryable", eg. HTTP 503""" if issubclass(exc, requests.exceptions.RequestException): code = exc.response.status_code elif issubclass(exc, HttpError): code = exc.resp.status else: return False return code in retryable_codes def retry_on_exception( exceptions=(HttpError, requests.exceptions.RequestException), retryable=is_retryable, tries=NUM_RETRIES, delay=RETRY_DELAY): """Retry call if function raises retryable exception""" def wrap(f): def wrapped_f(*args, **kwargs): mtries = tries while mtries > 1: try: return f(*args, **kwargs) except exceptions as err: # Reraise if non-retryable error if not retryable(err): raise _logger.warn("Retrying in %.2f seconds ...", delay) time.sleep(delay) mtries -= 1 # Only one last try left return f(*args, **kwargs) return wrapped_f return wrap
import logging import time import requests from googleapiclient.errors import HttpError # Number of retries in case of API errors NUM_RETRIES = 3 # Delay between retry attempts, seconds RETRY_DELAY = 1 _logger = logging.getLogger(__name__) def is_retryable(exc): retryable_codes = [500, 502, 503, 504] """Returns True if exception is "retryable", eg. HTTP 503""" if isinstance(exc, requests.exceptions.RequestException): code = exc.response.status_code elif isinstance(exc, HttpError): code = exc.resp.status else: return False return code in retryable_codes def retry_on_exception( retryable=is_retryable, tries=NUM_RETRIES, delay=RETRY_DELAY): """Retry call if function raises retryable exception""" def wrap(f): def wrapped_f(*args, **kwargs): mtries = tries while mtries > 1: try: return f(*args, **kwargs) except Exception as err: # Re-raise if non-retryable error if not retryable(err): raise _logger.warn("Retrying in %.2f seconds ...", delay) time.sleep(delay) mtries -= 1 # Only one last try left return f(*args, **kwargs) return wrapped_f return wrap
Remove unnecessary parameter, fix type detection bug
Remove unnecessary parameter, fix type detection bug
Python
apache-2.0
notapresent/rtrss,notapresent/rtrss,notapresent/rtrss,notapresent/rtrss
import logging import time import requests from googleapiclient.errors import HttpError # Number of retries in case of API errors NUM_RETRIES = 3 # Delay between retry attempts, seconds RETRY_DELAY = 1 _logger = logging.getLogger(__name__) def is_retryable(exc): retryable_codes = [500, 502, 503, 504] """Returns True if exception is "retryable", eg. HTTP 503""" if issubclass(exc, requests.exceptions.RequestException): code = exc.response.status_code elif issubclass(exc, HttpError): code = exc.resp.status else: return False return code in retryable_codes def retry_on_exception( exceptions=(HttpError, requests.exceptions.RequestException), retryable=is_retryable, tries=NUM_RETRIES, delay=RETRY_DELAY): """Retry call if function raises retryable exception""" def wrap(f): def wrapped_f(*args, **kwargs): mtries = tries while mtries > 1: try: return f(*args, **kwargs) except exceptions as err: # Reraise if non-retryable error if not retryable(err): raise _logger.warn("Retrying in %.2f seconds ...", delay) time.sleep(delay) mtries -= 1 # Only one last try left return f(*args, **kwargs) return wrapped_f return wrap Remove unnecessary parameter, fix type detection bug
import logging import time import requests from googleapiclient.errors import HttpError # Number of retries in case of API errors NUM_RETRIES = 3 # Delay between retry attempts, seconds RETRY_DELAY = 1 _logger = logging.getLogger(__name__) def is_retryable(exc): retryable_codes = [500, 502, 503, 504] """Returns True if exception is "retryable", eg. HTTP 503""" if isinstance(exc, requests.exceptions.RequestException): code = exc.response.status_code elif isinstance(exc, HttpError): code = exc.resp.status else: return False return code in retryable_codes def retry_on_exception( retryable=is_retryable, tries=NUM_RETRIES, delay=RETRY_DELAY): """Retry call if function raises retryable exception""" def wrap(f): def wrapped_f(*args, **kwargs): mtries = tries while mtries > 1: try: return f(*args, **kwargs) except Exception as err: # Re-raise if non-retryable error if not retryable(err): raise _logger.warn("Retrying in %.2f seconds ...", delay) time.sleep(delay) mtries -= 1 # Only one last try left return f(*args, **kwargs) return wrapped_f return wrap
<commit_before>import logging import time import requests from googleapiclient.errors import HttpError # Number of retries in case of API errors NUM_RETRIES = 3 # Delay between retry attempts, seconds RETRY_DELAY = 1 _logger = logging.getLogger(__name__) def is_retryable(exc): retryable_codes = [500, 502, 503, 504] """Returns True if exception is "retryable", eg. HTTP 503""" if issubclass(exc, requests.exceptions.RequestException): code = exc.response.status_code elif issubclass(exc, HttpError): code = exc.resp.status else: return False return code in retryable_codes def retry_on_exception( exceptions=(HttpError, requests.exceptions.RequestException), retryable=is_retryable, tries=NUM_RETRIES, delay=RETRY_DELAY): """Retry call if function raises retryable exception""" def wrap(f): def wrapped_f(*args, **kwargs): mtries = tries while mtries > 1: try: return f(*args, **kwargs) except exceptions as err: # Reraise if non-retryable error if not retryable(err): raise _logger.warn("Retrying in %.2f seconds ...", delay) time.sleep(delay) mtries -= 1 # Only one last try left return f(*args, **kwargs) return wrapped_f return wrap <commit_msg>Remove unnecessary parameter, fix type detection bug<commit_after>
import logging import time import requests from googleapiclient.errors import HttpError # Number of retries in case of API errors NUM_RETRIES = 3 # Delay between retry attempts, seconds RETRY_DELAY = 1 _logger = logging.getLogger(__name__) def is_retryable(exc): retryable_codes = [500, 502, 503, 504] """Returns True if exception is "retryable", eg. HTTP 503""" if isinstance(exc, requests.exceptions.RequestException): code = exc.response.status_code elif isinstance(exc, HttpError): code = exc.resp.status else: return False return code in retryable_codes def retry_on_exception( retryable=is_retryable, tries=NUM_RETRIES, delay=RETRY_DELAY): """Retry call if function raises retryable exception""" def wrap(f): def wrapped_f(*args, **kwargs): mtries = tries while mtries > 1: try: return f(*args, **kwargs) except Exception as err: # Re-raise if non-retryable error if not retryable(err): raise _logger.warn("Retrying in %.2f seconds ...", delay) time.sleep(delay) mtries -= 1 # Only one last try left return f(*args, **kwargs) return wrapped_f return wrap
import logging import time import requests from googleapiclient.errors import HttpError # Number of retries in case of API errors NUM_RETRIES = 3 # Delay between retry attempts, seconds RETRY_DELAY = 1 _logger = logging.getLogger(__name__) def is_retryable(exc): retryable_codes = [500, 502, 503, 504] """Returns True if exception is "retryable", eg. HTTP 503""" if issubclass(exc, requests.exceptions.RequestException): code = exc.response.status_code elif issubclass(exc, HttpError): code = exc.resp.status else: return False return code in retryable_codes def retry_on_exception( exceptions=(HttpError, requests.exceptions.RequestException), retryable=is_retryable, tries=NUM_RETRIES, delay=RETRY_DELAY): """Retry call if function raises retryable exception""" def wrap(f): def wrapped_f(*args, **kwargs): mtries = tries while mtries > 1: try: return f(*args, **kwargs) except exceptions as err: # Reraise if non-retryable error if not retryable(err): raise _logger.warn("Retrying in %.2f seconds ...", delay) time.sleep(delay) mtries -= 1 # Only one last try left return f(*args, **kwargs) return wrapped_f return wrap Remove unnecessary parameter, fix type detection bugimport logging import time import requests from googleapiclient.errors import HttpError # Number of retries in case of API errors NUM_RETRIES = 3 # Delay between retry attempts, seconds RETRY_DELAY = 1 _logger = logging.getLogger(__name__) def is_retryable(exc): retryable_codes = [500, 502, 503, 504] """Returns True if exception is "retryable", eg. HTTP 503""" if isinstance(exc, requests.exceptions.RequestException): code = exc.response.status_code elif isinstance(exc, HttpError): code = exc.resp.status else: return False return code in retryable_codes def retry_on_exception( retryable=is_retryable, tries=NUM_RETRIES, delay=RETRY_DELAY): """Retry call if function raises retryable exception""" def wrap(f): def wrapped_f(*args, **kwargs): mtries = tries while mtries > 1: try: return f(*args, **kwargs) except Exception as err: # Re-raise if non-retryable error if not retryable(err): raise _logger.warn("Retrying in %.2f seconds ...", delay) time.sleep(delay) mtries -= 1 # Only one last try left return f(*args, **kwargs) return wrapped_f return wrap
<commit_before>import logging import time import requests from googleapiclient.errors import HttpError # Number of retries in case of API errors NUM_RETRIES = 3 # Delay between retry attempts, seconds RETRY_DELAY = 1 _logger = logging.getLogger(__name__) def is_retryable(exc): retryable_codes = [500, 502, 503, 504] """Returns True if exception is "retryable", eg. HTTP 503""" if issubclass(exc, requests.exceptions.RequestException): code = exc.response.status_code elif issubclass(exc, HttpError): code = exc.resp.status else: return False return code in retryable_codes def retry_on_exception( exceptions=(HttpError, requests.exceptions.RequestException), retryable=is_retryable, tries=NUM_RETRIES, delay=RETRY_DELAY): """Retry call if function raises retryable exception""" def wrap(f): def wrapped_f(*args, **kwargs): mtries = tries while mtries > 1: try: return f(*args, **kwargs) except exceptions as err: # Reraise if non-retryable error if not retryable(err): raise _logger.warn("Retrying in %.2f seconds ...", delay) time.sleep(delay) mtries -= 1 # Only one last try left return f(*args, **kwargs) return wrapped_f return wrap <commit_msg>Remove unnecessary parameter, fix type detection bug<commit_after>import logging import time import requests from googleapiclient.errors import HttpError # Number of retries in case of API errors NUM_RETRIES = 3 # Delay between retry attempts, seconds RETRY_DELAY = 1 _logger = logging.getLogger(__name__) def is_retryable(exc): retryable_codes = [500, 502, 503, 504] """Returns True if exception is "retryable", eg. HTTP 503""" if isinstance(exc, requests.exceptions.RequestException): code = exc.response.status_code elif isinstance(exc, HttpError): code = exc.resp.status else: return False return code in retryable_codes def retry_on_exception( retryable=is_retryable, tries=NUM_RETRIES, delay=RETRY_DELAY): """Retry call if function raises retryable exception""" def wrap(f): def wrapped_f(*args, **kwargs): mtries = tries while mtries > 1: try: return f(*args, **kwargs) except Exception as err: # Re-raise if non-retryable error if not retryable(err): raise _logger.warn("Retrying in %.2f seconds ...", delay) time.sleep(delay) mtries -= 1 # Only one last try left return f(*args, **kwargs) return wrapped_f return wrap
5a3cdba13cb4898b76d83c949fc3ab6895c267ff
scripts/cpuflags.py
scripts/cpuflags.py
import errno class CPUFlags: def __init__(self): self.flags = set() try: self.flags = self.__parse_cpuinfo() except IOError as e: if e.errno == errno.ENOENT: return raise def __contains__(self, name): return name in self.flags def __parse_cpuinfo(self): def get_flags(): with open('/proc/cpuinfo', 'r') as f: for line in f: if line.startswith('flags'): return line line = get_flags().split() del line[:2] # remove tokens "flags", ":" return set(line) def main(): import sys flags = CPUFlags() if len(sys.argv) == 2: if sys.argv[1] in flags: print "present" if __name__ == '__main__': main()
import errno import sys class CPUFlags: def __init__(self): self.flags = set() try: self.flags = self.__parse_cpuinfo() except IOError as e: if e.errno == errno.ENOENT: return raise def __contains__(self, name): return name in self.flags def __parse_cpuinfo(self): def get_flags(): with open('/proc/cpuinfo', 'r') as f: for line in f: if line.startswith('flags'): return line line = get_flags().split() del line[:2] # remove tokens "flags", ":" return set(line) def main(): import sys flags = CPUFlags() if len(sys.argv) == 2: if sys.argv[1] in flags: print "present" return 0 return 1 if __name__ == '__main__': sys.exit(main())
Check script returns an exit code
Check script returns an exit code
Python
bsd-2-clause
WojciechMula/toys,WojciechMula/toys,WojciechMula/toys,WojciechMula/toys,WojciechMula/toys
import errno class CPUFlags: def __init__(self): self.flags = set() try: self.flags = self.__parse_cpuinfo() except IOError as e: if e.errno == errno.ENOENT: return raise def __contains__(self, name): return name in self.flags def __parse_cpuinfo(self): def get_flags(): with open('/proc/cpuinfo', 'r') as f: for line in f: if line.startswith('flags'): return line line = get_flags().split() del line[:2] # remove tokens "flags", ":" return set(line) def main(): import sys flags = CPUFlags() if len(sys.argv) == 2: if sys.argv[1] in flags: print "present" if __name__ == '__main__': main() Check script returns an exit code
import errno import sys class CPUFlags: def __init__(self): self.flags = set() try: self.flags = self.__parse_cpuinfo() except IOError as e: if e.errno == errno.ENOENT: return raise def __contains__(self, name): return name in self.flags def __parse_cpuinfo(self): def get_flags(): with open('/proc/cpuinfo', 'r') as f: for line in f: if line.startswith('flags'): return line line = get_flags().split() del line[:2] # remove tokens "flags", ":" return set(line) def main(): import sys flags = CPUFlags() if len(sys.argv) == 2: if sys.argv[1] in flags: print "present" return 0 return 1 if __name__ == '__main__': sys.exit(main())
<commit_before>import errno class CPUFlags: def __init__(self): self.flags = set() try: self.flags = self.__parse_cpuinfo() except IOError as e: if e.errno == errno.ENOENT: return raise def __contains__(self, name): return name in self.flags def __parse_cpuinfo(self): def get_flags(): with open('/proc/cpuinfo', 'r') as f: for line in f: if line.startswith('flags'): return line line = get_flags().split() del line[:2] # remove tokens "flags", ":" return set(line) def main(): import sys flags = CPUFlags() if len(sys.argv) == 2: if sys.argv[1] in flags: print "present" if __name__ == '__main__': main() <commit_msg>Check script returns an exit code<commit_after>
import errno import sys class CPUFlags: def __init__(self): self.flags = set() try: self.flags = self.__parse_cpuinfo() except IOError as e: if e.errno == errno.ENOENT: return raise def __contains__(self, name): return name in self.flags def __parse_cpuinfo(self): def get_flags(): with open('/proc/cpuinfo', 'r') as f: for line in f: if line.startswith('flags'): return line line = get_flags().split() del line[:2] # remove tokens "flags", ":" return set(line) def main(): import sys flags = CPUFlags() if len(sys.argv) == 2: if sys.argv[1] in flags: print "present" return 0 return 1 if __name__ == '__main__': sys.exit(main())
import errno class CPUFlags: def __init__(self): self.flags = set() try: self.flags = self.__parse_cpuinfo() except IOError as e: if e.errno == errno.ENOENT: return raise def __contains__(self, name): return name in self.flags def __parse_cpuinfo(self): def get_flags(): with open('/proc/cpuinfo', 'r') as f: for line in f: if line.startswith('flags'): return line line = get_flags().split() del line[:2] # remove tokens "flags", ":" return set(line) def main(): import sys flags = CPUFlags() if len(sys.argv) == 2: if sys.argv[1] in flags: print "present" if __name__ == '__main__': main() Check script returns an exit codeimport errno import sys class CPUFlags: def __init__(self): self.flags = set() try: self.flags = self.__parse_cpuinfo() except IOError as e: if e.errno == errno.ENOENT: return raise def __contains__(self, name): return name in self.flags def __parse_cpuinfo(self): def get_flags(): with open('/proc/cpuinfo', 'r') as f: for line in f: if line.startswith('flags'): return line line = get_flags().split() del line[:2] # remove tokens "flags", ":" return set(line) def main(): import sys flags = CPUFlags() if len(sys.argv) == 2: if sys.argv[1] in flags: print "present" return 0 return 1 if __name__ == '__main__': sys.exit(main())
<commit_before>import errno class CPUFlags: def __init__(self): self.flags = set() try: self.flags = self.__parse_cpuinfo() except IOError as e: if e.errno == errno.ENOENT: return raise def __contains__(self, name): return name in self.flags def __parse_cpuinfo(self): def get_flags(): with open('/proc/cpuinfo', 'r') as f: for line in f: if line.startswith('flags'): return line line = get_flags().split() del line[:2] # remove tokens "flags", ":" return set(line) def main(): import sys flags = CPUFlags() if len(sys.argv) == 2: if sys.argv[1] in flags: print "present" if __name__ == '__main__': main() <commit_msg>Check script returns an exit code<commit_after>import errno import sys class CPUFlags: def __init__(self): self.flags = set() try: self.flags = self.__parse_cpuinfo() except IOError as e: if e.errno == errno.ENOENT: return raise def __contains__(self, name): return name in self.flags def __parse_cpuinfo(self): def get_flags(): with open('/proc/cpuinfo', 'r') as f: for line in f: if line.startswith('flags'): return line line = get_flags().split() del line[:2] # remove tokens "flags", ":" return set(line) def main(): import sys flags = CPUFlags() if len(sys.argv) == 2: if sys.argv[1] in flags: print "present" return 0 return 1 if __name__ == '__main__': sys.exit(main())
ea62a1cd9642dbff69cbfae3f8b540604a8a8fca
mine/__init__.py
mine/__init__.py
#!/usr/bin/env python """Package for mine.""" import sys __project__ = 'mine' __version__ = '0.1' CLI = 'mine' VERSION = __project__ + '-' + __version__ DESCRIPTION = "Manages running applications across multiple computers." PYTHON_VERSION = 3, 3 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
#!/usr/bin/env python """Package for mine.""" import sys __project__ = 'mine' __version__ = '0.1' CLI = 'mine' VERSION = __project__ + '-' + __version__ DESCRIPTION = "For applications that haven't learned to share." PYTHON_VERSION = 3, 3 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
Update headline to match GitHub
Update headline to match GitHub
Python
mit
jacebrowning/mine
#!/usr/bin/env python """Package for mine.""" import sys __project__ = 'mine' __version__ = '0.1' CLI = 'mine' VERSION = __project__ + '-' + __version__ DESCRIPTION = "Manages running applications across multiple computers." PYTHON_VERSION = 3, 3 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION)) Update headline to match GitHub
#!/usr/bin/env python """Package for mine.""" import sys __project__ = 'mine' __version__ = '0.1' CLI = 'mine' VERSION = __project__ + '-' + __version__ DESCRIPTION = "For applications that haven't learned to share." PYTHON_VERSION = 3, 3 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
<commit_before>#!/usr/bin/env python """Package for mine.""" import sys __project__ = 'mine' __version__ = '0.1' CLI = 'mine' VERSION = __project__ + '-' + __version__ DESCRIPTION = "Manages running applications across multiple computers." PYTHON_VERSION = 3, 3 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION)) <commit_msg>Update headline to match GitHub<commit_after>
#!/usr/bin/env python """Package for mine.""" import sys __project__ = 'mine' __version__ = '0.1' CLI = 'mine' VERSION = __project__ + '-' + __version__ DESCRIPTION = "For applications that haven't learned to share." PYTHON_VERSION = 3, 3 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
#!/usr/bin/env python """Package for mine.""" import sys __project__ = 'mine' __version__ = '0.1' CLI = 'mine' VERSION = __project__ + '-' + __version__ DESCRIPTION = "Manages running applications across multiple computers." PYTHON_VERSION = 3, 3 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION)) Update headline to match GitHub#!/usr/bin/env python """Package for mine.""" import sys __project__ = 'mine' __version__ = '0.1' CLI = 'mine' VERSION = __project__ + '-' + __version__ DESCRIPTION = "For applications that haven't learned to share." PYTHON_VERSION = 3, 3 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
<commit_before>#!/usr/bin/env python """Package for mine.""" import sys __project__ = 'mine' __version__ = '0.1' CLI = 'mine' VERSION = __project__ + '-' + __version__ DESCRIPTION = "Manages running applications across multiple computers." PYTHON_VERSION = 3, 3 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION)) <commit_msg>Update headline to match GitHub<commit_after>#!/usr/bin/env python """Package for mine.""" import sys __project__ = 'mine' __version__ = '0.1' CLI = 'mine' VERSION = __project__ + '-' + __version__ DESCRIPTION = "For applications that haven't learned to share." PYTHON_VERSION = 3, 3 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
f95460070e80e1d83394fa6ed08bad9dad34802c
ovp_users/emails.py
ovp_users/emails.py
from ovp_core.emails import BaseMail class UserMail(BaseMail): """ This class is responsible for firing emails for Users """ def __init__(self, user, async_mail=None): super(UserMail, self).__init__(user.email, async_mail) def sendWelcome(self, context={}): """ Sent when user registers """ return self.sendEmail('welcome', 'Welcome', context) def sendRecoveryToken(self, context): """ Sent when volunteer requests recovery token """ context.update({ 'user_email': self.email_address }) return self.sendEmail('recoveryToken', 'Password recovery', context)
from ovp_core.emails import BaseMail class UserMail(BaseMail): """ This class is responsible for firing emails for Users """ def __init__(self, user, async_mail=None): super(UserMail, self).__init__(user.email, async_mail) def sendWelcome(self, context={}): """ Sent when user registers """ return self.sendEmail('welcome', 'Welcome', context) def sendRecoveryToken(self, context): """ Sent when volunteer requests recovery token """ context.update({ 'user_email': user.email }) return self.sendEmail('recoveryToken', 'Password recovery', context)
Revert "fix getting user email for recoveryToken"
Revert "fix getting user email for recoveryToken" This reverts commit a47b098e4d644391213958f9e05c179a7410208d.
Python
agpl-3.0
OpenVolunteeringPlatform/django-ovp-users,OpenVolunteeringPlatform/django-ovp-users
from ovp_core.emails import BaseMail class UserMail(BaseMail): """ This class is responsible for firing emails for Users """ def __init__(self, user, async_mail=None): super(UserMail, self).__init__(user.email, async_mail) def sendWelcome(self, context={}): """ Sent when user registers """ return self.sendEmail('welcome', 'Welcome', context) def sendRecoveryToken(self, context): """ Sent when volunteer requests recovery token """ context.update({ 'user_email': self.email_address }) return self.sendEmail('recoveryToken', 'Password recovery', context) Revert "fix getting user email for recoveryToken" This reverts commit a47b098e4d644391213958f9e05c179a7410208d.
from ovp_core.emails import BaseMail class UserMail(BaseMail): """ This class is responsible for firing emails for Users """ def __init__(self, user, async_mail=None): super(UserMail, self).__init__(user.email, async_mail) def sendWelcome(self, context={}): """ Sent when user registers """ return self.sendEmail('welcome', 'Welcome', context) def sendRecoveryToken(self, context): """ Sent when volunteer requests recovery token """ context.update({ 'user_email': user.email }) return self.sendEmail('recoveryToken', 'Password recovery', context)
<commit_before>from ovp_core.emails import BaseMail class UserMail(BaseMail): """ This class is responsible for firing emails for Users """ def __init__(self, user, async_mail=None): super(UserMail, self).__init__(user.email, async_mail) def sendWelcome(self, context={}): """ Sent when user registers """ return self.sendEmail('welcome', 'Welcome', context) def sendRecoveryToken(self, context): """ Sent when volunteer requests recovery token """ context.update({ 'user_email': self.email_address }) return self.sendEmail('recoveryToken', 'Password recovery', context) <commit_msg>Revert "fix getting user email for recoveryToken" This reverts commit a47b098e4d644391213958f9e05c179a7410208d.<commit_after>
from ovp_core.emails import BaseMail class UserMail(BaseMail): """ This class is responsible for firing emails for Users """ def __init__(self, user, async_mail=None): super(UserMail, self).__init__(user.email, async_mail) def sendWelcome(self, context={}): """ Sent when user registers """ return self.sendEmail('welcome', 'Welcome', context) def sendRecoveryToken(self, context): """ Sent when volunteer requests recovery token """ context.update({ 'user_email': user.email }) return self.sendEmail('recoveryToken', 'Password recovery', context)
from ovp_core.emails import BaseMail class UserMail(BaseMail): """ This class is responsible for firing emails for Users """ def __init__(self, user, async_mail=None): super(UserMail, self).__init__(user.email, async_mail) def sendWelcome(self, context={}): """ Sent when user registers """ return self.sendEmail('welcome', 'Welcome', context) def sendRecoveryToken(self, context): """ Sent when volunteer requests recovery token """ context.update({ 'user_email': self.email_address }) return self.sendEmail('recoveryToken', 'Password recovery', context) Revert "fix getting user email for recoveryToken" This reverts commit a47b098e4d644391213958f9e05c179a7410208d.from ovp_core.emails import BaseMail class UserMail(BaseMail): """ This class is responsible for firing emails for Users """ def __init__(self, user, async_mail=None): super(UserMail, self).__init__(user.email, async_mail) def sendWelcome(self, context={}): """ Sent when user registers """ return self.sendEmail('welcome', 'Welcome', context) def sendRecoveryToken(self, context): """ Sent when volunteer requests recovery token """ context.update({ 'user_email': user.email }) return self.sendEmail('recoveryToken', 'Password recovery', context)
<commit_before>from ovp_core.emails import BaseMail class UserMail(BaseMail): """ This class is responsible for firing emails for Users """ def __init__(self, user, async_mail=None): super(UserMail, self).__init__(user.email, async_mail) def sendWelcome(self, context={}): """ Sent when user registers """ return self.sendEmail('welcome', 'Welcome', context) def sendRecoveryToken(self, context): """ Sent when volunteer requests recovery token """ context.update({ 'user_email': self.email_address }) return self.sendEmail('recoveryToken', 'Password recovery', context) <commit_msg>Revert "fix getting user email for recoveryToken" This reverts commit a47b098e4d644391213958f9e05c179a7410208d.<commit_after>from ovp_core.emails import BaseMail class UserMail(BaseMail): """ This class is responsible for firing emails for Users """ def __init__(self, user, async_mail=None): super(UserMail, self).__init__(user.email, async_mail) def sendWelcome(self, context={}): """ Sent when user registers """ return self.sendEmail('welcome', 'Welcome', context) def sendRecoveryToken(self, context): """ Sent when volunteer requests recovery token """ context.update({ 'user_email': user.email }) return self.sendEmail('recoveryToken', 'Password recovery', context)
839a0cafca1d172f7a061dcec5f6a4eca6d725c8
superlists/lists/tests.py
superlists/lists/tests.py
from django.test import TestCase # Create your tests here.
from django.test import TestCase class SmokeTest(TestCase): def test_bad_maths(self): self.assertEqual(1 + 1, 3)
Add app for lists, with deliberately failing unit test
Add app for lists, with deliberately failing unit test
Python
apache-2.0
Alfawuhn/test-driven-python
from django.test import TestCase # Create your tests here. Add app for lists, with deliberately failing unit test
from django.test import TestCase class SmokeTest(TestCase): def test_bad_maths(self): self.assertEqual(1 + 1, 3)
<commit_before>from django.test import TestCase # Create your tests here. <commit_msg>Add app for lists, with deliberately failing unit test<commit_after>
from django.test import TestCase class SmokeTest(TestCase): def test_bad_maths(self): self.assertEqual(1 + 1, 3)
from django.test import TestCase # Create your tests here. Add app for lists, with deliberately failing unit testfrom django.test import TestCase class SmokeTest(TestCase): def test_bad_maths(self): self.assertEqual(1 + 1, 3)
<commit_before>from django.test import TestCase # Create your tests here. <commit_msg>Add app for lists, with deliberately failing unit test<commit_after>from django.test import TestCase class SmokeTest(TestCase): def test_bad_maths(self): self.assertEqual(1 + 1, 3)
13d9cf933e49849a3c5343e7bdbf887b9aee6097
busbus/entity.py
busbus/entity.py
from busbus import util class LazyEntityProperty(object): def __init__(self, f, *args, **kwargs): self.f = f self.args = args self.kwargs = kwargs def __call__(self): return self.f(*self.args, **self.kwargs) class BaseEntity(object): def __init__(self, provider, **kwargs): self._provider = provider self._lazy_properties = {} for attr in getattr(self, '__attrs__', []): if isinstance(kwargs.get(attr, None), LazyEntityProperty): self._lazy_properties[attr] = kwargs[attr] else: setattr(self, attr, kwargs.get(attr, None)) provider._new_entity(self) def __repr__(self, args=['id']): return u'<{0}({1})>'.format( util.clsname(self), ','.join('{0}={1!r}'.format(i, getattr(self, i)) for i in args)) def __getattr__(self, name): if name in self._lazy_properties: value = self._lazy_properties[name]() del self._lazy_properties[name] setattr(self, name, value) return value else: raise AttributeError(name) def to_dict(self): return dict((attr, getattr(self, attr)) for attr in self.__attrs__ if getattr(self, attr))
from busbus import util class LazyEntityProperty(object): def __init__(self, f, *args, **kwargs): self.f = f self.args = args self.kwargs = kwargs def __call__(self): return self.f(*self.args, **self.kwargs) class BaseEntity(object): __repr_attrs__ = ('id',) def __init__(self, provider, **kwargs): self._provider = provider self._lazy_properties = {} for attr in getattr(self, '__attrs__', []): if isinstance(kwargs.get(attr, None), LazyEntityProperty): self._lazy_properties[attr] = kwargs[attr] else: setattr(self, attr, kwargs.get(attr, None)) provider._new_entity(self) def __repr__(self): return u'<{0}({1})>'.format( util.clsname(self), ','.join( '{0}={1!r}'.format(i, getattr(self, i)) for i in self.__repr_attrs__)) def __getattr__(self, name): if name in self._lazy_properties: value = self._lazy_properties[name]() del self._lazy_properties[name] setattr(self, name, value) return value else: raise AttributeError(name) def to_dict(self): return dict((attr, getattr(self, attr)) for attr in self.__attrs__ if getattr(self, attr))
Use an instance variable instead of a non-standard argument to __repr__
Use an instance variable instead of a non-standard argument to __repr__
Python
mit
spaceboats/busbus
from busbus import util class LazyEntityProperty(object): def __init__(self, f, *args, **kwargs): self.f = f self.args = args self.kwargs = kwargs def __call__(self): return self.f(*self.args, **self.kwargs) class BaseEntity(object): def __init__(self, provider, **kwargs): self._provider = provider self._lazy_properties = {} for attr in getattr(self, '__attrs__', []): if isinstance(kwargs.get(attr, None), LazyEntityProperty): self._lazy_properties[attr] = kwargs[attr] else: setattr(self, attr, kwargs.get(attr, None)) provider._new_entity(self) def __repr__(self, args=['id']): return u'<{0}({1})>'.format( util.clsname(self), ','.join('{0}={1!r}'.format(i, getattr(self, i)) for i in args)) def __getattr__(self, name): if name in self._lazy_properties: value = self._lazy_properties[name]() del self._lazy_properties[name] setattr(self, name, value) return value else: raise AttributeError(name) def to_dict(self): return dict((attr, getattr(self, attr)) for attr in self.__attrs__ if getattr(self, attr)) Use an instance variable instead of a non-standard argument to __repr__
from busbus import util class LazyEntityProperty(object): def __init__(self, f, *args, **kwargs): self.f = f self.args = args self.kwargs = kwargs def __call__(self): return self.f(*self.args, **self.kwargs) class BaseEntity(object): __repr_attrs__ = ('id',) def __init__(self, provider, **kwargs): self._provider = provider self._lazy_properties = {} for attr in getattr(self, '__attrs__', []): if isinstance(kwargs.get(attr, None), LazyEntityProperty): self._lazy_properties[attr] = kwargs[attr] else: setattr(self, attr, kwargs.get(attr, None)) provider._new_entity(self) def __repr__(self): return u'<{0}({1})>'.format( util.clsname(self), ','.join( '{0}={1!r}'.format(i, getattr(self, i)) for i in self.__repr_attrs__)) def __getattr__(self, name): if name in self._lazy_properties: value = self._lazy_properties[name]() del self._lazy_properties[name] setattr(self, name, value) return value else: raise AttributeError(name) def to_dict(self): return dict((attr, getattr(self, attr)) for attr in self.__attrs__ if getattr(self, attr))
<commit_before>from busbus import util class LazyEntityProperty(object): def __init__(self, f, *args, **kwargs): self.f = f self.args = args self.kwargs = kwargs def __call__(self): return self.f(*self.args, **self.kwargs) class BaseEntity(object): def __init__(self, provider, **kwargs): self._provider = provider self._lazy_properties = {} for attr in getattr(self, '__attrs__', []): if isinstance(kwargs.get(attr, None), LazyEntityProperty): self._lazy_properties[attr] = kwargs[attr] else: setattr(self, attr, kwargs.get(attr, None)) provider._new_entity(self) def __repr__(self, args=['id']): return u'<{0}({1})>'.format( util.clsname(self), ','.join('{0}={1!r}'.format(i, getattr(self, i)) for i in args)) def __getattr__(self, name): if name in self._lazy_properties: value = self._lazy_properties[name]() del self._lazy_properties[name] setattr(self, name, value) return value else: raise AttributeError(name) def to_dict(self): return dict((attr, getattr(self, attr)) for attr in self.__attrs__ if getattr(self, attr)) <commit_msg>Use an instance variable instead of a non-standard argument to __repr__<commit_after>
from busbus import util class LazyEntityProperty(object): def __init__(self, f, *args, **kwargs): self.f = f self.args = args self.kwargs = kwargs def __call__(self): return self.f(*self.args, **self.kwargs) class BaseEntity(object): __repr_attrs__ = ('id',) def __init__(self, provider, **kwargs): self._provider = provider self._lazy_properties = {} for attr in getattr(self, '__attrs__', []): if isinstance(kwargs.get(attr, None), LazyEntityProperty): self._lazy_properties[attr] = kwargs[attr] else: setattr(self, attr, kwargs.get(attr, None)) provider._new_entity(self) def __repr__(self): return u'<{0}({1})>'.format( util.clsname(self), ','.join( '{0}={1!r}'.format(i, getattr(self, i)) for i in self.__repr_attrs__)) def __getattr__(self, name): if name in self._lazy_properties: value = self._lazy_properties[name]() del self._lazy_properties[name] setattr(self, name, value) return value else: raise AttributeError(name) def to_dict(self): return dict((attr, getattr(self, attr)) for attr in self.__attrs__ if getattr(self, attr))
from busbus import util class LazyEntityProperty(object): def __init__(self, f, *args, **kwargs): self.f = f self.args = args self.kwargs = kwargs def __call__(self): return self.f(*self.args, **self.kwargs) class BaseEntity(object): def __init__(self, provider, **kwargs): self._provider = provider self._lazy_properties = {} for attr in getattr(self, '__attrs__', []): if isinstance(kwargs.get(attr, None), LazyEntityProperty): self._lazy_properties[attr] = kwargs[attr] else: setattr(self, attr, kwargs.get(attr, None)) provider._new_entity(self) def __repr__(self, args=['id']): return u'<{0}({1})>'.format( util.clsname(self), ','.join('{0}={1!r}'.format(i, getattr(self, i)) for i in args)) def __getattr__(self, name): if name in self._lazy_properties: value = self._lazy_properties[name]() del self._lazy_properties[name] setattr(self, name, value) return value else: raise AttributeError(name) def to_dict(self): return dict((attr, getattr(self, attr)) for attr in self.__attrs__ if getattr(self, attr)) Use an instance variable instead of a non-standard argument to __repr__from busbus import util class LazyEntityProperty(object): def __init__(self, f, *args, **kwargs): self.f = f self.args = args self.kwargs = kwargs def __call__(self): return self.f(*self.args, **self.kwargs) class BaseEntity(object): __repr_attrs__ = ('id',) def __init__(self, provider, **kwargs): self._provider = provider self._lazy_properties = {} for attr in getattr(self, '__attrs__', []): if isinstance(kwargs.get(attr, None), LazyEntityProperty): self._lazy_properties[attr] = kwargs[attr] else: setattr(self, attr, kwargs.get(attr, None)) provider._new_entity(self) def __repr__(self): return u'<{0}({1})>'.format( util.clsname(self), ','.join( '{0}={1!r}'.format(i, getattr(self, i)) for i in self.__repr_attrs__)) def __getattr__(self, name): if name in self._lazy_properties: value = self._lazy_properties[name]() del self._lazy_properties[name] setattr(self, name, value) return value else: raise AttributeError(name) def to_dict(self): return dict((attr, getattr(self, attr)) for attr in self.__attrs__ if getattr(self, attr))
<commit_before>from busbus import util class LazyEntityProperty(object): def __init__(self, f, *args, **kwargs): self.f = f self.args = args self.kwargs = kwargs def __call__(self): return self.f(*self.args, **self.kwargs) class BaseEntity(object): def __init__(self, provider, **kwargs): self._provider = provider self._lazy_properties = {} for attr in getattr(self, '__attrs__', []): if isinstance(kwargs.get(attr, None), LazyEntityProperty): self._lazy_properties[attr] = kwargs[attr] else: setattr(self, attr, kwargs.get(attr, None)) provider._new_entity(self) def __repr__(self, args=['id']): return u'<{0}({1})>'.format( util.clsname(self), ','.join('{0}={1!r}'.format(i, getattr(self, i)) for i in args)) def __getattr__(self, name): if name in self._lazy_properties: value = self._lazy_properties[name]() del self._lazy_properties[name] setattr(self, name, value) return value else: raise AttributeError(name) def to_dict(self): return dict((attr, getattr(self, attr)) for attr in self.__attrs__ if getattr(self, attr)) <commit_msg>Use an instance variable instead of a non-standard argument to __repr__<commit_after>from busbus import util class LazyEntityProperty(object): def __init__(self, f, *args, **kwargs): self.f = f self.args = args self.kwargs = kwargs def __call__(self): return self.f(*self.args, **self.kwargs) class BaseEntity(object): __repr_attrs__ = ('id',) def __init__(self, provider, **kwargs): self._provider = provider self._lazy_properties = {} for attr in getattr(self, '__attrs__', []): if isinstance(kwargs.get(attr, None), LazyEntityProperty): self._lazy_properties[attr] = kwargs[attr] else: setattr(self, attr, kwargs.get(attr, None)) provider._new_entity(self) def __repr__(self): return u'<{0}({1})>'.format( util.clsname(self), ','.join( '{0}={1!r}'.format(i, getattr(self, i)) for i in self.__repr_attrs__)) def __getattr__(self, name): if name in self._lazy_properties: value = self._lazy_properties[name]() del self._lazy_properties[name] setattr(self, name, value) return value else: raise AttributeError(name) def to_dict(self): return dict((attr, getattr(self, attr)) for attr in self.__attrs__ if getattr(self, attr))
5b6823ec19185ed5b413d1c01d3afeb5b1716778
taca/server_status/cli.py
taca/server_status/cli.py
import click import logging from taca.server_status import server_status as status from taca.utils.config import CONFIG from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception @click.group(name='server_status') def server_status(): """ Monitor server status """ if not CONFIG.get('server_status', ''): logging.warning("Configuration missing required entries: server_status") # server status subcommands @server_status.command() @click.option('--statusdb', is_flag=True, help="Update the statusdb") def nases(statusdb): """ Checks the available space on all the nases """ disk_space = status.get_nases_disk_space() if statusdb: status.update_status_db(disk_space, server_type='nas') @server_status.command() def cronjobs(): """ Monitors cronjobs and updates statusdb """ cj.update_cronjob_db()
import click import logging from taca.server_status import server_status as status from taca.utils.config import CONFIG from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception @click.group(name='server_status') def server_status(): """ Monitor server status """ # server status subcommands @server_status.command() @click.option('--statusdb', is_flag=True, help="Update the statusdb") def nases(statusdb): """ Checks the available space on all the nases """ if not CONFIG.get('server_status', ''): logging.warning("Configuration missing required entries: server_status") disk_space = status.get_nases_disk_space() if statusdb: status.update_status_db(disk_space, server_type='nas') @server_status.command() def cronjobs(): """ Monitors cronjobs and updates statusdb """ cj.update_cronjob_db()
Move warning about missing config entry to relevant subcommand
Move warning about missing config entry to relevant subcommand
Python
mit
SciLifeLab/TACA,SciLifeLab/TACA,SciLifeLab/TACA
import click import logging from taca.server_status import server_status as status from taca.utils.config import CONFIG from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception @click.group(name='server_status') def server_status(): """ Monitor server status """ if not CONFIG.get('server_status', ''): logging.warning("Configuration missing required entries: server_status") # server status subcommands @server_status.command() @click.option('--statusdb', is_flag=True, help="Update the statusdb") def nases(statusdb): """ Checks the available space on all the nases """ disk_space = status.get_nases_disk_space() if statusdb: status.update_status_db(disk_space, server_type='nas') @server_status.command() def cronjobs(): """ Monitors cronjobs and updates statusdb """ cj.update_cronjob_db() Move warning about missing config entry to relevant subcommand
import click import logging from taca.server_status import server_status as status from taca.utils.config import CONFIG from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception @click.group(name='server_status') def server_status(): """ Monitor server status """ # server status subcommands @server_status.command() @click.option('--statusdb', is_flag=True, help="Update the statusdb") def nases(statusdb): """ Checks the available space on all the nases """ if not CONFIG.get('server_status', ''): logging.warning("Configuration missing required entries: server_status") disk_space = status.get_nases_disk_space() if statusdb: status.update_status_db(disk_space, server_type='nas') @server_status.command() def cronjobs(): """ Monitors cronjobs and updates statusdb """ cj.update_cronjob_db()
<commit_before>import click import logging from taca.server_status import server_status as status from taca.utils.config import CONFIG from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception @click.group(name='server_status') def server_status(): """ Monitor server status """ if not CONFIG.get('server_status', ''): logging.warning("Configuration missing required entries: server_status") # server status subcommands @server_status.command() @click.option('--statusdb', is_flag=True, help="Update the statusdb") def nases(statusdb): """ Checks the available space on all the nases """ disk_space = status.get_nases_disk_space() if statusdb: status.update_status_db(disk_space, server_type='nas') @server_status.command() def cronjobs(): """ Monitors cronjobs and updates statusdb """ cj.update_cronjob_db() <commit_msg>Move warning about missing config entry to relevant subcommand<commit_after>
import click import logging from taca.server_status import server_status as status from taca.utils.config import CONFIG from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception @click.group(name='server_status') def server_status(): """ Monitor server status """ # server status subcommands @server_status.command() @click.option('--statusdb', is_flag=True, help="Update the statusdb") def nases(statusdb): """ Checks the available space on all the nases """ if not CONFIG.get('server_status', ''): logging.warning("Configuration missing required entries: server_status") disk_space = status.get_nases_disk_space() if statusdb: status.update_status_db(disk_space, server_type='nas') @server_status.command() def cronjobs(): """ Monitors cronjobs and updates statusdb """ cj.update_cronjob_db()
import click import logging from taca.server_status import server_status as status from taca.utils.config import CONFIG from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception @click.group(name='server_status') def server_status(): """ Monitor server status """ if not CONFIG.get('server_status', ''): logging.warning("Configuration missing required entries: server_status") # server status subcommands @server_status.command() @click.option('--statusdb', is_flag=True, help="Update the statusdb") def nases(statusdb): """ Checks the available space on all the nases """ disk_space = status.get_nases_disk_space() if statusdb: status.update_status_db(disk_space, server_type='nas') @server_status.command() def cronjobs(): """ Monitors cronjobs and updates statusdb """ cj.update_cronjob_db() Move warning about missing config entry to relevant subcommandimport click import logging from taca.server_status import server_status as status from taca.utils.config import CONFIG from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception @click.group(name='server_status') def server_status(): """ Monitor server status """ # server status subcommands @server_status.command() @click.option('--statusdb', is_flag=True, help="Update the statusdb") def nases(statusdb): """ Checks the available space on all the nases """ if not CONFIG.get('server_status', ''): logging.warning("Configuration missing required entries: server_status") disk_space = status.get_nases_disk_space() if statusdb: status.update_status_db(disk_space, server_type='nas') @server_status.command() def cronjobs(): """ Monitors cronjobs and updates statusdb """ cj.update_cronjob_db()
<commit_before>import click import logging from taca.server_status import server_status as status from taca.utils.config import CONFIG from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception @click.group(name='server_status') def server_status(): """ Monitor server status """ if not CONFIG.get('server_status', ''): logging.warning("Configuration missing required entries: server_status") # server status subcommands @server_status.command() @click.option('--statusdb', is_flag=True, help="Update the statusdb") def nases(statusdb): """ Checks the available space on all the nases """ disk_space = status.get_nases_disk_space() if statusdb: status.update_status_db(disk_space, server_type='nas') @server_status.command() def cronjobs(): """ Monitors cronjobs and updates statusdb """ cj.update_cronjob_db() <commit_msg>Move warning about missing config entry to relevant subcommand<commit_after>import click import logging from taca.server_status import server_status as status from taca.utils.config import CONFIG from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception @click.group(name='server_status') def server_status(): """ Monitor server status """ # server status subcommands @server_status.command() @click.option('--statusdb', is_flag=True, help="Update the statusdb") def nases(statusdb): """ Checks the available space on all the nases """ if not CONFIG.get('server_status', ''): logging.warning("Configuration missing required entries: server_status") disk_space = status.get_nases_disk_space() if statusdb: status.update_status_db(disk_space, server_type='nas') @server_status.command() def cronjobs(): """ Monitors cronjobs and updates statusdb """ cj.update_cronjob_db()
d20347f4a57bb195291ebc79fc1ca0858b3f1d65
PyLunch/pylunch/specials/models.py
PyLunch/pylunch/specials/models.py
from django.db import models MAX_PRICE_FORMAT = { 'max_digits': 5, 'decimal_places': 2 } SPECIAL_TYPES = ( ('LU', 'Lunch'), ('BR', 'Breakfast'), ('DI', 'Dinner'), ) MAX_RESTAURANT_NAME_LENGTH = 50 MAX_DESCRIPTION_LENGTH = 500 class Restaurant(models.Model): name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) class Special(models.Model): restaurant = models.ForeignKey(Restaurant) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES) special_price = models.DecimalField(**MAX_PRICE_FORMAT) normal_price = models.DecimalField(**MAX_PRICE_FORMAT)
from django.db import models MAX_PRICE_FORMAT = { 'max_digits': 5, 'decimal_places': 2 } SPECIAL_TYPES = ( ('LU', 'Lunch'), ('BR', 'Breakfast'), ('DI', 'Dinner'), ) MAX_RESTAURANT_NAME_LENGTH = 50 MAX_DESCRIPTION_LENGTH = 500 class Restaurant(models.Model): name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) def __unicode__(self): return self.name class Special(models.Model): restaurant = models.ForeignKey(Restaurant) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES) special_price = models.DecimalField(**MAX_PRICE_FORMAT) normal_price = models.DecimalField(**MAX_PRICE_FORMAT) valid_from = models.DateField() valid_until = models.DateField() def __unicode__(self): return "%s: %s" % (self.restaurant.name, self.description)
Add fields to Special model
Add fields to Special model
Python
unlicense
wiehan-a/pylunch
from django.db import models MAX_PRICE_FORMAT = { 'max_digits': 5, 'decimal_places': 2 } SPECIAL_TYPES = ( ('LU', 'Lunch'), ('BR', 'Breakfast'), ('DI', 'Dinner'), ) MAX_RESTAURANT_NAME_LENGTH = 50 MAX_DESCRIPTION_LENGTH = 500 class Restaurant(models.Model): name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) class Special(models.Model): restaurant = models.ForeignKey(Restaurant) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES) special_price = models.DecimalField(**MAX_PRICE_FORMAT) normal_price = models.DecimalField(**MAX_PRICE_FORMAT)Add fields to Special model
from django.db import models MAX_PRICE_FORMAT = { 'max_digits': 5, 'decimal_places': 2 } SPECIAL_TYPES = ( ('LU', 'Lunch'), ('BR', 'Breakfast'), ('DI', 'Dinner'), ) MAX_RESTAURANT_NAME_LENGTH = 50 MAX_DESCRIPTION_LENGTH = 500 class Restaurant(models.Model): name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) def __unicode__(self): return self.name class Special(models.Model): restaurant = models.ForeignKey(Restaurant) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES) special_price = models.DecimalField(**MAX_PRICE_FORMAT) normal_price = models.DecimalField(**MAX_PRICE_FORMAT) valid_from = models.DateField() valid_until = models.DateField() def __unicode__(self): return "%s: %s" % (self.restaurant.name, self.description)
<commit_before>from django.db import models MAX_PRICE_FORMAT = { 'max_digits': 5, 'decimal_places': 2 } SPECIAL_TYPES = ( ('LU', 'Lunch'), ('BR', 'Breakfast'), ('DI', 'Dinner'), ) MAX_RESTAURANT_NAME_LENGTH = 50 MAX_DESCRIPTION_LENGTH = 500 class Restaurant(models.Model): name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) class Special(models.Model): restaurant = models.ForeignKey(Restaurant) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES) special_price = models.DecimalField(**MAX_PRICE_FORMAT) normal_price = models.DecimalField(**MAX_PRICE_FORMAT)<commit_msg>Add fields to Special model<commit_after>
from django.db import models MAX_PRICE_FORMAT = { 'max_digits': 5, 'decimal_places': 2 } SPECIAL_TYPES = ( ('LU', 'Lunch'), ('BR', 'Breakfast'), ('DI', 'Dinner'), ) MAX_RESTAURANT_NAME_LENGTH = 50 MAX_DESCRIPTION_LENGTH = 500 class Restaurant(models.Model): name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) def __unicode__(self): return self.name class Special(models.Model): restaurant = models.ForeignKey(Restaurant) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES) special_price = models.DecimalField(**MAX_PRICE_FORMAT) normal_price = models.DecimalField(**MAX_PRICE_FORMAT) valid_from = models.DateField() valid_until = models.DateField() def __unicode__(self): return "%s: %s" % (self.restaurant.name, self.description)
from django.db import models MAX_PRICE_FORMAT = { 'max_digits': 5, 'decimal_places': 2 } SPECIAL_TYPES = ( ('LU', 'Lunch'), ('BR', 'Breakfast'), ('DI', 'Dinner'), ) MAX_RESTAURANT_NAME_LENGTH = 50 MAX_DESCRIPTION_LENGTH = 500 class Restaurant(models.Model): name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) class Special(models.Model): restaurant = models.ForeignKey(Restaurant) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES) special_price = models.DecimalField(**MAX_PRICE_FORMAT) normal_price = models.DecimalField(**MAX_PRICE_FORMAT)Add fields to Special modelfrom django.db import models MAX_PRICE_FORMAT = { 'max_digits': 5, 'decimal_places': 2 } SPECIAL_TYPES = ( ('LU', 'Lunch'), ('BR', 'Breakfast'), ('DI', 'Dinner'), ) MAX_RESTAURANT_NAME_LENGTH = 50 MAX_DESCRIPTION_LENGTH = 500 class Restaurant(models.Model): name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) def __unicode__(self): return self.name class Special(models.Model): restaurant = models.ForeignKey(Restaurant) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES) special_price = models.DecimalField(**MAX_PRICE_FORMAT) normal_price = models.DecimalField(**MAX_PRICE_FORMAT) valid_from = models.DateField() valid_until = models.DateField() def __unicode__(self): return "%s: %s" % (self.restaurant.name, self.description)
<commit_before>from django.db import models MAX_PRICE_FORMAT = { 'max_digits': 5, 'decimal_places': 2 } SPECIAL_TYPES = ( ('LU', 'Lunch'), ('BR', 'Breakfast'), ('DI', 'Dinner'), ) MAX_RESTAURANT_NAME_LENGTH = 50 MAX_DESCRIPTION_LENGTH = 500 class Restaurant(models.Model): name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) class Special(models.Model): restaurant = models.ForeignKey(Restaurant) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES) special_price = models.DecimalField(**MAX_PRICE_FORMAT) normal_price = models.DecimalField(**MAX_PRICE_FORMAT)<commit_msg>Add fields to Special model<commit_after>from django.db import models MAX_PRICE_FORMAT = { 'max_digits': 5, 'decimal_places': 2 } SPECIAL_TYPES = ( ('LU', 'Lunch'), ('BR', 'Breakfast'), ('DI', 'Dinner'), ) MAX_RESTAURANT_NAME_LENGTH = 50 MAX_DESCRIPTION_LENGTH = 500 class Restaurant(models.Model): name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) def __unicode__(self): return self.name class Special(models.Model): restaurant = models.ForeignKey(Restaurant) description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH) special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES) special_price = models.DecimalField(**MAX_PRICE_FORMAT) normal_price = models.DecimalField(**MAX_PRICE_FORMAT) valid_from = models.DateField() valid_until = models.DateField() def __unicode__(self): return "%s: %s" % (self.restaurant.name, self.description)
d72e34de631e3f6984a1810cbd8ec2b128a196de
sigma_core/tests/factories.py
sigma_core/tests/factories.py
import factory from django.utils.text import slugify from sigma_core.models.user import User from sigma_core.models.group import Group class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User lastname = factory.Faker('last_name') firstname = factory.Faker('first_name') email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname))) class AdminUserFactory(UserFactory): is_staff = True class GroupFactory(factory.django.DjangoModelFactory): class Meta: model = Group name = factory.Sequence(lambda n: 'Group %d' % n)
import factory from django.utils.text import slugify from sigma_core.models.user import User from sigma_core.models.group import Group from sigma_core.models.user_group import UserGroup class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User lastname = factory.Faker('last_name') firstname = factory.Faker('first_name') email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname))) class AdminUserFactory(UserFactory): is_staff = True class GroupFactory(factory.django.DjangoModelFactory): class Meta: model = Group name = factory.Sequence(lambda n: 'Group %d' % n) class UserGroupFactory(factory.django.DjangoModelFactory): class Meta: model = UserGroup user = factory.SubFactory(UserFactory) group = factory.SubFactory(GroupFactory) join_date = factory.Faker('date')
Add UserGroupFactory for future tests
Add UserGroupFactory for future tests
Python
agpl-3.0
ProjetSigma/backend,ProjetSigma/backend
import factory from django.utils.text import slugify from sigma_core.models.user import User from sigma_core.models.group import Group class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User lastname = factory.Faker('last_name') firstname = factory.Faker('first_name') email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname))) class AdminUserFactory(UserFactory): is_staff = True class GroupFactory(factory.django.DjangoModelFactory): class Meta: model = Group name = factory.Sequence(lambda n: 'Group %d' % n) Add UserGroupFactory for future tests
import factory from django.utils.text import slugify from sigma_core.models.user import User from sigma_core.models.group import Group from sigma_core.models.user_group import UserGroup class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User lastname = factory.Faker('last_name') firstname = factory.Faker('first_name') email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname))) class AdminUserFactory(UserFactory): is_staff = True class GroupFactory(factory.django.DjangoModelFactory): class Meta: model = Group name = factory.Sequence(lambda n: 'Group %d' % n) class UserGroupFactory(factory.django.DjangoModelFactory): class Meta: model = UserGroup user = factory.SubFactory(UserFactory) group = factory.SubFactory(GroupFactory) join_date = factory.Faker('date')
<commit_before>import factory from django.utils.text import slugify from sigma_core.models.user import User from sigma_core.models.group import Group class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User lastname = factory.Faker('last_name') firstname = factory.Faker('first_name') email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname))) class AdminUserFactory(UserFactory): is_staff = True class GroupFactory(factory.django.DjangoModelFactory): class Meta: model = Group name = factory.Sequence(lambda n: 'Group %d' % n) <commit_msg>Add UserGroupFactory for future tests<commit_after>
import factory from django.utils.text import slugify from sigma_core.models.user import User from sigma_core.models.group import Group from sigma_core.models.user_group import UserGroup class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User lastname = factory.Faker('last_name') firstname = factory.Faker('first_name') email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname))) class AdminUserFactory(UserFactory): is_staff = True class GroupFactory(factory.django.DjangoModelFactory): class Meta: model = Group name = factory.Sequence(lambda n: 'Group %d' % n) class UserGroupFactory(factory.django.DjangoModelFactory): class Meta: model = UserGroup user = factory.SubFactory(UserFactory) group = factory.SubFactory(GroupFactory) join_date = factory.Faker('date')
import factory from django.utils.text import slugify from sigma_core.models.user import User from sigma_core.models.group import Group class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User lastname = factory.Faker('last_name') firstname = factory.Faker('first_name') email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname))) class AdminUserFactory(UserFactory): is_staff = True class GroupFactory(factory.django.DjangoModelFactory): class Meta: model = Group name = factory.Sequence(lambda n: 'Group %d' % n) Add UserGroupFactory for future testsimport factory from django.utils.text import slugify from sigma_core.models.user import User from sigma_core.models.group import Group from sigma_core.models.user_group import UserGroup class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User lastname = factory.Faker('last_name') firstname = factory.Faker('first_name') email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname))) class AdminUserFactory(UserFactory): is_staff = True class GroupFactory(factory.django.DjangoModelFactory): class Meta: model = Group name = factory.Sequence(lambda n: 'Group %d' % n) class UserGroupFactory(factory.django.DjangoModelFactory): class Meta: model = UserGroup user = factory.SubFactory(UserFactory) group = factory.SubFactory(GroupFactory) join_date = factory.Faker('date')
<commit_before>import factory from django.utils.text import slugify from sigma_core.models.user import User from sigma_core.models.group import Group class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User lastname = factory.Faker('last_name') firstname = factory.Faker('first_name') email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname))) class AdminUserFactory(UserFactory): is_staff = True class GroupFactory(factory.django.DjangoModelFactory): class Meta: model = Group name = factory.Sequence(lambda n: 'Group %d' % n) <commit_msg>Add UserGroupFactory for future tests<commit_after>import factory from django.utils.text import slugify from sigma_core.models.user import User from sigma_core.models.group import Group from sigma_core.models.user_group import UserGroup class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User lastname = factory.Faker('last_name') firstname = factory.Faker('first_name') email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname))) class AdminUserFactory(UserFactory): is_staff = True class GroupFactory(factory.django.DjangoModelFactory): class Meta: model = Group name = factory.Sequence(lambda n: 'Group %d' % n) class UserGroupFactory(factory.django.DjangoModelFactory): class Meta: model = UserGroup user = factory.SubFactory(UserFactory) group = factory.SubFactory(GroupFactory) join_date = factory.Faker('date')
70f69f7b801404f7091e91b6ed997602709f9f42
commands/globaladd.py
commands/globaladd.py
from devbot import chat def call(message: str, name, protocol, cfg, commands): if message is '': chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd'))) return if ' ' in message: chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name)) return chat.say('/msg {} Invited {} to GlobalChat'.format(name, message)) chat.say_wrap('/msg {}'.format(message), 'You have been added to global chat. Use /g GlobalChat to speak in the group, and /e to exit.') chat.say('/nlip GlobalChat {}'.format(message))
from devbot import chat def call(message: str, name, protocol, cfg, commands): if message is '': chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd'))) return if ' ' in message: chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name)) return chat.say('/msg {} Invited {} to GlobalChat'.format(name, message)) chat.say( '/msg {} You have been invited to global chat. Use /g GlobalChat to enter, and /e to exit.'.format( message)) chat.say('/nlip GlobalChat {}'.format(message))
Fix gadd not sending tutorial
Fix gadd not sending tutorial
Python
mit
Ameliorate/DevotedBot,Ameliorate/DevotedBot
from devbot import chat def call(message: str, name, protocol, cfg, commands): if message is '': chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd'))) return if ' ' in message: chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name)) return chat.say('/msg {} Invited {} to GlobalChat'.format(name, message)) chat.say_wrap('/msg {}'.format(message), 'You have been added to global chat. Use /g GlobalChat to speak in the group, and /e to exit.') chat.say('/nlip GlobalChat {}'.format(message)) Fix gadd not sending tutorial
from devbot import chat def call(message: str, name, protocol, cfg, commands): if message is '': chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd'))) return if ' ' in message: chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name)) return chat.say('/msg {} Invited {} to GlobalChat'.format(name, message)) chat.say( '/msg {} You have been invited to global chat. Use /g GlobalChat to enter, and /e to exit.'.format( message)) chat.say('/nlip GlobalChat {}'.format(message))
<commit_before>from devbot import chat def call(message: str, name, protocol, cfg, commands): if message is '': chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd'))) return if ' ' in message: chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name)) return chat.say('/msg {} Invited {} to GlobalChat'.format(name, message)) chat.say_wrap('/msg {}'.format(message), 'You have been added to global chat. Use /g GlobalChat to speak in the group, and /e to exit.') chat.say('/nlip GlobalChat {}'.format(message)) <commit_msg>Fix gadd not sending tutorial<commit_after>
from devbot import chat def call(message: str, name, protocol, cfg, commands): if message is '': chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd'))) return if ' ' in message: chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name)) return chat.say('/msg {} Invited {} to GlobalChat'.format(name, message)) chat.say( '/msg {} You have been invited to global chat. Use /g GlobalChat to enter, and /e to exit.'.format( message)) chat.say('/nlip GlobalChat {}'.format(message))
from devbot import chat def call(message: str, name, protocol, cfg, commands): if message is '': chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd'))) return if ' ' in message: chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name)) return chat.say('/msg {} Invited {} to GlobalChat'.format(name, message)) chat.say_wrap('/msg {}'.format(message), 'You have been added to global chat. Use /g GlobalChat to speak in the group, and /e to exit.') chat.say('/nlip GlobalChat {}'.format(message)) Fix gadd not sending tutorialfrom devbot import chat def call(message: str, name, protocol, cfg, commands): if message is '': chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd'))) return if ' ' in message: chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name)) return chat.say('/msg {} Invited {} to GlobalChat'.format(name, message)) chat.say( '/msg {} You have been invited to global chat. Use /g GlobalChat to enter, and /e to exit.'.format( message)) chat.say('/nlip GlobalChat {}'.format(message))
<commit_before>from devbot import chat def call(message: str, name, protocol, cfg, commands): if message is '': chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd'))) return if ' ' in message: chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name)) return chat.say('/msg {} Invited {} to GlobalChat'.format(name, message)) chat.say_wrap('/msg {}'.format(message), 'You have been added to global chat. Use /g GlobalChat to speak in the group, and /e to exit.') chat.say('/nlip GlobalChat {}'.format(message)) <commit_msg>Fix gadd not sending tutorial<commit_after>from devbot import chat def call(message: str, name, protocol, cfg, commands): if message is '': chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd'))) return if ' ' in message: chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name)) return chat.say('/msg {} Invited {} to GlobalChat'.format(name, message)) chat.say( '/msg {} You have been invited to global chat. Use /g GlobalChat to enter, and /e to exit.'.format( message)) chat.say('/nlip GlobalChat {}'.format(message))
559f3c18a7e27e4bb1147b03a27ec083a66749d0
was/photo/models.py
was/photo/models.py
from django.db import models class Photo(models.Model): artist = models.ForeignKey('artists.Artists') picture = models.ImageField(null=True, blank=True, upload_to="art_picture/") comment = models.TextField(max_length=500)
from django.db import models class Photo(models.Model): artist = models.ForeignKey('artists.Artists') picture = models.ImageField(null=True, blank=True, upload_to="art_picture/") comment = models.TextField(max_length=500) def __str__(self): return '{}'.format(self.picture)
Define a '__str__' method for photo model
Define a '__str__' method for photo model
Python
mit
KeserOner/where-artists-share,KeserOner/where-artists-share
from django.db import models class Photo(models.Model): artist = models.ForeignKey('artists.Artists') picture = models.ImageField(null=True, blank=True, upload_to="art_picture/") comment = models.TextField(max_length=500)Define a '__str__' method for photo model
from django.db import models class Photo(models.Model): artist = models.ForeignKey('artists.Artists') picture = models.ImageField(null=True, blank=True, upload_to="art_picture/") comment = models.TextField(max_length=500) def __str__(self): return '{}'.format(self.picture)
<commit_before>from django.db import models class Photo(models.Model): artist = models.ForeignKey('artists.Artists') picture = models.ImageField(null=True, blank=True, upload_to="art_picture/") comment = models.TextField(max_length=500)<commit_msg>Define a '__str__' method for photo model<commit_after>
from django.db import models class Photo(models.Model): artist = models.ForeignKey('artists.Artists') picture = models.ImageField(null=True, blank=True, upload_to="art_picture/") comment = models.TextField(max_length=500) def __str__(self): return '{}'.format(self.picture)
from django.db import models class Photo(models.Model): artist = models.ForeignKey('artists.Artists') picture = models.ImageField(null=True, blank=True, upload_to="art_picture/") comment = models.TextField(max_length=500)Define a '__str__' method for photo modelfrom django.db import models class Photo(models.Model): artist = models.ForeignKey('artists.Artists') picture = models.ImageField(null=True, blank=True, upload_to="art_picture/") comment = models.TextField(max_length=500) def __str__(self): return '{}'.format(self.picture)
<commit_before>from django.db import models class Photo(models.Model): artist = models.ForeignKey('artists.Artists') picture = models.ImageField(null=True, blank=True, upload_to="art_picture/") comment = models.TextField(max_length=500)<commit_msg>Define a '__str__' method for photo model<commit_after>from django.db import models class Photo(models.Model): artist = models.ForeignKey('artists.Artists') picture = models.ImageField(null=True, blank=True, upload_to="art_picture/") comment = models.TextField(max_length=500) def __str__(self): return '{}'.format(self.picture)
c967776872e9a5c6fa840d202f25a067020c148f
cyder/cydhcp/range/forms.py
cyder/cydhcp/range/forms.py
from django import forms from cyder.base.eav.forms import get_eav_form from cyder.base.mixins import UsabilityFormMixin from cyder.cydhcp.range.models import Range, RangeAV from cyder.cydns.forms import ViewChoiceForm class RangeForm(ViewChoiceForm, UsabilityFormMixin): class Meta: model = Range exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower') fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str', 'domain', 'is_reserved', 'allow', 'views', 'dhcpd_raw_include', 'dhcp_enabled') widgets = {'views': forms.CheckboxSelectMultiple, 'range_type': forms.RadioSelect, 'ip_type': forms.RadioSelect} exclude = 'range_usage' def __init__(self, *args, **kwargs): super(RangeForm, self).__init__(*args, **kwargs) self.fields['dhcpd_raw_include'].label = "DHCP Config Extras" self.fields['dhcpd_raw_include'].widget.attrs.update( {'cols': '80', 'style': 'display: none;width: 680px'}) RangeAVForm = get_eav_form(RangeAV, Range)
from django import forms from cyder.base.eav.forms import get_eav_form from cyder.base.mixins import UsabilityFormMixin from cyder.cydhcp.range.models import Range, RangeAV from cyder.cydns.forms import ViewChoiceForm class RangeForm(ViewChoiceForm, UsabilityFormMixin): class Meta: model = Range exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower') fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str', 'domain', 'is_reserved', 'allow', 'views', 'dhcpd_raw_include', 'dhcp_enabled', 'name') widgets = {'views': forms.CheckboxSelectMultiple, 'range_type': forms.RadioSelect, 'ip_type': forms.RadioSelect} exclude = 'range_usage' def __init__(self, *args, **kwargs): super(RangeForm, self).__init__(*args, **kwargs) self.fields['dhcpd_raw_include'].label = "DHCP Config Extras" self.fields['dhcpd_raw_include'].widget.attrs.update( {'cols': '80', 'style': 'display: none;width: 680px'}) RangeAVForm = get_eav_form(RangeAV, Range)
Include name field in range form
Include name field in range form
Python
bsd-3-clause
OSU-Net/cyder,zeeman/cyder,drkitty/cyder,OSU-Net/cyder,akeym/cyder,murrown/cyder,murrown/cyder,akeym/cyder,zeeman/cyder,akeym/cyder,murrown/cyder,murrown/cyder,drkitty/cyder,zeeman/cyder,akeym/cyder,zeeman/cyder,OSU-Net/cyder,drkitty/cyder,drkitty/cyder,OSU-Net/cyder
from django import forms from cyder.base.eav.forms import get_eav_form from cyder.base.mixins import UsabilityFormMixin from cyder.cydhcp.range.models import Range, RangeAV from cyder.cydns.forms import ViewChoiceForm class RangeForm(ViewChoiceForm, UsabilityFormMixin): class Meta: model = Range exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower') fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str', 'domain', 'is_reserved', 'allow', 'views', 'dhcpd_raw_include', 'dhcp_enabled') widgets = {'views': forms.CheckboxSelectMultiple, 'range_type': forms.RadioSelect, 'ip_type': forms.RadioSelect} exclude = 'range_usage' def __init__(self, *args, **kwargs): super(RangeForm, self).__init__(*args, **kwargs) self.fields['dhcpd_raw_include'].label = "DHCP Config Extras" self.fields['dhcpd_raw_include'].widget.attrs.update( {'cols': '80', 'style': 'display: none;width: 680px'}) RangeAVForm = get_eav_form(RangeAV, Range) Include name field in range form
from django import forms from cyder.base.eav.forms import get_eav_form from cyder.base.mixins import UsabilityFormMixin from cyder.cydhcp.range.models import Range, RangeAV from cyder.cydns.forms import ViewChoiceForm class RangeForm(ViewChoiceForm, UsabilityFormMixin): class Meta: model = Range exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower') fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str', 'domain', 'is_reserved', 'allow', 'views', 'dhcpd_raw_include', 'dhcp_enabled', 'name') widgets = {'views': forms.CheckboxSelectMultiple, 'range_type': forms.RadioSelect, 'ip_type': forms.RadioSelect} exclude = 'range_usage' def __init__(self, *args, **kwargs): super(RangeForm, self).__init__(*args, **kwargs) self.fields['dhcpd_raw_include'].label = "DHCP Config Extras" self.fields['dhcpd_raw_include'].widget.attrs.update( {'cols': '80', 'style': 'display: none;width: 680px'}) RangeAVForm = get_eav_form(RangeAV, Range)
<commit_before>from django import forms from cyder.base.eav.forms import get_eav_form from cyder.base.mixins import UsabilityFormMixin from cyder.cydhcp.range.models import Range, RangeAV from cyder.cydns.forms import ViewChoiceForm class RangeForm(ViewChoiceForm, UsabilityFormMixin): class Meta: model = Range exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower') fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str', 'domain', 'is_reserved', 'allow', 'views', 'dhcpd_raw_include', 'dhcp_enabled') widgets = {'views': forms.CheckboxSelectMultiple, 'range_type': forms.RadioSelect, 'ip_type': forms.RadioSelect} exclude = 'range_usage' def __init__(self, *args, **kwargs): super(RangeForm, self).__init__(*args, **kwargs) self.fields['dhcpd_raw_include'].label = "DHCP Config Extras" self.fields['dhcpd_raw_include'].widget.attrs.update( {'cols': '80', 'style': 'display: none;width: 680px'}) RangeAVForm = get_eav_form(RangeAV, Range) <commit_msg>Include name field in range form<commit_after>
from django import forms from cyder.base.eav.forms import get_eav_form from cyder.base.mixins import UsabilityFormMixin from cyder.cydhcp.range.models import Range, RangeAV from cyder.cydns.forms import ViewChoiceForm class RangeForm(ViewChoiceForm, UsabilityFormMixin): class Meta: model = Range exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower') fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str', 'domain', 'is_reserved', 'allow', 'views', 'dhcpd_raw_include', 'dhcp_enabled', 'name') widgets = {'views': forms.CheckboxSelectMultiple, 'range_type': forms.RadioSelect, 'ip_type': forms.RadioSelect} exclude = 'range_usage' def __init__(self, *args, **kwargs): super(RangeForm, self).__init__(*args, **kwargs) self.fields['dhcpd_raw_include'].label = "DHCP Config Extras" self.fields['dhcpd_raw_include'].widget.attrs.update( {'cols': '80', 'style': 'display: none;width: 680px'}) RangeAVForm = get_eav_form(RangeAV, Range)
from django import forms from cyder.base.eav.forms import get_eav_form from cyder.base.mixins import UsabilityFormMixin from cyder.cydhcp.range.models import Range, RangeAV from cyder.cydns.forms import ViewChoiceForm class RangeForm(ViewChoiceForm, UsabilityFormMixin): class Meta: model = Range exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower') fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str', 'domain', 'is_reserved', 'allow', 'views', 'dhcpd_raw_include', 'dhcp_enabled') widgets = {'views': forms.CheckboxSelectMultiple, 'range_type': forms.RadioSelect, 'ip_type': forms.RadioSelect} exclude = 'range_usage' def __init__(self, *args, **kwargs): super(RangeForm, self).__init__(*args, **kwargs) self.fields['dhcpd_raw_include'].label = "DHCP Config Extras" self.fields['dhcpd_raw_include'].widget.attrs.update( {'cols': '80', 'style': 'display: none;width: 680px'}) RangeAVForm = get_eav_form(RangeAV, Range) Include name field in range formfrom django import forms from cyder.base.eav.forms import get_eav_form from cyder.base.mixins import UsabilityFormMixin from cyder.cydhcp.range.models import Range, RangeAV from cyder.cydns.forms import ViewChoiceForm class RangeForm(ViewChoiceForm, UsabilityFormMixin): class Meta: model = Range exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower') fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str', 'domain', 'is_reserved', 'allow', 'views', 'dhcpd_raw_include', 'dhcp_enabled', 'name') widgets = {'views': forms.CheckboxSelectMultiple, 'range_type': forms.RadioSelect, 'ip_type': forms.RadioSelect} exclude = 'range_usage' def __init__(self, *args, **kwargs): super(RangeForm, self).__init__(*args, **kwargs) self.fields['dhcpd_raw_include'].label = "DHCP Config Extras" self.fields['dhcpd_raw_include'].widget.attrs.update( {'cols': '80', 'style': 'display: none;width: 680px'}) RangeAVForm = get_eav_form(RangeAV, Range)
<commit_before>from django import forms from cyder.base.eav.forms import get_eav_form from cyder.base.mixins import UsabilityFormMixin from cyder.cydhcp.range.models import Range, RangeAV from cyder.cydns.forms import ViewChoiceForm class RangeForm(ViewChoiceForm, UsabilityFormMixin): class Meta: model = Range exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower') fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str', 'domain', 'is_reserved', 'allow', 'views', 'dhcpd_raw_include', 'dhcp_enabled') widgets = {'views': forms.CheckboxSelectMultiple, 'range_type': forms.RadioSelect, 'ip_type': forms.RadioSelect} exclude = 'range_usage' def __init__(self, *args, **kwargs): super(RangeForm, self).__init__(*args, **kwargs) self.fields['dhcpd_raw_include'].label = "DHCP Config Extras" self.fields['dhcpd_raw_include'].widget.attrs.update( {'cols': '80', 'style': 'display: none;width: 680px'}) RangeAVForm = get_eav_form(RangeAV, Range) <commit_msg>Include name field in range form<commit_after>from django import forms from cyder.base.eav.forms import get_eav_form from cyder.base.mixins import UsabilityFormMixin from cyder.cydhcp.range.models import Range, RangeAV from cyder.cydns.forms import ViewChoiceForm class RangeForm(ViewChoiceForm, UsabilityFormMixin): class Meta: model = Range exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower') fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str', 'domain', 'is_reserved', 'allow', 'views', 'dhcpd_raw_include', 'dhcp_enabled', 'name') widgets = {'views': forms.CheckboxSelectMultiple, 'range_type': forms.RadioSelect, 'ip_type': forms.RadioSelect} exclude = 'range_usage' def __init__(self, *args, **kwargs): super(RangeForm, self).__init__(*args, **kwargs) self.fields['dhcpd_raw_include'].label = "DHCP Config Extras" self.fields['dhcpd_raw_include'].widget.attrs.update( {'cols': '80', 'style': 'display: none;width: 680px'}) RangeAVForm = get_eav_form(RangeAV, Range)
13ee0e2084765dcf958f4dbc844da54750878242
snapshottest/django.py
snapshottest/django.py
from __future__ import absolute_import from django.test import TestCase as dTestCase from django.test.runner import DiscoverRunner from snapshottest.reporting import reporting_lines from .unittest import TestCase as uTestCase class TestRunner(DiscoverRunner): separator1 = "=" * 70 separator2 = "-" * 70 def __init__(self, snapshot_update=False, **kwargs): super(TestRunner, self).__init__(**kwargs) TestCase.snapshot_should_update = snapshot_update @classmethod def add_arguments(cls, parser): super(TestRunner, cls).add_arguments(parser) parser.add_argument( '--snapshot-update', default=False, action='store_true', dest='snapshot_update', help='Update the snapshots automatically.', ) def run_tests(self, test_labels, extra_tests=None, **kwargs): result = super(TestRunner, self).run_tests( test_labels=test_labels, extra_tests=extra_tests, **kwargs ) self.print_report() return result def print_report(self): print("\n" + self.separator1) print('SnapshotTest summary') print(self.separator2) for line in reporting_lines('python manage.py test'): print(line) print(self.separator1) class TestCase(uTestCase, dTestCase): pass
from __future__ import absolute_import from django.test import TestCase as dTestCase from django.test import SimpleTestCase as dSimpleTestCase from django.test.runner import DiscoverRunner from snapshottest.reporting import reporting_lines from .unittest import TestCase as uTestCase class TestRunner(DiscoverRunner): separator1 = "=" * 70 separator2 = "-" * 70 def __init__(self, snapshot_update=False, **kwargs): super(TestRunner, self).__init__(**kwargs) uTestCase.snapshot_should_update = snapshot_update @classmethod def add_arguments(cls, parser): super(TestRunner, cls).add_arguments(parser) parser.add_argument( '--snapshot-update', default=False, action='store_true', dest='snapshot_update', help='Update the snapshots automatically.', ) def run_tests(self, test_labels, extra_tests=None, **kwargs): result = super(TestRunner, self).run_tests( test_labels=test_labels, extra_tests=extra_tests, **kwargs ) self.print_report() return result def print_report(self): print("\n" + self.separator1) print('SnapshotTest summary') print(self.separator2) for line in reporting_lines('python manage.py test'): print(line) print(self.separator1) class TestCase(uTestCase, dTestCase): pass class SimpleTestCase(uTestCase, dSimpleTestCase): pass
Allow use of alternate Django test cases
Allow use of alternate Django test cases
Python
mit
syrusakbary/snapshottest
from __future__ import absolute_import from django.test import TestCase as dTestCase from django.test.runner import DiscoverRunner from snapshottest.reporting import reporting_lines from .unittest import TestCase as uTestCase class TestRunner(DiscoverRunner): separator1 = "=" * 70 separator2 = "-" * 70 def __init__(self, snapshot_update=False, **kwargs): super(TestRunner, self).__init__(**kwargs) TestCase.snapshot_should_update = snapshot_update @classmethod def add_arguments(cls, parser): super(TestRunner, cls).add_arguments(parser) parser.add_argument( '--snapshot-update', default=False, action='store_true', dest='snapshot_update', help='Update the snapshots automatically.', ) def run_tests(self, test_labels, extra_tests=None, **kwargs): result = super(TestRunner, self).run_tests( test_labels=test_labels, extra_tests=extra_tests, **kwargs ) self.print_report() return result def print_report(self): print("\n" + self.separator1) print('SnapshotTest summary') print(self.separator2) for line in reporting_lines('python manage.py test'): print(line) print(self.separator1) class TestCase(uTestCase, dTestCase): pass Allow use of alternate Django test cases
from __future__ import absolute_import from django.test import TestCase as dTestCase from django.test import SimpleTestCase as dSimpleTestCase from django.test.runner import DiscoverRunner from snapshottest.reporting import reporting_lines from .unittest import TestCase as uTestCase class TestRunner(DiscoverRunner): separator1 = "=" * 70 separator2 = "-" * 70 def __init__(self, snapshot_update=False, **kwargs): super(TestRunner, self).__init__(**kwargs) uTestCase.snapshot_should_update = snapshot_update @classmethod def add_arguments(cls, parser): super(TestRunner, cls).add_arguments(parser) parser.add_argument( '--snapshot-update', default=False, action='store_true', dest='snapshot_update', help='Update the snapshots automatically.', ) def run_tests(self, test_labels, extra_tests=None, **kwargs): result = super(TestRunner, self).run_tests( test_labels=test_labels, extra_tests=extra_tests, **kwargs ) self.print_report() return result def print_report(self): print("\n" + self.separator1) print('SnapshotTest summary') print(self.separator2) for line in reporting_lines('python manage.py test'): print(line) print(self.separator1) class TestCase(uTestCase, dTestCase): pass class SimpleTestCase(uTestCase, dSimpleTestCase): pass
<commit_before>from __future__ import absolute_import from django.test import TestCase as dTestCase from django.test.runner import DiscoverRunner from snapshottest.reporting import reporting_lines from .unittest import TestCase as uTestCase class TestRunner(DiscoverRunner): separator1 = "=" * 70 separator2 = "-" * 70 def __init__(self, snapshot_update=False, **kwargs): super(TestRunner, self).__init__(**kwargs) TestCase.snapshot_should_update = snapshot_update @classmethod def add_arguments(cls, parser): super(TestRunner, cls).add_arguments(parser) parser.add_argument( '--snapshot-update', default=False, action='store_true', dest='snapshot_update', help='Update the snapshots automatically.', ) def run_tests(self, test_labels, extra_tests=None, **kwargs): result = super(TestRunner, self).run_tests( test_labels=test_labels, extra_tests=extra_tests, **kwargs ) self.print_report() return result def print_report(self): print("\n" + self.separator1) print('SnapshotTest summary') print(self.separator2) for line in reporting_lines('python manage.py test'): print(line) print(self.separator1) class TestCase(uTestCase, dTestCase): pass <commit_msg>Allow use of alternate Django test cases<commit_after>
from __future__ import absolute_import from django.test import TestCase as dTestCase from django.test import SimpleTestCase as dSimpleTestCase from django.test.runner import DiscoverRunner from snapshottest.reporting import reporting_lines from .unittest import TestCase as uTestCase class TestRunner(DiscoverRunner): separator1 = "=" * 70 separator2 = "-" * 70 def __init__(self, snapshot_update=False, **kwargs): super(TestRunner, self).__init__(**kwargs) uTestCase.snapshot_should_update = snapshot_update @classmethod def add_arguments(cls, parser): super(TestRunner, cls).add_arguments(parser) parser.add_argument( '--snapshot-update', default=False, action='store_true', dest='snapshot_update', help='Update the snapshots automatically.', ) def run_tests(self, test_labels, extra_tests=None, **kwargs): result = super(TestRunner, self).run_tests( test_labels=test_labels, extra_tests=extra_tests, **kwargs ) self.print_report() return result def print_report(self): print("\n" + self.separator1) print('SnapshotTest summary') print(self.separator2) for line in reporting_lines('python manage.py test'): print(line) print(self.separator1) class TestCase(uTestCase, dTestCase): pass class SimpleTestCase(uTestCase, dSimpleTestCase): pass
from __future__ import absolute_import from django.test import TestCase as dTestCase from django.test.runner import DiscoverRunner from snapshottest.reporting import reporting_lines from .unittest import TestCase as uTestCase class TestRunner(DiscoverRunner): separator1 = "=" * 70 separator2 = "-" * 70 def __init__(self, snapshot_update=False, **kwargs): super(TestRunner, self).__init__(**kwargs) TestCase.snapshot_should_update = snapshot_update @classmethod def add_arguments(cls, parser): super(TestRunner, cls).add_arguments(parser) parser.add_argument( '--snapshot-update', default=False, action='store_true', dest='snapshot_update', help='Update the snapshots automatically.', ) def run_tests(self, test_labels, extra_tests=None, **kwargs): result = super(TestRunner, self).run_tests( test_labels=test_labels, extra_tests=extra_tests, **kwargs ) self.print_report() return result def print_report(self): print("\n" + self.separator1) print('SnapshotTest summary') print(self.separator2) for line in reporting_lines('python manage.py test'): print(line) print(self.separator1) class TestCase(uTestCase, dTestCase): pass Allow use of alternate Django test casesfrom __future__ import absolute_import from django.test import TestCase as dTestCase from django.test import SimpleTestCase as dSimpleTestCase from django.test.runner import DiscoverRunner from snapshottest.reporting import reporting_lines from .unittest import TestCase as uTestCase class TestRunner(DiscoverRunner): separator1 = "=" * 70 separator2 = "-" * 70 def __init__(self, snapshot_update=False, **kwargs): super(TestRunner, self).__init__(**kwargs) uTestCase.snapshot_should_update = snapshot_update @classmethod def add_arguments(cls, parser): super(TestRunner, cls).add_arguments(parser) parser.add_argument( '--snapshot-update', default=False, action='store_true', dest='snapshot_update', help='Update the snapshots automatically.', ) def run_tests(self, test_labels, extra_tests=None, **kwargs): result = super(TestRunner, self).run_tests( test_labels=test_labels, extra_tests=extra_tests, **kwargs ) self.print_report() return result def print_report(self): print("\n" + self.separator1) print('SnapshotTest summary') print(self.separator2) for line in reporting_lines('python manage.py test'): print(line) print(self.separator1) class TestCase(uTestCase, dTestCase): pass class SimpleTestCase(uTestCase, dSimpleTestCase): pass
<commit_before>from __future__ import absolute_import from django.test import TestCase as dTestCase from django.test.runner import DiscoverRunner from snapshottest.reporting import reporting_lines from .unittest import TestCase as uTestCase class TestRunner(DiscoverRunner): separator1 = "=" * 70 separator2 = "-" * 70 def __init__(self, snapshot_update=False, **kwargs): super(TestRunner, self).__init__(**kwargs) TestCase.snapshot_should_update = snapshot_update @classmethod def add_arguments(cls, parser): super(TestRunner, cls).add_arguments(parser) parser.add_argument( '--snapshot-update', default=False, action='store_true', dest='snapshot_update', help='Update the snapshots automatically.', ) def run_tests(self, test_labels, extra_tests=None, **kwargs): result = super(TestRunner, self).run_tests( test_labels=test_labels, extra_tests=extra_tests, **kwargs ) self.print_report() return result def print_report(self): print("\n" + self.separator1) print('SnapshotTest summary') print(self.separator2) for line in reporting_lines('python manage.py test'): print(line) print(self.separator1) class TestCase(uTestCase, dTestCase): pass <commit_msg>Allow use of alternate Django test cases<commit_after>from __future__ import absolute_import from django.test import TestCase as dTestCase from django.test import SimpleTestCase as dSimpleTestCase from django.test.runner import DiscoverRunner from snapshottest.reporting import reporting_lines from .unittest import TestCase as uTestCase class TestRunner(DiscoverRunner): separator1 = "=" * 70 separator2 = "-" * 70 def __init__(self, snapshot_update=False, **kwargs): super(TestRunner, self).__init__(**kwargs) uTestCase.snapshot_should_update = snapshot_update @classmethod def add_arguments(cls, parser): super(TestRunner, cls).add_arguments(parser) parser.add_argument( '--snapshot-update', default=False, action='store_true', dest='snapshot_update', help='Update the snapshots automatically.', ) def run_tests(self, test_labels, extra_tests=None, **kwargs): result = super(TestRunner, self).run_tests( test_labels=test_labels, extra_tests=extra_tests, **kwargs ) self.print_report() return result def print_report(self): print("\n" + self.separator1) print('SnapshotTest summary') print(self.separator2) for line in reporting_lines('python manage.py test'): print(line) print(self.separator1) class TestCase(uTestCase, dTestCase): pass class SimpleTestCase(uTestCase, dSimpleTestCase): pass
7c591a38bc89350ea2586fb83a6880cdf71b4a9a
passwd_change.py
passwd_change.py
#!/usr/bin/env python3 import sys _args = sys.argv if __name__ == "__main__": if len(_args) == 4: keys_file = _args[1] target_file = _args[2] result_file = _args[3] with open(keys_file, 'r') as k: keys = k.readlines() keys = [key.strip() for key in keys] keys = [key for key in keys if key != ''] with open(target_file, 'r') as t: target_lines = t.readlines() with open(result_file, 'w') as r: for line in target_lines: if line.split(':')[0] in keys: r.write(line) else: print('./passwd_change.py keys_file.txt passwd_file result_file')
#!/usr/bin/env python3 import sys _args = sys.argv if __name__ == "__main__": if len(_args) == 4: keys_file = _args[1] target_file = _args[2] result_file = _args[3] try: with open(keys_file, 'r') as k: keys = k.readlines() keys = [key.strip().split('@')[0] for key in keys] keys = [key for key in keys if key != ''] with open(target_file, 'r') as t: target_lines = t.readlines() with open(result_file, 'w') as r: for line in target_lines: if line.split(':')[0] in keys or line.split(':')[3] != '12': r.write(line) except Exception as e: print(str(e)) sys.exit() else: print('./passwd_change.py keys_file.txt passwd_file result_file')
Add Exception to all with's block.
Add Exception to all with's block.
Python
mit
maxsocl/oldmailer
#!/usr/bin/env python3 import sys _args = sys.argv if __name__ == "__main__": if len(_args) == 4: keys_file = _args[1] target_file = _args[2] result_file = _args[3] with open(keys_file, 'r') as k: keys = k.readlines() keys = [key.strip() for key in keys] keys = [key for key in keys if key != ''] with open(target_file, 'r') as t: target_lines = t.readlines() with open(result_file, 'w') as r: for line in target_lines: if line.split(':')[0] in keys: r.write(line) else: print('./passwd_change.py keys_file.txt passwd_file result_file') Add Exception to all with's block.
#!/usr/bin/env python3 import sys _args = sys.argv if __name__ == "__main__": if len(_args) == 4: keys_file = _args[1] target_file = _args[2] result_file = _args[3] try: with open(keys_file, 'r') as k: keys = k.readlines() keys = [key.strip().split('@')[0] for key in keys] keys = [key for key in keys if key != ''] with open(target_file, 'r') as t: target_lines = t.readlines() with open(result_file, 'w') as r: for line in target_lines: if line.split(':')[0] in keys or line.split(':')[3] != '12': r.write(line) except Exception as e: print(str(e)) sys.exit() else: print('./passwd_change.py keys_file.txt passwd_file result_file')
<commit_before>#!/usr/bin/env python3 import sys _args = sys.argv if __name__ == "__main__": if len(_args) == 4: keys_file = _args[1] target_file = _args[2] result_file = _args[3] with open(keys_file, 'r') as k: keys = k.readlines() keys = [key.strip() for key in keys] keys = [key for key in keys if key != ''] with open(target_file, 'r') as t: target_lines = t.readlines() with open(result_file, 'w') as r: for line in target_lines: if line.split(':')[0] in keys: r.write(line) else: print('./passwd_change.py keys_file.txt passwd_file result_file') <commit_msg>Add Exception to all with's block.<commit_after>
#!/usr/bin/env python3 import sys _args = sys.argv if __name__ == "__main__": if len(_args) == 4: keys_file = _args[1] target_file = _args[2] result_file = _args[3] try: with open(keys_file, 'r') as k: keys = k.readlines() keys = [key.strip().split('@')[0] for key in keys] keys = [key for key in keys if key != ''] with open(target_file, 'r') as t: target_lines = t.readlines() with open(result_file, 'w') as r: for line in target_lines: if line.split(':')[0] in keys or line.split(':')[3] != '12': r.write(line) except Exception as e: print(str(e)) sys.exit() else: print('./passwd_change.py keys_file.txt passwd_file result_file')
#!/usr/bin/env python3 import sys _args = sys.argv if __name__ == "__main__": if len(_args) == 4: keys_file = _args[1] target_file = _args[2] result_file = _args[3] with open(keys_file, 'r') as k: keys = k.readlines() keys = [key.strip() for key in keys] keys = [key for key in keys if key != ''] with open(target_file, 'r') as t: target_lines = t.readlines() with open(result_file, 'w') as r: for line in target_lines: if line.split(':')[0] in keys: r.write(line) else: print('./passwd_change.py keys_file.txt passwd_file result_file') Add Exception to all with's block.#!/usr/bin/env python3 import sys _args = sys.argv if __name__ == "__main__": if len(_args) == 4: keys_file = _args[1] target_file = _args[2] result_file = _args[3] try: with open(keys_file, 'r') as k: keys = k.readlines() keys = [key.strip().split('@')[0] for key in keys] keys = [key for key in keys if key != ''] with open(target_file, 'r') as t: target_lines = t.readlines() with open(result_file, 'w') as r: for line in target_lines: if line.split(':')[0] in keys or line.split(':')[3] != '12': r.write(line) except Exception as e: print(str(e)) sys.exit() else: print('./passwd_change.py keys_file.txt passwd_file result_file')
<commit_before>#!/usr/bin/env python3 import sys _args = sys.argv if __name__ == "__main__": if len(_args) == 4: keys_file = _args[1] target_file = _args[2] result_file = _args[3] with open(keys_file, 'r') as k: keys = k.readlines() keys = [key.strip() for key in keys] keys = [key for key in keys if key != ''] with open(target_file, 'r') as t: target_lines = t.readlines() with open(result_file, 'w') as r: for line in target_lines: if line.split(':')[0] in keys: r.write(line) else: print('./passwd_change.py keys_file.txt passwd_file result_file') <commit_msg>Add Exception to all with's block.<commit_after>#!/usr/bin/env python3 import sys _args = sys.argv if __name__ == "__main__": if len(_args) == 4: keys_file = _args[1] target_file = _args[2] result_file = _args[3] try: with open(keys_file, 'r') as k: keys = k.readlines() keys = [key.strip().split('@')[0] for key in keys] keys = [key for key in keys if key != ''] with open(target_file, 'r') as t: target_lines = t.readlines() with open(result_file, 'w') as r: for line in target_lines: if line.split(':')[0] in keys or line.split(':')[3] != '12': r.write(line) except Exception as e: print(str(e)) sys.exit() else: print('./passwd_change.py keys_file.txt passwd_file result_file')
c79c573c93a96bf5b631472c5e7efccc60102813
yatsm/log_yatsm.py
yatsm/log_yatsm.py
import logging FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s' logging.basicConfig(format=FORMAT, level=logging.INFO, datefmt='%H:%M:%S') logger = logging.getLogger('yatsm')
import logging _FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s' _formatter = logging.Formatter(_FORMAT) _handler = logging.StreamHandler() _handler.setFormatter(_formatter) logger = logging.getLogger('yatsm') logger.addHandler(_handler) logger.setLevel(logging.INFO)
Change logger to be more friendly and play nice
Change logger to be more friendly and play nice
Python
mit
ceholden/yatsm,ceholden/yatsm,valpasq/yatsm,c11/yatsm,valpasq/yatsm,c11/yatsm
import logging FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s' logging.basicConfig(format=FORMAT, level=logging.INFO, datefmt='%H:%M:%S') logger = logging.getLogger('yatsm') Change logger to be more friendly and play nice
import logging _FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s' _formatter = logging.Formatter(_FORMAT) _handler = logging.StreamHandler() _handler.setFormatter(_formatter) logger = logging.getLogger('yatsm') logger.addHandler(_handler) logger.setLevel(logging.INFO)
<commit_before>import logging FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s' logging.basicConfig(format=FORMAT, level=logging.INFO, datefmt='%H:%M:%S') logger = logging.getLogger('yatsm') <commit_msg>Change logger to be more friendly and play nice<commit_after>
import logging _FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s' _formatter = logging.Formatter(_FORMAT) _handler = logging.StreamHandler() _handler.setFormatter(_formatter) logger = logging.getLogger('yatsm') logger.addHandler(_handler) logger.setLevel(logging.INFO)
import logging FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s' logging.basicConfig(format=FORMAT, level=logging.INFO, datefmt='%H:%M:%S') logger = logging.getLogger('yatsm') Change logger to be more friendly and play niceimport logging _FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s' _formatter = logging.Formatter(_FORMAT) _handler = logging.StreamHandler() _handler.setFormatter(_formatter) logger = logging.getLogger('yatsm') logger.addHandler(_handler) logger.setLevel(logging.INFO)
<commit_before>import logging FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s' logging.basicConfig(format=FORMAT, level=logging.INFO, datefmt='%H:%M:%S') logger = logging.getLogger('yatsm') <commit_msg>Change logger to be more friendly and play nice<commit_after>import logging _FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s' _formatter = logging.Formatter(_FORMAT) _handler = logging.StreamHandler() _handler.setFormatter(_formatter) logger = logging.getLogger('yatsm') logger.addHandler(_handler) logger.setLevel(logging.INFO)
03ebfe0518a7ac39f9414b3e8d8638c9dcba917c
tests/auth/test_models.py
tests/auth/test_models.py
# -*- coding: utf-8 -*- from django.core.urlresolvers import reverse from django.test import TestCase from django.utils import unittest from bakery.auth.models import BakeryUser class TestBakeryUserModel(TestCase): @unittest.skip('Not yet implemented') def test_get_absolute_url(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_absolute_url(), reverse('user-detail-view')) def test_get_full_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_full_name(), 'John Doe') def test_get_short_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_short_name(), 'John Doe')
# -*- coding: utf-8 -*- from django.test import TestCase from bakery.auth.models import BakeryUser class TestBakeryUserModel(TestCase): def test_get_absolute_url(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_absolute_url(), '/profile/user/') def test_get_full_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_full_name(), 'John Doe') def test_get_short_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_short_name(), 'John Doe')
Adjust test (refers prev commit)
Adjust test (refers prev commit)
Python
bsd-3-clause
muffins-on-dope/bakery,muffins-on-dope/bakery,muffins-on-dope/bakery
# -*- coding: utf-8 -*- from django.core.urlresolvers import reverse from django.test import TestCase from django.utils import unittest from bakery.auth.models import BakeryUser class TestBakeryUserModel(TestCase): @unittest.skip('Not yet implemented') def test_get_absolute_url(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_absolute_url(), reverse('user-detail-view')) def test_get_full_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_full_name(), 'John Doe') def test_get_short_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_short_name(), 'John Doe') Adjust test (refers prev commit)
# -*- coding: utf-8 -*- from django.test import TestCase from bakery.auth.models import BakeryUser class TestBakeryUserModel(TestCase): def test_get_absolute_url(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_absolute_url(), '/profile/user/') def test_get_full_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_full_name(), 'John Doe') def test_get_short_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_short_name(), 'John Doe')
<commit_before># -*- coding: utf-8 -*- from django.core.urlresolvers import reverse from django.test import TestCase from django.utils import unittest from bakery.auth.models import BakeryUser class TestBakeryUserModel(TestCase): @unittest.skip('Not yet implemented') def test_get_absolute_url(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_absolute_url(), reverse('user-detail-view')) def test_get_full_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_full_name(), 'John Doe') def test_get_short_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_short_name(), 'John Doe') <commit_msg>Adjust test (refers prev commit)<commit_after>
# -*- coding: utf-8 -*- from django.test import TestCase from bakery.auth.models import BakeryUser class TestBakeryUserModel(TestCase): def test_get_absolute_url(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_absolute_url(), '/profile/user/') def test_get_full_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_full_name(), 'John Doe') def test_get_short_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_short_name(), 'John Doe')
# -*- coding: utf-8 -*- from django.core.urlresolvers import reverse from django.test import TestCase from django.utils import unittest from bakery.auth.models import BakeryUser class TestBakeryUserModel(TestCase): @unittest.skip('Not yet implemented') def test_get_absolute_url(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_absolute_url(), reverse('user-detail-view')) def test_get_full_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_full_name(), 'John Doe') def test_get_short_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_short_name(), 'John Doe') Adjust test (refers prev commit)# -*- coding: utf-8 -*- from django.test import TestCase from bakery.auth.models import BakeryUser class TestBakeryUserModel(TestCase): def test_get_absolute_url(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_absolute_url(), '/profile/user/') def test_get_full_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_full_name(), 'John Doe') def test_get_short_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_short_name(), 'John Doe')
<commit_before># -*- coding: utf-8 -*- from django.core.urlresolvers import reverse from django.test import TestCase from django.utils import unittest from bakery.auth.models import BakeryUser class TestBakeryUserModel(TestCase): @unittest.skip('Not yet implemented') def test_get_absolute_url(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_absolute_url(), reverse('user-detail-view')) def test_get_full_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_full_name(), 'John Doe') def test_get_short_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_short_name(), 'John Doe') <commit_msg>Adjust test (refers prev commit)<commit_after># -*- coding: utf-8 -*- from django.test import TestCase from bakery.auth.models import BakeryUser class TestBakeryUserModel(TestCase): def test_get_absolute_url(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_absolute_url(), '/profile/user/') def test_get_full_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_full_name(), 'John Doe') def test_get_short_name(self): user = BakeryUser.objects.create_user('user', 'password') user.name = 'John Doe' self.assertEqual(user.get_short_name(), 'John Doe')
853eb4896315c7fc60b1cbd7c87be9f7674f01ba
urls.py
urls.py
from django.conf.urls.defaults import * from django.contrib import admin from django.contrib.auth.decorators import login_required from django.views.generic.simple import direct_to_template import settings admin.autodiscover() urlpatterns = patterns('', (r'^admin/', include(admin.site.urls)), (r'^idp/', include('authentic.idp.urls')), (r'^accounts/', include('registration.urls')), (r'^$', login_required(direct_to_template), { 'template': 'index.html' }, 'index'), ) if settings.AUTH_OPENID: urlpatterns += patterns('', (r'^openid/', include('django_authopenid.urls')), ) if settings.AUTH_SSL: urlpatterns += patterns('', url(r'^sslauth/$', 'authentic.sslauth.login_ssl.process_request', name='user_signin_ssl'), url(r'^error_ssl/$', direct_to_template, {'template': 'error_ssl.html'}, 'error_ssl'), ) if settings.STATIC_SERVE: urlpatterns += patterns('', url( regex = r'^media/(?P<path>.*)$', view = 'django.views.static.serve', kwargs = {'document_root': settings.MEDIA_ROOT}), )
from django.conf.urls.defaults import * from django.contrib import admin from django.contrib.auth.decorators import login_required from django.views.generic.simple import direct_to_template import settings admin.autodiscover() urlpatterns = patterns('', (r'^admin/', include(admin.site.urls)), (r'^idp/', include('authentic.idp.urls')), (r'^$', login_required(direct_to_template), { 'template': 'index.html' }, 'index'), ) if settings.AUTH_OPENID: urlpatterns += patterns('', (r'^accounts/openid/', include('django_authopenid.urls')), ) urlpatterns += patterns('', (r'^accounts/', include('registration.urls')), ) if settings.AUTH_SSL: urlpatterns += patterns('', url(r'^sslauth/$', 'authentic.sslauth.login_ssl.process_request', name='user_signin_ssl'), url(r'^error_ssl/$', direct_to_template, {'template': 'error_ssl.html'}, 'error_ssl'), ) if settings.STATIC_SERVE: urlpatterns += patterns('', url( regex = r'^media/(?P<path>.*)$', view = 'django.views.static.serve', kwargs = {'document_root': settings.MEDIA_ROOT}), )
Move OpenID stuff under /accounts/openid/
Move OpenID stuff under /accounts/openid/
Python
agpl-3.0
adieu/authentic2,adieu/authentic2,pu239ppy/authentic2,incuna/authentic,incuna/authentic,BryceLohr/authentic,BryceLohr/authentic,pu239ppy/authentic2,incuna/authentic,adieu/authentic2,adieu/authentic2,BryceLohr/authentic,BryceLohr/authentic,incuna/authentic,incuna/authentic,pu239ppy/authentic2,pu239ppy/authentic2
from django.conf.urls.defaults import * from django.contrib import admin from django.contrib.auth.decorators import login_required from django.views.generic.simple import direct_to_template import settings admin.autodiscover() urlpatterns = patterns('', (r'^admin/', include(admin.site.urls)), (r'^idp/', include('authentic.idp.urls')), (r'^accounts/', include('registration.urls')), (r'^$', login_required(direct_to_template), { 'template': 'index.html' }, 'index'), ) if settings.AUTH_OPENID: urlpatterns += patterns('', (r'^openid/', include('django_authopenid.urls')), ) if settings.AUTH_SSL: urlpatterns += patterns('', url(r'^sslauth/$', 'authentic.sslauth.login_ssl.process_request', name='user_signin_ssl'), url(r'^error_ssl/$', direct_to_template, {'template': 'error_ssl.html'}, 'error_ssl'), ) if settings.STATIC_SERVE: urlpatterns += patterns('', url( regex = r'^media/(?P<path>.*)$', view = 'django.views.static.serve', kwargs = {'document_root': settings.MEDIA_ROOT}), ) Move OpenID stuff under /accounts/openid/
from django.conf.urls.defaults import * from django.contrib import admin from django.contrib.auth.decorators import login_required from django.views.generic.simple import direct_to_template import settings admin.autodiscover() urlpatterns = patterns('', (r'^admin/', include(admin.site.urls)), (r'^idp/', include('authentic.idp.urls')), (r'^$', login_required(direct_to_template), { 'template': 'index.html' }, 'index'), ) if settings.AUTH_OPENID: urlpatterns += patterns('', (r'^accounts/openid/', include('django_authopenid.urls')), ) urlpatterns += patterns('', (r'^accounts/', include('registration.urls')), ) if settings.AUTH_SSL: urlpatterns += patterns('', url(r'^sslauth/$', 'authentic.sslauth.login_ssl.process_request', name='user_signin_ssl'), url(r'^error_ssl/$', direct_to_template, {'template': 'error_ssl.html'}, 'error_ssl'), ) if settings.STATIC_SERVE: urlpatterns += patterns('', url( regex = r'^media/(?P<path>.*)$', view = 'django.views.static.serve', kwargs = {'document_root': settings.MEDIA_ROOT}), )
<commit_before>from django.conf.urls.defaults import * from django.contrib import admin from django.contrib.auth.decorators import login_required from django.views.generic.simple import direct_to_template import settings admin.autodiscover() urlpatterns = patterns('', (r'^admin/', include(admin.site.urls)), (r'^idp/', include('authentic.idp.urls')), (r'^accounts/', include('registration.urls')), (r'^$', login_required(direct_to_template), { 'template': 'index.html' }, 'index'), ) if settings.AUTH_OPENID: urlpatterns += patterns('', (r'^openid/', include('django_authopenid.urls')), ) if settings.AUTH_SSL: urlpatterns += patterns('', url(r'^sslauth/$', 'authentic.sslauth.login_ssl.process_request', name='user_signin_ssl'), url(r'^error_ssl/$', direct_to_template, {'template': 'error_ssl.html'}, 'error_ssl'), ) if settings.STATIC_SERVE: urlpatterns += patterns('', url( regex = r'^media/(?P<path>.*)$', view = 'django.views.static.serve', kwargs = {'document_root': settings.MEDIA_ROOT}), ) <commit_msg>Move OpenID stuff under /accounts/openid/<commit_after>
from django.conf.urls.defaults import * from django.contrib import admin from django.contrib.auth.decorators import login_required from django.views.generic.simple import direct_to_template import settings admin.autodiscover() urlpatterns = patterns('', (r'^admin/', include(admin.site.urls)), (r'^idp/', include('authentic.idp.urls')), (r'^$', login_required(direct_to_template), { 'template': 'index.html' }, 'index'), ) if settings.AUTH_OPENID: urlpatterns += patterns('', (r'^accounts/openid/', include('django_authopenid.urls')), ) urlpatterns += patterns('', (r'^accounts/', include('registration.urls')), ) if settings.AUTH_SSL: urlpatterns += patterns('', url(r'^sslauth/$', 'authentic.sslauth.login_ssl.process_request', name='user_signin_ssl'), url(r'^error_ssl/$', direct_to_template, {'template': 'error_ssl.html'}, 'error_ssl'), ) if settings.STATIC_SERVE: urlpatterns += patterns('', url( regex = r'^media/(?P<path>.*)$', view = 'django.views.static.serve', kwargs = {'document_root': settings.MEDIA_ROOT}), )
from django.conf.urls.defaults import * from django.contrib import admin from django.contrib.auth.decorators import login_required from django.views.generic.simple import direct_to_template import settings admin.autodiscover() urlpatterns = patterns('', (r'^admin/', include(admin.site.urls)), (r'^idp/', include('authentic.idp.urls')), (r'^accounts/', include('registration.urls')), (r'^$', login_required(direct_to_template), { 'template': 'index.html' }, 'index'), ) if settings.AUTH_OPENID: urlpatterns += patterns('', (r'^openid/', include('django_authopenid.urls')), ) if settings.AUTH_SSL: urlpatterns += patterns('', url(r'^sslauth/$', 'authentic.sslauth.login_ssl.process_request', name='user_signin_ssl'), url(r'^error_ssl/$', direct_to_template, {'template': 'error_ssl.html'}, 'error_ssl'), ) if settings.STATIC_SERVE: urlpatterns += patterns('', url( regex = r'^media/(?P<path>.*)$', view = 'django.views.static.serve', kwargs = {'document_root': settings.MEDIA_ROOT}), ) Move OpenID stuff under /accounts/openid/from django.conf.urls.defaults import * from django.contrib import admin from django.contrib.auth.decorators import login_required from django.views.generic.simple import direct_to_template import settings admin.autodiscover() urlpatterns = patterns('', (r'^admin/', include(admin.site.urls)), (r'^idp/', include('authentic.idp.urls')), (r'^$', login_required(direct_to_template), { 'template': 'index.html' }, 'index'), ) if settings.AUTH_OPENID: urlpatterns += patterns('', (r'^accounts/openid/', include('django_authopenid.urls')), ) urlpatterns += patterns('', (r'^accounts/', include('registration.urls')), ) if settings.AUTH_SSL: urlpatterns += patterns('', url(r'^sslauth/$', 'authentic.sslauth.login_ssl.process_request', name='user_signin_ssl'), url(r'^error_ssl/$', direct_to_template, {'template': 'error_ssl.html'}, 'error_ssl'), ) if settings.STATIC_SERVE: urlpatterns += patterns('', url( regex = r'^media/(?P<path>.*)$', view = 'django.views.static.serve', kwargs = {'document_root': settings.MEDIA_ROOT}), )
<commit_before>from django.conf.urls.defaults import * from django.contrib import admin from django.contrib.auth.decorators import login_required from django.views.generic.simple import direct_to_template import settings admin.autodiscover() urlpatterns = patterns('', (r'^admin/', include(admin.site.urls)), (r'^idp/', include('authentic.idp.urls')), (r'^accounts/', include('registration.urls')), (r'^$', login_required(direct_to_template), { 'template': 'index.html' }, 'index'), ) if settings.AUTH_OPENID: urlpatterns += patterns('', (r'^openid/', include('django_authopenid.urls')), ) if settings.AUTH_SSL: urlpatterns += patterns('', url(r'^sslauth/$', 'authentic.sslauth.login_ssl.process_request', name='user_signin_ssl'), url(r'^error_ssl/$', direct_to_template, {'template': 'error_ssl.html'}, 'error_ssl'), ) if settings.STATIC_SERVE: urlpatterns += patterns('', url( regex = r'^media/(?P<path>.*)$', view = 'django.views.static.serve', kwargs = {'document_root': settings.MEDIA_ROOT}), ) <commit_msg>Move OpenID stuff under /accounts/openid/<commit_after>from django.conf.urls.defaults import * from django.contrib import admin from django.contrib.auth.decorators import login_required from django.views.generic.simple import direct_to_template import settings admin.autodiscover() urlpatterns = patterns('', (r'^admin/', include(admin.site.urls)), (r'^idp/', include('authentic.idp.urls')), (r'^$', login_required(direct_to_template), { 'template': 'index.html' }, 'index'), ) if settings.AUTH_OPENID: urlpatterns += patterns('', (r'^accounts/openid/', include('django_authopenid.urls')), ) urlpatterns += patterns('', (r'^accounts/', include('registration.urls')), ) if settings.AUTH_SSL: urlpatterns += patterns('', url(r'^sslauth/$', 'authentic.sslauth.login_ssl.process_request', name='user_signin_ssl'), url(r'^error_ssl/$', direct_to_template, {'template': 'error_ssl.html'}, 'error_ssl'), ) if settings.STATIC_SERVE: urlpatterns += patterns('', url( regex = r'^media/(?P<path>.*)$', view = 'django.views.static.serve', kwargs = {'document_root': settings.MEDIA_ROOT}), )
73af0eed3ce746154b957af5c05137f9e432c7a3
tests/test_pkgmanifest.py
tests/test_pkgmanifest.py
# Copyright (C) Ivan Kravets <me@ikravets.com> # See LICENSE for details. import requests from platformio.util import get_api_result def pytest_generate_tests(metafunc): if "package_data" not in metafunc.fixturenames: return pkgs_manifest = get_api_result("/packages") assert isinstance(pkgs_manifest, dict) packages = [] for _, variants in pkgs_manifest.iteritems(): for item in variants: packages.append(item) metafunc.parametrize("package_data", packages) def validate_response(req): assert req.status_code == 200 assert int(req.headers['Content-Length']) > 0 def validate_package(url): r = requests.head(url, allow_redirects=True) validate_response(r) assert r.headers['Content-Type'] == "application/x-gzip" def test_package(package_data): assert package_data['url'].endswith("%d.tar.gz" % package_data['version']) validate_package(package_data['url'])
# Copyright (C) Ivan Kravets <me@ikravets.com> # See LICENSE for details. import requests from platformio.util import get_api_result def pytest_generate_tests(metafunc): if "package_data" not in metafunc.fixturenames: return pkgs_manifest = get_api_result("/packages") assert isinstance(pkgs_manifest, dict) packages = [] for _, variants in pkgs_manifest.iteritems(): for item in variants: packages.append(item) metafunc.parametrize("package_data", packages) def validate_response(req): assert req.status_code == 200 assert int(req.headers['Content-Length']) > 0 def validate_package(url): r = requests.head(url, allow_redirects=True) validate_response(r) assert r.headers['Content-Type'] in ("application/x-gzip", "application/octet-stream") def test_package(package_data): assert package_data['url'].endswith("%d.tar.gz" % package_data['version']) validate_package(package_data['url'])
Add "application/octet-stream" mime type for package
Add "application/octet-stream" mime type for package
Python
apache-2.0
bkudria/platformio,awong1900/platformio,bkudria/platformio,platformio/platformio,jrobeson/platformio,platformio/platformio-core,TimJay/platformio,jrobeson/platformio,eiginn/platformio,platformio/platformio-core,bkudria/platformio,TimJay/platformio,TimJay/platformio,awong1900/platformio,valeros/platformio,ZachMassia/platformio,TimJay/platformio,mseroczynski/platformio,awong1900/platformio,jrobeson/platformio,mcanthony/platformio,dkuku/platformio,mplewis/platformio,TimJay/platformio,bkudria/platformio,jrobeson/platformio,atyenoria/platformio
# Copyright (C) Ivan Kravets <me@ikravets.com> # See LICENSE for details. import requests from platformio.util import get_api_result def pytest_generate_tests(metafunc): if "package_data" not in metafunc.fixturenames: return pkgs_manifest = get_api_result("/packages") assert isinstance(pkgs_manifest, dict) packages = [] for _, variants in pkgs_manifest.iteritems(): for item in variants: packages.append(item) metafunc.parametrize("package_data", packages) def validate_response(req): assert req.status_code == 200 assert int(req.headers['Content-Length']) > 0 def validate_package(url): r = requests.head(url, allow_redirects=True) validate_response(r) assert r.headers['Content-Type'] == "application/x-gzip" def test_package(package_data): assert package_data['url'].endswith("%d.tar.gz" % package_data['version']) validate_package(package_data['url']) Add "application/octet-stream" mime type for package
# Copyright (C) Ivan Kravets <me@ikravets.com> # See LICENSE for details. import requests from platformio.util import get_api_result def pytest_generate_tests(metafunc): if "package_data" not in metafunc.fixturenames: return pkgs_manifest = get_api_result("/packages") assert isinstance(pkgs_manifest, dict) packages = [] for _, variants in pkgs_manifest.iteritems(): for item in variants: packages.append(item) metafunc.parametrize("package_data", packages) def validate_response(req): assert req.status_code == 200 assert int(req.headers['Content-Length']) > 0 def validate_package(url): r = requests.head(url, allow_redirects=True) validate_response(r) assert r.headers['Content-Type'] in ("application/x-gzip", "application/octet-stream") def test_package(package_data): assert package_data['url'].endswith("%d.tar.gz" % package_data['version']) validate_package(package_data['url'])
<commit_before># Copyright (C) Ivan Kravets <me@ikravets.com> # See LICENSE for details. import requests from platformio.util import get_api_result def pytest_generate_tests(metafunc): if "package_data" not in metafunc.fixturenames: return pkgs_manifest = get_api_result("/packages") assert isinstance(pkgs_manifest, dict) packages = [] for _, variants in pkgs_manifest.iteritems(): for item in variants: packages.append(item) metafunc.parametrize("package_data", packages) def validate_response(req): assert req.status_code == 200 assert int(req.headers['Content-Length']) > 0 def validate_package(url): r = requests.head(url, allow_redirects=True) validate_response(r) assert r.headers['Content-Type'] == "application/x-gzip" def test_package(package_data): assert package_data['url'].endswith("%d.tar.gz" % package_data['version']) validate_package(package_data['url']) <commit_msg>Add "application/octet-stream" mime type for package<commit_after>
# Copyright (C) Ivan Kravets <me@ikravets.com> # See LICENSE for details. import requests from platformio.util import get_api_result def pytest_generate_tests(metafunc): if "package_data" not in metafunc.fixturenames: return pkgs_manifest = get_api_result("/packages") assert isinstance(pkgs_manifest, dict) packages = [] for _, variants in pkgs_manifest.iteritems(): for item in variants: packages.append(item) metafunc.parametrize("package_data", packages) def validate_response(req): assert req.status_code == 200 assert int(req.headers['Content-Length']) > 0 def validate_package(url): r = requests.head(url, allow_redirects=True) validate_response(r) assert r.headers['Content-Type'] in ("application/x-gzip", "application/octet-stream") def test_package(package_data): assert package_data['url'].endswith("%d.tar.gz" % package_data['version']) validate_package(package_data['url'])
# Copyright (C) Ivan Kravets <me@ikravets.com> # See LICENSE for details. import requests from platformio.util import get_api_result def pytest_generate_tests(metafunc): if "package_data" not in metafunc.fixturenames: return pkgs_manifest = get_api_result("/packages") assert isinstance(pkgs_manifest, dict) packages = [] for _, variants in pkgs_manifest.iteritems(): for item in variants: packages.append(item) metafunc.parametrize("package_data", packages) def validate_response(req): assert req.status_code == 200 assert int(req.headers['Content-Length']) > 0 def validate_package(url): r = requests.head(url, allow_redirects=True) validate_response(r) assert r.headers['Content-Type'] == "application/x-gzip" def test_package(package_data): assert package_data['url'].endswith("%d.tar.gz" % package_data['version']) validate_package(package_data['url']) Add "application/octet-stream" mime type for package# Copyright (C) Ivan Kravets <me@ikravets.com> # See LICENSE for details. import requests from platformio.util import get_api_result def pytest_generate_tests(metafunc): if "package_data" not in metafunc.fixturenames: return pkgs_manifest = get_api_result("/packages") assert isinstance(pkgs_manifest, dict) packages = [] for _, variants in pkgs_manifest.iteritems(): for item in variants: packages.append(item) metafunc.parametrize("package_data", packages) def validate_response(req): assert req.status_code == 200 assert int(req.headers['Content-Length']) > 0 def validate_package(url): r = requests.head(url, allow_redirects=True) validate_response(r) assert r.headers['Content-Type'] in ("application/x-gzip", "application/octet-stream") def test_package(package_data): assert package_data['url'].endswith("%d.tar.gz" % package_data['version']) validate_package(package_data['url'])
<commit_before># Copyright (C) Ivan Kravets <me@ikravets.com> # See LICENSE for details. import requests from platformio.util import get_api_result def pytest_generate_tests(metafunc): if "package_data" not in metafunc.fixturenames: return pkgs_manifest = get_api_result("/packages") assert isinstance(pkgs_manifest, dict) packages = [] for _, variants in pkgs_manifest.iteritems(): for item in variants: packages.append(item) metafunc.parametrize("package_data", packages) def validate_response(req): assert req.status_code == 200 assert int(req.headers['Content-Length']) > 0 def validate_package(url): r = requests.head(url, allow_redirects=True) validate_response(r) assert r.headers['Content-Type'] == "application/x-gzip" def test_package(package_data): assert package_data['url'].endswith("%d.tar.gz" % package_data['version']) validate_package(package_data['url']) <commit_msg>Add "application/octet-stream" mime type for package<commit_after># Copyright (C) Ivan Kravets <me@ikravets.com> # See LICENSE for details. import requests from platformio.util import get_api_result def pytest_generate_tests(metafunc): if "package_data" not in metafunc.fixturenames: return pkgs_manifest = get_api_result("/packages") assert isinstance(pkgs_manifest, dict) packages = [] for _, variants in pkgs_manifest.iteritems(): for item in variants: packages.append(item) metafunc.parametrize("package_data", packages) def validate_response(req): assert req.status_code == 200 assert int(req.headers['Content-Length']) > 0 def validate_package(url): r = requests.head(url, allow_redirects=True) validate_response(r) assert r.headers['Content-Type'] in ("application/x-gzip", "application/octet-stream") def test_package(package_data): assert package_data['url'].endswith("%d.tar.gz" % package_data['version']) validate_package(package_data['url'])
76dbf84facac62345ba02f17b33844a41d326d1f
tg/tests/test_testutil.py
tg/tests/test_testutil.py
"""Tests for the testutil module""" from unittest import TestCase from sqlalchemy import create_engine from sqlalchemy.exc import DBAPIError import transaction from tg.testutil import DBTest from tg.tests.fixtures import model # Ideally, we would have defined several different descendants of DBTest, # in order to test its behavior in different situations, but there seem to be # a problem in unittests and grand-grandchildren of TestCase won't work. You # may try this code if you want: http://paste.turbogears.org/paste/4721 # or http://paste.turbogears.org/paste/4724 class BaseModelTest(DBTest): database = create_engine("sqlite:///:memory:") model = model class TestGroup(BaseModelTest): """Test case for the Group model. This should tell us whether the setUp() and tearDown() of DBTest work as expected. """ def test_group_creation(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.save(group) model.DBSession.flush() transaction.commit() def test_this_group_was_already_removed(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.save(group) model.DBSession.flush() transaction.commit()
"""Tests for the testutil module""" from unittest import TestCase from sqlalchemy import create_engine from sqlalchemy.exc import DBAPIError import transaction from tg.testutil import DBTest from tg.tests.fixtures import model # Ideally, we would have defined several different descendants of DBTest, # in order to test its behavior in different situations, but there seem to be # a problem in unittests and grand-grandchildren of TestCase won't work. You # may try this code if you want: http://paste.turbogears.org/paste/4721 # or http://paste.turbogears.org/paste/4724 class BaseModelTest(DBTest): database = create_engine("sqlite:///:memory:") model = model class TestGroup(BaseModelTest): """Test case for the Group model. This should tell us whether the setUp() and tearDown() of DBTest work as expected. """ def test_group_creation(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.add(group) model.DBSession.flush() transaction.commit() def test_this_group_was_already_removed(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.add(group) model.DBSession.flush() transaction.commit()
Fix from cdevienne for SA Deprecation warnings
Fix from cdevienne for SA Deprecation warnings --HG-- extra : convert_revision : svn%3A77541ad4-5f01-0410-9ede-a1b63cd9a898/trunk%405878
Python
mit
lucius-feng/tg2,lucius-feng/tg2
"""Tests for the testutil module""" from unittest import TestCase from sqlalchemy import create_engine from sqlalchemy.exc import DBAPIError import transaction from tg.testutil import DBTest from tg.tests.fixtures import model # Ideally, we would have defined several different descendants of DBTest, # in order to test its behavior in different situations, but there seem to be # a problem in unittests and grand-grandchildren of TestCase won't work. You # may try this code if you want: http://paste.turbogears.org/paste/4721 # or http://paste.turbogears.org/paste/4724 class BaseModelTest(DBTest): database = create_engine("sqlite:///:memory:") model = model class TestGroup(BaseModelTest): """Test case for the Group model. This should tell us whether the setUp() and tearDown() of DBTest work as expected. """ def test_group_creation(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.save(group) model.DBSession.flush() transaction.commit() def test_this_group_was_already_removed(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.save(group) model.DBSession.flush() transaction.commit() Fix from cdevienne for SA Deprecation warnings --HG-- extra : convert_revision : svn%3A77541ad4-5f01-0410-9ede-a1b63cd9a898/trunk%405878
"""Tests for the testutil module""" from unittest import TestCase from sqlalchemy import create_engine from sqlalchemy.exc import DBAPIError import transaction from tg.testutil import DBTest from tg.tests.fixtures import model # Ideally, we would have defined several different descendants of DBTest, # in order to test its behavior in different situations, but there seem to be # a problem in unittests and grand-grandchildren of TestCase won't work. You # may try this code if you want: http://paste.turbogears.org/paste/4721 # or http://paste.turbogears.org/paste/4724 class BaseModelTest(DBTest): database = create_engine("sqlite:///:memory:") model = model class TestGroup(BaseModelTest): """Test case for the Group model. This should tell us whether the setUp() and tearDown() of DBTest work as expected. """ def test_group_creation(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.add(group) model.DBSession.flush() transaction.commit() def test_this_group_was_already_removed(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.add(group) model.DBSession.flush() transaction.commit()
<commit_before>"""Tests for the testutil module""" from unittest import TestCase from sqlalchemy import create_engine from sqlalchemy.exc import DBAPIError import transaction from tg.testutil import DBTest from tg.tests.fixtures import model # Ideally, we would have defined several different descendants of DBTest, # in order to test its behavior in different situations, but there seem to be # a problem in unittests and grand-grandchildren of TestCase won't work. You # may try this code if you want: http://paste.turbogears.org/paste/4721 # or http://paste.turbogears.org/paste/4724 class BaseModelTest(DBTest): database = create_engine("sqlite:///:memory:") model = model class TestGroup(BaseModelTest): """Test case for the Group model. This should tell us whether the setUp() and tearDown() of DBTest work as expected. """ def test_group_creation(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.save(group) model.DBSession.flush() transaction.commit() def test_this_group_was_already_removed(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.save(group) model.DBSession.flush() transaction.commit() <commit_msg>Fix from cdevienne for SA Deprecation warnings --HG-- extra : convert_revision : svn%3A77541ad4-5f01-0410-9ede-a1b63cd9a898/trunk%405878<commit_after>
"""Tests for the testutil module""" from unittest import TestCase from sqlalchemy import create_engine from sqlalchemy.exc import DBAPIError import transaction from tg.testutil import DBTest from tg.tests.fixtures import model # Ideally, we would have defined several different descendants of DBTest, # in order to test its behavior in different situations, but there seem to be # a problem in unittests and grand-grandchildren of TestCase won't work. You # may try this code if you want: http://paste.turbogears.org/paste/4721 # or http://paste.turbogears.org/paste/4724 class BaseModelTest(DBTest): database = create_engine("sqlite:///:memory:") model = model class TestGroup(BaseModelTest): """Test case for the Group model. This should tell us whether the setUp() and tearDown() of DBTest work as expected. """ def test_group_creation(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.add(group) model.DBSession.flush() transaction.commit() def test_this_group_was_already_removed(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.add(group) model.DBSession.flush() transaction.commit()
"""Tests for the testutil module""" from unittest import TestCase from sqlalchemy import create_engine from sqlalchemy.exc import DBAPIError import transaction from tg.testutil import DBTest from tg.tests.fixtures import model # Ideally, we would have defined several different descendants of DBTest, # in order to test its behavior in different situations, but there seem to be # a problem in unittests and grand-grandchildren of TestCase won't work. You # may try this code if you want: http://paste.turbogears.org/paste/4721 # or http://paste.turbogears.org/paste/4724 class BaseModelTest(DBTest): database = create_engine("sqlite:///:memory:") model = model class TestGroup(BaseModelTest): """Test case for the Group model. This should tell us whether the setUp() and tearDown() of DBTest work as expected. """ def test_group_creation(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.save(group) model.DBSession.flush() transaction.commit() def test_this_group_was_already_removed(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.save(group) model.DBSession.flush() transaction.commit() Fix from cdevienne for SA Deprecation warnings --HG-- extra : convert_revision : svn%3A77541ad4-5f01-0410-9ede-a1b63cd9a898/trunk%405878"""Tests for the testutil module""" from unittest import TestCase from sqlalchemy import create_engine from sqlalchemy.exc import DBAPIError import transaction from tg.testutil import DBTest from tg.tests.fixtures import model # Ideally, we would have defined several different descendants of DBTest, # in order to test its behavior in different situations, but there seem to be # a problem in unittests and grand-grandchildren of TestCase won't work. You # may try this code if you want: http://paste.turbogears.org/paste/4721 # or http://paste.turbogears.org/paste/4724 class BaseModelTest(DBTest): database = create_engine("sqlite:///:memory:") model = model class TestGroup(BaseModelTest): """Test case for the Group model. This should tell us whether the setUp() and tearDown() of DBTest work as expected. """ def test_group_creation(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.add(group) model.DBSession.flush() transaction.commit() def test_this_group_was_already_removed(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.add(group) model.DBSession.flush() transaction.commit()
<commit_before>"""Tests for the testutil module""" from unittest import TestCase from sqlalchemy import create_engine from sqlalchemy.exc import DBAPIError import transaction from tg.testutil import DBTest from tg.tests.fixtures import model # Ideally, we would have defined several different descendants of DBTest, # in order to test its behavior in different situations, but there seem to be # a problem in unittests and grand-grandchildren of TestCase won't work. You # may try this code if you want: http://paste.turbogears.org/paste/4721 # or http://paste.turbogears.org/paste/4724 class BaseModelTest(DBTest): database = create_engine("sqlite:///:memory:") model = model class TestGroup(BaseModelTest): """Test case for the Group model. This should tell us whether the setUp() and tearDown() of DBTest work as expected. """ def test_group_creation(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.save(group) model.DBSession.flush() transaction.commit() def test_this_group_was_already_removed(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.save(group) model.DBSession.flush() transaction.commit() <commit_msg>Fix from cdevienne for SA Deprecation warnings --HG-- extra : convert_revision : svn%3A77541ad4-5f01-0410-9ede-a1b63cd9a898/trunk%405878<commit_after>"""Tests for the testutil module""" from unittest import TestCase from sqlalchemy import create_engine from sqlalchemy.exc import DBAPIError import transaction from tg.testutil import DBTest from tg.tests.fixtures import model # Ideally, we would have defined several different descendants of DBTest, # in order to test its behavior in different situations, but there seem to be # a problem in unittests and grand-grandchildren of TestCase won't work. You # may try this code if you want: http://paste.turbogears.org/paste/4721 # or http://paste.turbogears.org/paste/4724 class BaseModelTest(DBTest): database = create_engine("sqlite:///:memory:") model = model class TestGroup(BaseModelTest): """Test case for the Group model. This should tell us whether the setUp() and tearDown() of DBTest work as expected. """ def test_group_creation(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.add(group) model.DBSession.flush() transaction.commit() def test_this_group_was_already_removed(self): group = model.Group() group.group_name = u"turbogears" group.display_name = u"The TurboGears Team" model.DBSession.add(group) model.DBSession.flush() transaction.commit()
daa4021011778f7511ad2c97648155bb17539d98
tests/func/test_examples.py
tests/func/test_examples.py
import pytest # noqa import os import sys import glob import imp def test_examples(): examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../../examples/*/*.py') # Filter out __init__.py examples = [f for f in glob.glob(examples_pat) if not any([x in f for x in ['__init__.py', 'molecular', 'custom_table_caching']])] for e in examples: example_dir = os.path.dirname(e) sys.path.insert(0, example_dir) (module_name, _) = os.path.splitext(os.path.basename(e)) (module_file, module_path, desc) = \ imp.find_module(module_name, [example_dir]) m = imp.load_module(module_name, module_file, module_path, desc) if hasattr(m, 'main'): m.main(debug=False)
import pytest # noqa import os import sys import glob import importlib def test_examples(): examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../../examples/*/*.py') # Filter out __init__.py examples = [f for f in glob.glob(examples_pat) if not any([x in f for x in ['__init__.py', 'molecular', 'custom_table_caching']])] for e in examples: example_dir = os.path.dirname(e) sys.path.insert(0, example_dir) (module_name, _) = os.path.splitext(os.path.basename(e)) m = importlib.import_module(module_name) if hasattr(m, 'main'): m.main(debug=False)
Replace deprecated imp with importlib
Replace deprecated imp with importlib
Python
mit
igordejanovic/parglare,igordejanovic/parglare
import pytest # noqa import os import sys import glob import imp def test_examples(): examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../../examples/*/*.py') # Filter out __init__.py examples = [f for f in glob.glob(examples_pat) if not any([x in f for x in ['__init__.py', 'molecular', 'custom_table_caching']])] for e in examples: example_dir = os.path.dirname(e) sys.path.insert(0, example_dir) (module_name, _) = os.path.splitext(os.path.basename(e)) (module_file, module_path, desc) = \ imp.find_module(module_name, [example_dir]) m = imp.load_module(module_name, module_file, module_path, desc) if hasattr(m, 'main'): m.main(debug=False) Replace deprecated imp with importlib
import pytest # noqa import os import sys import glob import importlib def test_examples(): examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../../examples/*/*.py') # Filter out __init__.py examples = [f for f in glob.glob(examples_pat) if not any([x in f for x in ['__init__.py', 'molecular', 'custom_table_caching']])] for e in examples: example_dir = os.path.dirname(e) sys.path.insert(0, example_dir) (module_name, _) = os.path.splitext(os.path.basename(e)) m = importlib.import_module(module_name) if hasattr(m, 'main'): m.main(debug=False)
<commit_before>import pytest # noqa import os import sys import glob import imp def test_examples(): examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../../examples/*/*.py') # Filter out __init__.py examples = [f for f in glob.glob(examples_pat) if not any([x in f for x in ['__init__.py', 'molecular', 'custom_table_caching']])] for e in examples: example_dir = os.path.dirname(e) sys.path.insert(0, example_dir) (module_name, _) = os.path.splitext(os.path.basename(e)) (module_file, module_path, desc) = \ imp.find_module(module_name, [example_dir]) m = imp.load_module(module_name, module_file, module_path, desc) if hasattr(m, 'main'): m.main(debug=False) <commit_msg>Replace deprecated imp with importlib<commit_after>
import pytest # noqa import os import sys import glob import importlib def test_examples(): examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../../examples/*/*.py') # Filter out __init__.py examples = [f for f in glob.glob(examples_pat) if not any([x in f for x in ['__init__.py', 'molecular', 'custom_table_caching']])] for e in examples: example_dir = os.path.dirname(e) sys.path.insert(0, example_dir) (module_name, _) = os.path.splitext(os.path.basename(e)) m = importlib.import_module(module_name) if hasattr(m, 'main'): m.main(debug=False)
import pytest # noqa import os import sys import glob import imp def test_examples(): examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../../examples/*/*.py') # Filter out __init__.py examples = [f for f in glob.glob(examples_pat) if not any([x in f for x in ['__init__.py', 'molecular', 'custom_table_caching']])] for e in examples: example_dir = os.path.dirname(e) sys.path.insert(0, example_dir) (module_name, _) = os.path.splitext(os.path.basename(e)) (module_file, module_path, desc) = \ imp.find_module(module_name, [example_dir]) m = imp.load_module(module_name, module_file, module_path, desc) if hasattr(m, 'main'): m.main(debug=False) Replace deprecated imp with importlibimport pytest # noqa import os import sys import glob import importlib def test_examples(): examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../../examples/*/*.py') # Filter out __init__.py examples = [f for f in glob.glob(examples_pat) if not any([x in f for x in ['__init__.py', 'molecular', 'custom_table_caching']])] for e in examples: example_dir = os.path.dirname(e) sys.path.insert(0, example_dir) (module_name, _) = os.path.splitext(os.path.basename(e)) m = importlib.import_module(module_name) if hasattr(m, 'main'): m.main(debug=False)
<commit_before>import pytest # noqa import os import sys import glob import imp def test_examples(): examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../../examples/*/*.py') # Filter out __init__.py examples = [f for f in glob.glob(examples_pat) if not any([x in f for x in ['__init__.py', 'molecular', 'custom_table_caching']])] for e in examples: example_dir = os.path.dirname(e) sys.path.insert(0, example_dir) (module_name, _) = os.path.splitext(os.path.basename(e)) (module_file, module_path, desc) = \ imp.find_module(module_name, [example_dir]) m = imp.load_module(module_name, module_file, module_path, desc) if hasattr(m, 'main'): m.main(debug=False) <commit_msg>Replace deprecated imp with importlib<commit_after>import pytest # noqa import os import sys import glob import importlib def test_examples(): examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)), '../../examples/*/*.py') # Filter out __init__.py examples = [f for f in glob.glob(examples_pat) if not any([x in f for x in ['__init__.py', 'molecular', 'custom_table_caching']])] for e in examples: example_dir = os.path.dirname(e) sys.path.insert(0, example_dir) (module_name, _) = os.path.splitext(os.path.basename(e)) m = importlib.import_module(module_name) if hasattr(m, 'main'): m.main(debug=False)
fe98a627943c235ba24fc6de781deec69e7fd02e
relayer/__init__.py
relayer/__init__.py
from kafka import KafkaProducer from .event_emitter import EventEmitter from .exceptions import ConfigurationError __version__ = '0.1.3' class Relayer(object): def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''): self.logging_topic = logging_topic if not kafka_hosts: raise ConfigurationError() if source == '': self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix) else: self.source = source producer = KafkaProducer(bootstrap_servers=kafka_hosts) emitter = EventEmitter(producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix) self.context = context_handler_class(emitter) def emit(self, event_type, event_subtype, payload, partition_key=None): payload = { 'source': self.source, 'event_type': event_type, 'event_subtype': event_subtype, 'payload': payload } self.context.emit(event_type, payload, partition_key) def emit_raw(self, topic, message, partition_key=None): self.context.emit(topic, message, partition_key) def log(self, log_level, payload): message = { 'log_level': log_level, 'payload': payload } self.context.log(message) def flush(self): self.emitter.flush()
from kafka import KafkaProducer from .event_emitter import EventEmitter from .exceptions import ConfigurationError __version__ = '0.1.3' class Relayer(object): def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''): self.logging_topic = logging_topic if not kafka_hosts: raise ConfigurationError() if source == '': self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix) else: self.source = source self._producer = KafkaProducer(bootstrap_servers=kafka_hosts) self._emitter = EventEmitter(self._producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix) self.context = context_handler_class(self._emitter) def emit(self, event_type, event_subtype, payload, partition_key=None): payload = { 'source': self.source, 'event_type': event_type, 'event_subtype': event_subtype, 'payload': payload } self.context.emit(event_type, payload, partition_key) def emit_raw(self, topic, message, partition_key=None): self.context.emit(topic, message, partition_key) def log(self, log_level, payload): message = { 'log_level': log_level, 'payload': payload } self.context.log(message) def flush(self): self._emitter.flush()
Save event emitter y producer reference in relayer instance
Save event emitter y producer reference in relayer instance
Python
mit
wizeline/relayer
from kafka import KafkaProducer from .event_emitter import EventEmitter from .exceptions import ConfigurationError __version__ = '0.1.3' class Relayer(object): def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''): self.logging_topic = logging_topic if not kafka_hosts: raise ConfigurationError() if source == '': self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix) else: self.source = source producer = KafkaProducer(bootstrap_servers=kafka_hosts) emitter = EventEmitter(producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix) self.context = context_handler_class(emitter) def emit(self, event_type, event_subtype, payload, partition_key=None): payload = { 'source': self.source, 'event_type': event_type, 'event_subtype': event_subtype, 'payload': payload } self.context.emit(event_type, payload, partition_key) def emit_raw(self, topic, message, partition_key=None): self.context.emit(topic, message, partition_key) def log(self, log_level, payload): message = { 'log_level': log_level, 'payload': payload } self.context.log(message) def flush(self): self.emitter.flush() Save event emitter y producer reference in relayer instance
from kafka import KafkaProducer from .event_emitter import EventEmitter from .exceptions import ConfigurationError __version__ = '0.1.3' class Relayer(object): def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''): self.logging_topic = logging_topic if not kafka_hosts: raise ConfigurationError() if source == '': self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix) else: self.source = source self._producer = KafkaProducer(bootstrap_servers=kafka_hosts) self._emitter = EventEmitter(self._producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix) self.context = context_handler_class(self._emitter) def emit(self, event_type, event_subtype, payload, partition_key=None): payload = { 'source': self.source, 'event_type': event_type, 'event_subtype': event_subtype, 'payload': payload } self.context.emit(event_type, payload, partition_key) def emit_raw(self, topic, message, partition_key=None): self.context.emit(topic, message, partition_key) def log(self, log_level, payload): message = { 'log_level': log_level, 'payload': payload } self.context.log(message) def flush(self): self._emitter.flush()
<commit_before>from kafka import KafkaProducer from .event_emitter import EventEmitter from .exceptions import ConfigurationError __version__ = '0.1.3' class Relayer(object): def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''): self.logging_topic = logging_topic if not kafka_hosts: raise ConfigurationError() if source == '': self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix) else: self.source = source producer = KafkaProducer(bootstrap_servers=kafka_hosts) emitter = EventEmitter(producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix) self.context = context_handler_class(emitter) def emit(self, event_type, event_subtype, payload, partition_key=None): payload = { 'source': self.source, 'event_type': event_type, 'event_subtype': event_subtype, 'payload': payload } self.context.emit(event_type, payload, partition_key) def emit_raw(self, topic, message, partition_key=None): self.context.emit(topic, message, partition_key) def log(self, log_level, payload): message = { 'log_level': log_level, 'payload': payload } self.context.log(message) def flush(self): self.emitter.flush() <commit_msg>Save event emitter y producer reference in relayer instance<commit_after>
from kafka import KafkaProducer from .event_emitter import EventEmitter from .exceptions import ConfigurationError __version__ = '0.1.3' class Relayer(object): def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''): self.logging_topic = logging_topic if not kafka_hosts: raise ConfigurationError() if source == '': self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix) else: self.source = source self._producer = KafkaProducer(bootstrap_servers=kafka_hosts) self._emitter = EventEmitter(self._producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix) self.context = context_handler_class(self._emitter) def emit(self, event_type, event_subtype, payload, partition_key=None): payload = { 'source': self.source, 'event_type': event_type, 'event_subtype': event_subtype, 'payload': payload } self.context.emit(event_type, payload, partition_key) def emit_raw(self, topic, message, partition_key=None): self.context.emit(topic, message, partition_key) def log(self, log_level, payload): message = { 'log_level': log_level, 'payload': payload } self.context.log(message) def flush(self): self._emitter.flush()
from kafka import KafkaProducer from .event_emitter import EventEmitter from .exceptions import ConfigurationError __version__ = '0.1.3' class Relayer(object): def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''): self.logging_topic = logging_topic if not kafka_hosts: raise ConfigurationError() if source == '': self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix) else: self.source = source producer = KafkaProducer(bootstrap_servers=kafka_hosts) emitter = EventEmitter(producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix) self.context = context_handler_class(emitter) def emit(self, event_type, event_subtype, payload, partition_key=None): payload = { 'source': self.source, 'event_type': event_type, 'event_subtype': event_subtype, 'payload': payload } self.context.emit(event_type, payload, partition_key) def emit_raw(self, topic, message, partition_key=None): self.context.emit(topic, message, partition_key) def log(self, log_level, payload): message = { 'log_level': log_level, 'payload': payload } self.context.log(message) def flush(self): self.emitter.flush() Save event emitter y producer reference in relayer instancefrom kafka import KafkaProducer from .event_emitter import EventEmitter from .exceptions import ConfigurationError __version__ = '0.1.3' class Relayer(object): def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''): self.logging_topic = logging_topic if not kafka_hosts: raise ConfigurationError() if source == '': self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix) else: self.source = source self._producer = KafkaProducer(bootstrap_servers=kafka_hosts) self._emitter = EventEmitter(self._producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix) self.context = context_handler_class(self._emitter) def emit(self, event_type, event_subtype, payload, partition_key=None): payload = { 'source': self.source, 'event_type': event_type, 'event_subtype': event_subtype, 'payload': payload } self.context.emit(event_type, payload, partition_key) def emit_raw(self, topic, message, partition_key=None): self.context.emit(topic, message, partition_key) def log(self, log_level, payload): message = { 'log_level': log_level, 'payload': payload } self.context.log(message) def flush(self): self._emitter.flush()
<commit_before>from kafka import KafkaProducer from .event_emitter import EventEmitter from .exceptions import ConfigurationError __version__ = '0.1.3' class Relayer(object): def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''): self.logging_topic = logging_topic if not kafka_hosts: raise ConfigurationError() if source == '': self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix) else: self.source = source producer = KafkaProducer(bootstrap_servers=kafka_hosts) emitter = EventEmitter(producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix) self.context = context_handler_class(emitter) def emit(self, event_type, event_subtype, payload, partition_key=None): payload = { 'source': self.source, 'event_type': event_type, 'event_subtype': event_subtype, 'payload': payload } self.context.emit(event_type, payload, partition_key) def emit_raw(self, topic, message, partition_key=None): self.context.emit(topic, message, partition_key) def log(self, log_level, payload): message = { 'log_level': log_level, 'payload': payload } self.context.log(message) def flush(self): self.emitter.flush() <commit_msg>Save event emitter y producer reference in relayer instance<commit_after>from kafka import KafkaProducer from .event_emitter import EventEmitter from .exceptions import ConfigurationError __version__ = '0.1.3' class Relayer(object): def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''): self.logging_topic = logging_topic if not kafka_hosts: raise ConfigurationError() if source == '': self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix) else: self.source = source self._producer = KafkaProducer(bootstrap_servers=kafka_hosts) self._emitter = EventEmitter(self._producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix) self.context = context_handler_class(self._emitter) def emit(self, event_type, event_subtype, payload, partition_key=None): payload = { 'source': self.source, 'event_type': event_type, 'event_subtype': event_subtype, 'payload': payload } self.context.emit(event_type, payload, partition_key) def emit_raw(self, topic, message, partition_key=None): self.context.emit(topic, message, partition_key) def log(self, log_level, payload): message = { 'log_level': log_level, 'payload': payload } self.context.log(message) def flush(self): self._emitter.flush()
64ac175029393c75a6b308bb36287081ab1bde8f
alignak_backend_import/__init__.py
alignak_backend_import/__init__.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Alignak backend import This module contains utility tools to import Nagios-like flat files configuration into an Alignak REST backend. """ # Application version and manifest VERSION = (0, 4, 3) __application__ = u"Alignak backend import" __short_version__ = '.'.join((str(each) for each in VERSION[:2])) __version__ = '.'.join((str(each) for each in VERSION[:4])) __author__ = u"Alignak team" __copyright__ = u"(c) 2015-2016, %s" % __author__ __license__ = u"GNU Affero General Public License, version 3" __description__ = u"Alignak backend import tools" __releasenotes__ = u"""Alignak Backend import tools""" __doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import" # Application manifest manifest = { 'name': __application__, 'version': __version__, 'author': __author__, 'description': __description__, 'copyright': __copyright__, 'license': __license__, 'release': __releasenotes__, 'doc': __doc_url__ }
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Alignak backend import This module contains utility tools to import Nagios-like flat files configuration into an Alignak REST backend. """ # Application version and manifest VERSION = (0, 4, 3) __application__ = u"Alignak backend import" __short_version__ = '.'.join((str(each) for each in VERSION[:2])) __version__ = '.'.join((str(each) for each in VERSION[:4])) __author__ = u"Alignak team" __copyright__ = u"(c) 2015-2016, %s" % __author__ __license__ = u"GNU Affero General Public License, version 3" __description__ = u"Alignak backend import tools" __releasenotes__ = u"""Alignak Backend import tools""" __doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import" # Application manifest manifest = { 'name': __application__, 'version': __version__, 'author': __author__, 'description': __description__, 'copyright': __copyright__, 'license': __license__, 'release': __releasenotes__, 'doc': __doc_url__ }
Fix bad indentation that broke PEP8 !
Fix bad indentation that broke PEP8 !
Python
agpl-3.0
Alignak-monitoring-contrib/alignak-backend-import,Alignak-monitoring-contrib/alignak-backend-import
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Alignak backend import This module contains utility tools to import Nagios-like flat files configuration into an Alignak REST backend. """ # Application version and manifest VERSION = (0, 4, 3) __application__ = u"Alignak backend import" __short_version__ = '.'.join((str(each) for each in VERSION[:2])) __version__ = '.'.join((str(each) for each in VERSION[:4])) __author__ = u"Alignak team" __copyright__ = u"(c) 2015-2016, %s" % __author__ __license__ = u"GNU Affero General Public License, version 3" __description__ = u"Alignak backend import tools" __releasenotes__ = u"""Alignak Backend import tools""" __doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import" # Application manifest manifest = { 'name': __application__, 'version': __version__, 'author': __author__, 'description': __description__, 'copyright': __copyright__, 'license': __license__, 'release': __releasenotes__, 'doc': __doc_url__ } Fix bad indentation that broke PEP8 !
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Alignak backend import This module contains utility tools to import Nagios-like flat files configuration into an Alignak REST backend. """ # Application version and manifest VERSION = (0, 4, 3) __application__ = u"Alignak backend import" __short_version__ = '.'.join((str(each) for each in VERSION[:2])) __version__ = '.'.join((str(each) for each in VERSION[:4])) __author__ = u"Alignak team" __copyright__ = u"(c) 2015-2016, %s" % __author__ __license__ = u"GNU Affero General Public License, version 3" __description__ = u"Alignak backend import tools" __releasenotes__ = u"""Alignak Backend import tools""" __doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import" # Application manifest manifest = { 'name': __application__, 'version': __version__, 'author': __author__, 'description': __description__, 'copyright': __copyright__, 'license': __license__, 'release': __releasenotes__, 'doc': __doc_url__ }
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """ Alignak backend import This module contains utility tools to import Nagios-like flat files configuration into an Alignak REST backend. """ # Application version and manifest VERSION = (0, 4, 3) __application__ = u"Alignak backend import" __short_version__ = '.'.join((str(each) for each in VERSION[:2])) __version__ = '.'.join((str(each) for each in VERSION[:4])) __author__ = u"Alignak team" __copyright__ = u"(c) 2015-2016, %s" % __author__ __license__ = u"GNU Affero General Public License, version 3" __description__ = u"Alignak backend import tools" __releasenotes__ = u"""Alignak Backend import tools""" __doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import" # Application manifest manifest = { 'name': __application__, 'version': __version__, 'author': __author__, 'description': __description__, 'copyright': __copyright__, 'license': __license__, 'release': __releasenotes__, 'doc': __doc_url__ } <commit_msg>Fix bad indentation that broke PEP8 !<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Alignak backend import This module contains utility tools to import Nagios-like flat files configuration into an Alignak REST backend. """ # Application version and manifest VERSION = (0, 4, 3) __application__ = u"Alignak backend import" __short_version__ = '.'.join((str(each) for each in VERSION[:2])) __version__ = '.'.join((str(each) for each in VERSION[:4])) __author__ = u"Alignak team" __copyright__ = u"(c) 2015-2016, %s" % __author__ __license__ = u"GNU Affero General Public License, version 3" __description__ = u"Alignak backend import tools" __releasenotes__ = u"""Alignak Backend import tools""" __doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import" # Application manifest manifest = { 'name': __application__, 'version': __version__, 'author': __author__, 'description': __description__, 'copyright': __copyright__, 'license': __license__, 'release': __releasenotes__, 'doc': __doc_url__ }
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Alignak backend import This module contains utility tools to import Nagios-like flat files configuration into an Alignak REST backend. """ # Application version and manifest VERSION = (0, 4, 3) __application__ = u"Alignak backend import" __short_version__ = '.'.join((str(each) for each in VERSION[:2])) __version__ = '.'.join((str(each) for each in VERSION[:4])) __author__ = u"Alignak team" __copyright__ = u"(c) 2015-2016, %s" % __author__ __license__ = u"GNU Affero General Public License, version 3" __description__ = u"Alignak backend import tools" __releasenotes__ = u"""Alignak Backend import tools""" __doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import" # Application manifest manifest = { 'name': __application__, 'version': __version__, 'author': __author__, 'description': __description__, 'copyright': __copyright__, 'license': __license__, 'release': __releasenotes__, 'doc': __doc_url__ } Fix bad indentation that broke PEP8 !#!/usr/bin/env python # -*- coding: utf-8 -*- """ Alignak backend import This module contains utility tools to import Nagios-like flat files configuration into an Alignak REST backend. """ # Application version and manifest VERSION = (0, 4, 3) __application__ = u"Alignak backend import" __short_version__ = '.'.join((str(each) for each in VERSION[:2])) __version__ = '.'.join((str(each) for each in VERSION[:4])) __author__ = u"Alignak team" __copyright__ = u"(c) 2015-2016, %s" % __author__ __license__ = u"GNU Affero General Public License, version 3" __description__ = u"Alignak backend import tools" __releasenotes__ = u"""Alignak Backend import tools""" __doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import" # Application manifest manifest = { 'name': __application__, 'version': __version__, 'author': __author__, 'description': __description__, 'copyright': __copyright__, 'license': __license__, 'release': __releasenotes__, 'doc': __doc_url__ }
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """ Alignak backend import This module contains utility tools to import Nagios-like flat files configuration into an Alignak REST backend. """ # Application version and manifest VERSION = (0, 4, 3) __application__ = u"Alignak backend import" __short_version__ = '.'.join((str(each) for each in VERSION[:2])) __version__ = '.'.join((str(each) for each in VERSION[:4])) __author__ = u"Alignak team" __copyright__ = u"(c) 2015-2016, %s" % __author__ __license__ = u"GNU Affero General Public License, version 3" __description__ = u"Alignak backend import tools" __releasenotes__ = u"""Alignak Backend import tools""" __doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import" # Application manifest manifest = { 'name': __application__, 'version': __version__, 'author': __author__, 'description': __description__, 'copyright': __copyright__, 'license': __license__, 'release': __releasenotes__, 'doc': __doc_url__ } <commit_msg>Fix bad indentation that broke PEP8 !<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- """ Alignak backend import This module contains utility tools to import Nagios-like flat files configuration into an Alignak REST backend. """ # Application version and manifest VERSION = (0, 4, 3) __application__ = u"Alignak backend import" __short_version__ = '.'.join((str(each) for each in VERSION[:2])) __version__ = '.'.join((str(each) for each in VERSION[:4])) __author__ = u"Alignak team" __copyright__ = u"(c) 2015-2016, %s" % __author__ __license__ = u"GNU Affero General Public License, version 3" __description__ = u"Alignak backend import tools" __releasenotes__ = u"""Alignak Backend import tools""" __doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import" # Application manifest manifest = { 'name': __application__, 'version': __version__, 'author': __author__, 'description': __description__, 'copyright': __copyright__, 'license': __license__, 'release': __releasenotes__, 'doc': __doc_url__ }
da447a06c65a421b8d2fe122cb7b6bee5d11614e
app.py
app.py
import sys from upload_s3 import set_metadata from flask import Flask, render_template from flask_frozen import Freezer app = Flask(__name__) freezer = Freezer(app) app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html' app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True app.config['FREEZER_BASE_URL'] = 'http://www.vpr.net/apps/test' # If project doesn't have it's own domain/subdomain, use BASE_URL # app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base' # If Flask is needed to generate URLs, use freezer.register_generator # see: http://pythonhosted.org/Frozen-Flask/#url-generators @app.route('/') def index(): return render_template('content.html') @app.route('/test/work') def work(): return render_template('content.html') if __name__ == '__main__': if len(sys.argv) > 1 and sys.argv[1] == 'build': freezer.freeze() set_metadata() else: app.run(debug=True)
import sys from upload_s3 import set_metadata from flask import Flask, render_template from flask_frozen import Freezer app = Flask(__name__) freezer = Freezer(app) app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html' app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True # If project doesn't have it's own domain/subdomain, use BASE_URL # app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base' # If Flask is needed to generate URLs, use freezer.register_generator # see: http://pythonhosted.org/Frozen-Flask/#url-generators @app.route('/') def index(): return render_template('content.html') if __name__ == '__main__': if len(sys.argv) > 1 and sys.argv[1] == 'build': freezer.freeze() set_metadata() else: app.run(debug=True)
Comment out optional base url configuration
Comment out optional base url configuration
Python
apache-2.0
vprnet/interactive-transcript-gov-peter-shumlins-2015-budget-speech,vprnet/google-s3-json,vprnet/app-template,vprnet/EOTS-iframe-widget,vprnet/soundcloud-podcast,vprnet/interactive-transcript-gov-peter-shumlins-2015-budget-speech,vprnet/timeline-dcf-systemic-failure,vprnet/live-from-the-fort,vprnet/live-from-the-fort,vprnet/EOTS-iframe-widget,vprnet/timeline-dcf-systemic-failure,vprnet/app-template,vprnet/old-app-template,vprnet/interactive-transcript-gov-peter-shumlins-third-inaugural-address,vprnet/live-from-the-fort,vprnet/EOTS-iframe-widget,vprnet/old-app-template,vprnet/app-template,vprnet/interactive-transcript-gov-peter-shumlins-third-inaugural-address
import sys from upload_s3 import set_metadata from flask import Flask, render_template from flask_frozen import Freezer app = Flask(__name__) freezer = Freezer(app) app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html' app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True app.config['FREEZER_BASE_URL'] = 'http://www.vpr.net/apps/test' # If project doesn't have it's own domain/subdomain, use BASE_URL # app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base' # If Flask is needed to generate URLs, use freezer.register_generator # see: http://pythonhosted.org/Frozen-Flask/#url-generators @app.route('/') def index(): return render_template('content.html') @app.route('/test/work') def work(): return render_template('content.html') if __name__ == '__main__': if len(sys.argv) > 1 and sys.argv[1] == 'build': freezer.freeze() set_metadata() else: app.run(debug=True) Comment out optional base url configuration
import sys from upload_s3 import set_metadata from flask import Flask, render_template from flask_frozen import Freezer app = Flask(__name__) freezer = Freezer(app) app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html' app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True # If project doesn't have it's own domain/subdomain, use BASE_URL # app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base' # If Flask is needed to generate URLs, use freezer.register_generator # see: http://pythonhosted.org/Frozen-Flask/#url-generators @app.route('/') def index(): return render_template('content.html') if __name__ == '__main__': if len(sys.argv) > 1 and sys.argv[1] == 'build': freezer.freeze() set_metadata() else: app.run(debug=True)
<commit_before>import sys from upload_s3 import set_metadata from flask import Flask, render_template from flask_frozen import Freezer app = Flask(__name__) freezer = Freezer(app) app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html' app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True app.config['FREEZER_BASE_URL'] = 'http://www.vpr.net/apps/test' # If project doesn't have it's own domain/subdomain, use BASE_URL # app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base' # If Flask is needed to generate URLs, use freezer.register_generator # see: http://pythonhosted.org/Frozen-Flask/#url-generators @app.route('/') def index(): return render_template('content.html') @app.route('/test/work') def work(): return render_template('content.html') if __name__ == '__main__': if len(sys.argv) > 1 and sys.argv[1] == 'build': freezer.freeze() set_metadata() else: app.run(debug=True) <commit_msg>Comment out optional base url configuration<commit_after>
import sys from upload_s3 import set_metadata from flask import Flask, render_template from flask_frozen import Freezer app = Flask(__name__) freezer = Freezer(app) app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html' app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True # If project doesn't have it's own domain/subdomain, use BASE_URL # app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base' # If Flask is needed to generate URLs, use freezer.register_generator # see: http://pythonhosted.org/Frozen-Flask/#url-generators @app.route('/') def index(): return render_template('content.html') if __name__ == '__main__': if len(sys.argv) > 1 and sys.argv[1] == 'build': freezer.freeze() set_metadata() else: app.run(debug=True)
import sys from upload_s3 import set_metadata from flask import Flask, render_template from flask_frozen import Freezer app = Flask(__name__) freezer = Freezer(app) app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html' app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True app.config['FREEZER_BASE_URL'] = 'http://www.vpr.net/apps/test' # If project doesn't have it's own domain/subdomain, use BASE_URL # app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base' # If Flask is needed to generate URLs, use freezer.register_generator # see: http://pythonhosted.org/Frozen-Flask/#url-generators @app.route('/') def index(): return render_template('content.html') @app.route('/test/work') def work(): return render_template('content.html') if __name__ == '__main__': if len(sys.argv) > 1 and sys.argv[1] == 'build': freezer.freeze() set_metadata() else: app.run(debug=True) Comment out optional base url configurationimport sys from upload_s3 import set_metadata from flask import Flask, render_template from flask_frozen import Freezer app = Flask(__name__) freezer = Freezer(app) app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html' app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True # If project doesn't have it's own domain/subdomain, use BASE_URL # app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base' # If Flask is needed to generate URLs, use freezer.register_generator # see: http://pythonhosted.org/Frozen-Flask/#url-generators @app.route('/') def index(): return render_template('content.html') if __name__ == '__main__': if len(sys.argv) > 1 and sys.argv[1] == 'build': freezer.freeze() set_metadata() else: app.run(debug=True)
<commit_before>import sys from upload_s3 import set_metadata from flask import Flask, render_template from flask_frozen import Freezer app = Flask(__name__) freezer = Freezer(app) app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html' app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True app.config['FREEZER_BASE_URL'] = 'http://www.vpr.net/apps/test' # If project doesn't have it's own domain/subdomain, use BASE_URL # app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base' # If Flask is needed to generate URLs, use freezer.register_generator # see: http://pythonhosted.org/Frozen-Flask/#url-generators @app.route('/') def index(): return render_template('content.html') @app.route('/test/work') def work(): return render_template('content.html') if __name__ == '__main__': if len(sys.argv) > 1 and sys.argv[1] == 'build': freezer.freeze() set_metadata() else: app.run(debug=True) <commit_msg>Comment out optional base url configuration<commit_after>import sys from upload_s3 import set_metadata from flask import Flask, render_template from flask_frozen import Freezer app = Flask(__name__) freezer = Freezer(app) app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html' app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True # If project doesn't have it's own domain/subdomain, use BASE_URL # app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base' # If Flask is needed to generate URLs, use freezer.register_generator # see: http://pythonhosted.org/Frozen-Flask/#url-generators @app.route('/') def index(): return render_template('content.html') if __name__ == '__main__': if len(sys.argv) > 1 and sys.argv[1] == 'build': freezer.freeze() set_metadata() else: app.run(debug=True)
07b22d0e4912678e3e2e5507e8a7fe2488ce2b3c
djangocms_installer/config/urls.py
djangocms_installer/config/urls.py
# -*- coding: utf-8 -*- from __future__ import absolute_import, print_function, unicode_literals from cms.sitemaps import CMSSitemap from django.conf import settings from django.conf.urls import include, patterns, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns admin.autodiscover() urlpatterns = i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^select2/', include('django_select2.urls')), url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA
# -*- coding: utf-8 -*- from __future__ import absolute_import, print_function, unicode_literals from cms.sitemaps import CMSSitemap from django.conf import settings from django.conf.urls import include, patterns, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns admin.autodiscover() urlpatterns = [ url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^select2/', include('django_select2.urls')), ] urlpatterns += i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA
Move sitemaps to non-language prefix url
Move sitemaps to non-language prefix url
Python
bsd-3-clause
nephila/djangocms-installer,nephila/djangocms-installer
# -*- coding: utf-8 -*- from __future__ import absolute_import, print_function, unicode_literals from cms.sitemaps import CMSSitemap from django.conf import settings from django.conf.urls import include, patterns, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns admin.autodiscover() urlpatterns = i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^select2/', include('django_select2.urls')), url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA Move sitemaps to non-language prefix url
# -*- coding: utf-8 -*- from __future__ import absolute_import, print_function, unicode_literals from cms.sitemaps import CMSSitemap from django.conf import settings from django.conf.urls import include, patterns, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns admin.autodiscover() urlpatterns = [ url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^select2/', include('django_select2.urls')), ] urlpatterns += i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA
<commit_before># -*- coding: utf-8 -*- from __future__ import absolute_import, print_function, unicode_literals from cms.sitemaps import CMSSitemap from django.conf import settings from django.conf.urls import include, patterns, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns admin.autodiscover() urlpatterns = i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^select2/', include('django_select2.urls')), url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA <commit_msg>Move sitemaps to non-language prefix url<commit_after>
# -*- coding: utf-8 -*- from __future__ import absolute_import, print_function, unicode_literals from cms.sitemaps import CMSSitemap from django.conf import settings from django.conf.urls import include, patterns, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns admin.autodiscover() urlpatterns = [ url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^select2/', include('django_select2.urls')), ] urlpatterns += i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA
# -*- coding: utf-8 -*- from __future__ import absolute_import, print_function, unicode_literals from cms.sitemaps import CMSSitemap from django.conf import settings from django.conf.urls import include, patterns, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns admin.autodiscover() urlpatterns = i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^select2/', include('django_select2.urls')), url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA Move sitemaps to non-language prefix url# -*- coding: utf-8 -*- from __future__ import absolute_import, print_function, unicode_literals from cms.sitemaps import CMSSitemap from django.conf import settings from django.conf.urls import include, patterns, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns admin.autodiscover() urlpatterns = [ url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^select2/', include('django_select2.urls')), ] urlpatterns += i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA
<commit_before># -*- coding: utf-8 -*- from __future__ import absolute_import, print_function, unicode_literals from cms.sitemaps import CMSSitemap from django.conf import settings from django.conf.urls import include, patterns, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns admin.autodiscover() urlpatterns = i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^select2/', include('django_select2.urls')), url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA <commit_msg>Move sitemaps to non-language prefix url<commit_after># -*- coding: utf-8 -*- from __future__ import absolute_import, print_function, unicode_literals from cms.sitemaps import CMSSitemap from django.conf import settings from django.conf.urls import include, patterns, url from django.conf.urls.i18n import i18n_patterns from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns admin.autodiscover() urlpatterns = [ url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^select2/', include('django_select2.urls')), ] urlpatterns += i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA
c2a2776f6bbff866078ed2ecaa30f9756af1a5dc
flow_workflow/historian/handler.py
flow_workflow/historian/handler.py
import logging import sys from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from flow_workflow.historian.messages import UpdateMessage LOG = logging.getLogger(__name__) class WorkflowHistorianMessageHandler(object): message_class = UpdateMessage def __init__(self, broker=None, storage=None, queue_name=None): self.broker = broker self.storage = storage self.queue_name = queue_name def __call__(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.") sys._exit()
import logging import os from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from flow_workflow.historian.messages import UpdateMessage LOG = logging.getLogger(__name__) class WorkflowHistorianMessageHandler(object): message_class = UpdateMessage def __init__(self, broker=None, storage=None, queue_name=None): self.broker = broker self.storage = storage self.queue_name = queue_name def __call__(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.") os._exit(1)
Use the more betterer os._exit
Use the more betterer os._exit
Python
agpl-3.0
genome/flow-workflow,genome/flow-workflow,genome/flow-workflow
import logging import sys from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from flow_workflow.historian.messages import UpdateMessage LOG = logging.getLogger(__name__) class WorkflowHistorianMessageHandler(object): message_class = UpdateMessage def __init__(self, broker=None, storage=None, queue_name=None): self.broker = broker self.storage = storage self.queue_name = queue_name def __call__(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.") sys._exit() Use the more betterer os._exit
import logging import os from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from flow_workflow.historian.messages import UpdateMessage LOG = logging.getLogger(__name__) class WorkflowHistorianMessageHandler(object): message_class = UpdateMessage def __init__(self, broker=None, storage=None, queue_name=None): self.broker = broker self.storage = storage self.queue_name = queue_name def __call__(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.") os._exit(1)
<commit_before>import logging import sys from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from flow_workflow.historian.messages import UpdateMessage LOG = logging.getLogger(__name__) class WorkflowHistorianMessageHandler(object): message_class = UpdateMessage def __init__(self, broker=None, storage=None, queue_name=None): self.broker = broker self.storage = storage self.queue_name = queue_name def __call__(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.") sys._exit() <commit_msg>Use the more betterer os._exit<commit_after>
import logging import os from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from flow_workflow.historian.messages import UpdateMessage LOG = logging.getLogger(__name__) class WorkflowHistorianMessageHandler(object): message_class = UpdateMessage def __init__(self, broker=None, storage=None, queue_name=None): self.broker = broker self.storage = storage self.queue_name = queue_name def __call__(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.") os._exit(1)
import logging import sys from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from flow_workflow.historian.messages import UpdateMessage LOG = logging.getLogger(__name__) class WorkflowHistorianMessageHandler(object): message_class = UpdateMessage def __init__(self, broker=None, storage=None, queue_name=None): self.broker = broker self.storage = storage self.queue_name = queue_name def __call__(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.") sys._exit() Use the more betterer os._exitimport logging import os from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from flow_workflow.historian.messages import UpdateMessage LOG = logging.getLogger(__name__) class WorkflowHistorianMessageHandler(object): message_class = UpdateMessage def __init__(self, broker=None, storage=None, queue_name=None): self.broker = broker self.storage = storage self.queue_name = queue_name def __call__(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.") os._exit(1)
<commit_before>import logging import sys from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from flow_workflow.historian.messages import UpdateMessage LOG = logging.getLogger(__name__) class WorkflowHistorianMessageHandler(object): message_class = UpdateMessage def __init__(self, broker=None, storage=None, queue_name=None): self.broker = broker self.storage = storage self.queue_name = queue_name def __call__(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.") sys._exit() <commit_msg>Use the more betterer os._exit<commit_after>import logging import os from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError from flow_workflow.historian.messages import UpdateMessage LOG = logging.getLogger(__name__) class WorkflowHistorianMessageHandler(object): message_class = UpdateMessage def __init__(self, broker=None, storage=None, queue_name=None): self.broker = broker self.storage = storage self.queue_name = queue_name def __call__(self, message): message_dict = message.to_dict() LOG.info("Updating [net_key='%s', operation_id='%s']: %r", message.net_key, message.operation_id, message_dict) try: self.storage.update(message_dict) except (ResourceClosedError, TimeoutError, DisconnectionError): LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.") os._exit(1)
b698c7ab1c13353d8e9538bb42797344049812c1
astro.py
astro.py
import ephem from datetime import datetime def const(planet_name): # function name and parameters planet_class = getattr(ephem, planet_name) # sets ephem object class date_class = datetime.now() planet = planet_class() # sets planet variable south_bend = ephem.Observer() # Creates the Observer object south_bend.lat = '41.40' # latitude south_bend.lon = '-86.15' south_bend.date = date_class # sets date parameter planet.compute(south_bend) # calculates the location data print date_class print planet.ra, planet.dec return ephem.constellation((planet.ra, planet.dec)) print const(raw_input('Planet: '))
import ephem from datetime import datetime def const(planet_name): # function name and parameters planet_class = getattr(ephem, planet_name) # sets ephem object class date_class = datetime.now() planet = planet_class() # sets planet variable south_bend = ephem.Observer() # Creates the Observer object south_bend.lat = '41.40' # latitude south_bend.lon = '-86.15' south_bend.date = date_class # sets date parameter planet.compute(south_bend) # calculates the location data print date_class print planet.ra, planet.dec print planet.alt, planet.az return ephem.constellation((planet.ra, planet.dec)) print const(raw_input('Planet: '))
Add to line 15 for testing.
Add to line 15 for testing.
Python
mit
bennettscience/PySky
import ephem from datetime import datetime def const(planet_name): # function name and parameters planet_class = getattr(ephem, planet_name) # sets ephem object class date_class = datetime.now() planet = planet_class() # sets planet variable south_bend = ephem.Observer() # Creates the Observer object south_bend.lat = '41.40' # latitude south_bend.lon = '-86.15' south_bend.date = date_class # sets date parameter planet.compute(south_bend) # calculates the location data print date_class print planet.ra, planet.dec return ephem.constellation((planet.ra, planet.dec)) print const(raw_input('Planet: ')) Add to line 15 for testing.
import ephem from datetime import datetime def const(planet_name): # function name and parameters planet_class = getattr(ephem, planet_name) # sets ephem object class date_class = datetime.now() planet = planet_class() # sets planet variable south_bend = ephem.Observer() # Creates the Observer object south_bend.lat = '41.40' # latitude south_bend.lon = '-86.15' south_bend.date = date_class # sets date parameter planet.compute(south_bend) # calculates the location data print date_class print planet.ra, planet.dec print planet.alt, planet.az return ephem.constellation((planet.ra, planet.dec)) print const(raw_input('Planet: '))
<commit_before>import ephem from datetime import datetime def const(planet_name): # function name and parameters planet_class = getattr(ephem, planet_name) # sets ephem object class date_class = datetime.now() planet = planet_class() # sets planet variable south_bend = ephem.Observer() # Creates the Observer object south_bend.lat = '41.40' # latitude south_bend.lon = '-86.15' south_bend.date = date_class # sets date parameter planet.compute(south_bend) # calculates the location data print date_class print planet.ra, planet.dec return ephem.constellation((planet.ra, planet.dec)) print const(raw_input('Planet: ')) <commit_msg>Add to line 15 for testing.<commit_after>
import ephem from datetime import datetime def const(planet_name): # function name and parameters planet_class = getattr(ephem, planet_name) # sets ephem object class date_class = datetime.now() planet = planet_class() # sets planet variable south_bend = ephem.Observer() # Creates the Observer object south_bend.lat = '41.40' # latitude south_bend.lon = '-86.15' south_bend.date = date_class # sets date parameter planet.compute(south_bend) # calculates the location data print date_class print planet.ra, planet.dec print planet.alt, planet.az return ephem.constellation((planet.ra, planet.dec)) print const(raw_input('Planet: '))
import ephem from datetime import datetime def const(planet_name): # function name and parameters planet_class = getattr(ephem, planet_name) # sets ephem object class date_class = datetime.now() planet = planet_class() # sets planet variable south_bend = ephem.Observer() # Creates the Observer object south_bend.lat = '41.40' # latitude south_bend.lon = '-86.15' south_bend.date = date_class # sets date parameter planet.compute(south_bend) # calculates the location data print date_class print planet.ra, planet.dec return ephem.constellation((planet.ra, planet.dec)) print const(raw_input('Planet: ')) Add to line 15 for testing.import ephem from datetime import datetime def const(planet_name): # function name and parameters planet_class = getattr(ephem, planet_name) # sets ephem object class date_class = datetime.now() planet = planet_class() # sets planet variable south_bend = ephem.Observer() # Creates the Observer object south_bend.lat = '41.40' # latitude south_bend.lon = '-86.15' south_bend.date = date_class # sets date parameter planet.compute(south_bend) # calculates the location data print date_class print planet.ra, planet.dec print planet.alt, planet.az return ephem.constellation((planet.ra, planet.dec)) print const(raw_input('Planet: '))
<commit_before>import ephem from datetime import datetime def const(planet_name): # function name and parameters planet_class = getattr(ephem, planet_name) # sets ephem object class date_class = datetime.now() planet = planet_class() # sets planet variable south_bend = ephem.Observer() # Creates the Observer object south_bend.lat = '41.40' # latitude south_bend.lon = '-86.15' south_bend.date = date_class # sets date parameter planet.compute(south_bend) # calculates the location data print date_class print planet.ra, planet.dec return ephem.constellation((planet.ra, planet.dec)) print const(raw_input('Planet: ')) <commit_msg>Add to line 15 for testing.<commit_after>import ephem from datetime import datetime def const(planet_name): # function name and parameters planet_class = getattr(ephem, planet_name) # sets ephem object class date_class = datetime.now() planet = planet_class() # sets planet variable south_bend = ephem.Observer() # Creates the Observer object south_bend.lat = '41.40' # latitude south_bend.lon = '-86.15' south_bend.date = date_class # sets date parameter planet.compute(south_bend) # calculates the location data print date_class print planet.ra, planet.dec print planet.alt, planet.az return ephem.constellation((planet.ra, planet.dec)) print const(raw_input('Planet: '))
7816fc20c2e46a9d1a1e7c11a061862163018069
tests/test_vfg_path.py
tests/test_vfg_path.py
import angr import logging import os l = logging.getLogger("angr_tests") test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)), '../../binaries/tests')) def test_vfg_paths(): p = angr.Project(os.path.join(test_location, "x86_64/track_user_input")) main_addr = p.loader.main_bin.get_symbol("main").addr printf_addr = 0x4005e1 # actually where it returns vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=4) paths = vfg.get_paths(main_addr, printf_addr) if __name__ == '__main__': test_vfg_paths()
import angr import logging import os l = logging.getLogger("angr_tests") test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)), '../../binaries/tests')) def test_vfg_paths(): p = angr.Project(os.path.join(test_location, "x86_64/track_user_input")) main_addr = p.loader.main_bin.get_symbol("main").addr printf_addr = 0x4005e1 # actually where it returns vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=5) paths = vfg.get_paths(main_addr, printf_addr) if __name__ == '__main__': test_vfg_paths()
Fix the VFG path test.
Fix the VFG path test.
Python
bsd-2-clause
tyb0807/angr,schieb/angr,schieb/angr,tyb0807/angr,chubbymaggie/angr,schieb/angr,chubbymaggie/angr,f-prettyland/angr,chubbymaggie/angr,angr/angr,angr/angr,axt/angr,axt/angr,axt/angr,f-prettyland/angr,angr/angr,tyb0807/angr,f-prettyland/angr,iamahuman/angr,iamahuman/angr,iamahuman/angr
import angr import logging import os l = logging.getLogger("angr_tests") test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)), '../../binaries/tests')) def test_vfg_paths(): p = angr.Project(os.path.join(test_location, "x86_64/track_user_input")) main_addr = p.loader.main_bin.get_symbol("main").addr printf_addr = 0x4005e1 # actually where it returns vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=4) paths = vfg.get_paths(main_addr, printf_addr) if __name__ == '__main__': test_vfg_paths() Fix the VFG path test.
import angr import logging import os l = logging.getLogger("angr_tests") test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)), '../../binaries/tests')) def test_vfg_paths(): p = angr.Project(os.path.join(test_location, "x86_64/track_user_input")) main_addr = p.loader.main_bin.get_symbol("main").addr printf_addr = 0x4005e1 # actually where it returns vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=5) paths = vfg.get_paths(main_addr, printf_addr) if __name__ == '__main__': test_vfg_paths()
<commit_before>import angr import logging import os l = logging.getLogger("angr_tests") test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)), '../../binaries/tests')) def test_vfg_paths(): p = angr.Project(os.path.join(test_location, "x86_64/track_user_input")) main_addr = p.loader.main_bin.get_symbol("main").addr printf_addr = 0x4005e1 # actually where it returns vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=4) paths = vfg.get_paths(main_addr, printf_addr) if __name__ == '__main__': test_vfg_paths() <commit_msg>Fix the VFG path test.<commit_after>
import angr import logging import os l = logging.getLogger("angr_tests") test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)), '../../binaries/tests')) def test_vfg_paths(): p = angr.Project(os.path.join(test_location, "x86_64/track_user_input")) main_addr = p.loader.main_bin.get_symbol("main").addr printf_addr = 0x4005e1 # actually where it returns vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=5) paths = vfg.get_paths(main_addr, printf_addr) if __name__ == '__main__': test_vfg_paths()
import angr import logging import os l = logging.getLogger("angr_tests") test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)), '../../binaries/tests')) def test_vfg_paths(): p = angr.Project(os.path.join(test_location, "x86_64/track_user_input")) main_addr = p.loader.main_bin.get_symbol("main").addr printf_addr = 0x4005e1 # actually where it returns vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=4) paths = vfg.get_paths(main_addr, printf_addr) if __name__ == '__main__': test_vfg_paths() Fix the VFG path test.import angr import logging import os l = logging.getLogger("angr_tests") test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)), '../../binaries/tests')) def test_vfg_paths(): p = angr.Project(os.path.join(test_location, "x86_64/track_user_input")) main_addr = p.loader.main_bin.get_symbol("main").addr printf_addr = 0x4005e1 # actually where it returns vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=5) paths = vfg.get_paths(main_addr, printf_addr) if __name__ == '__main__': test_vfg_paths()
<commit_before>import angr import logging import os l = logging.getLogger("angr_tests") test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)), '../../binaries/tests')) def test_vfg_paths(): p = angr.Project(os.path.join(test_location, "x86_64/track_user_input")) main_addr = p.loader.main_bin.get_symbol("main").addr printf_addr = 0x4005e1 # actually where it returns vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=4) paths = vfg.get_paths(main_addr, printf_addr) if __name__ == '__main__': test_vfg_paths() <commit_msg>Fix the VFG path test.<commit_after>import angr import logging import os l = logging.getLogger("angr_tests") test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)), '../../binaries/tests')) def test_vfg_paths(): p = angr.Project(os.path.join(test_location, "x86_64/track_user_input")) main_addr = p.loader.main_bin.get_symbol("main").addr printf_addr = 0x4005e1 # actually where it returns vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=5) paths = vfg.get_paths(main_addr, printf_addr) if __name__ == '__main__': test_vfg_paths()
d0018748cae3f0af4818106643926f7e8effe3c6
monasca_log_api_tempest/clients.py
monasca_log_api_tempest/clients.py
# Copyright 2015 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from tempest import clients from monasca_log_api_tempest.services import log_api_v2_client from monasca_log_api_tempest.services import log_api_v3_client from monasca_log_api_tempest.services import log_search_client class Manager(clients.Manager): def __init__(self, credentials=None, service=None): super(Manager, self).__init__(credentials, service) self.log_api_clients = { "v2": log_api_v2_client.LogApiV2Client( self.auth_provider, 'logs_v2', None ), "v3": log_api_v3_client.LogApiV3Client( self.auth_provider, 'logs', None ) } self.log_search_client = log_search_client.LogsSearchClient( self.auth_provider, 'logs-search', None )
# Copyright 2015-2016 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from tempest import clients from monasca_log_api_tempest.services import log_api_v2_client from monasca_log_api_tempest.services import log_api_v3_client from monasca_log_api_tempest.services import log_search_client class Manager(clients.Manager): def __init__(self, credentials=None): super(Manager, self).__init__(credentials) self.log_api_clients = { "v2": log_api_v2_client.LogApiV2Client( self.auth_provider, 'logs_v2', None ), "v3": log_api_v3_client.LogApiV3Client( self.auth_provider, 'logs', None ) } self.log_search_client = log_search_client.LogsSearchClient( self.auth_provider, 'logs-search', None )
Fix the Monasca Log API tempest tests
Fix the Monasca Log API tempest tests The Tempest Manager class must have changed and the service argument apparently no longer exists. Instead, it was being set as the scope which caused the catalog to not be retrieved See-also: If934bac4e2cd833fe4e381c373218383354969ec Change-Id: I43c023e91eb93e2c19096b0de812eabf7b2db62c
Python
apache-2.0
openstack/monasca-log-api,stackforge/monasca-log-api,stackforge/monasca-log-api,stackforge/monasca-log-api,openstack/monasca-log-api,openstack/monasca-log-api
# Copyright 2015 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from tempest import clients from monasca_log_api_tempest.services import log_api_v2_client from monasca_log_api_tempest.services import log_api_v3_client from monasca_log_api_tempest.services import log_search_client class Manager(clients.Manager): def __init__(self, credentials=None, service=None): super(Manager, self).__init__(credentials, service) self.log_api_clients = { "v2": log_api_v2_client.LogApiV2Client( self.auth_provider, 'logs_v2', None ), "v3": log_api_v3_client.LogApiV3Client( self.auth_provider, 'logs', None ) } self.log_search_client = log_search_client.LogsSearchClient( self.auth_provider, 'logs-search', None ) Fix the Monasca Log API tempest tests The Tempest Manager class must have changed and the service argument apparently no longer exists. Instead, it was being set as the scope which caused the catalog to not be retrieved See-also: If934bac4e2cd833fe4e381c373218383354969ec Change-Id: I43c023e91eb93e2c19096b0de812eabf7b2db62c
# Copyright 2015-2016 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from tempest import clients from monasca_log_api_tempest.services import log_api_v2_client from monasca_log_api_tempest.services import log_api_v3_client from monasca_log_api_tempest.services import log_search_client class Manager(clients.Manager): def __init__(self, credentials=None): super(Manager, self).__init__(credentials) self.log_api_clients = { "v2": log_api_v2_client.LogApiV2Client( self.auth_provider, 'logs_v2', None ), "v3": log_api_v3_client.LogApiV3Client( self.auth_provider, 'logs', None ) } self.log_search_client = log_search_client.LogsSearchClient( self.auth_provider, 'logs-search', None )
<commit_before># Copyright 2015 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from tempest import clients from monasca_log_api_tempest.services import log_api_v2_client from monasca_log_api_tempest.services import log_api_v3_client from monasca_log_api_tempest.services import log_search_client class Manager(clients.Manager): def __init__(self, credentials=None, service=None): super(Manager, self).__init__(credentials, service) self.log_api_clients = { "v2": log_api_v2_client.LogApiV2Client( self.auth_provider, 'logs_v2', None ), "v3": log_api_v3_client.LogApiV3Client( self.auth_provider, 'logs', None ) } self.log_search_client = log_search_client.LogsSearchClient( self.auth_provider, 'logs-search', None ) <commit_msg>Fix the Monasca Log API tempest tests The Tempest Manager class must have changed and the service argument apparently no longer exists. Instead, it was being set as the scope which caused the catalog to not be retrieved See-also: If934bac4e2cd833fe4e381c373218383354969ec Change-Id: I43c023e91eb93e2c19096b0de812eabf7b2db62c<commit_after>
# Copyright 2015-2016 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from tempest import clients from monasca_log_api_tempest.services import log_api_v2_client from monasca_log_api_tempest.services import log_api_v3_client from monasca_log_api_tempest.services import log_search_client class Manager(clients.Manager): def __init__(self, credentials=None): super(Manager, self).__init__(credentials) self.log_api_clients = { "v2": log_api_v2_client.LogApiV2Client( self.auth_provider, 'logs_v2', None ), "v3": log_api_v3_client.LogApiV3Client( self.auth_provider, 'logs', None ) } self.log_search_client = log_search_client.LogsSearchClient( self.auth_provider, 'logs-search', None )
# Copyright 2015 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from tempest import clients from monasca_log_api_tempest.services import log_api_v2_client from monasca_log_api_tempest.services import log_api_v3_client from monasca_log_api_tempest.services import log_search_client class Manager(clients.Manager): def __init__(self, credentials=None, service=None): super(Manager, self).__init__(credentials, service) self.log_api_clients = { "v2": log_api_v2_client.LogApiV2Client( self.auth_provider, 'logs_v2', None ), "v3": log_api_v3_client.LogApiV3Client( self.auth_provider, 'logs', None ) } self.log_search_client = log_search_client.LogsSearchClient( self.auth_provider, 'logs-search', None ) Fix the Monasca Log API tempest tests The Tempest Manager class must have changed and the service argument apparently no longer exists. Instead, it was being set as the scope which caused the catalog to not be retrieved See-also: If934bac4e2cd833fe4e381c373218383354969ec Change-Id: I43c023e91eb93e2c19096b0de812eabf7b2db62c# Copyright 2015-2016 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from tempest import clients from monasca_log_api_tempest.services import log_api_v2_client from monasca_log_api_tempest.services import log_api_v3_client from monasca_log_api_tempest.services import log_search_client class Manager(clients.Manager): def __init__(self, credentials=None): super(Manager, self).__init__(credentials) self.log_api_clients = { "v2": log_api_v2_client.LogApiV2Client( self.auth_provider, 'logs_v2', None ), "v3": log_api_v3_client.LogApiV3Client( self.auth_provider, 'logs', None ) } self.log_search_client = log_search_client.LogsSearchClient( self.auth_provider, 'logs-search', None )
<commit_before># Copyright 2015 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from tempest import clients from monasca_log_api_tempest.services import log_api_v2_client from monasca_log_api_tempest.services import log_api_v3_client from monasca_log_api_tempest.services import log_search_client class Manager(clients.Manager): def __init__(self, credentials=None, service=None): super(Manager, self).__init__(credentials, service) self.log_api_clients = { "v2": log_api_v2_client.LogApiV2Client( self.auth_provider, 'logs_v2', None ), "v3": log_api_v3_client.LogApiV3Client( self.auth_provider, 'logs', None ) } self.log_search_client = log_search_client.LogsSearchClient( self.auth_provider, 'logs-search', None ) <commit_msg>Fix the Monasca Log API tempest tests The Tempest Manager class must have changed and the service argument apparently no longer exists. Instead, it was being set as the scope which caused the catalog to not be retrieved See-also: If934bac4e2cd833fe4e381c373218383354969ec Change-Id: I43c023e91eb93e2c19096b0de812eabf7b2db62c<commit_after># Copyright 2015-2016 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from tempest import clients from monasca_log_api_tempest.services import log_api_v2_client from monasca_log_api_tempest.services import log_api_v3_client from monasca_log_api_tempest.services import log_search_client class Manager(clients.Manager): def __init__(self, credentials=None): super(Manager, self).__init__(credentials) self.log_api_clients = { "v2": log_api_v2_client.LogApiV2Client( self.auth_provider, 'logs_v2', None ), "v3": log_api_v3_client.LogApiV3Client( self.auth_provider, 'logs', None ) } self.log_search_client = log_search_client.LogsSearchClient( self.auth_provider, 'logs-search', None )
66fe6f98c079490d2d5de4c161da1d8b3801cda4
monasca_persister/conf/influxdb.py
monasca_persister/conf/influxdb.py
# (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP # Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from oslo_config import cfg influxdb_opts = [ cfg.StrOpt('database_name', help='database name where metrics are stored', default='mon'), cfg.IPOpt('ip_address', help='ip address to influxdb'), cfg.PortOpt('port', help='port to influxdb', default=8086), cfg.StrOpt('user', help='influxdb user ', default='mon_persister'), cfg.StrOpt('password', secret=True, help='influxdb password')] influxdb_group = cfg.OptGroup(name='influxdb', title='influxdb') def register_opts(conf): conf.register_group(influxdb_group) conf.register_opts(influxdb_opts, influxdb_group) def list_opts(): return influxdb_group, influxdb_opts
# (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP # Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from oslo_config import cfg influxdb_opts = [ cfg.StrOpt('database_name', help='database name where metrics are stored', default='mon'), cfg.HostAddressOpt('ip_address', help='Valid IP address or hostname ' 'to InfluxDB instance'), cfg.PortOpt('port', help='port to influxdb', default=8086), cfg.StrOpt('user', help='influxdb user ', default='mon_persister'), cfg.StrOpt('password', secret=True, help='influxdb password')] influxdb_group = cfg.OptGroup(name='influxdb', title='influxdb') def register_opts(conf): conf.register_group(influxdb_group) conf.register_opts(influxdb_opts, influxdb_group) def list_opts(): return influxdb_group, influxdb_opts
Allow hostnames to be used as ip_address
Allow hostnames to be used as ip_address Previously introduced change for monasca-persister had enforced the IPAddress as the only type one can configure influxdb.ip_address property with. Following change makes it possible to use also hostname. Using IPAdress is still possible. Change-Id: Ib0d7f19b3ac2dcb7c84923872d94f180cda58b2b
Python
apache-2.0
stackforge/monasca-persister,openstack/monasca-persister,stackforge/monasca-persister,stackforge/monasca-persister,openstack/monasca-persister,openstack/monasca-persister
# (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP # Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from oslo_config import cfg influxdb_opts = [ cfg.StrOpt('database_name', help='database name where metrics are stored', default='mon'), cfg.IPOpt('ip_address', help='ip address to influxdb'), cfg.PortOpt('port', help='port to influxdb', default=8086), cfg.StrOpt('user', help='influxdb user ', default='mon_persister'), cfg.StrOpt('password', secret=True, help='influxdb password')] influxdb_group = cfg.OptGroup(name='influxdb', title='influxdb') def register_opts(conf): conf.register_group(influxdb_group) conf.register_opts(influxdb_opts, influxdb_group) def list_opts(): return influxdb_group, influxdb_opts Allow hostnames to be used as ip_address Previously introduced change for monasca-persister had enforced the IPAddress as the only type one can configure influxdb.ip_address property with. Following change makes it possible to use also hostname. Using IPAdress is still possible. Change-Id: Ib0d7f19b3ac2dcb7c84923872d94f180cda58b2b
# (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP # Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from oslo_config import cfg influxdb_opts = [ cfg.StrOpt('database_name', help='database name where metrics are stored', default='mon'), cfg.HostAddressOpt('ip_address', help='Valid IP address or hostname ' 'to InfluxDB instance'), cfg.PortOpt('port', help='port to influxdb', default=8086), cfg.StrOpt('user', help='influxdb user ', default='mon_persister'), cfg.StrOpt('password', secret=True, help='influxdb password')] influxdb_group = cfg.OptGroup(name='influxdb', title='influxdb') def register_opts(conf): conf.register_group(influxdb_group) conf.register_opts(influxdb_opts, influxdb_group) def list_opts(): return influxdb_group, influxdb_opts
<commit_before># (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP # Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from oslo_config import cfg influxdb_opts = [ cfg.StrOpt('database_name', help='database name where metrics are stored', default='mon'), cfg.IPOpt('ip_address', help='ip address to influxdb'), cfg.PortOpt('port', help='port to influxdb', default=8086), cfg.StrOpt('user', help='influxdb user ', default='mon_persister'), cfg.StrOpt('password', secret=True, help='influxdb password')] influxdb_group = cfg.OptGroup(name='influxdb', title='influxdb') def register_opts(conf): conf.register_group(influxdb_group) conf.register_opts(influxdb_opts, influxdb_group) def list_opts(): return influxdb_group, influxdb_opts <commit_msg>Allow hostnames to be used as ip_address Previously introduced change for monasca-persister had enforced the IPAddress as the only type one can configure influxdb.ip_address property with. Following change makes it possible to use also hostname. Using IPAdress is still possible. Change-Id: Ib0d7f19b3ac2dcb7c84923872d94f180cda58b2b<commit_after>
# (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP # Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from oslo_config import cfg influxdb_opts = [ cfg.StrOpt('database_name', help='database name where metrics are stored', default='mon'), cfg.HostAddressOpt('ip_address', help='Valid IP address or hostname ' 'to InfluxDB instance'), cfg.PortOpt('port', help='port to influxdb', default=8086), cfg.StrOpt('user', help='influxdb user ', default='mon_persister'), cfg.StrOpt('password', secret=True, help='influxdb password')] influxdb_group = cfg.OptGroup(name='influxdb', title='influxdb') def register_opts(conf): conf.register_group(influxdb_group) conf.register_opts(influxdb_opts, influxdb_group) def list_opts(): return influxdb_group, influxdb_opts
# (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP # Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from oslo_config import cfg influxdb_opts = [ cfg.StrOpt('database_name', help='database name where metrics are stored', default='mon'), cfg.IPOpt('ip_address', help='ip address to influxdb'), cfg.PortOpt('port', help='port to influxdb', default=8086), cfg.StrOpt('user', help='influxdb user ', default='mon_persister'), cfg.StrOpt('password', secret=True, help='influxdb password')] influxdb_group = cfg.OptGroup(name='influxdb', title='influxdb') def register_opts(conf): conf.register_group(influxdb_group) conf.register_opts(influxdb_opts, influxdb_group) def list_opts(): return influxdb_group, influxdb_opts Allow hostnames to be used as ip_address Previously introduced change for monasca-persister had enforced the IPAddress as the only type one can configure influxdb.ip_address property with. Following change makes it possible to use also hostname. Using IPAdress is still possible. Change-Id: Ib0d7f19b3ac2dcb7c84923872d94f180cda58b2b# (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP # Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from oslo_config import cfg influxdb_opts = [ cfg.StrOpt('database_name', help='database name where metrics are stored', default='mon'), cfg.HostAddressOpt('ip_address', help='Valid IP address or hostname ' 'to InfluxDB instance'), cfg.PortOpt('port', help='port to influxdb', default=8086), cfg.StrOpt('user', help='influxdb user ', default='mon_persister'), cfg.StrOpt('password', secret=True, help='influxdb password')] influxdb_group = cfg.OptGroup(name='influxdb', title='influxdb') def register_opts(conf): conf.register_group(influxdb_group) conf.register_opts(influxdb_opts, influxdb_group) def list_opts(): return influxdb_group, influxdb_opts
<commit_before># (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP # Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from oslo_config import cfg influxdb_opts = [ cfg.StrOpt('database_name', help='database name where metrics are stored', default='mon'), cfg.IPOpt('ip_address', help='ip address to influxdb'), cfg.PortOpt('port', help='port to influxdb', default=8086), cfg.StrOpt('user', help='influxdb user ', default='mon_persister'), cfg.StrOpt('password', secret=True, help='influxdb password')] influxdb_group = cfg.OptGroup(name='influxdb', title='influxdb') def register_opts(conf): conf.register_group(influxdb_group) conf.register_opts(influxdb_opts, influxdb_group) def list_opts(): return influxdb_group, influxdb_opts <commit_msg>Allow hostnames to be used as ip_address Previously introduced change for monasca-persister had enforced the IPAddress as the only type one can configure influxdb.ip_address property with. Following change makes it possible to use also hostname. Using IPAdress is still possible. Change-Id: Ib0d7f19b3ac2dcb7c84923872d94f180cda58b2b<commit_after># (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP # Copyright 2017 FUJITSU LIMITED # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from oslo_config import cfg influxdb_opts = [ cfg.StrOpt('database_name', help='database name where metrics are stored', default='mon'), cfg.HostAddressOpt('ip_address', help='Valid IP address or hostname ' 'to InfluxDB instance'), cfg.PortOpt('port', help='port to influxdb', default=8086), cfg.StrOpt('user', help='influxdb user ', default='mon_persister'), cfg.StrOpt('password', secret=True, help='influxdb password')] influxdb_group = cfg.OptGroup(name='influxdb', title='influxdb') def register_opts(conf): conf.register_group(influxdb_group) conf.register_opts(influxdb_opts, influxdb_group) def list_opts(): return influxdb_group, influxdb_opts
3ecca9dfa3f79a4c42a386e0bfe27cdca7e46a69
tests/21-ct-clean-up-nc.py
tests/21-ct-clean-up-nc.py
import socket, sys if len(sys.argv) != 6: print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>') localport = int(sys.argv[1]) timeout = int(sys.argv[2]) serverAddr = sys.argv[3] serverPort = int(sys.argv[4]) httpPath = sys.argv[5] if ":" not in serverAddr: clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = serverAddr else: clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM) host = "["+serverAddr+"]" clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) clientsocket.bind(('', localport)) clientsocket.settimeout(timeout) clientsocket.connect((serverAddr, serverPort)) clientsocket.send('GET '+httpPath+' HTTP/1.0\r\nHost: '+host+ '\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n') data = clientsocket.recv(4096) print(data)
import socket, sys if len(sys.argv) != 6: print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>') localport = int(sys.argv[1]) timeout = int(sys.argv[2]) serverAddr = sys.argv[3] serverPort = int(sys.argv[4]) httpPath = sys.argv[5] if ":" not in serverAddr: clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = serverAddr else: clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM) host = "["+serverAddr+"]" clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) clientsocket.bind(('', localport)) clientsocket.settimeout(timeout) clientsocket.connect((serverAddr, serverPort)) clientsocket.send('GET '+httpPath+' HTTP/1.1\r\nHost: '+host+ '\r\nConnection: close\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n') data = clientsocket.recv(4096) print(data)
Use HTTP 1.1 instead of HTTP 1.0
tests: Use HTTP 1.1 instead of HTTP 1.0 Envoy does not support HTTP 1.0, so use HTTP 1.1 instead. Signed-off-by: Jarno Rajahalme <0f1ab0ac7dffd9db21aa539af2fd4bb04abc3ad4@covalent.io>
Python
apache-2.0
cilium/cilium,cilium/cilium,tgraf/cilium,michi-covalent/cilium,cilium-team/cilium,eloycoto/cilium,tgraf/cilium,eloycoto/cilium,cilium/cilium,eloycoto/cilium,scanf/cilium,cilium-team/cilium,scanf/cilium,tklauser/cilium,scanf/cilium,cilium/cilium,scanf/cilium,cilium/cilium,tgraf/cilium,tgraf/cilium,tklauser/cilium,eloycoto/cilium,eloycoto/cilium,tgraf/cilium,tklauser/cilium,tklauser/cilium,eloycoto/cilium,eloycoto/cilium,scanf/cilium,scanf/cilium,michi-covalent/cilium,scanf/cilium,michi-covalent/cilium,michi-covalent/cilium,tklauser/cilium,tgraf/cilium,michi-covalent/cilium
import socket, sys if len(sys.argv) != 6: print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>') localport = int(sys.argv[1]) timeout = int(sys.argv[2]) serverAddr = sys.argv[3] serverPort = int(sys.argv[4]) httpPath = sys.argv[5] if ":" not in serverAddr: clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = serverAddr else: clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM) host = "["+serverAddr+"]" clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) clientsocket.bind(('', localport)) clientsocket.settimeout(timeout) clientsocket.connect((serverAddr, serverPort)) clientsocket.send('GET '+httpPath+' HTTP/1.0\r\nHost: '+host+ '\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n') data = clientsocket.recv(4096) print(data) tests: Use HTTP 1.1 instead of HTTP 1.0 Envoy does not support HTTP 1.0, so use HTTP 1.1 instead. Signed-off-by: Jarno Rajahalme <0f1ab0ac7dffd9db21aa539af2fd4bb04abc3ad4@covalent.io>
import socket, sys if len(sys.argv) != 6: print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>') localport = int(sys.argv[1]) timeout = int(sys.argv[2]) serverAddr = sys.argv[3] serverPort = int(sys.argv[4]) httpPath = sys.argv[5] if ":" not in serverAddr: clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = serverAddr else: clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM) host = "["+serverAddr+"]" clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) clientsocket.bind(('', localport)) clientsocket.settimeout(timeout) clientsocket.connect((serverAddr, serverPort)) clientsocket.send('GET '+httpPath+' HTTP/1.1\r\nHost: '+host+ '\r\nConnection: close\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n') data = clientsocket.recv(4096) print(data)
<commit_before>import socket, sys if len(sys.argv) != 6: print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>') localport = int(sys.argv[1]) timeout = int(sys.argv[2]) serverAddr = sys.argv[3] serverPort = int(sys.argv[4]) httpPath = sys.argv[5] if ":" not in serverAddr: clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = serverAddr else: clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM) host = "["+serverAddr+"]" clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) clientsocket.bind(('', localport)) clientsocket.settimeout(timeout) clientsocket.connect((serverAddr, serverPort)) clientsocket.send('GET '+httpPath+' HTTP/1.0\r\nHost: '+host+ '\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n') data = clientsocket.recv(4096) print(data) <commit_msg>tests: Use HTTP 1.1 instead of HTTP 1.0 Envoy does not support HTTP 1.0, so use HTTP 1.1 instead. Signed-off-by: Jarno Rajahalme <0f1ab0ac7dffd9db21aa539af2fd4bb04abc3ad4@covalent.io><commit_after>
import socket, sys if len(sys.argv) != 6: print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>') localport = int(sys.argv[1]) timeout = int(sys.argv[2]) serverAddr = sys.argv[3] serverPort = int(sys.argv[4]) httpPath = sys.argv[5] if ":" not in serverAddr: clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = serverAddr else: clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM) host = "["+serverAddr+"]" clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) clientsocket.bind(('', localport)) clientsocket.settimeout(timeout) clientsocket.connect((serverAddr, serverPort)) clientsocket.send('GET '+httpPath+' HTTP/1.1\r\nHost: '+host+ '\r\nConnection: close\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n') data = clientsocket.recv(4096) print(data)
import socket, sys if len(sys.argv) != 6: print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>') localport = int(sys.argv[1]) timeout = int(sys.argv[2]) serverAddr = sys.argv[3] serverPort = int(sys.argv[4]) httpPath = sys.argv[5] if ":" not in serverAddr: clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = serverAddr else: clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM) host = "["+serverAddr+"]" clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) clientsocket.bind(('', localport)) clientsocket.settimeout(timeout) clientsocket.connect((serverAddr, serverPort)) clientsocket.send('GET '+httpPath+' HTTP/1.0\r\nHost: '+host+ '\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n') data = clientsocket.recv(4096) print(data) tests: Use HTTP 1.1 instead of HTTP 1.0 Envoy does not support HTTP 1.0, so use HTTP 1.1 instead. Signed-off-by: Jarno Rajahalme <0f1ab0ac7dffd9db21aa539af2fd4bb04abc3ad4@covalent.io>import socket, sys if len(sys.argv) != 6: print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>') localport = int(sys.argv[1]) timeout = int(sys.argv[2]) serverAddr = sys.argv[3] serverPort = int(sys.argv[4]) httpPath = sys.argv[5] if ":" not in serverAddr: clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = serverAddr else: clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM) host = "["+serverAddr+"]" clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) clientsocket.bind(('', localport)) clientsocket.settimeout(timeout) clientsocket.connect((serverAddr, serverPort)) clientsocket.send('GET '+httpPath+' HTTP/1.1\r\nHost: '+host+ '\r\nConnection: close\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n') data = clientsocket.recv(4096) print(data)
<commit_before>import socket, sys if len(sys.argv) != 6: print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>') localport = int(sys.argv[1]) timeout = int(sys.argv[2]) serverAddr = sys.argv[3] serverPort = int(sys.argv[4]) httpPath = sys.argv[5] if ":" not in serverAddr: clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = serverAddr else: clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM) host = "["+serverAddr+"]" clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) clientsocket.bind(('', localport)) clientsocket.settimeout(timeout) clientsocket.connect((serverAddr, serverPort)) clientsocket.send('GET '+httpPath+' HTTP/1.0\r\nHost: '+host+ '\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n') data = clientsocket.recv(4096) print(data) <commit_msg>tests: Use HTTP 1.1 instead of HTTP 1.0 Envoy does not support HTTP 1.0, so use HTTP 1.1 instead. Signed-off-by: Jarno Rajahalme <0f1ab0ac7dffd9db21aa539af2fd4bb04abc3ad4@covalent.io><commit_after>import socket, sys if len(sys.argv) != 6: print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>') localport = int(sys.argv[1]) timeout = int(sys.argv[2]) serverAddr = sys.argv[3] serverPort = int(sys.argv[4]) httpPath = sys.argv[5] if ":" not in serverAddr: clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = serverAddr else: clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM) host = "["+serverAddr+"]" clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) clientsocket.bind(('', localport)) clientsocket.settimeout(timeout) clientsocket.connect((serverAddr, serverPort)) clientsocket.send('GET '+httpPath+' HTTP/1.1\r\nHost: '+host+ '\r\nConnection: close\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n') data = clientsocket.recv(4096) print(data)
5dd5d4e6ac93ad3867f98bd35412de11dd6b1dc2
tests/test_generalwords.py
tests/test_generalwords.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_generalwords ---------------------------------- All the tests for the generalword module. Simple module, simple tests. """ import unittest from generalwords import get_word class TestGeneralwords(unittest.TestCase): def setUp(self): pass def test_get_word(self): self.assertIsNotNone(get_word) def test_get_word_is_somewhat_random(self): sample_size = 100 words = {get_word() for i in range(sample_size)} self.assertAlmostEqual(len(words), sample_size, delta=int((sample_size * 0.1))) def tearDown(self): pass if __name__ == '__main__': unittest.main()
#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_generalwords ---------------------------------- All the tests for the generalword module. Simple module, simple tests. """ import unittest from generalwords import get_word class TestGeneralwords(unittest.TestCase): def setUp(self): pass def test_get_word(self): self.assertIsNotNone(get_word) def test_get_word_is_somewhat_random(self): sample_size = 100 words = set(get_word() for i in range(sample_size)) self.assertAlmostEqual(len(words), sample_size, delta=int((sample_size * 0.1))) def tearDown(self): pass if __name__ == '__main__': unittest.main()
Remove set-comprehensions so that tests will pass on 2.6
Remove set-comprehensions so that tests will pass on 2.6
Python
bsd-3-clause
petrilli/generalwords
#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_generalwords ---------------------------------- All the tests for the generalword module. Simple module, simple tests. """ import unittest from generalwords import get_word class TestGeneralwords(unittest.TestCase): def setUp(self): pass def test_get_word(self): self.assertIsNotNone(get_word) def test_get_word_is_somewhat_random(self): sample_size = 100 words = {get_word() for i in range(sample_size)} self.assertAlmostEqual(len(words), sample_size, delta=int((sample_size * 0.1))) def tearDown(self): pass if __name__ == '__main__': unittest.main() Remove set-comprehensions so that tests will pass on 2.6
#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_generalwords ---------------------------------- All the tests for the generalword module. Simple module, simple tests. """ import unittest from generalwords import get_word class TestGeneralwords(unittest.TestCase): def setUp(self): pass def test_get_word(self): self.assertIsNotNone(get_word) def test_get_word_is_somewhat_random(self): sample_size = 100 words = set(get_word() for i in range(sample_size)) self.assertAlmostEqual(len(words), sample_size, delta=int((sample_size * 0.1))) def tearDown(self): pass if __name__ == '__main__': unittest.main()
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_generalwords ---------------------------------- All the tests for the generalword module. Simple module, simple tests. """ import unittest from generalwords import get_word class TestGeneralwords(unittest.TestCase): def setUp(self): pass def test_get_word(self): self.assertIsNotNone(get_word) def test_get_word_is_somewhat_random(self): sample_size = 100 words = {get_word() for i in range(sample_size)} self.assertAlmostEqual(len(words), sample_size, delta=int((sample_size * 0.1))) def tearDown(self): pass if __name__ == '__main__': unittest.main() <commit_msg>Remove set-comprehensions so that tests will pass on 2.6<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_generalwords ---------------------------------- All the tests for the generalword module. Simple module, simple tests. """ import unittest from generalwords import get_word class TestGeneralwords(unittest.TestCase): def setUp(self): pass def test_get_word(self): self.assertIsNotNone(get_word) def test_get_word_is_somewhat_random(self): sample_size = 100 words = set(get_word() for i in range(sample_size)) self.assertAlmostEqual(len(words), sample_size, delta=int((sample_size * 0.1))) def tearDown(self): pass if __name__ == '__main__': unittest.main()
#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_generalwords ---------------------------------- All the tests for the generalword module. Simple module, simple tests. """ import unittest from generalwords import get_word class TestGeneralwords(unittest.TestCase): def setUp(self): pass def test_get_word(self): self.assertIsNotNone(get_word) def test_get_word_is_somewhat_random(self): sample_size = 100 words = {get_word() for i in range(sample_size)} self.assertAlmostEqual(len(words), sample_size, delta=int((sample_size * 0.1))) def tearDown(self): pass if __name__ == '__main__': unittest.main() Remove set-comprehensions so that tests will pass on 2.6#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_generalwords ---------------------------------- All the tests for the generalword module. Simple module, simple tests. """ import unittest from generalwords import get_word class TestGeneralwords(unittest.TestCase): def setUp(self): pass def test_get_word(self): self.assertIsNotNone(get_word) def test_get_word_is_somewhat_random(self): sample_size = 100 words = set(get_word() for i in range(sample_size)) self.assertAlmostEqual(len(words), sample_size, delta=int((sample_size * 0.1))) def tearDown(self): pass if __name__ == '__main__': unittest.main()
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_generalwords ---------------------------------- All the tests for the generalword module. Simple module, simple tests. """ import unittest from generalwords import get_word class TestGeneralwords(unittest.TestCase): def setUp(self): pass def test_get_word(self): self.assertIsNotNone(get_word) def test_get_word_is_somewhat_random(self): sample_size = 100 words = {get_word() for i in range(sample_size)} self.assertAlmostEqual(len(words), sample_size, delta=int((sample_size * 0.1))) def tearDown(self): pass if __name__ == '__main__': unittest.main() <commit_msg>Remove set-comprehensions so that tests will pass on 2.6<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_generalwords ---------------------------------- All the tests for the generalword module. Simple module, simple tests. """ import unittest from generalwords import get_word class TestGeneralwords(unittest.TestCase): def setUp(self): pass def test_get_word(self): self.assertIsNotNone(get_word) def test_get_word_is_somewhat_random(self): sample_size = 100 words = set(get_word() for i in range(sample_size)) self.assertAlmostEqual(len(words), sample_size, delta=int((sample_size * 0.1))) def tearDown(self): pass if __name__ == '__main__': unittest.main()
fd5634902f079c000c870192bb28fa4c35b956ec
snippets/urls.py
snippets/urls.py
from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch # Apply funfactory monkeypatches. patch() # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() def robots_txt(request): permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow' return HttpResponse('User-agent: *\n{0}: /'.format(permission), mimetype='text/plain') urlpatterns = patterns('', url(r'', include('snippets.base.urls')), url(r'^admin/', include('smuggler.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^robots\.txt$', robots_txt) ) ## In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', { 'document_root': settings.MEDIA_ROOT, }), ) + staticfiles_urlpatterns()
from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from django.views.static import serve as static_serve from funfactory.monkeypatches import patch # Apply funfactory monkeypatches. patch() # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() def robots_txt(request): permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow' return HttpResponse('User-agent: *\n{0}: /'.format(permission), mimetype='text/plain') urlpatterns = patterns('', url(r'', include('snippets.base.urls')), url(r'^admin/', include('smuggler.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^robots\.txt$', robots_txt) ) ## In DEBUG mode, serve media files through Django. if settings.DEBUG: # Use custom serve function that adds necessary headers. def serve_media(*args, **kwargs): response = static_serve(*args, **kwargs) response['Access-Control-Allow-Origin'] = '*' return response urlpatterns += patterns('', url(r'^media/(?P<path>.*)$', serve_media, { 'document_root': settings.MEDIA_ROOT, }), ) + staticfiles_urlpatterns()
Add CORS headers to dev server media.
Add CORS headers to dev server media.
Python
mpl-2.0
mozilla/snippets-service,glogiotatidis/snippets-service,glogiotatidis/snippets-service,mozmar/snippets-service,schalkneethling/snippets-service,Osmose/snippets-service,bensternthal/snippets-service,bensternthal/snippets-service,Osmose/snippets-service,akatsoulas/snippets-service,schalkneethling/snippets-service,glogiotatidis/snippets-service,mozmar/snippets-service,akatsoulas/snippets-service,mozilla/snippets-service,mozilla/snippets-service,schalkneethling/snippets-service,schalkneethling/snippets-service,akatsoulas/snippets-service,bensternthal/snippets-service,Osmose/snippets-service,mozmar/snippets-service,bensternthal/snippets-service,mozilla/snippets-service,Osmose/snippets-service,glogiotatidis/snippets-service,mozmar/snippets-service,akatsoulas/snippets-service
from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch # Apply funfactory monkeypatches. patch() # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() def robots_txt(request): permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow' return HttpResponse('User-agent: *\n{0}: /'.format(permission), mimetype='text/plain') urlpatterns = patterns('', url(r'', include('snippets.base.urls')), url(r'^admin/', include('smuggler.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^robots\.txt$', robots_txt) ) ## In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', { 'document_root': settings.MEDIA_ROOT, }), ) + staticfiles_urlpatterns() Add CORS headers to dev server media.
from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from django.views.static import serve as static_serve from funfactory.monkeypatches import patch # Apply funfactory monkeypatches. patch() # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() def robots_txt(request): permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow' return HttpResponse('User-agent: *\n{0}: /'.format(permission), mimetype='text/plain') urlpatterns = patterns('', url(r'', include('snippets.base.urls')), url(r'^admin/', include('smuggler.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^robots\.txt$', robots_txt) ) ## In DEBUG mode, serve media files through Django. if settings.DEBUG: # Use custom serve function that adds necessary headers. def serve_media(*args, **kwargs): response = static_serve(*args, **kwargs) response['Access-Control-Allow-Origin'] = '*' return response urlpatterns += patterns('', url(r'^media/(?P<path>.*)$', serve_media, { 'document_root': settings.MEDIA_ROOT, }), ) + staticfiles_urlpatterns()
<commit_before>from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch # Apply funfactory monkeypatches. patch() # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() def robots_txt(request): permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow' return HttpResponse('User-agent: *\n{0}: /'.format(permission), mimetype='text/plain') urlpatterns = patterns('', url(r'', include('snippets.base.urls')), url(r'^admin/', include('smuggler.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^robots\.txt$', robots_txt) ) ## In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', { 'document_root': settings.MEDIA_ROOT, }), ) + staticfiles_urlpatterns() <commit_msg>Add CORS headers to dev server media.<commit_after>
from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from django.views.static import serve as static_serve from funfactory.monkeypatches import patch # Apply funfactory monkeypatches. patch() # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() def robots_txt(request): permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow' return HttpResponse('User-agent: *\n{0}: /'.format(permission), mimetype='text/plain') urlpatterns = patterns('', url(r'', include('snippets.base.urls')), url(r'^admin/', include('smuggler.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^robots\.txt$', robots_txt) ) ## In DEBUG mode, serve media files through Django. if settings.DEBUG: # Use custom serve function that adds necessary headers. def serve_media(*args, **kwargs): response = static_serve(*args, **kwargs) response['Access-Control-Allow-Origin'] = '*' return response urlpatterns += patterns('', url(r'^media/(?P<path>.*)$', serve_media, { 'document_root': settings.MEDIA_ROOT, }), ) + staticfiles_urlpatterns()
from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch # Apply funfactory monkeypatches. patch() # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() def robots_txt(request): permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow' return HttpResponse('User-agent: *\n{0}: /'.format(permission), mimetype='text/plain') urlpatterns = patterns('', url(r'', include('snippets.base.urls')), url(r'^admin/', include('smuggler.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^robots\.txt$', robots_txt) ) ## In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', { 'document_root': settings.MEDIA_ROOT, }), ) + staticfiles_urlpatterns() Add CORS headers to dev server media.from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from django.views.static import serve as static_serve from funfactory.monkeypatches import patch # Apply funfactory monkeypatches. patch() # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() def robots_txt(request): permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow' return HttpResponse('User-agent: *\n{0}: /'.format(permission), mimetype='text/plain') urlpatterns = patterns('', url(r'', include('snippets.base.urls')), url(r'^admin/', include('smuggler.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^robots\.txt$', robots_txt) ) ## In DEBUG mode, serve media files through Django. if settings.DEBUG: # Use custom serve function that adds necessary headers. def serve_media(*args, **kwargs): response = static_serve(*args, **kwargs) response['Access-Control-Allow-Origin'] = '*' return response urlpatterns += patterns('', url(r'^media/(?P<path>.*)$', serve_media, { 'document_root': settings.MEDIA_ROOT, }), ) + staticfiles_urlpatterns()
<commit_before>from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from funfactory.monkeypatches import patch # Apply funfactory monkeypatches. patch() # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() def robots_txt(request): permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow' return HttpResponse('User-agent: *\n{0}: /'.format(permission), mimetype='text/plain') urlpatterns = patterns('', url(r'', include('snippets.base.urls')), url(r'^admin/', include('smuggler.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^robots\.txt$', robots_txt) ) ## In DEBUG mode, serve media files through Django. if settings.DEBUG: urlpatterns += patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', { 'document_root': settings.MEDIA_ROOT, }), ) + staticfiles_urlpatterns() <commit_msg>Add CORS headers to dev server media.<commit_after>from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.http import HttpResponse from django.views.static import serve as static_serve from funfactory.monkeypatches import patch # Apply funfactory monkeypatches. patch() # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() def robots_txt(request): permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow' return HttpResponse('User-agent: *\n{0}: /'.format(permission), mimetype='text/plain') urlpatterns = patterns('', url(r'', include('snippets.base.urls')), url(r'^admin/', include('smuggler.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^robots\.txt$', robots_txt) ) ## In DEBUG mode, serve media files through Django. if settings.DEBUG: # Use custom serve function that adds necessary headers. def serve_media(*args, **kwargs): response = static_serve(*args, **kwargs) response['Access-Control-Allow-Origin'] = '*' return response urlpatterns += patterns('', url(r'^media/(?P<path>.*)$', serve_media, { 'document_root': settings.MEDIA_ROOT, }), ) + staticfiles_urlpatterns()
8d0d4704f62b223128bab193cd5f5cda8e978c19
polling_stations/apps/pollingstations/tests/test_urls.py
polling_stations/apps/pollingstations/tests/test_urls.py
import json from django.test import TestCase from django_extensions.management.commands.show_urls import Command class UrlTests(TestCase): def is_exception(self, url): exceptions = [".txt", ".ics", ".geojson"] for exception in exceptions: if exception in url: return True return False def test_trailing_slashes(self): c = Command() data = json.loads( c.handle( **{ "unsorted": False, "language": None, "decorator": [], "format_style": "json", "urlconf": "ROOT_URLCONF", "no_color": True, } ) ) urls = [rec["url"] for rec in data] for url in urls: if self.is_exception(url): continue assert url[-1] == "/" or ">", url + " does not end with /"
import json from django.test import TestCase from django_extensions.management.commands.show_urls import Command class UrlTests(TestCase): def is_exception(self, url): exceptions = [".txt", ".ics", ".geojson"] for exception in exceptions: if exception in url: return True return False def test_trailing_slashes(self): c = Command() data = json.loads( c.handle( **{ "unsorted": False, "language": None, "decorator": [], "format_style": "json", "urlconf": "ROOT_URLCONF", "no_color": True, } ) ) urls = [rec["url"] for rec in data] urls.remove("/admin/<url>") for url in urls: if self.is_exception(url): continue assert url[-1] == "/", url + " does not end with /"
Refactor test for new view added with upgrade
Refactor test for new view added with upgrade
Python
bsd-3-clause
DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations
import json from django.test import TestCase from django_extensions.management.commands.show_urls import Command class UrlTests(TestCase): def is_exception(self, url): exceptions = [".txt", ".ics", ".geojson"] for exception in exceptions: if exception in url: return True return False def test_trailing_slashes(self): c = Command() data = json.loads( c.handle( **{ "unsorted": False, "language": None, "decorator": [], "format_style": "json", "urlconf": "ROOT_URLCONF", "no_color": True, } ) ) urls = [rec["url"] for rec in data] for url in urls: if self.is_exception(url): continue assert url[-1] == "/" or ">", url + " does not end with /" Refactor test for new view added with upgrade
import json from django.test import TestCase from django_extensions.management.commands.show_urls import Command class UrlTests(TestCase): def is_exception(self, url): exceptions = [".txt", ".ics", ".geojson"] for exception in exceptions: if exception in url: return True return False def test_trailing_slashes(self): c = Command() data = json.loads( c.handle( **{ "unsorted": False, "language": None, "decorator": [], "format_style": "json", "urlconf": "ROOT_URLCONF", "no_color": True, } ) ) urls = [rec["url"] for rec in data] urls.remove("/admin/<url>") for url in urls: if self.is_exception(url): continue assert url[-1] == "/", url + " does not end with /"
<commit_before>import json from django.test import TestCase from django_extensions.management.commands.show_urls import Command class UrlTests(TestCase): def is_exception(self, url): exceptions = [".txt", ".ics", ".geojson"] for exception in exceptions: if exception in url: return True return False def test_trailing_slashes(self): c = Command() data = json.loads( c.handle( **{ "unsorted": False, "language": None, "decorator": [], "format_style": "json", "urlconf": "ROOT_URLCONF", "no_color": True, } ) ) urls = [rec["url"] for rec in data] for url in urls: if self.is_exception(url): continue assert url[-1] == "/" or ">", url + " does not end with /" <commit_msg>Refactor test for new view added with upgrade<commit_after>
import json from django.test import TestCase from django_extensions.management.commands.show_urls import Command class UrlTests(TestCase): def is_exception(self, url): exceptions = [".txt", ".ics", ".geojson"] for exception in exceptions: if exception in url: return True return False def test_trailing_slashes(self): c = Command() data = json.loads( c.handle( **{ "unsorted": False, "language": None, "decorator": [], "format_style": "json", "urlconf": "ROOT_URLCONF", "no_color": True, } ) ) urls = [rec["url"] for rec in data] urls.remove("/admin/<url>") for url in urls: if self.is_exception(url): continue assert url[-1] == "/", url + " does not end with /"
import json from django.test import TestCase from django_extensions.management.commands.show_urls import Command class UrlTests(TestCase): def is_exception(self, url): exceptions = [".txt", ".ics", ".geojson"] for exception in exceptions: if exception in url: return True return False def test_trailing_slashes(self): c = Command() data = json.loads( c.handle( **{ "unsorted": False, "language": None, "decorator": [], "format_style": "json", "urlconf": "ROOT_URLCONF", "no_color": True, } ) ) urls = [rec["url"] for rec in data] for url in urls: if self.is_exception(url): continue assert url[-1] == "/" or ">", url + " does not end with /" Refactor test for new view added with upgradeimport json from django.test import TestCase from django_extensions.management.commands.show_urls import Command class UrlTests(TestCase): def is_exception(self, url): exceptions = [".txt", ".ics", ".geojson"] for exception in exceptions: if exception in url: return True return False def test_trailing_slashes(self): c = Command() data = json.loads( c.handle( **{ "unsorted": False, "language": None, "decorator": [], "format_style": "json", "urlconf": "ROOT_URLCONF", "no_color": True, } ) ) urls = [rec["url"] for rec in data] urls.remove("/admin/<url>") for url in urls: if self.is_exception(url): continue assert url[-1] == "/", url + " does not end with /"
<commit_before>import json from django.test import TestCase from django_extensions.management.commands.show_urls import Command class UrlTests(TestCase): def is_exception(self, url): exceptions = [".txt", ".ics", ".geojson"] for exception in exceptions: if exception in url: return True return False def test_trailing_slashes(self): c = Command() data = json.loads( c.handle( **{ "unsorted": False, "language": None, "decorator": [], "format_style": "json", "urlconf": "ROOT_URLCONF", "no_color": True, } ) ) urls = [rec["url"] for rec in data] for url in urls: if self.is_exception(url): continue assert url[-1] == "/" or ">", url + " does not end with /" <commit_msg>Refactor test for new view added with upgrade<commit_after>import json from django.test import TestCase from django_extensions.management.commands.show_urls import Command class UrlTests(TestCase): def is_exception(self, url): exceptions = [".txt", ".ics", ".geojson"] for exception in exceptions: if exception in url: return True return False def test_trailing_slashes(self): c = Command() data = json.loads( c.handle( **{ "unsorted": False, "language": None, "decorator": [], "format_style": "json", "urlconf": "ROOT_URLCONF", "no_color": True, } ) ) urls = [rec["url"] for rec in data] urls.remove("/admin/<url>") for url in urls: if self.is_exception(url): continue assert url[-1] == "/", url + " does not end with /"
fb675239ae79adcdc5f050bcf8403effb067a59b
smsgateway/utils.py
smsgateway/utils.py
import logging logger = logging.getLogger(__name__) def strspn(source, allowed): newchrs = [] for c in source: if c in allowed: newchrs.append(c) return u''.join(newchrs) def check_cell_phone_number(number): cleaned_number = strspn(number, u'+0123456789') if not u'+' in cleaned_number[:1]: cleaned_number = u'+%s' % cleaned_number return cleaned_number def truncate_sms(text, max_length=160): if len(text) <= max_length: if len(text) > 140: logger.warning("SMS possibly too long (>140 chars): %s", text) return text else: logger.error("SMS is too long (>160 chars): %s", text) return text[:max_length-3] + '...' def parse_sms(content): content = content.upper().strip() from smsgateway.backends.base import hook for keyword, subkeywords in hook.iteritems(): if content[:len(keyword)] == unicode(keyword): remainder = content[len(keyword):].strip() if '*' in subkeywords: parts = remainder.split(u' ') subkeyword = parts[0].strip() if subkeyword in subkeywords: return [keyword] + parts return keyword, remainder else: for subkeyword in subkeywords: if remainder[:len(subkeyword)] == unicode(subkeyword): subremainder = remainder[len(subkeyword):].strip() return [keyword, subkeyword] + subremainder.split() return None
import logging logger = logging.getLogger(__name__) def strspn(source, allowed): newchrs = [] for c in source: if c in allowed: newchrs.append(c) return u''.join(newchrs) def check_cell_phone_number(number): cleaned_number = strspn(number, u'+0123456789') if not u'+' in cleaned_number[:1]: cleaned_number = u'+%s' % cleaned_number return cleaned_number def truncate_sms(text, max_length=160): if len(text) <= max_length: return text else: logger.error("Trying to send an SMS that is too long: %s", text) return text[:max_length-3] + '...' def parse_sms(content): content = content.upper().strip() from smsgateway.backends.base import hook for keyword, subkeywords in hook.iteritems(): if content[:len(keyword)] == unicode(keyword): remainder = content[len(keyword):].strip() if '*' in subkeywords: parts = remainder.split(u' ') subkeyword = parts[0].strip() if subkeyword in subkeywords: return [keyword] + parts return keyword, remainder else: for subkeyword in subkeywords: if remainder[:len(subkeyword)] == unicode(subkeyword): subremainder = remainder[len(subkeyword):].strip() return [keyword, subkeyword] + subremainder.split() return None
Remove warning for long messages
Remove warning for long messages
Python
bsd-3-clause
mvpoland/django-smsgateway,peterayeni/django-smsgateway,peterayeni/django-smsgateway,peterayeni/django-smsgateway,mvpoland/django-smsgateway,peterayeni/django-smsgateway,mvpoland/django-smsgateway
import logging logger = logging.getLogger(__name__) def strspn(source, allowed): newchrs = [] for c in source: if c in allowed: newchrs.append(c) return u''.join(newchrs) def check_cell_phone_number(number): cleaned_number = strspn(number, u'+0123456789') if not u'+' in cleaned_number[:1]: cleaned_number = u'+%s' % cleaned_number return cleaned_number def truncate_sms(text, max_length=160): if len(text) <= max_length: if len(text) > 140: logger.warning("SMS possibly too long (>140 chars): %s", text) return text else: logger.error("SMS is too long (>160 chars): %s", text) return text[:max_length-3] + '...' def parse_sms(content): content = content.upper().strip() from smsgateway.backends.base import hook for keyword, subkeywords in hook.iteritems(): if content[:len(keyword)] == unicode(keyword): remainder = content[len(keyword):].strip() if '*' in subkeywords: parts = remainder.split(u' ') subkeyword = parts[0].strip() if subkeyword in subkeywords: return [keyword] + parts return keyword, remainder else: for subkeyword in subkeywords: if remainder[:len(subkeyword)] == unicode(subkeyword): subremainder = remainder[len(subkeyword):].strip() return [keyword, subkeyword] + subremainder.split() return None Remove warning for long messages
import logging logger = logging.getLogger(__name__) def strspn(source, allowed): newchrs = [] for c in source: if c in allowed: newchrs.append(c) return u''.join(newchrs) def check_cell_phone_number(number): cleaned_number = strspn(number, u'+0123456789') if not u'+' in cleaned_number[:1]: cleaned_number = u'+%s' % cleaned_number return cleaned_number def truncate_sms(text, max_length=160): if len(text) <= max_length: return text else: logger.error("Trying to send an SMS that is too long: %s", text) return text[:max_length-3] + '...' def parse_sms(content): content = content.upper().strip() from smsgateway.backends.base import hook for keyword, subkeywords in hook.iteritems(): if content[:len(keyword)] == unicode(keyword): remainder = content[len(keyword):].strip() if '*' in subkeywords: parts = remainder.split(u' ') subkeyword = parts[0].strip() if subkeyword in subkeywords: return [keyword] + parts return keyword, remainder else: for subkeyword in subkeywords: if remainder[:len(subkeyword)] == unicode(subkeyword): subremainder = remainder[len(subkeyword):].strip() return [keyword, subkeyword] + subremainder.split() return None
<commit_before>import logging logger = logging.getLogger(__name__) def strspn(source, allowed): newchrs = [] for c in source: if c in allowed: newchrs.append(c) return u''.join(newchrs) def check_cell_phone_number(number): cleaned_number = strspn(number, u'+0123456789') if not u'+' in cleaned_number[:1]: cleaned_number = u'+%s' % cleaned_number return cleaned_number def truncate_sms(text, max_length=160): if len(text) <= max_length: if len(text) > 140: logger.warning("SMS possibly too long (>140 chars): %s", text) return text else: logger.error("SMS is too long (>160 chars): %s", text) return text[:max_length-3] + '...' def parse_sms(content): content = content.upper().strip() from smsgateway.backends.base import hook for keyword, subkeywords in hook.iteritems(): if content[:len(keyword)] == unicode(keyword): remainder = content[len(keyword):].strip() if '*' in subkeywords: parts = remainder.split(u' ') subkeyword = parts[0].strip() if subkeyword in subkeywords: return [keyword] + parts return keyword, remainder else: for subkeyword in subkeywords: if remainder[:len(subkeyword)] == unicode(subkeyword): subremainder = remainder[len(subkeyword):].strip() return [keyword, subkeyword] + subremainder.split() return None <commit_msg>Remove warning for long messages<commit_after>
import logging logger = logging.getLogger(__name__) def strspn(source, allowed): newchrs = [] for c in source: if c in allowed: newchrs.append(c) return u''.join(newchrs) def check_cell_phone_number(number): cleaned_number = strspn(number, u'+0123456789') if not u'+' in cleaned_number[:1]: cleaned_number = u'+%s' % cleaned_number return cleaned_number def truncate_sms(text, max_length=160): if len(text) <= max_length: return text else: logger.error("Trying to send an SMS that is too long: %s", text) return text[:max_length-3] + '...' def parse_sms(content): content = content.upper().strip() from smsgateway.backends.base import hook for keyword, subkeywords in hook.iteritems(): if content[:len(keyword)] == unicode(keyword): remainder = content[len(keyword):].strip() if '*' in subkeywords: parts = remainder.split(u' ') subkeyword = parts[0].strip() if subkeyword in subkeywords: return [keyword] + parts return keyword, remainder else: for subkeyword in subkeywords: if remainder[:len(subkeyword)] == unicode(subkeyword): subremainder = remainder[len(subkeyword):].strip() return [keyword, subkeyword] + subremainder.split() return None
import logging logger = logging.getLogger(__name__) def strspn(source, allowed): newchrs = [] for c in source: if c in allowed: newchrs.append(c) return u''.join(newchrs) def check_cell_phone_number(number): cleaned_number = strspn(number, u'+0123456789') if not u'+' in cleaned_number[:1]: cleaned_number = u'+%s' % cleaned_number return cleaned_number def truncate_sms(text, max_length=160): if len(text) <= max_length: if len(text) > 140: logger.warning("SMS possibly too long (>140 chars): %s", text) return text else: logger.error("SMS is too long (>160 chars): %s", text) return text[:max_length-3] + '...' def parse_sms(content): content = content.upper().strip() from smsgateway.backends.base import hook for keyword, subkeywords in hook.iteritems(): if content[:len(keyword)] == unicode(keyword): remainder = content[len(keyword):].strip() if '*' in subkeywords: parts = remainder.split(u' ') subkeyword = parts[0].strip() if subkeyword in subkeywords: return [keyword] + parts return keyword, remainder else: for subkeyword in subkeywords: if remainder[:len(subkeyword)] == unicode(subkeyword): subremainder = remainder[len(subkeyword):].strip() return [keyword, subkeyword] + subremainder.split() return None Remove warning for long messagesimport logging logger = logging.getLogger(__name__) def strspn(source, allowed): newchrs = [] for c in source: if c in allowed: newchrs.append(c) return u''.join(newchrs) def check_cell_phone_number(number): cleaned_number = strspn(number, u'+0123456789') if not u'+' in cleaned_number[:1]: cleaned_number = u'+%s' % cleaned_number return cleaned_number def truncate_sms(text, max_length=160): if len(text) <= max_length: return text else: logger.error("Trying to send an SMS that is too long: %s", text) return text[:max_length-3] + '...' def parse_sms(content): content = content.upper().strip() from smsgateway.backends.base import hook for keyword, subkeywords in hook.iteritems(): if content[:len(keyword)] == unicode(keyword): remainder = content[len(keyword):].strip() if '*' in subkeywords: parts = remainder.split(u' ') subkeyword = parts[0].strip() if subkeyword in subkeywords: return [keyword] + parts return keyword, remainder else: for subkeyword in subkeywords: if remainder[:len(subkeyword)] == unicode(subkeyword): subremainder = remainder[len(subkeyword):].strip() return [keyword, subkeyword] + subremainder.split() return None
<commit_before>import logging logger = logging.getLogger(__name__) def strspn(source, allowed): newchrs = [] for c in source: if c in allowed: newchrs.append(c) return u''.join(newchrs) def check_cell_phone_number(number): cleaned_number = strspn(number, u'+0123456789') if not u'+' in cleaned_number[:1]: cleaned_number = u'+%s' % cleaned_number return cleaned_number def truncate_sms(text, max_length=160): if len(text) <= max_length: if len(text) > 140: logger.warning("SMS possibly too long (>140 chars): %s", text) return text else: logger.error("SMS is too long (>160 chars): %s", text) return text[:max_length-3] + '...' def parse_sms(content): content = content.upper().strip() from smsgateway.backends.base import hook for keyword, subkeywords in hook.iteritems(): if content[:len(keyword)] == unicode(keyword): remainder = content[len(keyword):].strip() if '*' in subkeywords: parts = remainder.split(u' ') subkeyword = parts[0].strip() if subkeyword in subkeywords: return [keyword] + parts return keyword, remainder else: for subkeyword in subkeywords: if remainder[:len(subkeyword)] == unicode(subkeyword): subremainder = remainder[len(subkeyword):].strip() return [keyword, subkeyword] + subremainder.split() return None <commit_msg>Remove warning for long messages<commit_after>import logging logger = logging.getLogger(__name__) def strspn(source, allowed): newchrs = [] for c in source: if c in allowed: newchrs.append(c) return u''.join(newchrs) def check_cell_phone_number(number): cleaned_number = strspn(number, u'+0123456789') if not u'+' in cleaned_number[:1]: cleaned_number = u'+%s' % cleaned_number return cleaned_number def truncate_sms(text, max_length=160): if len(text) <= max_length: return text else: logger.error("Trying to send an SMS that is too long: %s", text) return text[:max_length-3] + '...' def parse_sms(content): content = content.upper().strip() from smsgateway.backends.base import hook for keyword, subkeywords in hook.iteritems(): if content[:len(keyword)] == unicode(keyword): remainder = content[len(keyword):].strip() if '*' in subkeywords: parts = remainder.split(u' ') subkeyword = parts[0].strip() if subkeyword in subkeywords: return [keyword] + parts return keyword, remainder else: for subkeyword in subkeywords: if remainder[:len(subkeyword)] == unicode(subkeyword): subremainder = remainder[len(subkeyword):].strip() return [keyword, subkeyword] + subremainder.split() return None
e2831c6241e18d6f611eaaf51854b75b970867d7
examples/demo.py
examples/demo.py
#------------------------------------------------------------------------------- # # Copyright (c) 2009, Enthought, Inc. # All rights reserved. # # This software is provided without warranty under the terms of the BSD # license included in enthought/LICENSE.txt and may be redistributed only # under the conditions described in the aforementioned license. The license # is also available online at http://www.enthought.com/licenses/BSD.txt # # Thanks for using Enthought open source! # # Author: Vibha Srinivasan # Date: 02/03/2009 # #------------------------------------------------------------------------------- """ Run the Chaco demo. """ from enthought.traits.ui.extras.demo import demo # Uncomment the config_filename portion to see a tree editor based on the # examples.cfg file. demo(use_files=True, # config_filename='examples.cfg' )
#------------------------------------------------------------------------------- # # Copyright (c) 2009, Enthought, Inc. # All rights reserved. # # This software is provided without warranty under the terms of the BSD # license included in enthought/LICENSE.txt and may be redistributed only # under the conditions described in the aforementioned license. The license # is also available online at http://www.enthought.com/licenses/BSD.txt # # Thanks for using Enthought open source! # # Author: Vibha Srinivasan # Date: 02/03/2009 # #------------------------------------------------------------------------------- """ Run the Chaco demo. """ from enthought.traits.ui.extras.demo import demo # Uncomment the config_filename portion to see a tree editor based on the # examples.cfg file. demo(use_files=True, config_filename='examples.cfg' )
Use config files to construct the tree of examples.
Use config files to construct the tree of examples.
Python
bsd-3-clause
ContinuumIO/chaco,ContinuumIO/chaco,tommy-u/chaco,ContinuumIO/chaco,burnpanck/chaco,tommy-u/chaco,tommy-u/chaco,burnpanck/chaco,burnpanck/chaco,ContinuumIO/chaco
#------------------------------------------------------------------------------- # # Copyright (c) 2009, Enthought, Inc. # All rights reserved. # # This software is provided without warranty under the terms of the BSD # license included in enthought/LICENSE.txt and may be redistributed only # under the conditions described in the aforementioned license. The license # is also available online at http://www.enthought.com/licenses/BSD.txt # # Thanks for using Enthought open source! # # Author: Vibha Srinivasan # Date: 02/03/2009 # #------------------------------------------------------------------------------- """ Run the Chaco demo. """ from enthought.traits.ui.extras.demo import demo # Uncomment the config_filename portion to see a tree editor based on the # examples.cfg file. demo(use_files=True, # config_filename='examples.cfg' ) Use config files to construct the tree of examples.
#------------------------------------------------------------------------------- # # Copyright (c) 2009, Enthought, Inc. # All rights reserved. # # This software is provided without warranty under the terms of the BSD # license included in enthought/LICENSE.txt and may be redistributed only # under the conditions described in the aforementioned license. The license # is also available online at http://www.enthought.com/licenses/BSD.txt # # Thanks for using Enthought open source! # # Author: Vibha Srinivasan # Date: 02/03/2009 # #------------------------------------------------------------------------------- """ Run the Chaco demo. """ from enthought.traits.ui.extras.demo import demo # Uncomment the config_filename portion to see a tree editor based on the # examples.cfg file. demo(use_files=True, config_filename='examples.cfg' )
<commit_before>#------------------------------------------------------------------------------- # # Copyright (c) 2009, Enthought, Inc. # All rights reserved. # # This software is provided without warranty under the terms of the BSD # license included in enthought/LICENSE.txt and may be redistributed only # under the conditions described in the aforementioned license. The license # is also available online at http://www.enthought.com/licenses/BSD.txt # # Thanks for using Enthought open source! # # Author: Vibha Srinivasan # Date: 02/03/2009 # #------------------------------------------------------------------------------- """ Run the Chaco demo. """ from enthought.traits.ui.extras.demo import demo # Uncomment the config_filename portion to see a tree editor based on the # examples.cfg file. demo(use_files=True, # config_filename='examples.cfg' ) <commit_msg>Use config files to construct the tree of examples.<commit_after>
#------------------------------------------------------------------------------- # # Copyright (c) 2009, Enthought, Inc. # All rights reserved. # # This software is provided without warranty under the terms of the BSD # license included in enthought/LICENSE.txt and may be redistributed only # under the conditions described in the aforementioned license. The license # is also available online at http://www.enthought.com/licenses/BSD.txt # # Thanks for using Enthought open source! # # Author: Vibha Srinivasan # Date: 02/03/2009 # #------------------------------------------------------------------------------- """ Run the Chaco demo. """ from enthought.traits.ui.extras.demo import demo # Uncomment the config_filename portion to see a tree editor based on the # examples.cfg file. demo(use_files=True, config_filename='examples.cfg' )
#------------------------------------------------------------------------------- # # Copyright (c) 2009, Enthought, Inc. # All rights reserved. # # This software is provided without warranty under the terms of the BSD # license included in enthought/LICENSE.txt and may be redistributed only # under the conditions described in the aforementioned license. The license # is also available online at http://www.enthought.com/licenses/BSD.txt # # Thanks for using Enthought open source! # # Author: Vibha Srinivasan # Date: 02/03/2009 # #------------------------------------------------------------------------------- """ Run the Chaco demo. """ from enthought.traits.ui.extras.demo import demo # Uncomment the config_filename portion to see a tree editor based on the # examples.cfg file. demo(use_files=True, # config_filename='examples.cfg' ) Use config files to construct the tree of examples.#------------------------------------------------------------------------------- # # Copyright (c) 2009, Enthought, Inc. # All rights reserved. # # This software is provided without warranty under the terms of the BSD # license included in enthought/LICENSE.txt and may be redistributed only # under the conditions described in the aforementioned license. The license # is also available online at http://www.enthought.com/licenses/BSD.txt # # Thanks for using Enthought open source! # # Author: Vibha Srinivasan # Date: 02/03/2009 # #------------------------------------------------------------------------------- """ Run the Chaco demo. """ from enthought.traits.ui.extras.demo import demo # Uncomment the config_filename portion to see a tree editor based on the # examples.cfg file. demo(use_files=True, config_filename='examples.cfg' )
<commit_before>#------------------------------------------------------------------------------- # # Copyright (c) 2009, Enthought, Inc. # All rights reserved. # # This software is provided without warranty under the terms of the BSD # license included in enthought/LICENSE.txt and may be redistributed only # under the conditions described in the aforementioned license. The license # is also available online at http://www.enthought.com/licenses/BSD.txt # # Thanks for using Enthought open source! # # Author: Vibha Srinivasan # Date: 02/03/2009 # #------------------------------------------------------------------------------- """ Run the Chaco demo. """ from enthought.traits.ui.extras.demo import demo # Uncomment the config_filename portion to see a tree editor based on the # examples.cfg file. demo(use_files=True, # config_filename='examples.cfg' ) <commit_msg>Use config files to construct the tree of examples.<commit_after>#------------------------------------------------------------------------------- # # Copyright (c) 2009, Enthought, Inc. # All rights reserved. # # This software is provided without warranty under the terms of the BSD # license included in enthought/LICENSE.txt and may be redistributed only # under the conditions described in the aforementioned license. The license # is also available online at http://www.enthought.com/licenses/BSD.txt # # Thanks for using Enthought open source! # # Author: Vibha Srinivasan # Date: 02/03/2009 # #------------------------------------------------------------------------------- """ Run the Chaco demo. """ from enthought.traits.ui.extras.demo import demo # Uncomment the config_filename portion to see a tree editor based on the # examples.cfg file. demo(use_files=True, config_filename='examples.cfg' )
f9c9cd4505e9055a2905a87f91cdaab399352b27
dthm4kaiako/config/__init__.py
dthm4kaiako/config/__init__.py
"""Configuration for Django system.""" __version__ = "0.15.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
"""Configuration for Django system.""" __version__ = "0.16.0" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
Increment version number to 0.16.0
Increment version number to 0.16.0
Python
mit
uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers
"""Configuration for Django system.""" __version__ = "0.15.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) Increment version number to 0.16.0
"""Configuration for Django system.""" __version__ = "0.16.0" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
<commit_before>"""Configuration for Django system.""" __version__ = "0.15.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) <commit_msg>Increment version number to 0.16.0<commit_after>
"""Configuration for Django system.""" __version__ = "0.16.0" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
"""Configuration for Django system.""" __version__ = "0.15.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) Increment version number to 0.16.0"""Configuration for Django system.""" __version__ = "0.16.0" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
<commit_before>"""Configuration for Django system.""" __version__ = "0.15.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) <commit_msg>Increment version number to 0.16.0<commit_after>"""Configuration for Django system.""" __version__ = "0.16.0" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
e93ec5939d5a47e20501b4812e02a6cdac4a3ba6
rplugin/python3/deoplete/sources/LanguageClientSource.py
rplugin/python3/deoplete/sources/LanguageClientSource.py
from .base import Base import re CompleteOutputs = "g:LanguageClient_omniCompleteResults" class Source(Base): def __init__(self, vim): super().__init__(vim) self.name = "LanguageClient" self.mark = "[LC]" self.rank = 1000 self.min_pattern_length = 1 self.filetypes = vim.eval( "get(g:, 'LanguageClient_serverCommands', {})").keys() self.input_pattern += r'(\.|::|->)\w*$' self.complete_pos = re.compile(r"\w*$") def get_complete_position(self, context): m = self.complete_pos.search(context['input']) return m.start() if m else -1 def gather_candidates(self, context): if context["is_async"]: outputs = self.vim.eval(CompleteOutputs) if len(outputs) != 0: context["is_async"] = False # TODO: error handling. candidates = outputs[0].get("result", []) # log(str(candidates)) return candidates else: context["is_async"] = True self.vim.command("let {} = []".format(CompleteOutputs)) self.vim.funcs.LanguageClient_omniComplete({ "character": context["complete_position"], }) return [] # f = open("/tmp/deoplete.log", "w") # def log(message): # f.writelines([message]) # f.flush()
from .base import Base import re CompleteOutputs = "g:LanguageClient_omniCompleteResults" class Source(Base): def __init__(self, vim): super().__init__(vim) self.name = "LanguageClient" self.mark = "[LC]" self.rank = 1000 self.min_pattern_length = 0 self.filetypes = vim.eval( "get(g:, 'LanguageClient_serverCommands', {})").keys() self.input_pattern += r'(\.|::|->)\w*$' self.complete_pos = re.compile(r"\w*$") def get_complete_position(self, context): m = self.complete_pos.search(context['input']) return m.start() if m else -1 def gather_candidates(self, context): if context["is_async"]: outputs = self.vim.eval(CompleteOutputs) if len(outputs) != 0: context["is_async"] = False # TODO: error handling. candidates = outputs[0].get("result", []) # log(str(candidates)) return candidates else: context["is_async"] = True self.vim.command("let {} = []".format(CompleteOutputs)) self.vim.funcs.LanguageClient_omniComplete({ "character": context["complete_position"] + len(context["complete_str"]), }) return [] # f = open("/tmp/deoplete.log", "w") # def log(message): # f.writelines([message]) # f.flush()
Fix deoplete source compete position sent to language servers.
Fix deoplete source compete position sent to language servers.
Python
mit
autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim
from .base import Base import re CompleteOutputs = "g:LanguageClient_omniCompleteResults" class Source(Base): def __init__(self, vim): super().__init__(vim) self.name = "LanguageClient" self.mark = "[LC]" self.rank = 1000 self.min_pattern_length = 1 self.filetypes = vim.eval( "get(g:, 'LanguageClient_serverCommands', {})").keys() self.input_pattern += r'(\.|::|->)\w*$' self.complete_pos = re.compile(r"\w*$") def get_complete_position(self, context): m = self.complete_pos.search(context['input']) return m.start() if m else -1 def gather_candidates(self, context): if context["is_async"]: outputs = self.vim.eval(CompleteOutputs) if len(outputs) != 0: context["is_async"] = False # TODO: error handling. candidates = outputs[0].get("result", []) # log(str(candidates)) return candidates else: context["is_async"] = True self.vim.command("let {} = []".format(CompleteOutputs)) self.vim.funcs.LanguageClient_omniComplete({ "character": context["complete_position"], }) return [] # f = open("/tmp/deoplete.log", "w") # def log(message): # f.writelines([message]) # f.flush() Fix deoplete source compete position sent to language servers.
from .base import Base import re CompleteOutputs = "g:LanguageClient_omniCompleteResults" class Source(Base): def __init__(self, vim): super().__init__(vim) self.name = "LanguageClient" self.mark = "[LC]" self.rank = 1000 self.min_pattern_length = 0 self.filetypes = vim.eval( "get(g:, 'LanguageClient_serverCommands', {})").keys() self.input_pattern += r'(\.|::|->)\w*$' self.complete_pos = re.compile(r"\w*$") def get_complete_position(self, context): m = self.complete_pos.search(context['input']) return m.start() if m else -1 def gather_candidates(self, context): if context["is_async"]: outputs = self.vim.eval(CompleteOutputs) if len(outputs) != 0: context["is_async"] = False # TODO: error handling. candidates = outputs[0].get("result", []) # log(str(candidates)) return candidates else: context["is_async"] = True self.vim.command("let {} = []".format(CompleteOutputs)) self.vim.funcs.LanguageClient_omniComplete({ "character": context["complete_position"] + len(context["complete_str"]), }) return [] # f = open("/tmp/deoplete.log", "w") # def log(message): # f.writelines([message]) # f.flush()
<commit_before>from .base import Base import re CompleteOutputs = "g:LanguageClient_omniCompleteResults" class Source(Base): def __init__(self, vim): super().__init__(vim) self.name = "LanguageClient" self.mark = "[LC]" self.rank = 1000 self.min_pattern_length = 1 self.filetypes = vim.eval( "get(g:, 'LanguageClient_serverCommands', {})").keys() self.input_pattern += r'(\.|::|->)\w*$' self.complete_pos = re.compile(r"\w*$") def get_complete_position(self, context): m = self.complete_pos.search(context['input']) return m.start() if m else -1 def gather_candidates(self, context): if context["is_async"]: outputs = self.vim.eval(CompleteOutputs) if len(outputs) != 0: context["is_async"] = False # TODO: error handling. candidates = outputs[0].get("result", []) # log(str(candidates)) return candidates else: context["is_async"] = True self.vim.command("let {} = []".format(CompleteOutputs)) self.vim.funcs.LanguageClient_omniComplete({ "character": context["complete_position"], }) return [] # f = open("/tmp/deoplete.log", "w") # def log(message): # f.writelines([message]) # f.flush() <commit_msg>Fix deoplete source compete position sent to language servers.<commit_after>
from .base import Base import re CompleteOutputs = "g:LanguageClient_omniCompleteResults" class Source(Base): def __init__(self, vim): super().__init__(vim) self.name = "LanguageClient" self.mark = "[LC]" self.rank = 1000 self.min_pattern_length = 0 self.filetypes = vim.eval( "get(g:, 'LanguageClient_serverCommands', {})").keys() self.input_pattern += r'(\.|::|->)\w*$' self.complete_pos = re.compile(r"\w*$") def get_complete_position(self, context): m = self.complete_pos.search(context['input']) return m.start() if m else -1 def gather_candidates(self, context): if context["is_async"]: outputs = self.vim.eval(CompleteOutputs) if len(outputs) != 0: context["is_async"] = False # TODO: error handling. candidates = outputs[0].get("result", []) # log(str(candidates)) return candidates else: context["is_async"] = True self.vim.command("let {} = []".format(CompleteOutputs)) self.vim.funcs.LanguageClient_omniComplete({ "character": context["complete_position"] + len(context["complete_str"]), }) return [] # f = open("/tmp/deoplete.log", "w") # def log(message): # f.writelines([message]) # f.flush()
from .base import Base import re CompleteOutputs = "g:LanguageClient_omniCompleteResults" class Source(Base): def __init__(self, vim): super().__init__(vim) self.name = "LanguageClient" self.mark = "[LC]" self.rank = 1000 self.min_pattern_length = 1 self.filetypes = vim.eval( "get(g:, 'LanguageClient_serverCommands', {})").keys() self.input_pattern += r'(\.|::|->)\w*$' self.complete_pos = re.compile(r"\w*$") def get_complete_position(self, context): m = self.complete_pos.search(context['input']) return m.start() if m else -1 def gather_candidates(self, context): if context["is_async"]: outputs = self.vim.eval(CompleteOutputs) if len(outputs) != 0: context["is_async"] = False # TODO: error handling. candidates = outputs[0].get("result", []) # log(str(candidates)) return candidates else: context["is_async"] = True self.vim.command("let {} = []".format(CompleteOutputs)) self.vim.funcs.LanguageClient_omniComplete({ "character": context["complete_position"], }) return [] # f = open("/tmp/deoplete.log", "w") # def log(message): # f.writelines([message]) # f.flush() Fix deoplete source compete position sent to language servers.from .base import Base import re CompleteOutputs = "g:LanguageClient_omniCompleteResults" class Source(Base): def __init__(self, vim): super().__init__(vim) self.name = "LanguageClient" self.mark = "[LC]" self.rank = 1000 self.min_pattern_length = 0 self.filetypes = vim.eval( "get(g:, 'LanguageClient_serverCommands', {})").keys() self.input_pattern += r'(\.|::|->)\w*$' self.complete_pos = re.compile(r"\w*$") def get_complete_position(self, context): m = self.complete_pos.search(context['input']) return m.start() if m else -1 def gather_candidates(self, context): if context["is_async"]: outputs = self.vim.eval(CompleteOutputs) if len(outputs) != 0: context["is_async"] = False # TODO: error handling. candidates = outputs[0].get("result", []) # log(str(candidates)) return candidates else: context["is_async"] = True self.vim.command("let {} = []".format(CompleteOutputs)) self.vim.funcs.LanguageClient_omniComplete({ "character": context["complete_position"] + len(context["complete_str"]), }) return [] # f = open("/tmp/deoplete.log", "w") # def log(message): # f.writelines([message]) # f.flush()
<commit_before>from .base import Base import re CompleteOutputs = "g:LanguageClient_omniCompleteResults" class Source(Base): def __init__(self, vim): super().__init__(vim) self.name = "LanguageClient" self.mark = "[LC]" self.rank = 1000 self.min_pattern_length = 1 self.filetypes = vim.eval( "get(g:, 'LanguageClient_serverCommands', {})").keys() self.input_pattern += r'(\.|::|->)\w*$' self.complete_pos = re.compile(r"\w*$") def get_complete_position(self, context): m = self.complete_pos.search(context['input']) return m.start() if m else -1 def gather_candidates(self, context): if context["is_async"]: outputs = self.vim.eval(CompleteOutputs) if len(outputs) != 0: context["is_async"] = False # TODO: error handling. candidates = outputs[0].get("result", []) # log(str(candidates)) return candidates else: context["is_async"] = True self.vim.command("let {} = []".format(CompleteOutputs)) self.vim.funcs.LanguageClient_omniComplete({ "character": context["complete_position"], }) return [] # f = open("/tmp/deoplete.log", "w") # def log(message): # f.writelines([message]) # f.flush() <commit_msg>Fix deoplete source compete position sent to language servers.<commit_after>from .base import Base import re CompleteOutputs = "g:LanguageClient_omniCompleteResults" class Source(Base): def __init__(self, vim): super().__init__(vim) self.name = "LanguageClient" self.mark = "[LC]" self.rank = 1000 self.min_pattern_length = 0 self.filetypes = vim.eval( "get(g:, 'LanguageClient_serverCommands', {})").keys() self.input_pattern += r'(\.|::|->)\w*$' self.complete_pos = re.compile(r"\w*$") def get_complete_position(self, context): m = self.complete_pos.search(context['input']) return m.start() if m else -1 def gather_candidates(self, context): if context["is_async"]: outputs = self.vim.eval(CompleteOutputs) if len(outputs) != 0: context["is_async"] = False # TODO: error handling. candidates = outputs[0].get("result", []) # log(str(candidates)) return candidates else: context["is_async"] = True self.vim.command("let {} = []".format(CompleteOutputs)) self.vim.funcs.LanguageClient_omniComplete({ "character": context["complete_position"] + len(context["complete_str"]), }) return [] # f = open("/tmp/deoplete.log", "w") # def log(message): # f.writelines([message]) # f.flush()
e25f085025f881ccf0a0da2e620b09787819507a
sub.py
sub.py
import csv import sys import threading from time import sleep from datetime import datetime import msgpack import zmq context = zmq.Context() socket = context.socket(zmq.SUB) socket.setsockopt(zmq.SUBSCRIBE, b'') socket.bind("tcp://*:4200") terminate = threading.Event() def go(): global terminate writer = None firsttime = True with open('ani.csv', 'w', newline='') as csvfile: while not terminate.is_set(): try: msg = socket.recv(flags=zmq.NOBLOCK) except zmq.Again as e: # No message received continue orig, msgpackdata = msg.split(b' ', 1) unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8') if not isinstance(unpacked, dict): print("Message garbled: {}", unpacked) continue unpacked.update({'datetime': str(datetime.now())}) if firsttime: writer = csv.DictWriter(csvfile, fieldnames=list(unpacked.keys())) writer.writeheader() firsttime = False writer.writerow(unpacked) print(msgpackdata, unpacked) anithread = threading.Thread(target=go) anithread.start() while True: try: sleep(1) except KeyboardInterrupt: terminate.set() anithread.join() break
import csv import sys import threading from time import sleep from datetime import datetime import msgpack import zmq context = zmq.Context() socket = context.socket(zmq.SUB) socket.setsockopt(zmq.SUBSCRIBE, b'') socket.bind("tcp://*:4200") terminate = threading.Event() def go(): global terminate writer = None firsttime = True with open('ani.csv', 'w', newline='') as csvfile: while not terminate.is_set(): try: msg = socket.recv(flags=zmq.NOBLOCK) except zmq.Again as e: # No message received continue orig, msgpackdata = msg.split(b' ', 1) unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8') if not isinstance(unpacked, dict): print("Message garbled: {}", unpacked) continue if firsttime: headers = ['datetime'] + list(unpacked.keys()) writer = csv.DictWriter(csvfile, fieldnames=headers) writer.writeheader() firsttime = False unpacked.update({'datetime': str(datetime.now())}) writer.writerow(unpacked) print(msgpackdata, unpacked) anithread = threading.Thread(target=go) anithread.start() while True: try: sleep(1) except KeyboardInterrupt: terminate.set() anithread.join() break
Move date/time to the first csv column.
Move date/time to the first csv column.
Python
isc
jaj42/hsmedstream,jaj42/phystream
import csv import sys import threading from time import sleep from datetime import datetime import msgpack import zmq context = zmq.Context() socket = context.socket(zmq.SUB) socket.setsockopt(zmq.SUBSCRIBE, b'') socket.bind("tcp://*:4200") terminate = threading.Event() def go(): global terminate writer = None firsttime = True with open('ani.csv', 'w', newline='') as csvfile: while not terminate.is_set(): try: msg = socket.recv(flags=zmq.NOBLOCK) except zmq.Again as e: # No message received continue orig, msgpackdata = msg.split(b' ', 1) unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8') if not isinstance(unpacked, dict): print("Message garbled: {}", unpacked) continue unpacked.update({'datetime': str(datetime.now())}) if firsttime: writer = csv.DictWriter(csvfile, fieldnames=list(unpacked.keys())) writer.writeheader() firsttime = False writer.writerow(unpacked) print(msgpackdata, unpacked) anithread = threading.Thread(target=go) anithread.start() while True: try: sleep(1) except KeyboardInterrupt: terminate.set() anithread.join() break Move date/time to the first csv column.
import csv import sys import threading from time import sleep from datetime import datetime import msgpack import zmq context = zmq.Context() socket = context.socket(zmq.SUB) socket.setsockopt(zmq.SUBSCRIBE, b'') socket.bind("tcp://*:4200") terminate = threading.Event() def go(): global terminate writer = None firsttime = True with open('ani.csv', 'w', newline='') as csvfile: while not terminate.is_set(): try: msg = socket.recv(flags=zmq.NOBLOCK) except zmq.Again as e: # No message received continue orig, msgpackdata = msg.split(b' ', 1) unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8') if not isinstance(unpacked, dict): print("Message garbled: {}", unpacked) continue if firsttime: headers = ['datetime'] + list(unpacked.keys()) writer = csv.DictWriter(csvfile, fieldnames=headers) writer.writeheader() firsttime = False unpacked.update({'datetime': str(datetime.now())}) writer.writerow(unpacked) print(msgpackdata, unpacked) anithread = threading.Thread(target=go) anithread.start() while True: try: sleep(1) except KeyboardInterrupt: terminate.set() anithread.join() break
<commit_before>import csv import sys import threading from time import sleep from datetime import datetime import msgpack import zmq context = zmq.Context() socket = context.socket(zmq.SUB) socket.setsockopt(zmq.SUBSCRIBE, b'') socket.bind("tcp://*:4200") terminate = threading.Event() def go(): global terminate writer = None firsttime = True with open('ani.csv', 'w', newline='') as csvfile: while not terminate.is_set(): try: msg = socket.recv(flags=zmq.NOBLOCK) except zmq.Again as e: # No message received continue orig, msgpackdata = msg.split(b' ', 1) unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8') if not isinstance(unpacked, dict): print("Message garbled: {}", unpacked) continue unpacked.update({'datetime': str(datetime.now())}) if firsttime: writer = csv.DictWriter(csvfile, fieldnames=list(unpacked.keys())) writer.writeheader() firsttime = False writer.writerow(unpacked) print(msgpackdata, unpacked) anithread = threading.Thread(target=go) anithread.start() while True: try: sleep(1) except KeyboardInterrupt: terminate.set() anithread.join() break <commit_msg>Move date/time to the first csv column.<commit_after>
import csv import sys import threading from time import sleep from datetime import datetime import msgpack import zmq context = zmq.Context() socket = context.socket(zmq.SUB) socket.setsockopt(zmq.SUBSCRIBE, b'') socket.bind("tcp://*:4200") terminate = threading.Event() def go(): global terminate writer = None firsttime = True with open('ani.csv', 'w', newline='') as csvfile: while not terminate.is_set(): try: msg = socket.recv(flags=zmq.NOBLOCK) except zmq.Again as e: # No message received continue orig, msgpackdata = msg.split(b' ', 1) unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8') if not isinstance(unpacked, dict): print("Message garbled: {}", unpacked) continue if firsttime: headers = ['datetime'] + list(unpacked.keys()) writer = csv.DictWriter(csvfile, fieldnames=headers) writer.writeheader() firsttime = False unpacked.update({'datetime': str(datetime.now())}) writer.writerow(unpacked) print(msgpackdata, unpacked) anithread = threading.Thread(target=go) anithread.start() while True: try: sleep(1) except KeyboardInterrupt: terminate.set() anithread.join() break
import csv import sys import threading from time import sleep from datetime import datetime import msgpack import zmq context = zmq.Context() socket = context.socket(zmq.SUB) socket.setsockopt(zmq.SUBSCRIBE, b'') socket.bind("tcp://*:4200") terminate = threading.Event() def go(): global terminate writer = None firsttime = True with open('ani.csv', 'w', newline='') as csvfile: while not terminate.is_set(): try: msg = socket.recv(flags=zmq.NOBLOCK) except zmq.Again as e: # No message received continue orig, msgpackdata = msg.split(b' ', 1) unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8') if not isinstance(unpacked, dict): print("Message garbled: {}", unpacked) continue unpacked.update({'datetime': str(datetime.now())}) if firsttime: writer = csv.DictWriter(csvfile, fieldnames=list(unpacked.keys())) writer.writeheader() firsttime = False writer.writerow(unpacked) print(msgpackdata, unpacked) anithread = threading.Thread(target=go) anithread.start() while True: try: sleep(1) except KeyboardInterrupt: terminate.set() anithread.join() break Move date/time to the first csv column.import csv import sys import threading from time import sleep from datetime import datetime import msgpack import zmq context = zmq.Context() socket = context.socket(zmq.SUB) socket.setsockopt(zmq.SUBSCRIBE, b'') socket.bind("tcp://*:4200") terminate = threading.Event() def go(): global terminate writer = None firsttime = True with open('ani.csv', 'w', newline='') as csvfile: while not terminate.is_set(): try: msg = socket.recv(flags=zmq.NOBLOCK) except zmq.Again as e: # No message received continue orig, msgpackdata = msg.split(b' ', 1) unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8') if not isinstance(unpacked, dict): print("Message garbled: {}", unpacked) continue if firsttime: headers = ['datetime'] + list(unpacked.keys()) writer = csv.DictWriter(csvfile, fieldnames=headers) writer.writeheader() firsttime = False unpacked.update({'datetime': str(datetime.now())}) writer.writerow(unpacked) print(msgpackdata, unpacked) anithread = threading.Thread(target=go) anithread.start() while True: try: sleep(1) except KeyboardInterrupt: terminate.set() anithread.join() break
<commit_before>import csv import sys import threading from time import sleep from datetime import datetime import msgpack import zmq context = zmq.Context() socket = context.socket(zmq.SUB) socket.setsockopt(zmq.SUBSCRIBE, b'') socket.bind("tcp://*:4200") terminate = threading.Event() def go(): global terminate writer = None firsttime = True with open('ani.csv', 'w', newline='') as csvfile: while not terminate.is_set(): try: msg = socket.recv(flags=zmq.NOBLOCK) except zmq.Again as e: # No message received continue orig, msgpackdata = msg.split(b' ', 1) unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8') if not isinstance(unpacked, dict): print("Message garbled: {}", unpacked) continue unpacked.update({'datetime': str(datetime.now())}) if firsttime: writer = csv.DictWriter(csvfile, fieldnames=list(unpacked.keys())) writer.writeheader() firsttime = False writer.writerow(unpacked) print(msgpackdata, unpacked) anithread = threading.Thread(target=go) anithread.start() while True: try: sleep(1) except KeyboardInterrupt: terminate.set() anithread.join() break <commit_msg>Move date/time to the first csv column.<commit_after>import csv import sys import threading from time import sleep from datetime import datetime import msgpack import zmq context = zmq.Context() socket = context.socket(zmq.SUB) socket.setsockopt(zmq.SUBSCRIBE, b'') socket.bind("tcp://*:4200") terminate = threading.Event() def go(): global terminate writer = None firsttime = True with open('ani.csv', 'w', newline='') as csvfile: while not terminate.is_set(): try: msg = socket.recv(flags=zmq.NOBLOCK) except zmq.Again as e: # No message received continue orig, msgpackdata = msg.split(b' ', 1) unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8') if not isinstance(unpacked, dict): print("Message garbled: {}", unpacked) continue if firsttime: headers = ['datetime'] + list(unpacked.keys()) writer = csv.DictWriter(csvfile, fieldnames=headers) writer.writeheader() firsttime = False unpacked.update({'datetime': str(datetime.now())}) writer.writerow(unpacked) print(msgpackdata, unpacked) anithread = threading.Thread(target=go) anithread.start() while True: try: sleep(1) except KeyboardInterrupt: terminate.set() anithread.join() break
866026a5d2f89a8ac76a726720e4fbe812c94eb4
ds/providers/shell.py
ds/providers/shell.py
from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format({ 'environment': task.environment, 'sha': task.sha, 'ref': task.ref, 'task': task.name, }) return workspace.run(command)
from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format( environment=task.environment, sha=task.sha, ref=task.ref, task=task.name, ) return workspace.run(command)
Fix arg passing to command
Fix arg passing to command
Python
apache-2.0
rshk/freight,klynton/freight,jkimbo/freight,getsentry/freight,getsentry/freight,klynton/freight,getsentry/freight,klynton/freight,rshk/freight,klynton/freight,jkimbo/freight,getsentry/freight,rshk/freight,jkimbo/freight,jkimbo/freight,rshk/freight,getsentry/freight
from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format({ 'environment': task.environment, 'sha': task.sha, 'ref': task.ref, 'task': task.name, }) return workspace.run(command) Fix arg passing to command
from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format( environment=task.environment, sha=task.sha, ref=task.ref, task=task.name, ) return workspace.run(command)
<commit_before>from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format({ 'environment': task.environment, 'sha': task.sha, 'ref': task.ref, 'task': task.name, }) return workspace.run(command) <commit_msg>Fix arg passing to command<commit_after>
from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format( environment=task.environment, sha=task.sha, ref=task.ref, task=task.name, ) return workspace.run(command)
from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format({ 'environment': task.environment, 'sha': task.sha, 'ref': task.ref, 'task': task.name, }) return workspace.run(command) Fix arg passing to commandfrom __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format( environment=task.environment, sha=task.sha, ref=task.ref, task=task.name, ) return workspace.run(command)
<commit_before>from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format({ 'environment': task.environment, 'sha': task.sha, 'ref': task.ref, 'task': task.name, }) return workspace.run(command) <commit_msg>Fix arg passing to command<commit_after>from __future__ import absolute_import, unicode_literals __all__ = ['ShellProvider'] from .base import Provider class ShellProvider(Provider): def get_options(self): return { 'command': {'required': True}, } def execute(self, workspace, task): command = task.provider_config['command'].format( environment=task.environment, sha=task.sha, ref=task.ref, task=task.name, ) return workspace.run(command)
b05d2666c834a9c4d151d0340612010851bd4610
eniric/Qcalculator.py
eniric/Qcalculator.py
""" Created on Mon Dec 29 00:14:56 2014 @author: pfigueira Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric. """ # from eniric.IOmodule import read_2col import numpy as np import pandas as pd c = 299792458 # m/s def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"): """ function that claculates the RV precision achievable on a spectrum """ data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64) wav, flux = data["wavelength"].values, data["flux"].values return [c / SqrtSumWis(wavelength, flux)] def SqrtSumWis(wavelength, flux): """ Calculation of the SquareRoot of the sum of the Wis for a spectrum """ delta_F = (np.array(flux[1:]) - np.array(flux[:-1])) delta_l = np.array(wavelength[:-1]) derivF_over_lambda = delta_F/delta_l return np.sqrt(np.sum(np.array(wavelength[:-1])**2.0 * derivF_over_lambda**2.0 / np.array(flux[:-1])))
""" Created on Mon Dec 29 00:14:56 2014 @author: pfigueira Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric. """ # from eniric.IOmodule import read_2col import numpy as np import pandas as pd c = 299792458 # m/s def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"): """ function that claculates the RV precision achievable on a spectrum """ data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64) wav, flux = data["wavelength"].values, data["flux"].values return [c / SqrtSumWis(wavelength, flux)] def SqrtSumWis(wavelength, flux): """ Calculation of the SquareRoot of the sum of the Wis for a spectrum """ delta_F = np.diff(flux) delta_l = wavelength[:-1] derivF_over_lambda = delta_F/delta_l return np.sqrt(np.sum(wavelength[:-1]**2.0 * derivF_over_lambda**2.0 / flux[:-1]))
Update RVprec_calculation use numpy.diff() and remove unnecessary array calls.
Update RVprec_calculation use numpy.diff() and remove unnecessary array calls. Former-commit-id: 646ff0cea061feb87c08b819a47d8e9f3dd55b55
Python
mit
jason-neal/eniric,jason-neal/eniric
""" Created on Mon Dec 29 00:14:56 2014 @author: pfigueira Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric. """ # from eniric.IOmodule import read_2col import numpy as np import pandas as pd c = 299792458 # m/s def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"): """ function that claculates the RV precision achievable on a spectrum """ data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64) wav, flux = data["wavelength"].values, data["flux"].values return [c / SqrtSumWis(wavelength, flux)] def SqrtSumWis(wavelength, flux): """ Calculation of the SquareRoot of the sum of the Wis for a spectrum """ delta_F = (np.array(flux[1:]) - np.array(flux[:-1])) delta_l = np.array(wavelength[:-1]) derivF_over_lambda = delta_F/delta_l return np.sqrt(np.sum(np.array(wavelength[:-1])**2.0 * derivF_over_lambda**2.0 / np.array(flux[:-1]))) Update RVprec_calculation use numpy.diff() and remove unnecessary array calls. Former-commit-id: 646ff0cea061feb87c08b819a47d8e9f3dd55b55
""" Created on Mon Dec 29 00:14:56 2014 @author: pfigueira Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric. """ # from eniric.IOmodule import read_2col import numpy as np import pandas as pd c = 299792458 # m/s def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"): """ function that claculates the RV precision achievable on a spectrum """ data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64) wav, flux = data["wavelength"].values, data["flux"].values return [c / SqrtSumWis(wavelength, flux)] def SqrtSumWis(wavelength, flux): """ Calculation of the SquareRoot of the sum of the Wis for a spectrum """ delta_F = np.diff(flux) delta_l = wavelength[:-1] derivF_over_lambda = delta_F/delta_l return np.sqrt(np.sum(wavelength[:-1]**2.0 * derivF_over_lambda**2.0 / flux[:-1]))
<commit_before>""" Created on Mon Dec 29 00:14:56 2014 @author: pfigueira Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric. """ # from eniric.IOmodule import read_2col import numpy as np import pandas as pd c = 299792458 # m/s def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"): """ function that claculates the RV precision achievable on a spectrum """ data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64) wav, flux = data["wavelength"].values, data["flux"].values return [c / SqrtSumWis(wavelength, flux)] def SqrtSumWis(wavelength, flux): """ Calculation of the SquareRoot of the sum of the Wis for a spectrum """ delta_F = (np.array(flux[1:]) - np.array(flux[:-1])) delta_l = np.array(wavelength[:-1]) derivF_over_lambda = delta_F/delta_l return np.sqrt(np.sum(np.array(wavelength[:-1])**2.0 * derivF_over_lambda**2.0 / np.array(flux[:-1]))) <commit_msg>Update RVprec_calculation use numpy.diff() and remove unnecessary array calls. Former-commit-id: 646ff0cea061feb87c08b819a47d8e9f3dd55b55<commit_after>
""" Created on Mon Dec 29 00:14:56 2014 @author: pfigueira Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric. """ # from eniric.IOmodule import read_2col import numpy as np import pandas as pd c = 299792458 # m/s def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"): """ function that claculates the RV precision achievable on a spectrum """ data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64) wav, flux = data["wavelength"].values, data["flux"].values return [c / SqrtSumWis(wavelength, flux)] def SqrtSumWis(wavelength, flux): """ Calculation of the SquareRoot of the sum of the Wis for a spectrum """ delta_F = np.diff(flux) delta_l = wavelength[:-1] derivF_over_lambda = delta_F/delta_l return np.sqrt(np.sum(wavelength[:-1]**2.0 * derivF_over_lambda**2.0 / flux[:-1]))
""" Created on Mon Dec 29 00:14:56 2014 @author: pfigueira Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric. """ # from eniric.IOmodule import read_2col import numpy as np import pandas as pd c = 299792458 # m/s def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"): """ function that claculates the RV precision achievable on a spectrum """ data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64) wav, flux = data["wavelength"].values, data["flux"].values return [c / SqrtSumWis(wavelength, flux)] def SqrtSumWis(wavelength, flux): """ Calculation of the SquareRoot of the sum of the Wis for a spectrum """ delta_F = (np.array(flux[1:]) - np.array(flux[:-1])) delta_l = np.array(wavelength[:-1]) derivF_over_lambda = delta_F/delta_l return np.sqrt(np.sum(np.array(wavelength[:-1])**2.0 * derivF_over_lambda**2.0 / np.array(flux[:-1]))) Update RVprec_calculation use numpy.diff() and remove unnecessary array calls. Former-commit-id: 646ff0cea061feb87c08b819a47d8e9f3dd55b55""" Created on Mon Dec 29 00:14:56 2014 @author: pfigueira Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric. """ # from eniric.IOmodule import read_2col import numpy as np import pandas as pd c = 299792458 # m/s def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"): """ function that claculates the RV precision achievable on a spectrum """ data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64) wav, flux = data["wavelength"].values, data["flux"].values return [c / SqrtSumWis(wavelength, flux)] def SqrtSumWis(wavelength, flux): """ Calculation of the SquareRoot of the sum of the Wis for a spectrum """ delta_F = np.diff(flux) delta_l = wavelength[:-1] derivF_over_lambda = delta_F/delta_l return np.sqrt(np.sum(wavelength[:-1]**2.0 * derivF_over_lambda**2.0 / flux[:-1]))
<commit_before>""" Created on Mon Dec 29 00:14:56 2014 @author: pfigueira Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric. """ # from eniric.IOmodule import read_2col import numpy as np import pandas as pd c = 299792458 # m/s def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"): """ function that claculates the RV precision achievable on a spectrum """ data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64) wav, flux = data["wavelength"].values, data["flux"].values return [c / SqrtSumWis(wavelength, flux)] def SqrtSumWis(wavelength, flux): """ Calculation of the SquareRoot of the sum of the Wis for a spectrum """ delta_F = (np.array(flux[1:]) - np.array(flux[:-1])) delta_l = np.array(wavelength[:-1]) derivF_over_lambda = delta_F/delta_l return np.sqrt(np.sum(np.array(wavelength[:-1])**2.0 * derivF_over_lambda**2.0 / np.array(flux[:-1]))) <commit_msg>Update RVprec_calculation use numpy.diff() and remove unnecessary array calls. Former-commit-id: 646ff0cea061feb87c08b819a47d8e9f3dd55b55<commit_after>""" Created on Mon Dec 29 00:14:56 2014 @author: pfigueira Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric. """ # from eniric.IOmodule import read_2col import numpy as np import pandas as pd c = 299792458 # m/s def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"): """ function that claculates the RV precision achievable on a spectrum """ data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64) wav, flux = data["wavelength"].values, data["flux"].values return [c / SqrtSumWis(wavelength, flux)] def SqrtSumWis(wavelength, flux): """ Calculation of the SquareRoot of the sum of the Wis for a spectrum """ delta_F = np.diff(flux) delta_l = wavelength[:-1] derivF_over_lambda = delta_F/delta_l return np.sqrt(np.sum(wavelength[:-1]**2.0 * derivF_over_lambda**2.0 / flux[:-1]))
7d0f3ba1aa82c2ea5a4a2eca2bbe842b63a82c72
wafer/talks/serializers.py
wafer/talks/serializers.py
from rest_framework import serializers from reversion import revisions from wafer.talks.models import Talk class TalkSerializer(serializers.ModelSerializer): class Meta: model = Talk exclude = ('_abstract_rendered', ) @revisions.create_revision() def create(self, validated_data): revisions.set_comment("Created via REST api") return super(TalkSerializer, self).create(validated_data) @revisions.create_revision() def update(self, talk, validated_data): revisions.set_comment("Changed via REST api") talk.abstract = validated_data['abstract'] talk.title = validated_data['title'] talk.status = validated_data['status'] talk.talk_type = validated_data['talk_type'] talk.notes = validated_data['notes'] talk.private_notes = validated_data['private_notes'] talk.save() return talk
from rest_framework import serializers from reversion import revisions from wafer.talks.models import Talk class TalkSerializer(serializers.ModelSerializer): class Meta: model = Talk # private_notes should possibly be accessible to # talk reviewers by the API, but certainly # not to the other users. # Similar considerations apply to notes, which should # not be generally accessible exclude = ('_abstract_rendered', 'private_notes', 'notes') @revisions.create_revision() def create(self, validated_data): revisions.set_comment("Created via REST api") return super(TalkSerializer, self).create(validated_data) @revisions.create_revision() def update(self, talk, validated_data): revisions.set_comment("Changed via REST api") talk.abstract = validated_data['abstract'] talk.title = validated_data['title'] talk.status = validated_data['status'] talk.talk_type = validated_data['talk_type'] talk.notes = validated_data['notes'] talk.private_notes = validated_data['private_notes'] talk.save() return talk
Exclude notes and private_notes from api for now
Exclude notes and private_notes from api for now
Python
isc
CTPUG/wafer,CTPUG/wafer,CTPUG/wafer,CTPUG/wafer
from rest_framework import serializers from reversion import revisions from wafer.talks.models import Talk class TalkSerializer(serializers.ModelSerializer): class Meta: model = Talk exclude = ('_abstract_rendered', ) @revisions.create_revision() def create(self, validated_data): revisions.set_comment("Created via REST api") return super(TalkSerializer, self).create(validated_data) @revisions.create_revision() def update(self, talk, validated_data): revisions.set_comment("Changed via REST api") talk.abstract = validated_data['abstract'] talk.title = validated_data['title'] talk.status = validated_data['status'] talk.talk_type = validated_data['talk_type'] talk.notes = validated_data['notes'] talk.private_notes = validated_data['private_notes'] talk.save() return talk Exclude notes and private_notes from api for now
from rest_framework import serializers from reversion import revisions from wafer.talks.models import Talk class TalkSerializer(serializers.ModelSerializer): class Meta: model = Talk # private_notes should possibly be accessible to # talk reviewers by the API, but certainly # not to the other users. # Similar considerations apply to notes, which should # not be generally accessible exclude = ('_abstract_rendered', 'private_notes', 'notes') @revisions.create_revision() def create(self, validated_data): revisions.set_comment("Created via REST api") return super(TalkSerializer, self).create(validated_data) @revisions.create_revision() def update(self, talk, validated_data): revisions.set_comment("Changed via REST api") talk.abstract = validated_data['abstract'] talk.title = validated_data['title'] talk.status = validated_data['status'] talk.talk_type = validated_data['talk_type'] talk.notes = validated_data['notes'] talk.private_notes = validated_data['private_notes'] talk.save() return talk
<commit_before>from rest_framework import serializers from reversion import revisions from wafer.talks.models import Talk class TalkSerializer(serializers.ModelSerializer): class Meta: model = Talk exclude = ('_abstract_rendered', ) @revisions.create_revision() def create(self, validated_data): revisions.set_comment("Created via REST api") return super(TalkSerializer, self).create(validated_data) @revisions.create_revision() def update(self, talk, validated_data): revisions.set_comment("Changed via REST api") talk.abstract = validated_data['abstract'] talk.title = validated_data['title'] talk.status = validated_data['status'] talk.talk_type = validated_data['talk_type'] talk.notes = validated_data['notes'] talk.private_notes = validated_data['private_notes'] talk.save() return talk <commit_msg>Exclude notes and private_notes from api for now<commit_after>
from rest_framework import serializers from reversion import revisions from wafer.talks.models import Talk class TalkSerializer(serializers.ModelSerializer): class Meta: model = Talk # private_notes should possibly be accessible to # talk reviewers by the API, but certainly # not to the other users. # Similar considerations apply to notes, which should # not be generally accessible exclude = ('_abstract_rendered', 'private_notes', 'notes') @revisions.create_revision() def create(self, validated_data): revisions.set_comment("Created via REST api") return super(TalkSerializer, self).create(validated_data) @revisions.create_revision() def update(self, talk, validated_data): revisions.set_comment("Changed via REST api") talk.abstract = validated_data['abstract'] talk.title = validated_data['title'] talk.status = validated_data['status'] talk.talk_type = validated_data['talk_type'] talk.notes = validated_data['notes'] talk.private_notes = validated_data['private_notes'] talk.save() return talk
from rest_framework import serializers from reversion import revisions from wafer.talks.models import Talk class TalkSerializer(serializers.ModelSerializer): class Meta: model = Talk exclude = ('_abstract_rendered', ) @revisions.create_revision() def create(self, validated_data): revisions.set_comment("Created via REST api") return super(TalkSerializer, self).create(validated_data) @revisions.create_revision() def update(self, talk, validated_data): revisions.set_comment("Changed via REST api") talk.abstract = validated_data['abstract'] talk.title = validated_data['title'] talk.status = validated_data['status'] talk.talk_type = validated_data['talk_type'] talk.notes = validated_data['notes'] talk.private_notes = validated_data['private_notes'] talk.save() return talk Exclude notes and private_notes from api for nowfrom rest_framework import serializers from reversion import revisions from wafer.talks.models import Talk class TalkSerializer(serializers.ModelSerializer): class Meta: model = Talk # private_notes should possibly be accessible to # talk reviewers by the API, but certainly # not to the other users. # Similar considerations apply to notes, which should # not be generally accessible exclude = ('_abstract_rendered', 'private_notes', 'notes') @revisions.create_revision() def create(self, validated_data): revisions.set_comment("Created via REST api") return super(TalkSerializer, self).create(validated_data) @revisions.create_revision() def update(self, talk, validated_data): revisions.set_comment("Changed via REST api") talk.abstract = validated_data['abstract'] talk.title = validated_data['title'] talk.status = validated_data['status'] talk.talk_type = validated_data['talk_type'] talk.notes = validated_data['notes'] talk.private_notes = validated_data['private_notes'] talk.save() return talk
<commit_before>from rest_framework import serializers from reversion import revisions from wafer.talks.models import Talk class TalkSerializer(serializers.ModelSerializer): class Meta: model = Talk exclude = ('_abstract_rendered', ) @revisions.create_revision() def create(self, validated_data): revisions.set_comment("Created via REST api") return super(TalkSerializer, self).create(validated_data) @revisions.create_revision() def update(self, talk, validated_data): revisions.set_comment("Changed via REST api") talk.abstract = validated_data['abstract'] talk.title = validated_data['title'] talk.status = validated_data['status'] talk.talk_type = validated_data['talk_type'] talk.notes = validated_data['notes'] talk.private_notes = validated_data['private_notes'] talk.save() return talk <commit_msg>Exclude notes and private_notes from api for now<commit_after>from rest_framework import serializers from reversion import revisions from wafer.talks.models import Talk class TalkSerializer(serializers.ModelSerializer): class Meta: model = Talk # private_notes should possibly be accessible to # talk reviewers by the API, but certainly # not to the other users. # Similar considerations apply to notes, which should # not be generally accessible exclude = ('_abstract_rendered', 'private_notes', 'notes') @revisions.create_revision() def create(self, validated_data): revisions.set_comment("Created via REST api") return super(TalkSerializer, self).create(validated_data) @revisions.create_revision() def update(self, talk, validated_data): revisions.set_comment("Changed via REST api") talk.abstract = validated_data['abstract'] talk.title = validated_data['title'] talk.status = validated_data['status'] talk.talk_type = validated_data['talk_type'] talk.notes = validated_data['notes'] talk.private_notes = validated_data['private_notes'] talk.save() return talk
0600f3a4f9e13ac0a2a2b4d542db949f90e4185c
challenge_2/python/ning/challenge_2.py
challenge_2/python/ning/challenge_2.py
def find_unique(sequence): item_counter = dict() uniques = list() for item in sequence: if item not in item_counter: item_counter[item] = 1 else: item_counter[item] += 1 for item, item_count in item_counter.items(): if item_count == 1: uniques.append(item) return uniques test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7] print(find_unique(test_sequence_list))
import collections def find_unique(sequence): counter_dict = collections.defaultdict(int) uniques = [] for item in sequence: counter_dict[item] += 1 for item, count in counter_dict.items(): if count == 1: uniques.append(item) return uniques test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7] print(find_unique(test_sequence_list))
Change logic to use defaultdict, lists initiate now with
Change logic to use defaultdict, lists initiate now with []
Python
mit
erocs/2017Challenges,popcornanachronism/2017Challenges,erocs/2017Challenges,erocs/2017Challenges,popcornanachronism/2017Challenges,DakRomo/2017Challenges,popcornanachronism/2017Challenges,mindm/2017Challenges,erocs/2017Challenges,DakRomo/2017Challenges,popcornanachronism/2017Challenges,mindm/2017Challenges,DakRomo/2017Challenges,popcornanachronism/2017Challenges,popcornanachronism/2017Challenges,erocs/2017Challenges,popcornanachronism/2017Challenges,erocs/2017Challenges,erocs/2017Challenges,mindm/2017Challenges,DakRomo/2017Challenges,erocs/2017Challenges,mindm/2017Challenges,DakRomo/2017Challenges,DakRomo/2017Challenges,DakRomo/2017Challenges,DakRomo/2017Challenges,mindm/2017Challenges,DakRomo/2017Challenges,popcornanachronism/2017Challenges,DakRomo/2017Challenges,mindm/2017Challenges,popcornanachronism/2017Challenges,mindm/2017Challenges,mindm/2017Challenges,erocs/2017Challenges,erocs/2017Challenges,mindm/2017Challenges,popcornanachronism/2017Challenges,popcornanachronism/2017Challenges,mindm/2017Challenges,mindm/2017Challenges,mindm/2017Challenges,mindm/2017Challenges,erocs/2017Challenges,popcornanachronism/2017Challenges,DakRomo/2017Challenges,DakRomo/2017Challenges,DakRomo/2017Challenges,popcornanachronism/2017Challenges,erocs/2017Challenges,DakRomo/2017Challenges,mindm/2017Challenges,erocs/2017Challenges,erocs/2017Challenges,popcornanachronism/2017Challenges
def find_unique(sequence): item_counter = dict() uniques = list() for item in sequence: if item not in item_counter: item_counter[item] = 1 else: item_counter[item] += 1 for item, item_count in item_counter.items(): if item_count == 1: uniques.append(item) return uniques test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7] print(find_unique(test_sequence_list)) Change logic to use defaultdict, lists initiate now with []
import collections def find_unique(sequence): counter_dict = collections.defaultdict(int) uniques = [] for item in sequence: counter_dict[item] += 1 for item, count in counter_dict.items(): if count == 1: uniques.append(item) return uniques test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7] print(find_unique(test_sequence_list))
<commit_before>def find_unique(sequence): item_counter = dict() uniques = list() for item in sequence: if item not in item_counter: item_counter[item] = 1 else: item_counter[item] += 1 for item, item_count in item_counter.items(): if item_count == 1: uniques.append(item) return uniques test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7] print(find_unique(test_sequence_list)) <commit_msg>Change logic to use defaultdict, lists initiate now with []<commit_after>
import collections def find_unique(sequence): counter_dict = collections.defaultdict(int) uniques = [] for item in sequence: counter_dict[item] += 1 for item, count in counter_dict.items(): if count == 1: uniques.append(item) return uniques test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7] print(find_unique(test_sequence_list))
def find_unique(sequence): item_counter = dict() uniques = list() for item in sequence: if item not in item_counter: item_counter[item] = 1 else: item_counter[item] += 1 for item, item_count in item_counter.items(): if item_count == 1: uniques.append(item) return uniques test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7] print(find_unique(test_sequence_list)) Change logic to use defaultdict, lists initiate now with []import collections def find_unique(sequence): counter_dict = collections.defaultdict(int) uniques = [] for item in sequence: counter_dict[item] += 1 for item, count in counter_dict.items(): if count == 1: uniques.append(item) return uniques test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7] print(find_unique(test_sequence_list))
<commit_before>def find_unique(sequence): item_counter = dict() uniques = list() for item in sequence: if item not in item_counter: item_counter[item] = 1 else: item_counter[item] += 1 for item, item_count in item_counter.items(): if item_count == 1: uniques.append(item) return uniques test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7] print(find_unique(test_sequence_list)) <commit_msg>Change logic to use defaultdict, lists initiate now with []<commit_after>import collections def find_unique(sequence): counter_dict = collections.defaultdict(int) uniques = [] for item in sequence: counter_dict[item] += 1 for item, count in counter_dict.items(): if count == 1: uniques.append(item) return uniques test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7] print(find_unique(test_sequence_list))
415c9bd14016453d474901fe16f2da8246b44e1b
dojango/__init__.py
dojango/__init__.py
VERSION = (0, 5, 3, 'final', 0) def get_version(): version = '%s.%s' % (VERSION[0], VERSION[1]) if VERSION[2]: version = '%s.%s' % (version, VERSION[2]) if VERSION[3:] == ('alpha', 0): version = '%s pre-alpha' % version else: if VERSION[3] != 'final': version = '%s %s %s' % (version, VERSION[3], VERSION[4]) #from django.utils.version import get_svn_revision #svn_rev = get_svn_revision() #if svn_rev != u'SVN-unknown': # version = "%s %s" % (version, svn_rev) return version
VERSION = (0, 5, 4, 'alpha', 0) def get_version(): version = '%s.%s' % (VERSION[0], VERSION[1]) if VERSION[2]: version = '%s.%s' % (version, VERSION[2]) if VERSION[3:] == ('alpha', 0): version = '%s pre-alpha' % version else: if VERSION[3] != 'final': version = '%s %s %s' % (version, VERSION[3], VERSION[4]) #from django.utils.version import get_svn_revision #svn_rev = get_svn_revision() #if svn_rev != u'SVN-unknown': # version = "%s %s" % (version, svn_rev) return version
Update to 0.5.4 pre-alpha (preparing next build version)
Update to 0.5.4 pre-alpha (preparing next build version)
Python
bsd-3-clause
ofirr/dojango,ofirr/dojango,william-gr/dojango,ofirr/dojango,william-gr/dojango,ricard33/dojango,ricard33/dojango,klipstein/dojango,ricard33/dojango,william-gr/dojango,klipstein/dojango
VERSION = (0, 5, 3, 'final', 0) def get_version(): version = '%s.%s' % (VERSION[0], VERSION[1]) if VERSION[2]: version = '%s.%s' % (version, VERSION[2]) if VERSION[3:] == ('alpha', 0): version = '%s pre-alpha' % version else: if VERSION[3] != 'final': version = '%s %s %s' % (version, VERSION[3], VERSION[4]) #from django.utils.version import get_svn_revision #svn_rev = get_svn_revision() #if svn_rev != u'SVN-unknown': # version = "%s %s" % (version, svn_rev) return version Update to 0.5.4 pre-alpha (preparing next build version)
VERSION = (0, 5, 4, 'alpha', 0) def get_version(): version = '%s.%s' % (VERSION[0], VERSION[1]) if VERSION[2]: version = '%s.%s' % (version, VERSION[2]) if VERSION[3:] == ('alpha', 0): version = '%s pre-alpha' % version else: if VERSION[3] != 'final': version = '%s %s %s' % (version, VERSION[3], VERSION[4]) #from django.utils.version import get_svn_revision #svn_rev = get_svn_revision() #if svn_rev != u'SVN-unknown': # version = "%s %s" % (version, svn_rev) return version
<commit_before>VERSION = (0, 5, 3, 'final', 0) def get_version(): version = '%s.%s' % (VERSION[0], VERSION[1]) if VERSION[2]: version = '%s.%s' % (version, VERSION[2]) if VERSION[3:] == ('alpha', 0): version = '%s pre-alpha' % version else: if VERSION[3] != 'final': version = '%s %s %s' % (version, VERSION[3], VERSION[4]) #from django.utils.version import get_svn_revision #svn_rev = get_svn_revision() #if svn_rev != u'SVN-unknown': # version = "%s %s" % (version, svn_rev) return version <commit_msg>Update to 0.5.4 pre-alpha (preparing next build version)<commit_after>
VERSION = (0, 5, 4, 'alpha', 0) def get_version(): version = '%s.%s' % (VERSION[0], VERSION[1]) if VERSION[2]: version = '%s.%s' % (version, VERSION[2]) if VERSION[3:] == ('alpha', 0): version = '%s pre-alpha' % version else: if VERSION[3] != 'final': version = '%s %s %s' % (version, VERSION[3], VERSION[4]) #from django.utils.version import get_svn_revision #svn_rev = get_svn_revision() #if svn_rev != u'SVN-unknown': # version = "%s %s" % (version, svn_rev) return version
VERSION = (0, 5, 3, 'final', 0) def get_version(): version = '%s.%s' % (VERSION[0], VERSION[1]) if VERSION[2]: version = '%s.%s' % (version, VERSION[2]) if VERSION[3:] == ('alpha', 0): version = '%s pre-alpha' % version else: if VERSION[3] != 'final': version = '%s %s %s' % (version, VERSION[3], VERSION[4]) #from django.utils.version import get_svn_revision #svn_rev = get_svn_revision() #if svn_rev != u'SVN-unknown': # version = "%s %s" % (version, svn_rev) return version Update to 0.5.4 pre-alpha (preparing next build version)VERSION = (0, 5, 4, 'alpha', 0) def get_version(): version = '%s.%s' % (VERSION[0], VERSION[1]) if VERSION[2]: version = '%s.%s' % (version, VERSION[2]) if VERSION[3:] == ('alpha', 0): version = '%s pre-alpha' % version else: if VERSION[3] != 'final': version = '%s %s %s' % (version, VERSION[3], VERSION[4]) #from django.utils.version import get_svn_revision #svn_rev = get_svn_revision() #if svn_rev != u'SVN-unknown': # version = "%s %s" % (version, svn_rev) return version
<commit_before>VERSION = (0, 5, 3, 'final', 0) def get_version(): version = '%s.%s' % (VERSION[0], VERSION[1]) if VERSION[2]: version = '%s.%s' % (version, VERSION[2]) if VERSION[3:] == ('alpha', 0): version = '%s pre-alpha' % version else: if VERSION[3] != 'final': version = '%s %s %s' % (version, VERSION[3], VERSION[4]) #from django.utils.version import get_svn_revision #svn_rev = get_svn_revision() #if svn_rev != u'SVN-unknown': # version = "%s %s" % (version, svn_rev) return version <commit_msg>Update to 0.5.4 pre-alpha (preparing next build version)<commit_after>VERSION = (0, 5, 4, 'alpha', 0) def get_version(): version = '%s.%s' % (VERSION[0], VERSION[1]) if VERSION[2]: version = '%s.%s' % (version, VERSION[2]) if VERSION[3:] == ('alpha', 0): version = '%s pre-alpha' % version else: if VERSION[3] != 'final': version = '%s %s %s' % (version, VERSION[3], VERSION[4]) #from django.utils.version import get_svn_revision #svn_rev = get_svn_revision() #if svn_rev != u'SVN-unknown': # version = "%s %s" % (version, svn_rev) return version
bd850668e2c599abcf110c5e3ebb65b79ee91e22
xpserver_web/models.py
xpserver_web/models.py
from django.db import models from django.contrib.auth.models import User class Profile(models.Model): user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True) activation_code = models.CharField(max_length=255, default="0000") fcm_registration_id = models.CharField(max_length=255, blank=True) def __str__(self): if self.user: if self.fcm_registration_id: return '%s - profile with activated mobile app.' % self.user.username else: return '%s - profile.' % self.user.username else: return 'Profile unknown user'
from django.db import models from django.contrib.auth.models import User class Profile(models.Model): user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True) activation_code = models.CharField(max_length=255, default="0000") fcm_registration_id = models.CharField(max_length=255, blank=True) def __str__(self): if self.user: if self.fcm_registration_id: return '%s - profile with activated mobile app.' % self.user.username else: return '%s - profile.' % self.user.username else: return 'Profile with unknown user'
Change str method for unknow user
Change str method for unknow user
Python
mit
xp2017-hackergarden/server,xp2017-hackergarden/server,xp2017-hackergarden/server,xp2017-hackergarden/server
from django.db import models from django.contrib.auth.models import User class Profile(models.Model): user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True) activation_code = models.CharField(max_length=255, default="0000") fcm_registration_id = models.CharField(max_length=255, blank=True) def __str__(self): if self.user: if self.fcm_registration_id: return '%s - profile with activated mobile app.' % self.user.username else: return '%s - profile.' % self.user.username else: return 'Profile unknown user' Change str method for unknow user
from django.db import models from django.contrib.auth.models import User class Profile(models.Model): user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True) activation_code = models.CharField(max_length=255, default="0000") fcm_registration_id = models.CharField(max_length=255, blank=True) def __str__(self): if self.user: if self.fcm_registration_id: return '%s - profile with activated mobile app.' % self.user.username else: return '%s - profile.' % self.user.username else: return 'Profile with unknown user'
<commit_before>from django.db import models from django.contrib.auth.models import User class Profile(models.Model): user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True) activation_code = models.CharField(max_length=255, default="0000") fcm_registration_id = models.CharField(max_length=255, blank=True) def __str__(self): if self.user: if self.fcm_registration_id: return '%s - profile with activated mobile app.' % self.user.username else: return '%s - profile.' % self.user.username else: return 'Profile unknown user' <commit_msg>Change str method for unknow user<commit_after>
from django.db import models from django.contrib.auth.models import User class Profile(models.Model): user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True) activation_code = models.CharField(max_length=255, default="0000") fcm_registration_id = models.CharField(max_length=255, blank=True) def __str__(self): if self.user: if self.fcm_registration_id: return '%s - profile with activated mobile app.' % self.user.username else: return '%s - profile.' % self.user.username else: return 'Profile with unknown user'
from django.db import models from django.contrib.auth.models import User class Profile(models.Model): user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True) activation_code = models.CharField(max_length=255, default="0000") fcm_registration_id = models.CharField(max_length=255, blank=True) def __str__(self): if self.user: if self.fcm_registration_id: return '%s - profile with activated mobile app.' % self.user.username else: return '%s - profile.' % self.user.username else: return 'Profile unknown user' Change str method for unknow userfrom django.db import models from django.contrib.auth.models import User class Profile(models.Model): user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True) activation_code = models.CharField(max_length=255, default="0000") fcm_registration_id = models.CharField(max_length=255, blank=True) def __str__(self): if self.user: if self.fcm_registration_id: return '%s - profile with activated mobile app.' % self.user.username else: return '%s - profile.' % self.user.username else: return 'Profile with unknown user'
<commit_before>from django.db import models from django.contrib.auth.models import User class Profile(models.Model): user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True) activation_code = models.CharField(max_length=255, default="0000") fcm_registration_id = models.CharField(max_length=255, blank=True) def __str__(self): if self.user: if self.fcm_registration_id: return '%s - profile with activated mobile app.' % self.user.username else: return '%s - profile.' % self.user.username else: return 'Profile unknown user' <commit_msg>Change str method for unknow user<commit_after>from django.db import models from django.contrib.auth.models import User class Profile(models.Model): user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True) activation_code = models.CharField(max_length=255, default="0000") fcm_registration_id = models.CharField(max_length=255, blank=True) def __str__(self): if self.user: if self.fcm_registration_id: return '%s - profile with activated mobile app.' % self.user.username else: return '%s - profile.' % self.user.username else: return 'Profile with unknown user'
49f3d4d168fa62b545aa308684667c355db5c9d8
examples/write_avi.py
examples/write_avi.py
#!/usr/bin/env python # Here we use the camhd module to write a single-frame AVI file, only # downloading data from the moov atom and one frame. This AVI file can # easily be converted to a PNG or losslessly to a raw YUV using ffmpeg. # # Timothy Crone (tjcrone@gmail.com) import camhd # remote file filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov' # write single frame to avi file frame_number = 4976 write_frame(filename, frame_number)
#!/usr/bin/env python # Here we use the camhd module to write a single-frame AVI file, only # downloading data from the moov atom and one frame. This AVI file can # easily be converted to a PNG or losslessly to a raw YUV using ffmpeg. # # Timothy Crone (tjcrone@gmail.com) import camhd # remote file filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov' # write single frame to avi file frame_number = 4976 camhd.write_frame(filename, frame_number)
Fix minor bug in example program
Fix minor bug in example program
Python
mit
tjcrone/pycamhd
#!/usr/bin/env python # Here we use the camhd module to write a single-frame AVI file, only # downloading data from the moov atom and one frame. This AVI file can # easily be converted to a PNG or losslessly to a raw YUV using ffmpeg. # # Timothy Crone (tjcrone@gmail.com) import camhd # remote file filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov' # write single frame to avi file frame_number = 4976 write_frame(filename, frame_number) Fix minor bug in example program
#!/usr/bin/env python # Here we use the camhd module to write a single-frame AVI file, only # downloading data from the moov atom and one frame. This AVI file can # easily be converted to a PNG or losslessly to a raw YUV using ffmpeg. # # Timothy Crone (tjcrone@gmail.com) import camhd # remote file filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov' # write single frame to avi file frame_number = 4976 camhd.write_frame(filename, frame_number)
<commit_before>#!/usr/bin/env python # Here we use the camhd module to write a single-frame AVI file, only # downloading data from the moov atom and one frame. This AVI file can # easily be converted to a PNG or losslessly to a raw YUV using ffmpeg. # # Timothy Crone (tjcrone@gmail.com) import camhd # remote file filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov' # write single frame to avi file frame_number = 4976 write_frame(filename, frame_number) <commit_msg>Fix minor bug in example program<commit_after>
#!/usr/bin/env python # Here we use the camhd module to write a single-frame AVI file, only # downloading data from the moov atom and one frame. This AVI file can # easily be converted to a PNG or losslessly to a raw YUV using ffmpeg. # # Timothy Crone (tjcrone@gmail.com) import camhd # remote file filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov' # write single frame to avi file frame_number = 4976 camhd.write_frame(filename, frame_number)
#!/usr/bin/env python # Here we use the camhd module to write a single-frame AVI file, only # downloading data from the moov atom and one frame. This AVI file can # easily be converted to a PNG or losslessly to a raw YUV using ffmpeg. # # Timothy Crone (tjcrone@gmail.com) import camhd # remote file filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov' # write single frame to avi file frame_number = 4976 write_frame(filename, frame_number) Fix minor bug in example program#!/usr/bin/env python # Here we use the camhd module to write a single-frame AVI file, only # downloading data from the moov atom and one frame. This AVI file can # easily be converted to a PNG or losslessly to a raw YUV using ffmpeg. # # Timothy Crone (tjcrone@gmail.com) import camhd # remote file filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov' # write single frame to avi file frame_number = 4976 camhd.write_frame(filename, frame_number)
<commit_before>#!/usr/bin/env python # Here we use the camhd module to write a single-frame AVI file, only # downloading data from the moov atom and one frame. This AVI file can # easily be converted to a PNG or losslessly to a raw YUV using ffmpeg. # # Timothy Crone (tjcrone@gmail.com) import camhd # remote file filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov' # write single frame to avi file frame_number = 4976 write_frame(filename, frame_number) <commit_msg>Fix minor bug in example program<commit_after>#!/usr/bin/env python # Here we use the camhd module to write a single-frame AVI file, only # downloading data from the moov atom and one frame. This AVI file can # easily be converted to a PNG or losslessly to a raw YUV using ffmpeg. # # Timothy Crone (tjcrone@gmail.com) import camhd # remote file filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov' # write single frame to avi file frame_number = 4976 camhd.write_frame(filename, frame_number)
4f70897d5a85f1822a93df9bc91979ea79594901
nose2/tests/unit/test_generators_plugin.py
nose2/tests/unit/test_generators_plugin.py
from nose2.plugins.loader import functions from nose2.tests._common import TestCase class TestGeneratorUnpack(TestCase): tags = ['unit'] def setUp(self): self.expect = [(0, ('call', (0, 1))), (1, ('call', (1, 2))), (2, ('call', (2, 3))),] def test_unpack_handles_nose_style_generators(self): def gen(): for i in range(0, 3): yield 'call', i, i + 1 out = list(functions.Functions().unpack(gen())) self.assertEqual(out, self.expect) def test_unpack_handles_unittest2_style_generators(self): def gen(): for i in range(0, 3): yield 'call', (i, i + 1) out = list(functions.Functions().unpack(gen())) self.assertEqual(out, self.expect)
from nose2 import events, loader, session from nose2.plugins.loader import generators from nose2.tests._common import TestCase class TestGeneratorUnpack(TestCase): tags = ['unit'] def setUp(self): self.session = session.Session() self.loader = loader.PluggableTestLoader(self.session) self.expect = [(0, ('call', (0, 1))), (1, ('call', (1, 2))), (2, ('call', (2, 3))),] self.plugin = generators.Generators(session=self.session) def test_unpack_handles_nose_style_generators(self): def gen(): for i in range(0, 3): yield 'call', i, i + 1 out = list(self.plugin.unpack(gen())) self.assertEqual(out, self.expect) def test_unpack_handles_unittest2_style_generators(self): def gen(): for i in range(0, 3): yield 'call', (i, i + 1) out = list(self.plugin.unpack(gen())) self.assertEqual(out, self.expect) def test_ignores_ordinary_functions(self): class Mod(object): pass def test(): pass m = Mod() m.test = test event = events.LoadFromModuleEvent(self.loader, m) self.session.hooks.loadTestsFromModule(event) self.assertEqual(len(event.extraTests), 0) def test_can_load_tests_from_generator_functions(self): class Mod(object): pass def check(x): assert x == 1 def test(): yield check, 1 yield check, 2 m = Mod() m.test = test event = events.LoadFromModuleEvent(self.loader, m) self.session.hooks.loadTestsFromModule(event) self.assertEqual(len(event.extraTests), 2)
Add initial tests for generators plugin
Add initial tests for generators plugin
Python
bsd-2-clause
ojengwa/nose2,ojengwa/nose2,leth/nose2,little-dude/nose2,ptthiem/nose2,ezigman/nose2,ezigman/nose2,ptthiem/nose2,leth/nose2,little-dude/nose2
from nose2.plugins.loader import functions from nose2.tests._common import TestCase class TestGeneratorUnpack(TestCase): tags = ['unit'] def setUp(self): self.expect = [(0, ('call', (0, 1))), (1, ('call', (1, 2))), (2, ('call', (2, 3))),] def test_unpack_handles_nose_style_generators(self): def gen(): for i in range(0, 3): yield 'call', i, i + 1 out = list(functions.Functions().unpack(gen())) self.assertEqual(out, self.expect) def test_unpack_handles_unittest2_style_generators(self): def gen(): for i in range(0, 3): yield 'call', (i, i + 1) out = list(functions.Functions().unpack(gen())) self.assertEqual(out, self.expect) Add initial tests for generators plugin
from nose2 import events, loader, session from nose2.plugins.loader import generators from nose2.tests._common import TestCase class TestGeneratorUnpack(TestCase): tags = ['unit'] def setUp(self): self.session = session.Session() self.loader = loader.PluggableTestLoader(self.session) self.expect = [(0, ('call', (0, 1))), (1, ('call', (1, 2))), (2, ('call', (2, 3))),] self.plugin = generators.Generators(session=self.session) def test_unpack_handles_nose_style_generators(self): def gen(): for i in range(0, 3): yield 'call', i, i + 1 out = list(self.plugin.unpack(gen())) self.assertEqual(out, self.expect) def test_unpack_handles_unittest2_style_generators(self): def gen(): for i in range(0, 3): yield 'call', (i, i + 1) out = list(self.plugin.unpack(gen())) self.assertEqual(out, self.expect) def test_ignores_ordinary_functions(self): class Mod(object): pass def test(): pass m = Mod() m.test = test event = events.LoadFromModuleEvent(self.loader, m) self.session.hooks.loadTestsFromModule(event) self.assertEqual(len(event.extraTests), 0) def test_can_load_tests_from_generator_functions(self): class Mod(object): pass def check(x): assert x == 1 def test(): yield check, 1 yield check, 2 m = Mod() m.test = test event = events.LoadFromModuleEvent(self.loader, m) self.session.hooks.loadTestsFromModule(event) self.assertEqual(len(event.extraTests), 2)
<commit_before>from nose2.plugins.loader import functions from nose2.tests._common import TestCase class TestGeneratorUnpack(TestCase): tags = ['unit'] def setUp(self): self.expect = [(0, ('call', (0, 1))), (1, ('call', (1, 2))), (2, ('call', (2, 3))),] def test_unpack_handles_nose_style_generators(self): def gen(): for i in range(0, 3): yield 'call', i, i + 1 out = list(functions.Functions().unpack(gen())) self.assertEqual(out, self.expect) def test_unpack_handles_unittest2_style_generators(self): def gen(): for i in range(0, 3): yield 'call', (i, i + 1) out = list(functions.Functions().unpack(gen())) self.assertEqual(out, self.expect) <commit_msg>Add initial tests for generators plugin<commit_after>
from nose2 import events, loader, session from nose2.plugins.loader import generators from nose2.tests._common import TestCase class TestGeneratorUnpack(TestCase): tags = ['unit'] def setUp(self): self.session = session.Session() self.loader = loader.PluggableTestLoader(self.session) self.expect = [(0, ('call', (0, 1))), (1, ('call', (1, 2))), (2, ('call', (2, 3))),] self.plugin = generators.Generators(session=self.session) def test_unpack_handles_nose_style_generators(self): def gen(): for i in range(0, 3): yield 'call', i, i + 1 out = list(self.plugin.unpack(gen())) self.assertEqual(out, self.expect) def test_unpack_handles_unittest2_style_generators(self): def gen(): for i in range(0, 3): yield 'call', (i, i + 1) out = list(self.plugin.unpack(gen())) self.assertEqual(out, self.expect) def test_ignores_ordinary_functions(self): class Mod(object): pass def test(): pass m = Mod() m.test = test event = events.LoadFromModuleEvent(self.loader, m) self.session.hooks.loadTestsFromModule(event) self.assertEqual(len(event.extraTests), 0) def test_can_load_tests_from_generator_functions(self): class Mod(object): pass def check(x): assert x == 1 def test(): yield check, 1 yield check, 2 m = Mod() m.test = test event = events.LoadFromModuleEvent(self.loader, m) self.session.hooks.loadTestsFromModule(event) self.assertEqual(len(event.extraTests), 2)
from nose2.plugins.loader import functions from nose2.tests._common import TestCase class TestGeneratorUnpack(TestCase): tags = ['unit'] def setUp(self): self.expect = [(0, ('call', (0, 1))), (1, ('call', (1, 2))), (2, ('call', (2, 3))),] def test_unpack_handles_nose_style_generators(self): def gen(): for i in range(0, 3): yield 'call', i, i + 1 out = list(functions.Functions().unpack(gen())) self.assertEqual(out, self.expect) def test_unpack_handles_unittest2_style_generators(self): def gen(): for i in range(0, 3): yield 'call', (i, i + 1) out = list(functions.Functions().unpack(gen())) self.assertEqual(out, self.expect) Add initial tests for generators pluginfrom nose2 import events, loader, session from nose2.plugins.loader import generators from nose2.tests._common import TestCase class TestGeneratorUnpack(TestCase): tags = ['unit'] def setUp(self): self.session = session.Session() self.loader = loader.PluggableTestLoader(self.session) self.expect = [(0, ('call', (0, 1))), (1, ('call', (1, 2))), (2, ('call', (2, 3))),] self.plugin = generators.Generators(session=self.session) def test_unpack_handles_nose_style_generators(self): def gen(): for i in range(0, 3): yield 'call', i, i + 1 out = list(self.plugin.unpack(gen())) self.assertEqual(out, self.expect) def test_unpack_handles_unittest2_style_generators(self): def gen(): for i in range(0, 3): yield 'call', (i, i + 1) out = list(self.plugin.unpack(gen())) self.assertEqual(out, self.expect) def test_ignores_ordinary_functions(self): class Mod(object): pass def test(): pass m = Mod() m.test = test event = events.LoadFromModuleEvent(self.loader, m) self.session.hooks.loadTestsFromModule(event) self.assertEqual(len(event.extraTests), 0) def test_can_load_tests_from_generator_functions(self): class Mod(object): pass def check(x): assert x == 1 def test(): yield check, 1 yield check, 2 m = Mod() m.test = test event = events.LoadFromModuleEvent(self.loader, m) self.session.hooks.loadTestsFromModule(event) self.assertEqual(len(event.extraTests), 2)
<commit_before>from nose2.plugins.loader import functions from nose2.tests._common import TestCase class TestGeneratorUnpack(TestCase): tags = ['unit'] def setUp(self): self.expect = [(0, ('call', (0, 1))), (1, ('call', (1, 2))), (2, ('call', (2, 3))),] def test_unpack_handles_nose_style_generators(self): def gen(): for i in range(0, 3): yield 'call', i, i + 1 out = list(functions.Functions().unpack(gen())) self.assertEqual(out, self.expect) def test_unpack_handles_unittest2_style_generators(self): def gen(): for i in range(0, 3): yield 'call', (i, i + 1) out = list(functions.Functions().unpack(gen())) self.assertEqual(out, self.expect) <commit_msg>Add initial tests for generators plugin<commit_after>from nose2 import events, loader, session from nose2.plugins.loader import generators from nose2.tests._common import TestCase class TestGeneratorUnpack(TestCase): tags = ['unit'] def setUp(self): self.session = session.Session() self.loader = loader.PluggableTestLoader(self.session) self.expect = [(0, ('call', (0, 1))), (1, ('call', (1, 2))), (2, ('call', (2, 3))),] self.plugin = generators.Generators(session=self.session) def test_unpack_handles_nose_style_generators(self): def gen(): for i in range(0, 3): yield 'call', i, i + 1 out = list(self.plugin.unpack(gen())) self.assertEqual(out, self.expect) def test_unpack_handles_unittest2_style_generators(self): def gen(): for i in range(0, 3): yield 'call', (i, i + 1) out = list(self.plugin.unpack(gen())) self.assertEqual(out, self.expect) def test_ignores_ordinary_functions(self): class Mod(object): pass def test(): pass m = Mod() m.test = test event = events.LoadFromModuleEvent(self.loader, m) self.session.hooks.loadTestsFromModule(event) self.assertEqual(len(event.extraTests), 0) def test_can_load_tests_from_generator_functions(self): class Mod(object): pass def check(x): assert x == 1 def test(): yield check, 1 yield check, 2 m = Mod() m.test = test event = events.LoadFromModuleEvent(self.loader, m) self.session.hooks.loadTestsFromModule(event) self.assertEqual(len(event.extraTests), 2)
8aa02a12883d058b498ac56673079c1e9b53fdf7
src/libcask/network.py
src/libcask/network.py
import subprocess class SetupNetworkMixin(object): def _setup_hostname(self): with self.get_attachment().attach(): subprocess.check_call(['hostname', self.hostname]) def _setup_virtual_ethernet(self): # Setup virtual ethernet interface on the host # TODO - Need to allocate virtual interface names to containers! subprocess.check_call([ 'ip', 'link', 'add', 'name', 'veth0', 'type', 'veth', 'peer', 'name', 'veth1', 'netns', str(self.pid()) ]) subprocess.check_call([ 'ifconfig', 'veth0', self.ipaddr_host, 'up', ]) # Set up virtual ethernet interface inside the container # TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're # not relying on the container having ifconfig. with self.get_attachment().attach(): subprocess.check_call([ 'ifconfig', 'veth1', self.ipaddr, 'up', ]) def setup_network(self): self._setup_hostname() self._setup_virtual_ethernet()
import subprocess class SetupNetworkMixin(object): def _setup_hostname(self): with self.get_attachment().attach(): subprocess.check_call(['hostname', self.hostname]) def _setup_virtual_ethernet(self): veth_name = 'veth-{hostname}'.format(hostname=self.hostname) veth_host_name = 'hveth-{hostname}'.format(hostname=self.hostname) # Create virtual ethernet pair subprocess.check_call([ 'ip', 'link', 'add', 'name', veth_host_name, 'type', 'veth', 'peer', 'name', veth_name, 'netns', str(self.pid()) ]) # Add the container's host IP address and bring the interface up subprocess.check_call(['ip', 'addr', 'add', self.ipaddr_host, 'dev', veth_host_name]) subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'up']) # Add the host interface to the bridge # Assuming here that `cask0` bridge interface exists. It should # be created and initialized by the Makefile. subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'master', 'cask0']) # Set up virtual ethernet interface inside the container # TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're # not relying on the container having ifconfig. with self.get_attachment().attach(): subprocess.check_call([ 'ifconfig', veth_name, self.ipaddr, 'up', ]) def setup_network(self): self._setup_hostname() self._setup_virtual_ethernet()
Set up separate veth interfaces for each container with a global bridge
Set up separate veth interfaces for each container with a global bridge
Python
mit
ianpreston/cask,ianpreston/cask
import subprocess class SetupNetworkMixin(object): def _setup_hostname(self): with self.get_attachment().attach(): subprocess.check_call(['hostname', self.hostname]) def _setup_virtual_ethernet(self): # Setup virtual ethernet interface on the host # TODO - Need to allocate virtual interface names to containers! subprocess.check_call([ 'ip', 'link', 'add', 'name', 'veth0', 'type', 'veth', 'peer', 'name', 'veth1', 'netns', str(self.pid()) ]) subprocess.check_call([ 'ifconfig', 'veth0', self.ipaddr_host, 'up', ]) # Set up virtual ethernet interface inside the container # TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're # not relying on the container having ifconfig. with self.get_attachment().attach(): subprocess.check_call([ 'ifconfig', 'veth1', self.ipaddr, 'up', ]) def setup_network(self): self._setup_hostname() self._setup_virtual_ethernet() Set up separate veth interfaces for each container with a global bridge
import subprocess class SetupNetworkMixin(object): def _setup_hostname(self): with self.get_attachment().attach(): subprocess.check_call(['hostname', self.hostname]) def _setup_virtual_ethernet(self): veth_name = 'veth-{hostname}'.format(hostname=self.hostname) veth_host_name = 'hveth-{hostname}'.format(hostname=self.hostname) # Create virtual ethernet pair subprocess.check_call([ 'ip', 'link', 'add', 'name', veth_host_name, 'type', 'veth', 'peer', 'name', veth_name, 'netns', str(self.pid()) ]) # Add the container's host IP address and bring the interface up subprocess.check_call(['ip', 'addr', 'add', self.ipaddr_host, 'dev', veth_host_name]) subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'up']) # Add the host interface to the bridge # Assuming here that `cask0` bridge interface exists. It should # be created and initialized by the Makefile. subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'master', 'cask0']) # Set up virtual ethernet interface inside the container # TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're # not relying on the container having ifconfig. with self.get_attachment().attach(): subprocess.check_call([ 'ifconfig', veth_name, self.ipaddr, 'up', ]) def setup_network(self): self._setup_hostname() self._setup_virtual_ethernet()
<commit_before>import subprocess class SetupNetworkMixin(object): def _setup_hostname(self): with self.get_attachment().attach(): subprocess.check_call(['hostname', self.hostname]) def _setup_virtual_ethernet(self): # Setup virtual ethernet interface on the host # TODO - Need to allocate virtual interface names to containers! subprocess.check_call([ 'ip', 'link', 'add', 'name', 'veth0', 'type', 'veth', 'peer', 'name', 'veth1', 'netns', str(self.pid()) ]) subprocess.check_call([ 'ifconfig', 'veth0', self.ipaddr_host, 'up', ]) # Set up virtual ethernet interface inside the container # TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're # not relying on the container having ifconfig. with self.get_attachment().attach(): subprocess.check_call([ 'ifconfig', 'veth1', self.ipaddr, 'up', ]) def setup_network(self): self._setup_hostname() self._setup_virtual_ethernet() <commit_msg>Set up separate veth interfaces for each container with a global bridge<commit_after>
import subprocess class SetupNetworkMixin(object): def _setup_hostname(self): with self.get_attachment().attach(): subprocess.check_call(['hostname', self.hostname]) def _setup_virtual_ethernet(self): veth_name = 'veth-{hostname}'.format(hostname=self.hostname) veth_host_name = 'hveth-{hostname}'.format(hostname=self.hostname) # Create virtual ethernet pair subprocess.check_call([ 'ip', 'link', 'add', 'name', veth_host_name, 'type', 'veth', 'peer', 'name', veth_name, 'netns', str(self.pid()) ]) # Add the container's host IP address and bring the interface up subprocess.check_call(['ip', 'addr', 'add', self.ipaddr_host, 'dev', veth_host_name]) subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'up']) # Add the host interface to the bridge # Assuming here that `cask0` bridge interface exists. It should # be created and initialized by the Makefile. subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'master', 'cask0']) # Set up virtual ethernet interface inside the container # TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're # not relying on the container having ifconfig. with self.get_attachment().attach(): subprocess.check_call([ 'ifconfig', veth_name, self.ipaddr, 'up', ]) def setup_network(self): self._setup_hostname() self._setup_virtual_ethernet()
import subprocess class SetupNetworkMixin(object): def _setup_hostname(self): with self.get_attachment().attach(): subprocess.check_call(['hostname', self.hostname]) def _setup_virtual_ethernet(self): # Setup virtual ethernet interface on the host # TODO - Need to allocate virtual interface names to containers! subprocess.check_call([ 'ip', 'link', 'add', 'name', 'veth0', 'type', 'veth', 'peer', 'name', 'veth1', 'netns', str(self.pid()) ]) subprocess.check_call([ 'ifconfig', 'veth0', self.ipaddr_host, 'up', ]) # Set up virtual ethernet interface inside the container # TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're # not relying on the container having ifconfig. with self.get_attachment().attach(): subprocess.check_call([ 'ifconfig', 'veth1', self.ipaddr, 'up', ]) def setup_network(self): self._setup_hostname() self._setup_virtual_ethernet() Set up separate veth interfaces for each container with a global bridgeimport subprocess class SetupNetworkMixin(object): def _setup_hostname(self): with self.get_attachment().attach(): subprocess.check_call(['hostname', self.hostname]) def _setup_virtual_ethernet(self): veth_name = 'veth-{hostname}'.format(hostname=self.hostname) veth_host_name = 'hveth-{hostname}'.format(hostname=self.hostname) # Create virtual ethernet pair subprocess.check_call([ 'ip', 'link', 'add', 'name', veth_host_name, 'type', 'veth', 'peer', 'name', veth_name, 'netns', str(self.pid()) ]) # Add the container's host IP address and bring the interface up subprocess.check_call(['ip', 'addr', 'add', self.ipaddr_host, 'dev', veth_host_name]) subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'up']) # Add the host interface to the bridge # Assuming here that `cask0` bridge interface exists. It should # be created and initialized by the Makefile. subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'master', 'cask0']) # Set up virtual ethernet interface inside the container # TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're # not relying on the container having ifconfig. with self.get_attachment().attach(): subprocess.check_call([ 'ifconfig', veth_name, self.ipaddr, 'up', ]) def setup_network(self): self._setup_hostname() self._setup_virtual_ethernet()
<commit_before>import subprocess class SetupNetworkMixin(object): def _setup_hostname(self): with self.get_attachment().attach(): subprocess.check_call(['hostname', self.hostname]) def _setup_virtual_ethernet(self): # Setup virtual ethernet interface on the host # TODO - Need to allocate virtual interface names to containers! subprocess.check_call([ 'ip', 'link', 'add', 'name', 'veth0', 'type', 'veth', 'peer', 'name', 'veth1', 'netns', str(self.pid()) ]) subprocess.check_call([ 'ifconfig', 'veth0', self.ipaddr_host, 'up', ]) # Set up virtual ethernet interface inside the container # TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're # not relying on the container having ifconfig. with self.get_attachment().attach(): subprocess.check_call([ 'ifconfig', 'veth1', self.ipaddr, 'up', ]) def setup_network(self): self._setup_hostname() self._setup_virtual_ethernet() <commit_msg>Set up separate veth interfaces for each container with a global bridge<commit_after>import subprocess class SetupNetworkMixin(object): def _setup_hostname(self): with self.get_attachment().attach(): subprocess.check_call(['hostname', self.hostname]) def _setup_virtual_ethernet(self): veth_name = 'veth-{hostname}'.format(hostname=self.hostname) veth_host_name = 'hveth-{hostname}'.format(hostname=self.hostname) # Create virtual ethernet pair subprocess.check_call([ 'ip', 'link', 'add', 'name', veth_host_name, 'type', 'veth', 'peer', 'name', veth_name, 'netns', str(self.pid()) ]) # Add the container's host IP address and bring the interface up subprocess.check_call(['ip', 'addr', 'add', self.ipaddr_host, 'dev', veth_host_name]) subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'up']) # Add the host interface to the bridge # Assuming here that `cask0` bridge interface exists. It should # be created and initialized by the Makefile. subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'master', 'cask0']) # Set up virtual ethernet interface inside the container # TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're # not relying on the container having ifconfig. with self.get_attachment().attach(): subprocess.check_call([ 'ifconfig', veth_name, self.ipaddr, 'up', ]) def setup_network(self): self._setup_hostname() self._setup_virtual_ethernet()
afdab20403a360508bced14f4750dd6ef4e6aa57
flask_apidoc/utils.py
flask_apidoc/utils.py
""" Helpers. """ def cached(f): """ Cache decorator for functions taking one or more arguments. :param f: The function to be cached. :return: The cached value. """ class CachedDict(dict): def __init__(self, f): self.f = f def __call__(self, *args): return self[args] def __missing__(self, key): ret = self[key] = self.f(*key) return ret return CachedDict(f)
""" Helpers. """ import functools def cached(f): """ Cache decorator for functions taking one or more arguments. :param f: The function to be cached. :return: The cached value. """ cache = f.cache = {} @functools.wraps(f) def decorator(*args, **kwargs): key = str(args) + str(kwargs) if key not in cache: cache[key] = f(*args, **kwargs) return cache[key] return decorator
Improve cached decorator to support class based methods
Improve cached decorator to support class based methods
Python
mit
viniciuschiele/flask-apidoc
""" Helpers. """ def cached(f): """ Cache decorator for functions taking one or more arguments. :param f: The function to be cached. :return: The cached value. """ class CachedDict(dict): def __init__(self, f): self.f = f def __call__(self, *args): return self[args] def __missing__(self, key): ret = self[key] = self.f(*key) return ret return CachedDict(f) Improve cached decorator to support class based methods
""" Helpers. """ import functools def cached(f): """ Cache decorator for functions taking one or more arguments. :param f: The function to be cached. :return: The cached value. """ cache = f.cache = {} @functools.wraps(f) def decorator(*args, **kwargs): key = str(args) + str(kwargs) if key not in cache: cache[key] = f(*args, **kwargs) return cache[key] return decorator
<commit_before>""" Helpers. """ def cached(f): """ Cache decorator for functions taking one or more arguments. :param f: The function to be cached. :return: The cached value. """ class CachedDict(dict): def __init__(self, f): self.f = f def __call__(self, *args): return self[args] def __missing__(self, key): ret = self[key] = self.f(*key) return ret return CachedDict(f) <commit_msg>Improve cached decorator to support class based methods<commit_after>
""" Helpers. """ import functools def cached(f): """ Cache decorator for functions taking one or more arguments. :param f: The function to be cached. :return: The cached value. """ cache = f.cache = {} @functools.wraps(f) def decorator(*args, **kwargs): key = str(args) + str(kwargs) if key not in cache: cache[key] = f(*args, **kwargs) return cache[key] return decorator
""" Helpers. """ def cached(f): """ Cache decorator for functions taking one or more arguments. :param f: The function to be cached. :return: The cached value. """ class CachedDict(dict): def __init__(self, f): self.f = f def __call__(self, *args): return self[args] def __missing__(self, key): ret = self[key] = self.f(*key) return ret return CachedDict(f) Improve cached decorator to support class based methods""" Helpers. """ import functools def cached(f): """ Cache decorator for functions taking one or more arguments. :param f: The function to be cached. :return: The cached value. """ cache = f.cache = {} @functools.wraps(f) def decorator(*args, **kwargs): key = str(args) + str(kwargs) if key not in cache: cache[key] = f(*args, **kwargs) return cache[key] return decorator
<commit_before>""" Helpers. """ def cached(f): """ Cache decorator for functions taking one or more arguments. :param f: The function to be cached. :return: The cached value. """ class CachedDict(dict): def __init__(self, f): self.f = f def __call__(self, *args): return self[args] def __missing__(self, key): ret = self[key] = self.f(*key) return ret return CachedDict(f) <commit_msg>Improve cached decorator to support class based methods<commit_after>""" Helpers. """ import functools def cached(f): """ Cache decorator for functions taking one or more arguments. :param f: The function to be cached. :return: The cached value. """ cache = f.cache = {} @functools.wraps(f) def decorator(*args, **kwargs): key = str(args) + str(kwargs) if key not in cache: cache[key] = f(*args, **kwargs) return cache[key] return decorator
540b7cbcf51745e7a7eddb5f0030bc9f52dab35a
Demo/scripts/makedir.py
Demo/scripts/makedir.py
#! /usr/bin/env python # Like mkdir, but also make intermediate directories if necessary. # It is not an error if the given directory already exists (as long # as it is a directory). # Errors are not treated specially -- you just get a Python exception. import sys, os def main(): for p in sys.argv[1:]: makedirs(p) def makedirs(p): if not os.path.isdir(p): head, tail = os.path.split(p) makedirs(head) os.mkdir(p, 0777) main()
#! /usr/bin/env python # Like mkdir, but also make intermediate directories if necessary. # It is not an error if the given directory already exists (as long # as it is a directory). # Errors are not treated specially -- you just get a Python exception. import sys, os def main(): for p in sys.argv[1:]: makedirs(p) def makedirs(p): if p and not os.path.isdir(p): head, tail = os.path.split(p) makedirs(head) os.mkdir(p, 0777) main()
Fix bug discovered by Klaus-Juergen Wolf -- it runs into infinite recursion!
Fix bug discovered by Klaus-Juergen Wolf -- it runs into infinite recursion!
Python
mit
sk-/python2.7-type-annotator,sk-/python2.7-type-annotator,sk-/python2.7-type-annotator
#! /usr/bin/env python # Like mkdir, but also make intermediate directories if necessary. # It is not an error if the given directory already exists (as long # as it is a directory). # Errors are not treated specially -- you just get a Python exception. import sys, os def main(): for p in sys.argv[1:]: makedirs(p) def makedirs(p): if not os.path.isdir(p): head, tail = os.path.split(p) makedirs(head) os.mkdir(p, 0777) main() Fix bug discovered by Klaus-Juergen Wolf -- it runs into infinite recursion!
#! /usr/bin/env python # Like mkdir, but also make intermediate directories if necessary. # It is not an error if the given directory already exists (as long # as it is a directory). # Errors are not treated specially -- you just get a Python exception. import sys, os def main(): for p in sys.argv[1:]: makedirs(p) def makedirs(p): if p and not os.path.isdir(p): head, tail = os.path.split(p) makedirs(head) os.mkdir(p, 0777) main()
<commit_before>#! /usr/bin/env python # Like mkdir, but also make intermediate directories if necessary. # It is not an error if the given directory already exists (as long # as it is a directory). # Errors are not treated specially -- you just get a Python exception. import sys, os def main(): for p in sys.argv[1:]: makedirs(p) def makedirs(p): if not os.path.isdir(p): head, tail = os.path.split(p) makedirs(head) os.mkdir(p, 0777) main() <commit_msg>Fix bug discovered by Klaus-Juergen Wolf -- it runs into infinite recursion!<commit_after>
#! /usr/bin/env python # Like mkdir, but also make intermediate directories if necessary. # It is not an error if the given directory already exists (as long # as it is a directory). # Errors are not treated specially -- you just get a Python exception. import sys, os def main(): for p in sys.argv[1:]: makedirs(p) def makedirs(p): if p and not os.path.isdir(p): head, tail = os.path.split(p) makedirs(head) os.mkdir(p, 0777) main()
#! /usr/bin/env python # Like mkdir, but also make intermediate directories if necessary. # It is not an error if the given directory already exists (as long # as it is a directory). # Errors are not treated specially -- you just get a Python exception. import sys, os def main(): for p in sys.argv[1:]: makedirs(p) def makedirs(p): if not os.path.isdir(p): head, tail = os.path.split(p) makedirs(head) os.mkdir(p, 0777) main() Fix bug discovered by Klaus-Juergen Wolf -- it runs into infinite recursion!#! /usr/bin/env python # Like mkdir, but also make intermediate directories if necessary. # It is not an error if the given directory already exists (as long # as it is a directory). # Errors are not treated specially -- you just get a Python exception. import sys, os def main(): for p in sys.argv[1:]: makedirs(p) def makedirs(p): if p and not os.path.isdir(p): head, tail = os.path.split(p) makedirs(head) os.mkdir(p, 0777) main()
<commit_before>#! /usr/bin/env python # Like mkdir, but also make intermediate directories if necessary. # It is not an error if the given directory already exists (as long # as it is a directory). # Errors are not treated specially -- you just get a Python exception. import sys, os def main(): for p in sys.argv[1:]: makedirs(p) def makedirs(p): if not os.path.isdir(p): head, tail = os.path.split(p) makedirs(head) os.mkdir(p, 0777) main() <commit_msg>Fix bug discovered by Klaus-Juergen Wolf -- it runs into infinite recursion!<commit_after>#! /usr/bin/env python # Like mkdir, but also make intermediate directories if necessary. # It is not an error if the given directory already exists (as long # as it is a directory). # Errors are not treated specially -- you just get a Python exception. import sys, os def main(): for p in sys.argv[1:]: makedirs(p) def makedirs(p): if p and not os.path.isdir(p): head, tail = os.path.split(p) makedirs(head) os.mkdir(p, 0777) main()
d66e44fa9fd9b8e8944907b2490d32102c3fba82
keystoneclient/hacking/checks.py
keystoneclient/hacking/checks.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """python-keystoneclient's pep8 extensions. In order to make the review process faster and easier for core devs we are adding some python-keystoneclient specific pep8 checks. This will catch common errors so that core devs don't have to. """ import re def check_oslo_namespace_imports(logical_line, blank_before, filename): oslo_namespace_imports = re.compile( r"(((from)|(import))\s+oslo\." "((config)|(serialization)|(utils)|(i18n)))|" "(from\s+oslo\s+import\s+((config)|(serialization)|(utils)|(i18n)))") if re.match(oslo_namespace_imports, logical_line): msg = ("K333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def factory(register): register(check_oslo_namespace_imports)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """python-keystoneclient's pep8 extensions. In order to make the review process faster and easier for core devs we are adding some python-keystoneclient specific pep8 checks. This will catch common errors so that core devs don't have to. """ import re def check_oslo_namespace_imports(logical_line, blank_before, filename): oslo_namespace_imports = re.compile( r"(((from)|(import))\s+oslo\.)|(from\s+oslo\s+import\s+)") if re.match(oslo_namespace_imports, logical_line): msg = ("K333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def factory(register): register(check_oslo_namespace_imports)
Change hacking check to verify all oslo imports
Change hacking check to verify all oslo imports The hacking check was verifying that specific oslo imports weren't using the oslo-namespaced package. Since all the oslo libraries used by keystoneclient are now changed to use the new package name the hacking check can be simplified. bp drop-namespace-packages Change-Id: I6466e857c6eda0add6918e9fb14dc9296ed98600
Python
apache-2.0
Mercador/python-keystoneclient,Mercador/python-keystoneclient,sdpp/python-keystoneclient,magic0704/python-keystoneclient,ging/python-keystoneclient,darren-wang/ksc,magic0704/python-keystoneclient,klmitch/python-keystoneclient,darren-wang/ksc,klmitch/python-keystoneclient,sdpp/python-keystoneclient,ging/python-keystoneclient
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """python-keystoneclient's pep8 extensions. In order to make the review process faster and easier for core devs we are adding some python-keystoneclient specific pep8 checks. This will catch common errors so that core devs don't have to. """ import re def check_oslo_namespace_imports(logical_line, blank_before, filename): oslo_namespace_imports = re.compile( r"(((from)|(import))\s+oslo\." "((config)|(serialization)|(utils)|(i18n)))|" "(from\s+oslo\s+import\s+((config)|(serialization)|(utils)|(i18n)))") if re.match(oslo_namespace_imports, logical_line): msg = ("K333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def factory(register): register(check_oslo_namespace_imports) Change hacking check to verify all oslo imports The hacking check was verifying that specific oslo imports weren't using the oslo-namespaced package. Since all the oslo libraries used by keystoneclient are now changed to use the new package name the hacking check can be simplified. bp drop-namespace-packages Change-Id: I6466e857c6eda0add6918e9fb14dc9296ed98600
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """python-keystoneclient's pep8 extensions. In order to make the review process faster and easier for core devs we are adding some python-keystoneclient specific pep8 checks. This will catch common errors so that core devs don't have to. """ import re def check_oslo_namespace_imports(logical_line, blank_before, filename): oslo_namespace_imports = re.compile( r"(((from)|(import))\s+oslo\.)|(from\s+oslo\s+import\s+)") if re.match(oslo_namespace_imports, logical_line): msg = ("K333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def factory(register): register(check_oslo_namespace_imports)
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """python-keystoneclient's pep8 extensions. In order to make the review process faster and easier for core devs we are adding some python-keystoneclient specific pep8 checks. This will catch common errors so that core devs don't have to. """ import re def check_oslo_namespace_imports(logical_line, blank_before, filename): oslo_namespace_imports = re.compile( r"(((from)|(import))\s+oslo\." "((config)|(serialization)|(utils)|(i18n)))|" "(from\s+oslo\s+import\s+((config)|(serialization)|(utils)|(i18n)))") if re.match(oslo_namespace_imports, logical_line): msg = ("K333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def factory(register): register(check_oslo_namespace_imports) <commit_msg>Change hacking check to verify all oslo imports The hacking check was verifying that specific oslo imports weren't using the oslo-namespaced package. Since all the oslo libraries used by keystoneclient are now changed to use the new package name the hacking check can be simplified. bp drop-namespace-packages Change-Id: I6466e857c6eda0add6918e9fb14dc9296ed98600<commit_after>
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """python-keystoneclient's pep8 extensions. In order to make the review process faster and easier for core devs we are adding some python-keystoneclient specific pep8 checks. This will catch common errors so that core devs don't have to. """ import re def check_oslo_namespace_imports(logical_line, blank_before, filename): oslo_namespace_imports = re.compile( r"(((from)|(import))\s+oslo\.)|(from\s+oslo\s+import\s+)") if re.match(oslo_namespace_imports, logical_line): msg = ("K333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def factory(register): register(check_oslo_namespace_imports)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """python-keystoneclient's pep8 extensions. In order to make the review process faster and easier for core devs we are adding some python-keystoneclient specific pep8 checks. This will catch common errors so that core devs don't have to. """ import re def check_oslo_namespace_imports(logical_line, blank_before, filename): oslo_namespace_imports = re.compile( r"(((from)|(import))\s+oslo\." "((config)|(serialization)|(utils)|(i18n)))|" "(from\s+oslo\s+import\s+((config)|(serialization)|(utils)|(i18n)))") if re.match(oslo_namespace_imports, logical_line): msg = ("K333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def factory(register): register(check_oslo_namespace_imports) Change hacking check to verify all oslo imports The hacking check was verifying that specific oslo imports weren't using the oslo-namespaced package. Since all the oslo libraries used by keystoneclient are now changed to use the new package name the hacking check can be simplified. bp drop-namespace-packages Change-Id: I6466e857c6eda0add6918e9fb14dc9296ed98600# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """python-keystoneclient's pep8 extensions. In order to make the review process faster and easier for core devs we are adding some python-keystoneclient specific pep8 checks. This will catch common errors so that core devs don't have to. """ import re def check_oslo_namespace_imports(logical_line, blank_before, filename): oslo_namespace_imports = re.compile( r"(((from)|(import))\s+oslo\.)|(from\s+oslo\s+import\s+)") if re.match(oslo_namespace_imports, logical_line): msg = ("K333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def factory(register): register(check_oslo_namespace_imports)
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """python-keystoneclient's pep8 extensions. In order to make the review process faster and easier for core devs we are adding some python-keystoneclient specific pep8 checks. This will catch common errors so that core devs don't have to. """ import re def check_oslo_namespace_imports(logical_line, blank_before, filename): oslo_namespace_imports = re.compile( r"(((from)|(import))\s+oslo\." "((config)|(serialization)|(utils)|(i18n)))|" "(from\s+oslo\s+import\s+((config)|(serialization)|(utils)|(i18n)))") if re.match(oslo_namespace_imports, logical_line): msg = ("K333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def factory(register): register(check_oslo_namespace_imports) <commit_msg>Change hacking check to verify all oslo imports The hacking check was verifying that specific oslo imports weren't using the oslo-namespaced package. Since all the oslo libraries used by keystoneclient are now changed to use the new package name the hacking check can be simplified. bp drop-namespace-packages Change-Id: I6466e857c6eda0add6918e9fb14dc9296ed98600<commit_after># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """python-keystoneclient's pep8 extensions. In order to make the review process faster and easier for core devs we are adding some python-keystoneclient specific pep8 checks. This will catch common errors so that core devs don't have to. """ import re def check_oslo_namespace_imports(logical_line, blank_before, filename): oslo_namespace_imports = re.compile( r"(((from)|(import))\s+oslo\.)|(from\s+oslo\s+import\s+)") if re.match(oslo_namespace_imports, logical_line): msg = ("K333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def factory(register): register(check_oslo_namespace_imports)
36e0821fcd871935e48ae10926be7594d42f13b8
knowledge_repo/converters/pdf.py
knowledge_repo/converters/pdf.py
from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf()
from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML, CSS html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf(stylesheets=[CSS(string='body { font-family: Helvetica, sans-serif !important }')])
Change PDF font to Helvetica
Change PDF font to Helvetica Changing the PDF font from the default to Helvetica
Python
apache-2.0
airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo
from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf() Change PDF font to Helvetica Changing the PDF font from the default to Helvetica
from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML, CSS html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf(stylesheets=[CSS(string='body { font-family: Helvetica, sans-serif !important }')])
<commit_before>from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf() <commit_msg>Change PDF font to Helvetica Changing the PDF font from the default to Helvetica<commit_after>
from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML, CSS html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf(stylesheets=[CSS(string='body { font-family: Helvetica, sans-serif !important }')])
from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf() Change PDF font to Helvetica Changing the PDF font from the default to Helveticafrom ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML, CSS html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf(stylesheets=[CSS(string='body { font-family: Helvetica, sans-serif !important }')])
<commit_before>from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf() <commit_msg>Change PDF font to Helvetica Changing the PDF font from the default to Helvetica<commit_after>from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML, CSS html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf(stylesheets=[CSS(string='body { font-family: Helvetica, sans-serif !important }')])
985198a9ea569cc6f418c5b337632b91cdda7e37
lib/rapidsms/backends/backend.py
lib/rapidsms/backends/backend.py
#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 class Backend(object): def log(self, level, message): self.router.log(level, message) def start(self): raise NotImplementedError def stop(self): raise NotImplementedError def send(self): raise NotImplementedError def receive(self): raise NotImplementedError
#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 class Backend(object): def __init__ (self, router): self.router = router def log(self, level, message): self.router.log(level, message) def start(self): raise NotImplementedError def stop(self): raise NotImplementedError def send(self): raise NotImplementedError def receive(self): raise NotImplementedError
Add a constructor method for Backend
Add a constructor method for Backend
Python
bsd-3-clause
dimagi/rapidsms-core-dev,peterayeni/rapidsms,rapidsms/rapidsms-core-dev,lsgunth/rapidsms,catalpainternational/rapidsms,ken-muturi/rapidsms,ehealthafrica-ci/rapidsms,eHealthAfrica/rapidsms,unicefuganda/edtrac,unicefuganda/edtrac,eHealthAfrica/rapidsms,catalpainternational/rapidsms,catalpainternational/rapidsms,eHealthAfrica/rapidsms,caktus/rapidsms,caktus/rapidsms,lsgunth/rapidsms,peterayeni/rapidsms,peterayeni/rapidsms,ken-muturi/rapidsms,rapidsms/rapidsms-core-dev,ehealthafrica-ci/rapidsms,dimagi/rapidsms,dimagi/rapidsms-core-dev,lsgunth/rapidsms,ken-muturi/rapidsms,catalpainternational/rapidsms,caktus/rapidsms,ehealthafrica-ci/rapidsms,peterayeni/rapidsms,unicefuganda/edtrac,lsgunth/rapidsms,dimagi/rapidsms
#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 class Backend(object): def log(self, level, message): self.router.log(level, message) def start(self): raise NotImplementedError def stop(self): raise NotImplementedError def send(self): raise NotImplementedError def receive(self): raise NotImplementedError Add a constructor method for Backend
#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 class Backend(object): def __init__ (self, router): self.router = router def log(self, level, message): self.router.log(level, message) def start(self): raise NotImplementedError def stop(self): raise NotImplementedError def send(self): raise NotImplementedError def receive(self): raise NotImplementedError
<commit_before>#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 class Backend(object): def log(self, level, message): self.router.log(level, message) def start(self): raise NotImplementedError def stop(self): raise NotImplementedError def send(self): raise NotImplementedError def receive(self): raise NotImplementedError <commit_msg>Add a constructor method for Backend<commit_after>
#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 class Backend(object): def __init__ (self, router): self.router = router def log(self, level, message): self.router.log(level, message) def start(self): raise NotImplementedError def stop(self): raise NotImplementedError def send(self): raise NotImplementedError def receive(self): raise NotImplementedError
#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 class Backend(object): def log(self, level, message): self.router.log(level, message) def start(self): raise NotImplementedError def stop(self): raise NotImplementedError def send(self): raise NotImplementedError def receive(self): raise NotImplementedError Add a constructor method for Backend#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 class Backend(object): def __init__ (self, router): self.router = router def log(self, level, message): self.router.log(level, message) def start(self): raise NotImplementedError def stop(self): raise NotImplementedError def send(self): raise NotImplementedError def receive(self): raise NotImplementedError
<commit_before>#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 class Backend(object): def log(self, level, message): self.router.log(level, message) def start(self): raise NotImplementedError def stop(self): raise NotImplementedError def send(self): raise NotImplementedError def receive(self): raise NotImplementedError <commit_msg>Add a constructor method for Backend<commit_after>#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 class Backend(object): def __init__ (self, router): self.router = router def log(self, level, message): self.router.log(level, message) def start(self): raise NotImplementedError def stop(self): raise NotImplementedError def send(self): raise NotImplementedError def receive(self): raise NotImplementedError
7d574570ad831a86758352ef98bf6192f7d3e6fa
setup.py
setup.py
import os from setuptools import setup install_requires = [ 'mysql-python>=1.2.3', 'psycopg2>=2.4.2', 'pyyaml>=3.10.0', 'argparse', 'pytz', ] if os.name == 'posix': install_requires.append('termcolor>=1.1.0') setup( name='py-mysql2pgsql', version='0.1.6', description='Tool for migrating/converting from mysql to postgresql.', long_description=open('README.rst').read(), license='MIT License', author='Philip Southam', author_email='philipsoutham@gmail.com', url='https://github.com/philipsoutham/py-mysql2pgsql', zip_safe=False, packages=['mysql2pgsql', 'mysql2pgsql.lib'], scripts=['bin/py-mysql2pgsql'], platforms='any', install_requires=install_requires, classifiers=[ 'License :: OSI Approved :: MIT License', 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: System Administrators', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Topic :: Database', 'Topic :: Utilities' ], keywords = 'mysql postgres postgresql pgsql psql migration', )
import os from setuptools import setup install_requires = [ 'mysql-python>=1.2.3', 'psycopg2>=2.4.2', 'pyyaml>=3.10.0', 'pytz', ] if os.name == 'posix': install_requires.append('termcolor>=1.1.0') if version < (2,7) or (3,0) <= version <= (3,1): install_requires += ['argparse'] setup( name='py-mysql2pgsql', version='0.1.6', description='Tool for migrating/converting from mysql to postgresql.', long_description=open('README.rst').read(), license='MIT License', author='Philip Southam', author_email='philipsoutham@gmail.com', url='https://github.com/philipsoutham/py-mysql2pgsql', zip_safe=False, packages=['mysql2pgsql', 'mysql2pgsql.lib'], scripts=['bin/py-mysql2pgsql'], platforms='any', install_requires=install_requires, classifiers=[ 'License :: OSI Approved :: MIT License', 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: System Administrators', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Topic :: Database', 'Topic :: Utilities' ], keywords = 'mysql postgres postgresql pgsql psql migration', )
Correct argparse dependency - argparse already is a part of base python as of 2.7 and 3.2.
Correct argparse dependency - argparse already is a part of base python as of 2.7 and 3.2.
Python
mit
Varsha-Arun/py-mysql2pgsql,philipsoutham/py-mysql2pgsql,EdNeg/DB-Project,leture/py-mysql2pgsql,mr-tron/py-mysql2pgsql,EdNeg/DB-Project,emailgregn/py-mysql2pgsql,kevgathuku/py-mysql2pgsql,andela-kndungu/py-mysql2pgsql,EdNeg/DB-Project,SpootDev/py-mysql2pgsql,edwardsamuel/py-mysql2pgsql
import os from setuptools import setup install_requires = [ 'mysql-python>=1.2.3', 'psycopg2>=2.4.2', 'pyyaml>=3.10.0', 'argparse', 'pytz', ] if os.name == 'posix': install_requires.append('termcolor>=1.1.0') setup( name='py-mysql2pgsql', version='0.1.6', description='Tool for migrating/converting from mysql to postgresql.', long_description=open('README.rst').read(), license='MIT License', author='Philip Southam', author_email='philipsoutham@gmail.com', url='https://github.com/philipsoutham/py-mysql2pgsql', zip_safe=False, packages=['mysql2pgsql', 'mysql2pgsql.lib'], scripts=['bin/py-mysql2pgsql'], platforms='any', install_requires=install_requires, classifiers=[ 'License :: OSI Approved :: MIT License', 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: System Administrators', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Topic :: Database', 'Topic :: Utilities' ], keywords = 'mysql postgres postgresql pgsql psql migration', ) Correct argparse dependency - argparse already is a part of base python as of 2.7 and 3.2.
import os from setuptools import setup install_requires = [ 'mysql-python>=1.2.3', 'psycopg2>=2.4.2', 'pyyaml>=3.10.0', 'pytz', ] if os.name == 'posix': install_requires.append('termcolor>=1.1.0') if version < (2,7) or (3,0) <= version <= (3,1): install_requires += ['argparse'] setup( name='py-mysql2pgsql', version='0.1.6', description='Tool for migrating/converting from mysql to postgresql.', long_description=open('README.rst').read(), license='MIT License', author='Philip Southam', author_email='philipsoutham@gmail.com', url='https://github.com/philipsoutham/py-mysql2pgsql', zip_safe=False, packages=['mysql2pgsql', 'mysql2pgsql.lib'], scripts=['bin/py-mysql2pgsql'], platforms='any', install_requires=install_requires, classifiers=[ 'License :: OSI Approved :: MIT License', 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: System Administrators', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Topic :: Database', 'Topic :: Utilities' ], keywords = 'mysql postgres postgresql pgsql psql migration', )
<commit_before>import os from setuptools import setup install_requires = [ 'mysql-python>=1.2.3', 'psycopg2>=2.4.2', 'pyyaml>=3.10.0', 'argparse', 'pytz', ] if os.name == 'posix': install_requires.append('termcolor>=1.1.0') setup( name='py-mysql2pgsql', version='0.1.6', description='Tool for migrating/converting from mysql to postgresql.', long_description=open('README.rst').read(), license='MIT License', author='Philip Southam', author_email='philipsoutham@gmail.com', url='https://github.com/philipsoutham/py-mysql2pgsql', zip_safe=False, packages=['mysql2pgsql', 'mysql2pgsql.lib'], scripts=['bin/py-mysql2pgsql'], platforms='any', install_requires=install_requires, classifiers=[ 'License :: OSI Approved :: MIT License', 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: System Administrators', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Topic :: Database', 'Topic :: Utilities' ], keywords = 'mysql postgres postgresql pgsql psql migration', ) <commit_msg>Correct argparse dependency - argparse already is a part of base python as of 2.7 and 3.2.<commit_after>
import os from setuptools import setup install_requires = [ 'mysql-python>=1.2.3', 'psycopg2>=2.4.2', 'pyyaml>=3.10.0', 'pytz', ] if os.name == 'posix': install_requires.append('termcolor>=1.1.0') if version < (2,7) or (3,0) <= version <= (3,1): install_requires += ['argparse'] setup( name='py-mysql2pgsql', version='0.1.6', description='Tool for migrating/converting from mysql to postgresql.', long_description=open('README.rst').read(), license='MIT License', author='Philip Southam', author_email='philipsoutham@gmail.com', url='https://github.com/philipsoutham/py-mysql2pgsql', zip_safe=False, packages=['mysql2pgsql', 'mysql2pgsql.lib'], scripts=['bin/py-mysql2pgsql'], platforms='any', install_requires=install_requires, classifiers=[ 'License :: OSI Approved :: MIT License', 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: System Administrators', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Topic :: Database', 'Topic :: Utilities' ], keywords = 'mysql postgres postgresql pgsql psql migration', )
import os from setuptools import setup install_requires = [ 'mysql-python>=1.2.3', 'psycopg2>=2.4.2', 'pyyaml>=3.10.0', 'argparse', 'pytz', ] if os.name == 'posix': install_requires.append('termcolor>=1.1.0') setup( name='py-mysql2pgsql', version='0.1.6', description='Tool for migrating/converting from mysql to postgresql.', long_description=open('README.rst').read(), license='MIT License', author='Philip Southam', author_email='philipsoutham@gmail.com', url='https://github.com/philipsoutham/py-mysql2pgsql', zip_safe=False, packages=['mysql2pgsql', 'mysql2pgsql.lib'], scripts=['bin/py-mysql2pgsql'], platforms='any', install_requires=install_requires, classifiers=[ 'License :: OSI Approved :: MIT License', 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: System Administrators', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Topic :: Database', 'Topic :: Utilities' ], keywords = 'mysql postgres postgresql pgsql psql migration', ) Correct argparse dependency - argparse already is a part of base python as of 2.7 and 3.2.import os from setuptools import setup install_requires = [ 'mysql-python>=1.2.3', 'psycopg2>=2.4.2', 'pyyaml>=3.10.0', 'pytz', ] if os.name == 'posix': install_requires.append('termcolor>=1.1.0') if version < (2,7) or (3,0) <= version <= (3,1): install_requires += ['argparse'] setup( name='py-mysql2pgsql', version='0.1.6', description='Tool for migrating/converting from mysql to postgresql.', long_description=open('README.rst').read(), license='MIT License', author='Philip Southam', author_email='philipsoutham@gmail.com', url='https://github.com/philipsoutham/py-mysql2pgsql', zip_safe=False, packages=['mysql2pgsql', 'mysql2pgsql.lib'], scripts=['bin/py-mysql2pgsql'], platforms='any', install_requires=install_requires, classifiers=[ 'License :: OSI Approved :: MIT License', 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: System Administrators', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Topic :: Database', 'Topic :: Utilities' ], keywords = 'mysql postgres postgresql pgsql psql migration', )
<commit_before>import os from setuptools import setup install_requires = [ 'mysql-python>=1.2.3', 'psycopg2>=2.4.2', 'pyyaml>=3.10.0', 'argparse', 'pytz', ] if os.name == 'posix': install_requires.append('termcolor>=1.1.0') setup( name='py-mysql2pgsql', version='0.1.6', description='Tool for migrating/converting from mysql to postgresql.', long_description=open('README.rst').read(), license='MIT License', author='Philip Southam', author_email='philipsoutham@gmail.com', url='https://github.com/philipsoutham/py-mysql2pgsql', zip_safe=False, packages=['mysql2pgsql', 'mysql2pgsql.lib'], scripts=['bin/py-mysql2pgsql'], platforms='any', install_requires=install_requires, classifiers=[ 'License :: OSI Approved :: MIT License', 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: System Administrators', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Topic :: Database', 'Topic :: Utilities' ], keywords = 'mysql postgres postgresql pgsql psql migration', ) <commit_msg>Correct argparse dependency - argparse already is a part of base python as of 2.7 and 3.2.<commit_after>import os from setuptools import setup install_requires = [ 'mysql-python>=1.2.3', 'psycopg2>=2.4.2', 'pyyaml>=3.10.0', 'pytz', ] if os.name == 'posix': install_requires.append('termcolor>=1.1.0') if version < (2,7) or (3,0) <= version <= (3,1): install_requires += ['argparse'] setup( name='py-mysql2pgsql', version='0.1.6', description='Tool for migrating/converting from mysql to postgresql.', long_description=open('README.rst').read(), license='MIT License', author='Philip Southam', author_email='philipsoutham@gmail.com', url='https://github.com/philipsoutham/py-mysql2pgsql', zip_safe=False, packages=['mysql2pgsql', 'mysql2pgsql.lib'], scripts=['bin/py-mysql2pgsql'], platforms='any', install_requires=install_requires, classifiers=[ 'License :: OSI Approved :: MIT License', 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: System Administrators', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Topic :: Database', 'Topic :: Utilities' ], keywords = 'mysql postgres postgresql pgsql psql migration', )
329fedb27fe54d593f192912beda56588faec214
tests/__init__.py
tests/__init__.py
from django.conf import settings from mock import Mock, patch from unittest2 import TestCase settings.configure() # Need to import this after configure() from django.db.models import ForeignKey class TestPreference(object): _meta = Mock(fields=[ForeignKey('user', name='user')]) objects = Mock() def __init__(self, name, value, user=None): self.name = name self.value = value self.user = user def __eq__(self, other): return self.__dict__ == other.__dict__ def __repr__(self): return '<{name}:{value}:{user}>'.format(**self.__dict__) def __cmp__(self, other): return cmp(self.name, other.name) class TestUser(object): @property def preferences(self): return Mock(all=Mock(return_value=self._preferences)) @preferences.setter def preferences(self, value): self._preferences = [ TestPreference(k, v) for k, v in value.iteritems()] class SerializerTestCase(TestCase): def patch_from_native(self): patcher = patch( 'madprops.serializers.ModelSerializer.from_native', new=lambda self, data, files: TestPreference( data['name'], data['value'], data.get('user')) ) self.patched_from_native = patcher.start() self.addCleanup(patcher.stop)
from django.conf import settings from mock import Mock, patch from unittest2 import TestCase settings.configure() # Need to import this after configure() from django.db.models import ForeignKey class TestPreference(object): _meta = Mock(fields=[ForeignKey('user', name='user')]) objects = Mock() def __init__(self, name, value, user=None): self.name = name self.value = value self.user = user def __eq__(self, other): return self.__dict__ == other.__dict__ def __repr__(self): return '<{name}:{value}:{user}>'.format(**self.__dict__) def __cmp__(self, other): return cmp(self.name, other.name) class TestUser(object): @property def preferences(self): return Mock(all=Mock(return_value=self._preferences)) @preferences.setter def preferences(self, value): self._preferences = [ TestPreference(k, v) for k, v in value.iteritems()] class SerializerTestCase(TestCase): def patch_from_native(self): def from_native(self, data, files): self._errors = {} if data: self.perform_validation(data) return TestPreference(data['name'], data['value'], data.get('user')) patcher = patch( 'madprops.serializers.ModelSerializer.from_native', new=from_native) self.patched_from_native = patcher.start() self.addCleanup(patcher.stop) # get_fields inspects the model's _meta, deeply patcher = patch( 'madprops.serializers.ModelSerializer.get_fields', new=lambda self: {}) self.patched_get_fields = patcher.start() self.addCleanup(patcher.stop)
Extend mocking to run validation
Extend mocking to run validation
Python
mit
yola/drf-madprops
from django.conf import settings from mock import Mock, patch from unittest2 import TestCase settings.configure() # Need to import this after configure() from django.db.models import ForeignKey class TestPreference(object): _meta = Mock(fields=[ForeignKey('user', name='user')]) objects = Mock() def __init__(self, name, value, user=None): self.name = name self.value = value self.user = user def __eq__(self, other): return self.__dict__ == other.__dict__ def __repr__(self): return '<{name}:{value}:{user}>'.format(**self.__dict__) def __cmp__(self, other): return cmp(self.name, other.name) class TestUser(object): @property def preferences(self): return Mock(all=Mock(return_value=self._preferences)) @preferences.setter def preferences(self, value): self._preferences = [ TestPreference(k, v) for k, v in value.iteritems()] class SerializerTestCase(TestCase): def patch_from_native(self): patcher = patch( 'madprops.serializers.ModelSerializer.from_native', new=lambda self, data, files: TestPreference( data['name'], data['value'], data.get('user')) ) self.patched_from_native = patcher.start() self.addCleanup(patcher.stop) Extend mocking to run validation
from django.conf import settings from mock import Mock, patch from unittest2 import TestCase settings.configure() # Need to import this after configure() from django.db.models import ForeignKey class TestPreference(object): _meta = Mock(fields=[ForeignKey('user', name='user')]) objects = Mock() def __init__(self, name, value, user=None): self.name = name self.value = value self.user = user def __eq__(self, other): return self.__dict__ == other.__dict__ def __repr__(self): return '<{name}:{value}:{user}>'.format(**self.__dict__) def __cmp__(self, other): return cmp(self.name, other.name) class TestUser(object): @property def preferences(self): return Mock(all=Mock(return_value=self._preferences)) @preferences.setter def preferences(self, value): self._preferences = [ TestPreference(k, v) for k, v in value.iteritems()] class SerializerTestCase(TestCase): def patch_from_native(self): def from_native(self, data, files): self._errors = {} if data: self.perform_validation(data) return TestPreference(data['name'], data['value'], data.get('user')) patcher = patch( 'madprops.serializers.ModelSerializer.from_native', new=from_native) self.patched_from_native = patcher.start() self.addCleanup(patcher.stop) # get_fields inspects the model's _meta, deeply patcher = patch( 'madprops.serializers.ModelSerializer.get_fields', new=lambda self: {}) self.patched_get_fields = patcher.start() self.addCleanup(patcher.stop)
<commit_before>from django.conf import settings from mock import Mock, patch from unittest2 import TestCase settings.configure() # Need to import this after configure() from django.db.models import ForeignKey class TestPreference(object): _meta = Mock(fields=[ForeignKey('user', name='user')]) objects = Mock() def __init__(self, name, value, user=None): self.name = name self.value = value self.user = user def __eq__(self, other): return self.__dict__ == other.__dict__ def __repr__(self): return '<{name}:{value}:{user}>'.format(**self.__dict__) def __cmp__(self, other): return cmp(self.name, other.name) class TestUser(object): @property def preferences(self): return Mock(all=Mock(return_value=self._preferences)) @preferences.setter def preferences(self, value): self._preferences = [ TestPreference(k, v) for k, v in value.iteritems()] class SerializerTestCase(TestCase): def patch_from_native(self): patcher = patch( 'madprops.serializers.ModelSerializer.from_native', new=lambda self, data, files: TestPreference( data['name'], data['value'], data.get('user')) ) self.patched_from_native = patcher.start() self.addCleanup(patcher.stop) <commit_msg>Extend mocking to run validation<commit_after>
from django.conf import settings from mock import Mock, patch from unittest2 import TestCase settings.configure() # Need to import this after configure() from django.db.models import ForeignKey class TestPreference(object): _meta = Mock(fields=[ForeignKey('user', name='user')]) objects = Mock() def __init__(self, name, value, user=None): self.name = name self.value = value self.user = user def __eq__(self, other): return self.__dict__ == other.__dict__ def __repr__(self): return '<{name}:{value}:{user}>'.format(**self.__dict__) def __cmp__(self, other): return cmp(self.name, other.name) class TestUser(object): @property def preferences(self): return Mock(all=Mock(return_value=self._preferences)) @preferences.setter def preferences(self, value): self._preferences = [ TestPreference(k, v) for k, v in value.iteritems()] class SerializerTestCase(TestCase): def patch_from_native(self): def from_native(self, data, files): self._errors = {} if data: self.perform_validation(data) return TestPreference(data['name'], data['value'], data.get('user')) patcher = patch( 'madprops.serializers.ModelSerializer.from_native', new=from_native) self.patched_from_native = patcher.start() self.addCleanup(patcher.stop) # get_fields inspects the model's _meta, deeply patcher = patch( 'madprops.serializers.ModelSerializer.get_fields', new=lambda self: {}) self.patched_get_fields = patcher.start() self.addCleanup(patcher.stop)
from django.conf import settings from mock import Mock, patch from unittest2 import TestCase settings.configure() # Need to import this after configure() from django.db.models import ForeignKey class TestPreference(object): _meta = Mock(fields=[ForeignKey('user', name='user')]) objects = Mock() def __init__(self, name, value, user=None): self.name = name self.value = value self.user = user def __eq__(self, other): return self.__dict__ == other.__dict__ def __repr__(self): return '<{name}:{value}:{user}>'.format(**self.__dict__) def __cmp__(self, other): return cmp(self.name, other.name) class TestUser(object): @property def preferences(self): return Mock(all=Mock(return_value=self._preferences)) @preferences.setter def preferences(self, value): self._preferences = [ TestPreference(k, v) for k, v in value.iteritems()] class SerializerTestCase(TestCase): def patch_from_native(self): patcher = patch( 'madprops.serializers.ModelSerializer.from_native', new=lambda self, data, files: TestPreference( data['name'], data['value'], data.get('user')) ) self.patched_from_native = patcher.start() self.addCleanup(patcher.stop) Extend mocking to run validationfrom django.conf import settings from mock import Mock, patch from unittest2 import TestCase settings.configure() # Need to import this after configure() from django.db.models import ForeignKey class TestPreference(object): _meta = Mock(fields=[ForeignKey('user', name='user')]) objects = Mock() def __init__(self, name, value, user=None): self.name = name self.value = value self.user = user def __eq__(self, other): return self.__dict__ == other.__dict__ def __repr__(self): return '<{name}:{value}:{user}>'.format(**self.__dict__) def __cmp__(self, other): return cmp(self.name, other.name) class TestUser(object): @property def preferences(self): return Mock(all=Mock(return_value=self._preferences)) @preferences.setter def preferences(self, value): self._preferences = [ TestPreference(k, v) for k, v in value.iteritems()] class SerializerTestCase(TestCase): def patch_from_native(self): def from_native(self, data, files): self._errors = {} if data: self.perform_validation(data) return TestPreference(data['name'], data['value'], data.get('user')) patcher = patch( 'madprops.serializers.ModelSerializer.from_native', new=from_native) self.patched_from_native = patcher.start() self.addCleanup(patcher.stop) # get_fields inspects the model's _meta, deeply patcher = patch( 'madprops.serializers.ModelSerializer.get_fields', new=lambda self: {}) self.patched_get_fields = patcher.start() self.addCleanup(patcher.stop)
<commit_before>from django.conf import settings from mock import Mock, patch from unittest2 import TestCase settings.configure() # Need to import this after configure() from django.db.models import ForeignKey class TestPreference(object): _meta = Mock(fields=[ForeignKey('user', name='user')]) objects = Mock() def __init__(self, name, value, user=None): self.name = name self.value = value self.user = user def __eq__(self, other): return self.__dict__ == other.__dict__ def __repr__(self): return '<{name}:{value}:{user}>'.format(**self.__dict__) def __cmp__(self, other): return cmp(self.name, other.name) class TestUser(object): @property def preferences(self): return Mock(all=Mock(return_value=self._preferences)) @preferences.setter def preferences(self, value): self._preferences = [ TestPreference(k, v) for k, v in value.iteritems()] class SerializerTestCase(TestCase): def patch_from_native(self): patcher = patch( 'madprops.serializers.ModelSerializer.from_native', new=lambda self, data, files: TestPreference( data['name'], data['value'], data.get('user')) ) self.patched_from_native = patcher.start() self.addCleanup(patcher.stop) <commit_msg>Extend mocking to run validation<commit_after>from django.conf import settings from mock import Mock, patch from unittest2 import TestCase settings.configure() # Need to import this after configure() from django.db.models import ForeignKey class TestPreference(object): _meta = Mock(fields=[ForeignKey('user', name='user')]) objects = Mock() def __init__(self, name, value, user=None): self.name = name self.value = value self.user = user def __eq__(self, other): return self.__dict__ == other.__dict__ def __repr__(self): return '<{name}:{value}:{user}>'.format(**self.__dict__) def __cmp__(self, other): return cmp(self.name, other.name) class TestUser(object): @property def preferences(self): return Mock(all=Mock(return_value=self._preferences)) @preferences.setter def preferences(self, value): self._preferences = [ TestPreference(k, v) for k, v in value.iteritems()] class SerializerTestCase(TestCase): def patch_from_native(self): def from_native(self, data, files): self._errors = {} if data: self.perform_validation(data) return TestPreference(data['name'], data['value'], data.get('user')) patcher = patch( 'madprops.serializers.ModelSerializer.from_native', new=from_native) self.patched_from_native = patcher.start() self.addCleanup(patcher.stop) # get_fields inspects the model's _meta, deeply patcher = patch( 'madprops.serializers.ModelSerializer.get_fields', new=lambda self: {}) self.patched_get_fields = patcher.start() self.addCleanup(patcher.stop)
6d90adebbcd9085de4f0af503b7e0b55dc9424b2
tests/conftest.py
tests/conftest.py
import numpy import pytest @pytest.fixture(scope="module") def rgen(): return numpy.random.RandomState(seed=52973992)
import numpy import pytest @pytest.fixture(scope="module") def rgen(): return numpy.random.RandomState(seed=3476583865)
Fix test by questionable means :)
Fix test by questionable means :)
Python
bsd-3-clause
dseuss/mpnum,dseuss/mpnum
import numpy import pytest @pytest.fixture(scope="module") def rgen(): return numpy.random.RandomState(seed=52973992) Fix test by questionable means :)
import numpy import pytest @pytest.fixture(scope="module") def rgen(): return numpy.random.RandomState(seed=3476583865)
<commit_before>import numpy import pytest @pytest.fixture(scope="module") def rgen(): return numpy.random.RandomState(seed=52973992) <commit_msg>Fix test by questionable means :)<commit_after>
import numpy import pytest @pytest.fixture(scope="module") def rgen(): return numpy.random.RandomState(seed=3476583865)
import numpy import pytest @pytest.fixture(scope="module") def rgen(): return numpy.random.RandomState(seed=52973992) Fix test by questionable means :)import numpy import pytest @pytest.fixture(scope="module") def rgen(): return numpy.random.RandomState(seed=3476583865)
<commit_before>import numpy import pytest @pytest.fixture(scope="module") def rgen(): return numpy.random.RandomState(seed=52973992) <commit_msg>Fix test by questionable means :)<commit_after>import numpy import pytest @pytest.fixture(scope="module") def rgen(): return numpy.random.RandomState(seed=3476583865)
817976878b584086bedc296e5fd6d264006c8dcd
tests/conftest.py
tests/conftest.py
from __future__ import absolute_import from __future__ import unicode_literals import os import subprocess import mock import pytest @pytest.yield_fixture def in_tmpdir(tmpdir): pwd = os.getcwd() os.chdir(tmpdir.strpath) try: yield finally: os.chdir(pwd) @pytest.yield_fixture def check_call_mock(): with mock.patch.object(subprocess, 'check_call') as mocked: yield mocked
from __future__ import absolute_import from __future__ import unicode_literals import os import subprocess import mock import pytest @pytest.fixture def in_tmpdir(tmpdir): pwd = os.getcwd() os.chdir(tmpdir.strpath) try: yield finally: os.chdir(pwd) @pytest.fixture def check_call_mock(): with mock.patch.object(subprocess, 'check_call') as mocked: yield mocked
Replace deprecated yield_fixture with fixture
Replace deprecated yield_fixture with fixture Committed via https://github.com/asottile/all-repos
Python
mit
asottile/css-explore
from __future__ import absolute_import from __future__ import unicode_literals import os import subprocess import mock import pytest @pytest.yield_fixture def in_tmpdir(tmpdir): pwd = os.getcwd() os.chdir(tmpdir.strpath) try: yield finally: os.chdir(pwd) @pytest.yield_fixture def check_call_mock(): with mock.patch.object(subprocess, 'check_call') as mocked: yield mocked Replace deprecated yield_fixture with fixture Committed via https://github.com/asottile/all-repos
from __future__ import absolute_import from __future__ import unicode_literals import os import subprocess import mock import pytest @pytest.fixture def in_tmpdir(tmpdir): pwd = os.getcwd() os.chdir(tmpdir.strpath) try: yield finally: os.chdir(pwd) @pytest.fixture def check_call_mock(): with mock.patch.object(subprocess, 'check_call') as mocked: yield mocked
<commit_before>from __future__ import absolute_import from __future__ import unicode_literals import os import subprocess import mock import pytest @pytest.yield_fixture def in_tmpdir(tmpdir): pwd = os.getcwd() os.chdir(tmpdir.strpath) try: yield finally: os.chdir(pwd) @pytest.yield_fixture def check_call_mock(): with mock.patch.object(subprocess, 'check_call') as mocked: yield mocked <commit_msg>Replace deprecated yield_fixture with fixture Committed via https://github.com/asottile/all-repos<commit_after>
from __future__ import absolute_import from __future__ import unicode_literals import os import subprocess import mock import pytest @pytest.fixture def in_tmpdir(tmpdir): pwd = os.getcwd() os.chdir(tmpdir.strpath) try: yield finally: os.chdir(pwd) @pytest.fixture def check_call_mock(): with mock.patch.object(subprocess, 'check_call') as mocked: yield mocked
from __future__ import absolute_import from __future__ import unicode_literals import os import subprocess import mock import pytest @pytest.yield_fixture def in_tmpdir(tmpdir): pwd = os.getcwd() os.chdir(tmpdir.strpath) try: yield finally: os.chdir(pwd) @pytest.yield_fixture def check_call_mock(): with mock.patch.object(subprocess, 'check_call') as mocked: yield mocked Replace deprecated yield_fixture with fixture Committed via https://github.com/asottile/all-reposfrom __future__ import absolute_import from __future__ import unicode_literals import os import subprocess import mock import pytest @pytest.fixture def in_tmpdir(tmpdir): pwd = os.getcwd() os.chdir(tmpdir.strpath) try: yield finally: os.chdir(pwd) @pytest.fixture def check_call_mock(): with mock.patch.object(subprocess, 'check_call') as mocked: yield mocked
<commit_before>from __future__ import absolute_import from __future__ import unicode_literals import os import subprocess import mock import pytest @pytest.yield_fixture def in_tmpdir(tmpdir): pwd = os.getcwd() os.chdir(tmpdir.strpath) try: yield finally: os.chdir(pwd) @pytest.yield_fixture def check_call_mock(): with mock.patch.object(subprocess, 'check_call') as mocked: yield mocked <commit_msg>Replace deprecated yield_fixture with fixture Committed via https://github.com/asottile/all-repos<commit_after>from __future__ import absolute_import from __future__ import unicode_literals import os import subprocess import mock import pytest @pytest.fixture def in_tmpdir(tmpdir): pwd = os.getcwd() os.chdir(tmpdir.strpath) try: yield finally: os.chdir(pwd) @pytest.fixture def check_call_mock(): with mock.patch.object(subprocess, 'check_call') as mocked: yield mocked
b778c0192cabc652fc06daf99f6b890b3297f83f
Lib/test/test_sqlite.py
Lib/test/test_sqlite.py
from test.support import run_unittest, import_module, verbose # Skip test if _sqlite3 module not installed import_module('_sqlite3') import sqlite3 from sqlite3.test import (dbapi, types, userfunctions, factory, transactions, hooks, regression, dump) def test_main(): if verbose: print("test_sqlite: testing with version", "{!r}, sqlite_version {!r}".format(sqlite3.version, sqlite3.sqlite_version)) run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(), factory.suite(), transactions.suite(), hooks.suite(), regression.suite(), dump.suite()) if __name__ == "__main__": test_main()
import test.support # Skip test if _sqlite3 module not installed test.support.import_module('_sqlite3') import sqlite3 from sqlite3.test import (dbapi, types, userfunctions, factory, transactions, hooks, regression, dump) def test_main(): if test.support.verbose: print("test_sqlite: testing with version", "{!r}, sqlite_version {!r}".format(sqlite3.version, sqlite3.sqlite_version)) test.support.run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(), factory.suite(), transactions.suite(), hooks.suite(), regression.suite(), dump.suite()) if __name__ == "__main__": test_main()
Make the printing of sqlite version in verbose mode work with regrtest -w.
Make the printing of sqlite version in verbose mode work with regrtest -w.
Python
mit
sk-/python2.7-type-annotator,sk-/python2.7-type-annotator,sk-/python2.7-type-annotator
from test.support import run_unittest, import_module, verbose # Skip test if _sqlite3 module not installed import_module('_sqlite3') import sqlite3 from sqlite3.test import (dbapi, types, userfunctions, factory, transactions, hooks, regression, dump) def test_main(): if verbose: print("test_sqlite: testing with version", "{!r}, sqlite_version {!r}".format(sqlite3.version, sqlite3.sqlite_version)) run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(), factory.suite(), transactions.suite(), hooks.suite(), regression.suite(), dump.suite()) if __name__ == "__main__": test_main() Make the printing of sqlite version in verbose mode work with regrtest -w.
import test.support # Skip test if _sqlite3 module not installed test.support.import_module('_sqlite3') import sqlite3 from sqlite3.test import (dbapi, types, userfunctions, factory, transactions, hooks, regression, dump) def test_main(): if test.support.verbose: print("test_sqlite: testing with version", "{!r}, sqlite_version {!r}".format(sqlite3.version, sqlite3.sqlite_version)) test.support.run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(), factory.suite(), transactions.suite(), hooks.suite(), regression.suite(), dump.suite()) if __name__ == "__main__": test_main()
<commit_before>from test.support import run_unittest, import_module, verbose # Skip test if _sqlite3 module not installed import_module('_sqlite3') import sqlite3 from sqlite3.test import (dbapi, types, userfunctions, factory, transactions, hooks, regression, dump) def test_main(): if verbose: print("test_sqlite: testing with version", "{!r}, sqlite_version {!r}".format(sqlite3.version, sqlite3.sqlite_version)) run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(), factory.suite(), transactions.suite(), hooks.suite(), regression.suite(), dump.suite()) if __name__ == "__main__": test_main() <commit_msg>Make the printing of sqlite version in verbose mode work with regrtest -w.<commit_after>
import test.support # Skip test if _sqlite3 module not installed test.support.import_module('_sqlite3') import sqlite3 from sqlite3.test import (dbapi, types, userfunctions, factory, transactions, hooks, regression, dump) def test_main(): if test.support.verbose: print("test_sqlite: testing with version", "{!r}, sqlite_version {!r}".format(sqlite3.version, sqlite3.sqlite_version)) test.support.run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(), factory.suite(), transactions.suite(), hooks.suite(), regression.suite(), dump.suite()) if __name__ == "__main__": test_main()
from test.support import run_unittest, import_module, verbose # Skip test if _sqlite3 module not installed import_module('_sqlite3') import sqlite3 from sqlite3.test import (dbapi, types, userfunctions, factory, transactions, hooks, regression, dump) def test_main(): if verbose: print("test_sqlite: testing with version", "{!r}, sqlite_version {!r}".format(sqlite3.version, sqlite3.sqlite_version)) run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(), factory.suite(), transactions.suite(), hooks.suite(), regression.suite(), dump.suite()) if __name__ == "__main__": test_main() Make the printing of sqlite version in verbose mode work with regrtest -w.import test.support # Skip test if _sqlite3 module not installed test.support.import_module('_sqlite3') import sqlite3 from sqlite3.test import (dbapi, types, userfunctions, factory, transactions, hooks, regression, dump) def test_main(): if test.support.verbose: print("test_sqlite: testing with version", "{!r}, sqlite_version {!r}".format(sqlite3.version, sqlite3.sqlite_version)) test.support.run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(), factory.suite(), transactions.suite(), hooks.suite(), regression.suite(), dump.suite()) if __name__ == "__main__": test_main()
<commit_before>from test.support import run_unittest, import_module, verbose # Skip test if _sqlite3 module not installed import_module('_sqlite3') import sqlite3 from sqlite3.test import (dbapi, types, userfunctions, factory, transactions, hooks, regression, dump) def test_main(): if verbose: print("test_sqlite: testing with version", "{!r}, sqlite_version {!r}".format(sqlite3.version, sqlite3.sqlite_version)) run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(), factory.suite(), transactions.suite(), hooks.suite(), regression.suite(), dump.suite()) if __name__ == "__main__": test_main() <commit_msg>Make the printing of sqlite version in verbose mode work with regrtest -w.<commit_after>import test.support # Skip test if _sqlite3 module not installed test.support.import_module('_sqlite3') import sqlite3 from sqlite3.test import (dbapi, types, userfunctions, factory, transactions, hooks, regression, dump) def test_main(): if test.support.verbose: print("test_sqlite: testing with version", "{!r}, sqlite_version {!r}".format(sqlite3.version, sqlite3.sqlite_version)) test.support.run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(), factory.suite(), transactions.suite(), hooks.suite(), regression.suite(), dump.suite()) if __name__ == "__main__": test_main()
d8e9201c627840c72a540a77425ec0c13ac48a22
tests/test_cmd.py
tests/test_cmd.py
import unittest from click.testing import CliRunner from scuevals_api.cmd import cli class CmdsTestCase(unittest.TestCase): @classmethod def setUpClass(cls): cls.runner = CliRunner() def cli_run(self, *cmds): return self.runner.invoke(cli, cmds) cls.cli_run = cli_run def test_initdb(self): result = self.cli_run('initdb') self.assertEqual(0, result.exit_code)
import unittest from click.testing import CliRunner from scuevals_api.cmd import cli class CmdsTestCase(unittest.TestCase): @classmethod def setUpClass(cls): cls.runner = CliRunner() def cli_run(self, *cmds): return self.runner.invoke(cli, cmds) cls.cli_run = cli_run def test_initdb(self): result = self.cli_run('initdb') self.assertEqual(0, result.exit_code, msg=result.output)
Add detailed error for CLI test failure
Add detailed error for CLI test failure
Python
agpl-3.0
SCUEvals/scuevals-api,SCUEvals/scuevals-api
import unittest from click.testing import CliRunner from scuevals_api.cmd import cli class CmdsTestCase(unittest.TestCase): @classmethod def setUpClass(cls): cls.runner = CliRunner() def cli_run(self, *cmds): return self.runner.invoke(cli, cmds) cls.cli_run = cli_run def test_initdb(self): result = self.cli_run('initdb') self.assertEqual(0, result.exit_code) Add detailed error for CLI test failure
import unittest from click.testing import CliRunner from scuevals_api.cmd import cli class CmdsTestCase(unittest.TestCase): @classmethod def setUpClass(cls): cls.runner = CliRunner() def cli_run(self, *cmds): return self.runner.invoke(cli, cmds) cls.cli_run = cli_run def test_initdb(self): result = self.cli_run('initdb') self.assertEqual(0, result.exit_code, msg=result.output)
<commit_before>import unittest from click.testing import CliRunner from scuevals_api.cmd import cli class CmdsTestCase(unittest.TestCase): @classmethod def setUpClass(cls): cls.runner = CliRunner() def cli_run(self, *cmds): return self.runner.invoke(cli, cmds) cls.cli_run = cli_run def test_initdb(self): result = self.cli_run('initdb') self.assertEqual(0, result.exit_code) <commit_msg>Add detailed error for CLI test failure<commit_after>
import unittest from click.testing import CliRunner from scuevals_api.cmd import cli class CmdsTestCase(unittest.TestCase): @classmethod def setUpClass(cls): cls.runner = CliRunner() def cli_run(self, *cmds): return self.runner.invoke(cli, cmds) cls.cli_run = cli_run def test_initdb(self): result = self.cli_run('initdb') self.assertEqual(0, result.exit_code, msg=result.output)
import unittest from click.testing import CliRunner from scuevals_api.cmd import cli class CmdsTestCase(unittest.TestCase): @classmethod def setUpClass(cls): cls.runner = CliRunner() def cli_run(self, *cmds): return self.runner.invoke(cli, cmds) cls.cli_run = cli_run def test_initdb(self): result = self.cli_run('initdb') self.assertEqual(0, result.exit_code) Add detailed error for CLI test failureimport unittest from click.testing import CliRunner from scuevals_api.cmd import cli class CmdsTestCase(unittest.TestCase): @classmethod def setUpClass(cls): cls.runner = CliRunner() def cli_run(self, *cmds): return self.runner.invoke(cli, cmds) cls.cli_run = cli_run def test_initdb(self): result = self.cli_run('initdb') self.assertEqual(0, result.exit_code, msg=result.output)
<commit_before>import unittest from click.testing import CliRunner from scuevals_api.cmd import cli class CmdsTestCase(unittest.TestCase): @classmethod def setUpClass(cls): cls.runner = CliRunner() def cli_run(self, *cmds): return self.runner.invoke(cli, cmds) cls.cli_run = cli_run def test_initdb(self): result = self.cli_run('initdb') self.assertEqual(0, result.exit_code) <commit_msg>Add detailed error for CLI test failure<commit_after>import unittest from click.testing import CliRunner from scuevals_api.cmd import cli class CmdsTestCase(unittest.TestCase): @classmethod def setUpClass(cls): cls.runner = CliRunner() def cli_run(self, *cmds): return self.runner.invoke(cli, cmds) cls.cli_run = cli_run def test_initdb(self): result = self.cli_run('initdb') self.assertEqual(0, result.exit_code, msg=result.output)
e7d42847284ae73befad8bdf2fa035a6f95a82bb
tests/test_dow.py
tests/test_dow.py
from datetime import datetime import pycron def test_parser(): now = datetime(2015, 6, 18, 16, 7) assert pycron.is_now('* * * * *', now) assert pycron.is_now('* * * * 4', now) assert pycron.is_now('* * * * */4', now) assert pycron.is_now('* * * * 0,3,4', now) assert pycron.is_now('* * * * 3', now) is False assert pycron.is_now('* * * * */3', now) is False assert pycron.is_now('* * * * 0,3,6', now) is False assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday' assert pycron.DOW_CHOICES[0][1] == 'Sunday' now = datetime(2015, 6, 21, 16, 7) assert pycron.is_now('* * * * 0', now)
from datetime import datetime, timedelta import pycron def test_parser(): now = datetime(2015, 6, 18, 16, 7) assert pycron.is_now('* * * * *', now) assert pycron.is_now('* * * * 4', now) assert pycron.is_now('* * * * */4', now) assert pycron.is_now('* * * * 0,3,4', now) assert pycron.is_now('* * * * 3', now) is False assert pycron.is_now('* * * * */3', now) is False assert pycron.is_now('* * * * 0,3,6', now) is False assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday' assert pycron.DOW_CHOICES[0][1] == 'Sunday' now = datetime(2015, 6, 20, 16, 7) for i in range(0, 7): # Test day matching from Sunday onwards... now += timedelta(days=1) assert pycron.is_now('* * * * %i' % (i), now) # Test weekdays assert pycron.is_now('* * * * 1,2,3,4,5', now) is (True if i not in [0, 6] else False) # Test weekends assert pycron.is_now('* * * * 0,6', now) is (True if i in [0, 6] else False)
Add more thorough testing of day of week.
Add more thorough testing of day of week.
Python
mit
kipe/pycron
from datetime import datetime import pycron def test_parser(): now = datetime(2015, 6, 18, 16, 7) assert pycron.is_now('* * * * *', now) assert pycron.is_now('* * * * 4', now) assert pycron.is_now('* * * * */4', now) assert pycron.is_now('* * * * 0,3,4', now) assert pycron.is_now('* * * * 3', now) is False assert pycron.is_now('* * * * */3', now) is False assert pycron.is_now('* * * * 0,3,6', now) is False assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday' assert pycron.DOW_CHOICES[0][1] == 'Sunday' now = datetime(2015, 6, 21, 16, 7) assert pycron.is_now('* * * * 0', now) Add more thorough testing of day of week.
from datetime import datetime, timedelta import pycron def test_parser(): now = datetime(2015, 6, 18, 16, 7) assert pycron.is_now('* * * * *', now) assert pycron.is_now('* * * * 4', now) assert pycron.is_now('* * * * */4', now) assert pycron.is_now('* * * * 0,3,4', now) assert pycron.is_now('* * * * 3', now) is False assert pycron.is_now('* * * * */3', now) is False assert pycron.is_now('* * * * 0,3,6', now) is False assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday' assert pycron.DOW_CHOICES[0][1] == 'Sunday' now = datetime(2015, 6, 20, 16, 7) for i in range(0, 7): # Test day matching from Sunday onwards... now += timedelta(days=1) assert pycron.is_now('* * * * %i' % (i), now) # Test weekdays assert pycron.is_now('* * * * 1,2,3,4,5', now) is (True if i not in [0, 6] else False) # Test weekends assert pycron.is_now('* * * * 0,6', now) is (True if i in [0, 6] else False)
<commit_before>from datetime import datetime import pycron def test_parser(): now = datetime(2015, 6, 18, 16, 7) assert pycron.is_now('* * * * *', now) assert pycron.is_now('* * * * 4', now) assert pycron.is_now('* * * * */4', now) assert pycron.is_now('* * * * 0,3,4', now) assert pycron.is_now('* * * * 3', now) is False assert pycron.is_now('* * * * */3', now) is False assert pycron.is_now('* * * * 0,3,6', now) is False assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday' assert pycron.DOW_CHOICES[0][1] == 'Sunday' now = datetime(2015, 6, 21, 16, 7) assert pycron.is_now('* * * * 0', now) <commit_msg>Add more thorough testing of day of week.<commit_after>
from datetime import datetime, timedelta import pycron def test_parser(): now = datetime(2015, 6, 18, 16, 7) assert pycron.is_now('* * * * *', now) assert pycron.is_now('* * * * 4', now) assert pycron.is_now('* * * * */4', now) assert pycron.is_now('* * * * 0,3,4', now) assert pycron.is_now('* * * * 3', now) is False assert pycron.is_now('* * * * */3', now) is False assert pycron.is_now('* * * * 0,3,6', now) is False assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday' assert pycron.DOW_CHOICES[0][1] == 'Sunday' now = datetime(2015, 6, 20, 16, 7) for i in range(0, 7): # Test day matching from Sunday onwards... now += timedelta(days=1) assert pycron.is_now('* * * * %i' % (i), now) # Test weekdays assert pycron.is_now('* * * * 1,2,3,4,5', now) is (True if i not in [0, 6] else False) # Test weekends assert pycron.is_now('* * * * 0,6', now) is (True if i in [0, 6] else False)
from datetime import datetime import pycron def test_parser(): now = datetime(2015, 6, 18, 16, 7) assert pycron.is_now('* * * * *', now) assert pycron.is_now('* * * * 4', now) assert pycron.is_now('* * * * */4', now) assert pycron.is_now('* * * * 0,3,4', now) assert pycron.is_now('* * * * 3', now) is False assert pycron.is_now('* * * * */3', now) is False assert pycron.is_now('* * * * 0,3,6', now) is False assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday' assert pycron.DOW_CHOICES[0][1] == 'Sunday' now = datetime(2015, 6, 21, 16, 7) assert pycron.is_now('* * * * 0', now) Add more thorough testing of day of week.from datetime import datetime, timedelta import pycron def test_parser(): now = datetime(2015, 6, 18, 16, 7) assert pycron.is_now('* * * * *', now) assert pycron.is_now('* * * * 4', now) assert pycron.is_now('* * * * */4', now) assert pycron.is_now('* * * * 0,3,4', now) assert pycron.is_now('* * * * 3', now) is False assert pycron.is_now('* * * * */3', now) is False assert pycron.is_now('* * * * 0,3,6', now) is False assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday' assert pycron.DOW_CHOICES[0][1] == 'Sunday' now = datetime(2015, 6, 20, 16, 7) for i in range(0, 7): # Test day matching from Sunday onwards... now += timedelta(days=1) assert pycron.is_now('* * * * %i' % (i), now) # Test weekdays assert pycron.is_now('* * * * 1,2,3,4,5', now) is (True if i not in [0, 6] else False) # Test weekends assert pycron.is_now('* * * * 0,6', now) is (True if i in [0, 6] else False)
<commit_before>from datetime import datetime import pycron def test_parser(): now = datetime(2015, 6, 18, 16, 7) assert pycron.is_now('* * * * *', now) assert pycron.is_now('* * * * 4', now) assert pycron.is_now('* * * * */4', now) assert pycron.is_now('* * * * 0,3,4', now) assert pycron.is_now('* * * * 3', now) is False assert pycron.is_now('* * * * */3', now) is False assert pycron.is_now('* * * * 0,3,6', now) is False assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday' assert pycron.DOW_CHOICES[0][1] == 'Sunday' now = datetime(2015, 6, 21, 16, 7) assert pycron.is_now('* * * * 0', now) <commit_msg>Add more thorough testing of day of week.<commit_after>from datetime import datetime, timedelta import pycron def test_parser(): now = datetime(2015, 6, 18, 16, 7) assert pycron.is_now('* * * * *', now) assert pycron.is_now('* * * * 4', now) assert pycron.is_now('* * * * */4', now) assert pycron.is_now('* * * * 0,3,4', now) assert pycron.is_now('* * * * 3', now) is False assert pycron.is_now('* * * * */3', now) is False assert pycron.is_now('* * * * 0,3,6', now) is False assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday' assert pycron.DOW_CHOICES[0][1] == 'Sunday' now = datetime(2015, 6, 20, 16, 7) for i in range(0, 7): # Test day matching from Sunday onwards... now += timedelta(days=1) assert pycron.is_now('* * * * %i' % (i), now) # Test weekdays assert pycron.is_now('* * * * 1,2,3,4,5', now) is (True if i not in [0, 6] else False) # Test weekends assert pycron.is_now('* * * * 0,6', now) is (True if i in [0, 6] else False)
a89e7f9f625427d558300eb5e5cbc2881cdcc207
get_a_job/__init__.py
get_a_job/__init__.py
from flask import Flask from flask.ext.restful import Api from .models import db from .api import configure_api def create_app(object_name): app = Flask(object_name) app.config.from_object(object_name) db.init_app(app) configure_api(app) return app
from flask import Flask from flask.ext.restful import Api from .models import db from .api import configure_api def create_app(object_name, **kwargs): app = Flask(object_name) app.config.from_object(object_name) app.config.update(kwargs) db.init_app(app) configure_api(app) return app
Add optional configuration customization of app.
Add optional configuration customization of app.
Python
mit
smoynes/get_a_job
from flask import Flask from flask.ext.restful import Api from .models import db from .api import configure_api def create_app(object_name): app = Flask(object_name) app.config.from_object(object_name) db.init_app(app) configure_api(app) return app Add optional configuration customization of app.
from flask import Flask from flask.ext.restful import Api from .models import db from .api import configure_api def create_app(object_name, **kwargs): app = Flask(object_name) app.config.from_object(object_name) app.config.update(kwargs) db.init_app(app) configure_api(app) return app
<commit_before>from flask import Flask from flask.ext.restful import Api from .models import db from .api import configure_api def create_app(object_name): app = Flask(object_name) app.config.from_object(object_name) db.init_app(app) configure_api(app) return app <commit_msg>Add optional configuration customization of app.<commit_after>
from flask import Flask from flask.ext.restful import Api from .models import db from .api import configure_api def create_app(object_name, **kwargs): app = Flask(object_name) app.config.from_object(object_name) app.config.update(kwargs) db.init_app(app) configure_api(app) return app
from flask import Flask from flask.ext.restful import Api from .models import db from .api import configure_api def create_app(object_name): app = Flask(object_name) app.config.from_object(object_name) db.init_app(app) configure_api(app) return app Add optional configuration customization of app.from flask import Flask from flask.ext.restful import Api from .models import db from .api import configure_api def create_app(object_name, **kwargs): app = Flask(object_name) app.config.from_object(object_name) app.config.update(kwargs) db.init_app(app) configure_api(app) return app
<commit_before>from flask import Flask from flask.ext.restful import Api from .models import db from .api import configure_api def create_app(object_name): app = Flask(object_name) app.config.from_object(object_name) db.init_app(app) configure_api(app) return app <commit_msg>Add optional configuration customization of app.<commit_after>from flask import Flask from flask.ext.restful import Api from .models import db from .api import configure_api def create_app(object_name, **kwargs): app = Flask(object_name) app.config.from_object(object_name) app.config.update(kwargs) db.init_app(app) configure_api(app) return app
d8cde079d6e8dd0dcd5a13a36a0bca9685ba7b1c
api/BucketListAPI.py
api/BucketListAPI.py
from flask import Flask, jsonify from modals.modals import User, Bucket, Item from api.__init__ import create_app, db app = create_app('DevelopmentEnv') @app.errorhandler(404) def page_not_found(e): response = jsonify({'error': 'The request can not be completed'}) response.status_code = 404 return response if __name__ == '__main__': app.run()
from flask import Flask, jsonify from modals.modals import User, Bucket, Item from api.__init__ import create_app, db app = create_app('DevelopmentEnv') @app.errorhandler(404) def page_not_found(e): response = jsonify({'error': 'The request can not be completed'}) response.status_code = 404 return response @app.errorhandler(401) def invalid_token(e): response = jsonify({'error': 'Invalid Token'}) response.status_code = 401 return response if __name__ == '__main__': app.run()
Add error handler for invalid token
Add error handler for invalid token
Python
mit
patlub/BucketListAPI,patlub/BucketListAPI
from flask import Flask, jsonify from modals.modals import User, Bucket, Item from api.__init__ import create_app, db app = create_app('DevelopmentEnv') @app.errorhandler(404) def page_not_found(e): response = jsonify({'error': 'The request can not be completed'}) response.status_code = 404 return response if __name__ == '__main__': app.run() Add error handler for invalid token
from flask import Flask, jsonify from modals.modals import User, Bucket, Item from api.__init__ import create_app, db app = create_app('DevelopmentEnv') @app.errorhandler(404) def page_not_found(e): response = jsonify({'error': 'The request can not be completed'}) response.status_code = 404 return response @app.errorhandler(401) def invalid_token(e): response = jsonify({'error': 'Invalid Token'}) response.status_code = 401 return response if __name__ == '__main__': app.run()
<commit_before>from flask import Flask, jsonify from modals.modals import User, Bucket, Item from api.__init__ import create_app, db app = create_app('DevelopmentEnv') @app.errorhandler(404) def page_not_found(e): response = jsonify({'error': 'The request can not be completed'}) response.status_code = 404 return response if __name__ == '__main__': app.run() <commit_msg>Add error handler for invalid token<commit_after>
from flask import Flask, jsonify from modals.modals import User, Bucket, Item from api.__init__ import create_app, db app = create_app('DevelopmentEnv') @app.errorhandler(404) def page_not_found(e): response = jsonify({'error': 'The request can not be completed'}) response.status_code = 404 return response @app.errorhandler(401) def invalid_token(e): response = jsonify({'error': 'Invalid Token'}) response.status_code = 401 return response if __name__ == '__main__': app.run()
from flask import Flask, jsonify from modals.modals import User, Bucket, Item from api.__init__ import create_app, db app = create_app('DevelopmentEnv') @app.errorhandler(404) def page_not_found(e): response = jsonify({'error': 'The request can not be completed'}) response.status_code = 404 return response if __name__ == '__main__': app.run() Add error handler for invalid tokenfrom flask import Flask, jsonify from modals.modals import User, Bucket, Item from api.__init__ import create_app, db app = create_app('DevelopmentEnv') @app.errorhandler(404) def page_not_found(e): response = jsonify({'error': 'The request can not be completed'}) response.status_code = 404 return response @app.errorhandler(401) def invalid_token(e): response = jsonify({'error': 'Invalid Token'}) response.status_code = 401 return response if __name__ == '__main__': app.run()
<commit_before>from flask import Flask, jsonify from modals.modals import User, Bucket, Item from api.__init__ import create_app, db app = create_app('DevelopmentEnv') @app.errorhandler(404) def page_not_found(e): response = jsonify({'error': 'The request can not be completed'}) response.status_code = 404 return response if __name__ == '__main__': app.run() <commit_msg>Add error handler for invalid token<commit_after>from flask import Flask, jsonify from modals.modals import User, Bucket, Item from api.__init__ import create_app, db app = create_app('DevelopmentEnv') @app.errorhandler(404) def page_not_found(e): response = jsonify({'error': 'The request can not be completed'}) response.status_code = 404 return response @app.errorhandler(401) def invalid_token(e): response = jsonify({'error': 'Invalid Token'}) response.status_code = 401 return response if __name__ == '__main__': app.run()
45254f3a7401b4b63d829f38c426c0635485f1e0
PRESUBMIT.py
PRESUBMIT.py
# Copyright (c) 2011 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Top-level presubmit script for GYP. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into gcl. """ def CheckChangeOnUpload(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) return report def CheckChangeOnCommit(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) report.extend(input_api.canned_checks.CheckTreeIsOpen( input_api, output_api, 'http://gyp-status.appspot.com/status', 'http://gyp-status.appspot.com/current')) return report def GetPreferredTrySlaves(): return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac']
# Copyright (c) 2011 Google Inc. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Top-level presubmit script for GYP. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into gcl. """ def CheckChangeOnUpload(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) return report def CheckChangeOnCommit(input_api, output_api): report = [] license = ( r'.*? Copyright \(c\) %(year)s Google Inc\. All rights reserved\.\n' r'.*? Use of this source code is governed by a BSD-style license that ' r'can be\n' r'.*? found in the LICENSE file\.\n' ) % { 'year': input_api.time.strftime('%Y'), } report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api, license_header=license)) report.extend(input_api.canned_checks.CheckTreeIsOpen( input_api, output_api, 'http://gyp-status.appspot.com/status', 'http://gyp-status.appspot.com/current')) return report def GetPreferredTrySlaves(): return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac']
Fix the license header regex.
Fix the license header regex. Most of the files are attributed to Google Inc so I used this instead of Chromium Authors. R=mark@chromium.org BUG= TEST= Review URL: http://codereview.chromium.org/7108074 git-svn-id: e7e1075985beda50ea81ac4472467b4f6e91fc78@936 78cadc50-ecff-11dd-a971-7dbc132099af
Python
bsd-3-clause
enkripsi/gyp,mistydemeo/gyp,cysp/gyp,IllusionRom-deprecated/android_platform_external_chromium_org_tools_gyp,LazyCodingCat/gyp,IllusionRom-deprecated/android_platform_external_chromium_org_tools_gyp,ttyangf/gyp,saghul/gyn,ttyangf/pdfium_gyp,lukeweber/gyp-override,carlTLR/gyp,amoikevin/gyp,erikge/watch_gyp,LazyCodingCat/gyp,omasanori/gyp,saghul/gyn,Omegaphora/external_chromium_org_tools_gyp,channing/gyp,yinquan529/platform-external-chromium_org-tools-gyp,Jack-Q/GYP-copy,duanhjlt/gyp,bnq4ever/gypgoogle,Danath/gyp,yangrongwei/gyp,Chilledheart/gyp,enkripsi/gyp,chromium/gyp,Chilledheart/gyp,bnoordhuis/gyp,svn2github/gyp,msc-/gyp,sanyaade-teachings/gyp,okumura/gyp,carlTLR/gyp,alexcrichton/gyp,trafi/gyp,brson/gyp,sanyaade-teachings/gyp,brson/gyp,sport-monkey/GYP,sanyaade-teachings/gyp,bnoordhuis/gyp,okumura/gyp,bulldy80/gyp_unofficial,saghul/gyn,mgamer/gyp,AWhetter/gyp,cchamberlain/gyp,bulldy80/gyp_unofficial,clar/gyp,lukeweber/gyp-override,adblockplus/gyp,bdarnell/gyp,cysp/gyp,Phuehvk/gyp,svn2github/gyp,mgamer/gyp,trafi/gyp,mistydemeo/gyp,amoikevin/gyp,sloanyang/gyp,azunite/gyp_20150930,bdarnell/gyp,turbulenz/gyp,mkrautz/gyp-libmumble,sdklite/gyp,mapbox/gyp,mgamer/gyp,pandaxcl/gyp,bpsinc-native/src_tools_gyp,sanyaade-teachings/gyp,msc-/gyp,ryfx/gyp,openpeer/webrtc-gyp,msc-/gyp,enkripsi/gyp,MIPS/external-chromium_org-tools-gyp,yinquan529/platform-external-chromium_org-tools-gyp,turbulenz/gyp,kevinchen3315/gyp-git,Jack-Q/GYP-copy,cysp/gyp,LazyCodingCat/gyp,chromium/gyp,svn2github/kgyp,ttyangf/gyp,cysp/gyp,azunite/gyp_20150930,lianliuwei/gyp,dougbeal/gyp,cchamberlain/gyp,adblockplus/gyp,mkrautz/gyp-libmumble,bnoordhuis/gyp,svn2github/gyp,Omegaphora/external_chromium_org_tools_gyp,mapbox/gyp,pyokagan/gyp,pyokagan/gyp,bulldy80/gyp_unofficial,bnoordhuis/gyp,sloanyang/gyp,mkrautz/gyp-libmumble,saghul/gyn,ttyangf/pdfium_gyp,alexcrichton/gyp,bdarnell/gyp,enkripsi/gyp,springmeyer/gyp,mapbox/gyp,trafi/gyp,svn2github/kgyp,luvit/gyp,duanhjlt/gyp,geekboxzone/lollipop_external_chromium_org_tools_gyp,amoikevin/gyp,lianliuwei/gyp,pandaxcl/gyp,clar/gyp,okwasi/gyp,carlTLR/gyp,turbulenz/gyp,AWhetter/gyp,ryfx/gyp,lukeweber/gyp-override,alexcrichton/gyp,lukeweber/gyp-override,sport-monkey/GYP,trafi/gyp,duanhjlt/gyp,adblockplus/gyp,chromium/gyp,brson/gyp,dougbeal/gyp,azunite/gyp_20150930,omasanori/gyp,sdklite/gyp,amoikevin/gyp,erikge/watch_gyp,svn2github/kgyp,pyokagan/gyp,amoikevin/gyp,AOSPU/external_chromium_org_tools_gyp,LazyCodingCat/gyp,android-ia/platform_external_chromium_org_tools_gyp,Danath/gyp,mumble-voip/libmumble-gyp,Jack-Q/GYP-copy,luvit/gyp,azunite/gyp_20150930,turbulenz/gyp,ryfx/gyp,svn2github/kgyp,bnq4ever/gypgoogle,android-ia/platform_external_chromium_org_tools_gyp,carlTLR/gyp,okwasi/gyp,omasanori/gyp,channing/gyp,enkripsi/gyp,Omegaphora/external_chromium_org_tools_gyp,springmeyer/gyp,yangrongwei/gyp,MIPS/external-chromium_org-tools-gyp,mumble-voip/libmumble-gyp,bulldy80/gyp_unofficial,okumura/gyp,tarc/gyp,luvit/gyp,ttyangf/pdfium_gyp,yinquan529/platform-external-chromium_org-tools-gyp,yjhjstz/gyp,ttyangf/pdfium_gyp,sport-monkey/GYP,geekboxzone/lollipop_external_chromium_org_tools_gyp,mapbox/gyp,sanyaade-teachings/gyp,ttyangf/pdfium_gyp,kevinchen3315/gyp-git,tarc/gyp,saghul/gyn,bnq4ever/gypgoogle,mgamer/gyp,bnq4ever/gypgoogle,channing/gyp,ryfx/gyp,bnoordhuis/gyp,azunite/gyp,trafi/gyp,AWhetter/gyp,erikge/watch_gyp,sloanyang/gyp,erikge/watch_gyp,xin3liang/platform_external_chromium_org_tools_gyp,springmeyer/gyp,omasanori/gyp,azunite/gyp,openpeer/webrtc-gyp,msc-/gyp,yjhjstz/gyp,tarc/gyp,mistydemeo/gyp,Danath/gyp,azunite/gyp,clar/gyp,mapbox/gyp,MIPS/external-chromium_org-tools-gyp,Danath/gyp,Phuehvk/gyp,ttyangf/gyp,mkrautz/gyp-libmumble,okumura/gyp,Omegaphora/external_chromium_org_tools_gyp,svn2github/gyp,sloanyang/gyp,AOSPU/external_chromium_org_tools_gyp,mistydemeo/gyp,dougbeal/gyp,bpsinc-native/src_tools_gyp,pandaxcl/gyp,Phuehvk/gyp,Jack-Q/GYP-copy,azunite/gyp,kevinchen3315/gyp-git,chromium/gyp,Chilledheart/gyp,mgamer/gyp,cchamberlain/gyp,okwasi/gyp,svn2github/kgyp,bnq4ever/gypgoogle,xin3liang/platform_external_chromium_org_tools_gyp,erikge/watch_gyp,azunite/gyp_20150930,luvit/gyp,adblockplus/gyp,alexcrichton/gyp,yinquan529/platform-external-chromium_org-tools-gyp,Phuehvk/gyp,Chilledheart/gyp,dougbeal/gyp,bdarnell/gyp,Chilledheart/gyp,cchamberlain/gyp,sdklite/gyp,pyokagan/gyp,okwasi/gyp,springmeyer/gyp,AWhetter/gyp,clar/gyp,sport-monkey/GYP,android-ia/platform_external_chromium_org_tools_gyp,Danath/gyp,carlTLR/gyp,cchamberlain/gyp,openpeer/webrtc-gyp,xin3liang/platform_external_chromium_org_tools_gyp,yjhjstz/gyp,AOSPU/external_chromium_org_tools_gyp,yangrongwei/gyp,Phuehvk/gyp,pandaxcl/gyp,yjhjstz/gyp,sport-monkey/GYP,bpsinc-native/src_tools_gyp,yangrongwei/gyp,AOSPU/external_chromium_org_tools_gyp,Jack-Q/GYP-copy,kevinchen3315/gyp-git,pyokagan/gyp,ryfx/gyp,brson/gyp,sdklite/gyp,bpsinc-native/src_tools_gyp,turbulenz/gyp,chromium/gyp,mumble-voip/libmumble-gyp,svn2github/gyp,channing/gyp,tarc/gyp,mumble-voip/libmumble-gyp,ttyangf/gyp,openpeer/webrtc-gyp,yjhjstz/gyp,adblockplus/gyp,xin3liang/platform_external_chromium_org_tools_gyp,MIPS/external-chromium_org-tools-gyp,android-ia/platform_external_chromium_org_tools_gyp,AWhetter/gyp,lianliuwei/gyp,bulldy80/gyp_unofficial,tarc/gyp,geekboxzone/lollipop_external_chromium_org_tools_gyp,IllusionRom-deprecated/android_platform_external_chromium_org_tools_gyp,msc-/gyp,cysp/gyp,duanhjlt/gyp,dougbeal/gyp,clar/gyp,geekboxzone/lollipop_external_chromium_org_tools_gyp,openpeer/webrtc-gyp,pandaxcl/gyp,springmeyer/gyp,sdklite/gyp,LazyCodingCat/gyp,ttyangf/gyp,azunite/gyp,duanhjlt/gyp,IllusionRom-deprecated/android_platform_external_chromium_org_tools_gyp,lianliuwei/gyp
# Copyright (c) 2011 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Top-level presubmit script for GYP. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into gcl. """ def CheckChangeOnUpload(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) return report def CheckChangeOnCommit(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) report.extend(input_api.canned_checks.CheckTreeIsOpen( input_api, output_api, 'http://gyp-status.appspot.com/status', 'http://gyp-status.appspot.com/current')) return report def GetPreferredTrySlaves(): return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac'] Fix the license header regex. Most of the files are attributed to Google Inc so I used this instead of Chromium Authors. R=mark@chromium.org BUG= TEST= Review URL: http://codereview.chromium.org/7108074 git-svn-id: e7e1075985beda50ea81ac4472467b4f6e91fc78@936 78cadc50-ecff-11dd-a971-7dbc132099af
# Copyright (c) 2011 Google Inc. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Top-level presubmit script for GYP. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into gcl. """ def CheckChangeOnUpload(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) return report def CheckChangeOnCommit(input_api, output_api): report = [] license = ( r'.*? Copyright \(c\) %(year)s Google Inc\. All rights reserved\.\n' r'.*? Use of this source code is governed by a BSD-style license that ' r'can be\n' r'.*? found in the LICENSE file\.\n' ) % { 'year': input_api.time.strftime('%Y'), } report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api, license_header=license)) report.extend(input_api.canned_checks.CheckTreeIsOpen( input_api, output_api, 'http://gyp-status.appspot.com/status', 'http://gyp-status.appspot.com/current')) return report def GetPreferredTrySlaves(): return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac']
<commit_before># Copyright (c) 2011 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Top-level presubmit script for GYP. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into gcl. """ def CheckChangeOnUpload(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) return report def CheckChangeOnCommit(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) report.extend(input_api.canned_checks.CheckTreeIsOpen( input_api, output_api, 'http://gyp-status.appspot.com/status', 'http://gyp-status.appspot.com/current')) return report def GetPreferredTrySlaves(): return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac'] <commit_msg>Fix the license header regex. Most of the files are attributed to Google Inc so I used this instead of Chromium Authors. R=mark@chromium.org BUG= TEST= Review URL: http://codereview.chromium.org/7108074 git-svn-id: e7e1075985beda50ea81ac4472467b4f6e91fc78@936 78cadc50-ecff-11dd-a971-7dbc132099af<commit_after>
# Copyright (c) 2011 Google Inc. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Top-level presubmit script for GYP. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into gcl. """ def CheckChangeOnUpload(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) return report def CheckChangeOnCommit(input_api, output_api): report = [] license = ( r'.*? Copyright \(c\) %(year)s Google Inc\. All rights reserved\.\n' r'.*? Use of this source code is governed by a BSD-style license that ' r'can be\n' r'.*? found in the LICENSE file\.\n' ) % { 'year': input_api.time.strftime('%Y'), } report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api, license_header=license)) report.extend(input_api.canned_checks.CheckTreeIsOpen( input_api, output_api, 'http://gyp-status.appspot.com/status', 'http://gyp-status.appspot.com/current')) return report def GetPreferredTrySlaves(): return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac']
# Copyright (c) 2011 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Top-level presubmit script for GYP. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into gcl. """ def CheckChangeOnUpload(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) return report def CheckChangeOnCommit(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) report.extend(input_api.canned_checks.CheckTreeIsOpen( input_api, output_api, 'http://gyp-status.appspot.com/status', 'http://gyp-status.appspot.com/current')) return report def GetPreferredTrySlaves(): return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac'] Fix the license header regex. Most of the files are attributed to Google Inc so I used this instead of Chromium Authors. R=mark@chromium.org BUG= TEST= Review URL: http://codereview.chromium.org/7108074 git-svn-id: e7e1075985beda50ea81ac4472467b4f6e91fc78@936 78cadc50-ecff-11dd-a971-7dbc132099af# Copyright (c) 2011 Google Inc. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Top-level presubmit script for GYP. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into gcl. """ def CheckChangeOnUpload(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) return report def CheckChangeOnCommit(input_api, output_api): report = [] license = ( r'.*? Copyright \(c\) %(year)s Google Inc\. All rights reserved\.\n' r'.*? Use of this source code is governed by a BSD-style license that ' r'can be\n' r'.*? found in the LICENSE file\.\n' ) % { 'year': input_api.time.strftime('%Y'), } report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api, license_header=license)) report.extend(input_api.canned_checks.CheckTreeIsOpen( input_api, output_api, 'http://gyp-status.appspot.com/status', 'http://gyp-status.appspot.com/current')) return report def GetPreferredTrySlaves(): return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac']
<commit_before># Copyright (c) 2011 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Top-level presubmit script for GYP. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into gcl. """ def CheckChangeOnUpload(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) return report def CheckChangeOnCommit(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) report.extend(input_api.canned_checks.CheckTreeIsOpen( input_api, output_api, 'http://gyp-status.appspot.com/status', 'http://gyp-status.appspot.com/current')) return report def GetPreferredTrySlaves(): return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac'] <commit_msg>Fix the license header regex. Most of the files are attributed to Google Inc so I used this instead of Chromium Authors. R=mark@chromium.org BUG= TEST= Review URL: http://codereview.chromium.org/7108074 git-svn-id: e7e1075985beda50ea81ac4472467b4f6e91fc78@936 78cadc50-ecff-11dd-a971-7dbc132099af<commit_after># Copyright (c) 2011 Google Inc. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Top-level presubmit script for GYP. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into gcl. """ def CheckChangeOnUpload(input_api, output_api): report = [] report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api)) return report def CheckChangeOnCommit(input_api, output_api): report = [] license = ( r'.*? Copyright \(c\) %(year)s Google Inc\. All rights reserved\.\n' r'.*? Use of this source code is governed by a BSD-style license that ' r'can be\n' r'.*? found in the LICENSE file\.\n' ) % { 'year': input_api.time.strftime('%Y'), } report.extend(input_api.canned_checks.PanProjectChecks( input_api, output_api, license_header=license)) report.extend(input_api.canned_checks.CheckTreeIsOpen( input_api, output_api, 'http://gyp-status.appspot.com/status', 'http://gyp-status.appspot.com/current')) return report def GetPreferredTrySlaves(): return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac']
ac3c855583a023fc76b8720aa7e38419b28a26d4
falcom/api/hathi.py
falcom/api/hathi.py
# Copyright (c) 2017 The Regents of the University of Michigan. # All Rights Reserved. Licensed according to the terms of the Revised # BSD License. See LICENSE.txt for details. import json def get_counts_from_item_list (items, htid): a = len([x for x in items if x["htid"] == htid]) b = len(items) - a return a, b def get_oclc_counts_from_json (json_data, htid = ""): try: data = json.loads(json_data) return get_counts_from_item_list(data["items"], htid) except: return 0, 0 def get_hathi_data_from_json (): return ()
# Copyright (c) 2017 The Regents of the University of Michigan. # All Rights Reserved. Licensed according to the terms of the Revised # BSD License. See LICENSE.txt for details. import json class HathiItems: def __init__ (self): pass def __len__ (self): return 0 def get_counts_from_item_list (items, htid): a = len([x for x in items if x["htid"] == htid]) b = len(items) - a return a, b def get_oclc_counts_from_json (json_data, htid = ""): try: data = json.loads(json_data) return get_counts_from_item_list(data["items"], htid) except: return 0, 0 def get_hathi_data_from_json (): return HathiItems()
Refactor empty tuple into empty object with len()
Refactor empty tuple into empty object with len()
Python
bsd-3-clause
mlibrary/image-conversion-and-validation,mlibrary/image-conversion-and-validation
# Copyright (c) 2017 The Regents of the University of Michigan. # All Rights Reserved. Licensed according to the terms of the Revised # BSD License. See LICENSE.txt for details. import json def get_counts_from_item_list (items, htid): a = len([x for x in items if x["htid"] == htid]) b = len(items) - a return a, b def get_oclc_counts_from_json (json_data, htid = ""): try: data = json.loads(json_data) return get_counts_from_item_list(data["items"], htid) except: return 0, 0 def get_hathi_data_from_json (): return () Refactor empty tuple into empty object with len()
# Copyright (c) 2017 The Regents of the University of Michigan. # All Rights Reserved. Licensed according to the terms of the Revised # BSD License. See LICENSE.txt for details. import json class HathiItems: def __init__ (self): pass def __len__ (self): return 0 def get_counts_from_item_list (items, htid): a = len([x for x in items if x["htid"] == htid]) b = len(items) - a return a, b def get_oclc_counts_from_json (json_data, htid = ""): try: data = json.loads(json_data) return get_counts_from_item_list(data["items"], htid) except: return 0, 0 def get_hathi_data_from_json (): return HathiItems()
<commit_before># Copyright (c) 2017 The Regents of the University of Michigan. # All Rights Reserved. Licensed according to the terms of the Revised # BSD License. See LICENSE.txt for details. import json def get_counts_from_item_list (items, htid): a = len([x for x in items if x["htid"] == htid]) b = len(items) - a return a, b def get_oclc_counts_from_json (json_data, htid = ""): try: data = json.loads(json_data) return get_counts_from_item_list(data["items"], htid) except: return 0, 0 def get_hathi_data_from_json (): return () <commit_msg>Refactor empty tuple into empty object with len()<commit_after>
# Copyright (c) 2017 The Regents of the University of Michigan. # All Rights Reserved. Licensed according to the terms of the Revised # BSD License. See LICENSE.txt for details. import json class HathiItems: def __init__ (self): pass def __len__ (self): return 0 def get_counts_from_item_list (items, htid): a = len([x for x in items if x["htid"] == htid]) b = len(items) - a return a, b def get_oclc_counts_from_json (json_data, htid = ""): try: data = json.loads(json_data) return get_counts_from_item_list(data["items"], htid) except: return 0, 0 def get_hathi_data_from_json (): return HathiItems()
# Copyright (c) 2017 The Regents of the University of Michigan. # All Rights Reserved. Licensed according to the terms of the Revised # BSD License. See LICENSE.txt for details. import json def get_counts_from_item_list (items, htid): a = len([x for x in items if x["htid"] == htid]) b = len(items) - a return a, b def get_oclc_counts_from_json (json_data, htid = ""): try: data = json.loads(json_data) return get_counts_from_item_list(data["items"], htid) except: return 0, 0 def get_hathi_data_from_json (): return () Refactor empty tuple into empty object with len()# Copyright (c) 2017 The Regents of the University of Michigan. # All Rights Reserved. Licensed according to the terms of the Revised # BSD License. See LICENSE.txt for details. import json class HathiItems: def __init__ (self): pass def __len__ (self): return 0 def get_counts_from_item_list (items, htid): a = len([x for x in items if x["htid"] == htid]) b = len(items) - a return a, b def get_oclc_counts_from_json (json_data, htid = ""): try: data = json.loads(json_data) return get_counts_from_item_list(data["items"], htid) except: return 0, 0 def get_hathi_data_from_json (): return HathiItems()
<commit_before># Copyright (c) 2017 The Regents of the University of Michigan. # All Rights Reserved. Licensed according to the terms of the Revised # BSD License. See LICENSE.txt for details. import json def get_counts_from_item_list (items, htid): a = len([x for x in items if x["htid"] == htid]) b = len(items) - a return a, b def get_oclc_counts_from_json (json_data, htid = ""): try: data = json.loads(json_data) return get_counts_from_item_list(data["items"], htid) except: return 0, 0 def get_hathi_data_from_json (): return () <commit_msg>Refactor empty tuple into empty object with len()<commit_after># Copyright (c) 2017 The Regents of the University of Michigan. # All Rights Reserved. Licensed according to the terms of the Revised # BSD License. See LICENSE.txt for details. import json class HathiItems: def __init__ (self): pass def __len__ (self): return 0 def get_counts_from_item_list (items, htid): a = len([x for x in items if x["htid"] == htid]) b = len(items) - a return a, b def get_oclc_counts_from_json (json_data, htid = ""): try: data = json.loads(json_data) return get_counts_from_item_list(data["items"], htid) except: return 0, 0 def get_hathi_data_from_json (): return HathiItems()
7872a2327f9dea7d4c1f5a3054b6be6bba25fdd4
scripts/migration/migrate_deleted_wikis.py
scripts/migration/migrate_deleted_wikis.py
import logging import sys from modularodm import Q from framework.transactions.context import TokuTransaction from website.app import init_app from website.models import NodeLog from scripts import utils as script_utils logger = logging.getLogger(__name__) def get_targets(): return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED)) def migrate(targets, dry_run=True): # iterate over targets for log in targets: node = log.node versions = node.wiki_pages_versions current = node.wiki_pages_current updated_versions = {} for wiki in versions: if wiki in current: updated_versions[wiki] = versions[wiki] with TokuTransaction(): node.wiki_pages_versions = updated_versions node.save() if dry_run: raise RuntimeError('Dry run, transaction rolled back.') def main(): dry_run = False if '--dry' in sys.argv: dry_run = True if not dry_run: script_utils.add_file_logger(logger, __file__) init_app(set_backends=True, routes=False) with TokuTransaction(): migrate(targets=get_targets(), dry_run=dry_run) if __name__ == "__main__": main()
import logging import sys from modularodm import Q from framework.transactions.context import TokuTransaction from website.app import init_app from website.models import NodeLog from scripts import utils as script_utils logger = logging.getLogger(__name__) def get_targets(): return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED)) def migrate(targets, dry_run=True): # iterate over targets for log in targets: node = log.node versions = node.wiki_pages_versions current = node.wiki_pages_current updated_versions = {} for wiki in versions: if wiki in current: updated_versions[wiki] = versions[wiki] node.wiki_pages_versions = updated_versions node.save() def main(): dry_run = False if '--dry' in sys.argv: dry_run = True if not dry_run: script_utils.add_file_logger(logger, __file__) init_app(set_backends=True, routes=False) with TokuTransaction(): migrate(targets=get_targets(), dry_run=dry_run) if dry_run: raise RuntimeError('Dry run, transaction rolled back.') if __name__ == "__main__": main()
Remove TokuTransaction in migrate function
Remove TokuTransaction in migrate function
Python
apache-2.0
hmoco/osf.io,samchrisinger/osf.io,hmoco/osf.io,icereval/osf.io,caneruguz/osf.io,cwisecarver/osf.io,chrisseto/osf.io,erinspace/osf.io,SSJohns/osf.io,monikagrabowska/osf.io,crcresearch/osf.io,crcresearch/osf.io,laurenrevere/osf.io,leb2dg/osf.io,crcresearch/osf.io,baylee-d/osf.io,leb2dg/osf.io,saradbowman/osf.io,sloria/osf.io,felliott/osf.io,mluke93/osf.io,adlius/osf.io,SSJohns/osf.io,mluke93/osf.io,binoculars/osf.io,mluo613/osf.io,felliott/osf.io,amyshi188/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,mfraezz/osf.io,DanielSBrown/osf.io,mluo613/osf.io,erinspace/osf.io,adlius/osf.io,mluke93/osf.io,acshi/osf.io,abought/osf.io,wearpants/osf.io,laurenrevere/osf.io,wearpants/osf.io,cslzchen/osf.io,mattclark/osf.io,rdhyee/osf.io,binoculars/osf.io,hmoco/osf.io,laurenrevere/osf.io,monikagrabowska/osf.io,baylee-d/osf.io,hmoco/osf.io,emetsger/osf.io,saradbowman/osf.io,mfraezz/osf.io,TomBaxter/osf.io,Nesiehr/osf.io,rdhyee/osf.io,abought/osf.io,abought/osf.io,rdhyee/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,aaxelb/osf.io,samchrisinger/osf.io,chrisseto/osf.io,caseyrollins/osf.io,felliott/osf.io,alexschiller/osf.io,DanielSBrown/osf.io,Nesiehr/osf.io,amyshi188/osf.io,chrisseto/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,mluo613/osf.io,rdhyee/osf.io,felliott/osf.io,abought/osf.io,samchrisinger/osf.io,alexschiller/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,brianjgeiger/osf.io,mluo613/osf.io,DanielSBrown/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,acshi/osf.io,TomBaxter/osf.io,emetsger/osf.io,adlius/osf.io,caseyrollins/osf.io,CenterForOpenScience/osf.io,wearpants/osf.io,emetsger/osf.io,monikagrabowska/osf.io,acshi/osf.io,aaxelb/osf.io,SSJohns/osf.io,brianjgeiger/osf.io,amyshi188/osf.io,chennan47/osf.io,sloria/osf.io,icereval/osf.io,mattclark/osf.io,alexschiller/osf.io,monikagrabowska/osf.io,binoculars/osf.io,wearpants/osf.io,samchrisinger/osf.io,HalcyonChimera/osf.io,amyshi188/osf.io,chrisseto/osf.io,cwisecarver/osf.io,aaxelb/osf.io,erinspace/osf.io,cslzchen/osf.io,aaxelb/osf.io,caneruguz/osf.io,pattisdr/osf.io,mluo613/osf.io,cwisecarver/osf.io,alexschiller/osf.io,chennan47/osf.io,pattisdr/osf.io,CenterForOpenScience/osf.io,CenterForOpenScience/osf.io,sloria/osf.io,mattclark/osf.io,mfraezz/osf.io,Nesiehr/osf.io,emetsger/osf.io,HalcyonChimera/osf.io,Johnetordoff/osf.io,leb2dg/osf.io,cslzchen/osf.io,pattisdr/osf.io,SSJohns/osf.io,icereval/osf.io,caneruguz/osf.io,alexschiller/osf.io,chennan47/osf.io,acshi/osf.io,caseyrollins/osf.io,mluke93/osf.io,CenterForOpenScience/osf.io,adlius/osf.io,HalcyonChimera/osf.io,Nesiehr/osf.io,acshi/osf.io,caneruguz/osf.io,mfraezz/osf.io,cwisecarver/osf.io
import logging import sys from modularodm import Q from framework.transactions.context import TokuTransaction from website.app import init_app from website.models import NodeLog from scripts import utils as script_utils logger = logging.getLogger(__name__) def get_targets(): return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED)) def migrate(targets, dry_run=True): # iterate over targets for log in targets: node = log.node versions = node.wiki_pages_versions current = node.wiki_pages_current updated_versions = {} for wiki in versions: if wiki in current: updated_versions[wiki] = versions[wiki] with TokuTransaction(): node.wiki_pages_versions = updated_versions node.save() if dry_run: raise RuntimeError('Dry run, transaction rolled back.') def main(): dry_run = False if '--dry' in sys.argv: dry_run = True if not dry_run: script_utils.add_file_logger(logger, __file__) init_app(set_backends=True, routes=False) with TokuTransaction(): migrate(targets=get_targets(), dry_run=dry_run) if __name__ == "__main__": main() Remove TokuTransaction in migrate function
import logging import sys from modularodm import Q from framework.transactions.context import TokuTransaction from website.app import init_app from website.models import NodeLog from scripts import utils as script_utils logger = logging.getLogger(__name__) def get_targets(): return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED)) def migrate(targets, dry_run=True): # iterate over targets for log in targets: node = log.node versions = node.wiki_pages_versions current = node.wiki_pages_current updated_versions = {} for wiki in versions: if wiki in current: updated_versions[wiki] = versions[wiki] node.wiki_pages_versions = updated_versions node.save() def main(): dry_run = False if '--dry' in sys.argv: dry_run = True if not dry_run: script_utils.add_file_logger(logger, __file__) init_app(set_backends=True, routes=False) with TokuTransaction(): migrate(targets=get_targets(), dry_run=dry_run) if dry_run: raise RuntimeError('Dry run, transaction rolled back.') if __name__ == "__main__": main()
<commit_before>import logging import sys from modularodm import Q from framework.transactions.context import TokuTransaction from website.app import init_app from website.models import NodeLog from scripts import utils as script_utils logger = logging.getLogger(__name__) def get_targets(): return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED)) def migrate(targets, dry_run=True): # iterate over targets for log in targets: node = log.node versions = node.wiki_pages_versions current = node.wiki_pages_current updated_versions = {} for wiki in versions: if wiki in current: updated_versions[wiki] = versions[wiki] with TokuTransaction(): node.wiki_pages_versions = updated_versions node.save() if dry_run: raise RuntimeError('Dry run, transaction rolled back.') def main(): dry_run = False if '--dry' in sys.argv: dry_run = True if not dry_run: script_utils.add_file_logger(logger, __file__) init_app(set_backends=True, routes=False) with TokuTransaction(): migrate(targets=get_targets(), dry_run=dry_run) if __name__ == "__main__": main() <commit_msg>Remove TokuTransaction in migrate function<commit_after>
import logging import sys from modularodm import Q from framework.transactions.context import TokuTransaction from website.app import init_app from website.models import NodeLog from scripts import utils as script_utils logger = logging.getLogger(__name__) def get_targets(): return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED)) def migrate(targets, dry_run=True): # iterate over targets for log in targets: node = log.node versions = node.wiki_pages_versions current = node.wiki_pages_current updated_versions = {} for wiki in versions: if wiki in current: updated_versions[wiki] = versions[wiki] node.wiki_pages_versions = updated_versions node.save() def main(): dry_run = False if '--dry' in sys.argv: dry_run = True if not dry_run: script_utils.add_file_logger(logger, __file__) init_app(set_backends=True, routes=False) with TokuTransaction(): migrate(targets=get_targets(), dry_run=dry_run) if dry_run: raise RuntimeError('Dry run, transaction rolled back.') if __name__ == "__main__": main()
import logging import sys from modularodm import Q from framework.transactions.context import TokuTransaction from website.app import init_app from website.models import NodeLog from scripts import utils as script_utils logger = logging.getLogger(__name__) def get_targets(): return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED)) def migrate(targets, dry_run=True): # iterate over targets for log in targets: node = log.node versions = node.wiki_pages_versions current = node.wiki_pages_current updated_versions = {} for wiki in versions: if wiki in current: updated_versions[wiki] = versions[wiki] with TokuTransaction(): node.wiki_pages_versions = updated_versions node.save() if dry_run: raise RuntimeError('Dry run, transaction rolled back.') def main(): dry_run = False if '--dry' in sys.argv: dry_run = True if not dry_run: script_utils.add_file_logger(logger, __file__) init_app(set_backends=True, routes=False) with TokuTransaction(): migrate(targets=get_targets(), dry_run=dry_run) if __name__ == "__main__": main() Remove TokuTransaction in migrate functionimport logging import sys from modularodm import Q from framework.transactions.context import TokuTransaction from website.app import init_app from website.models import NodeLog from scripts import utils as script_utils logger = logging.getLogger(__name__) def get_targets(): return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED)) def migrate(targets, dry_run=True): # iterate over targets for log in targets: node = log.node versions = node.wiki_pages_versions current = node.wiki_pages_current updated_versions = {} for wiki in versions: if wiki in current: updated_versions[wiki] = versions[wiki] node.wiki_pages_versions = updated_versions node.save() def main(): dry_run = False if '--dry' in sys.argv: dry_run = True if not dry_run: script_utils.add_file_logger(logger, __file__) init_app(set_backends=True, routes=False) with TokuTransaction(): migrate(targets=get_targets(), dry_run=dry_run) if dry_run: raise RuntimeError('Dry run, transaction rolled back.') if __name__ == "__main__": main()
<commit_before>import logging import sys from modularodm import Q from framework.transactions.context import TokuTransaction from website.app import init_app from website.models import NodeLog from scripts import utils as script_utils logger = logging.getLogger(__name__) def get_targets(): return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED)) def migrate(targets, dry_run=True): # iterate over targets for log in targets: node = log.node versions = node.wiki_pages_versions current = node.wiki_pages_current updated_versions = {} for wiki in versions: if wiki in current: updated_versions[wiki] = versions[wiki] with TokuTransaction(): node.wiki_pages_versions = updated_versions node.save() if dry_run: raise RuntimeError('Dry run, transaction rolled back.') def main(): dry_run = False if '--dry' in sys.argv: dry_run = True if not dry_run: script_utils.add_file_logger(logger, __file__) init_app(set_backends=True, routes=False) with TokuTransaction(): migrate(targets=get_targets(), dry_run=dry_run) if __name__ == "__main__": main() <commit_msg>Remove TokuTransaction in migrate function<commit_after>import logging import sys from modularodm import Q from framework.transactions.context import TokuTransaction from website.app import init_app from website.models import NodeLog from scripts import utils as script_utils logger = logging.getLogger(__name__) def get_targets(): return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED)) def migrate(targets, dry_run=True): # iterate over targets for log in targets: node = log.node versions = node.wiki_pages_versions current = node.wiki_pages_current updated_versions = {} for wiki in versions: if wiki in current: updated_versions[wiki] = versions[wiki] node.wiki_pages_versions = updated_versions node.save() def main(): dry_run = False if '--dry' in sys.argv: dry_run = True if not dry_run: script_utils.add_file_logger(logger, __file__) init_app(set_backends=True, routes=False) with TokuTransaction(): migrate(targets=get_targets(), dry_run=dry_run) if dry_run: raise RuntimeError('Dry run, transaction rolled back.') if __name__ == "__main__": main()
c69d99ff9c102926d94d8bd2d55c5de40f5e2072
application/senic/nuimo_hub/subprocess_run.py
application/senic/nuimo_hub/subprocess_run.py
"""Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`.""" try: from subprocess import run except ImportError: from collections import namedtuple from subprocess import check_output def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None): stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, encoding=encoding, errors=errors, timeout=timeout) Output = namedtuple('Output', ['stdout']) return Output(stdout=stdout_bytes)
"""Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`.""" try: from subprocess import run except ImportError: from collections import namedtuple from subprocess import check_output def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None): stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, timeout=timeout) Output = namedtuple('Output', ['stdout']) return Output(stdout=stdout_bytes)
Remove parameters that are not supported
Remove parameters that are not supported Apparently Python 3.4 doesn't have encoding and errors parameters
Python
mit
grunskis/nuimo-hub-backend,grunskis/senic-hub,grunskis/senic-hub,getsenic/senic-hub,grunskis/senic-hub,grunskis/senic-hub,grunskis/nuimo-hub-backend,grunskis/nuimo-hub-backend,grunskis/senic-hub,getsenic/senic-hub,grunskis/nuimo-hub-backend,grunskis/senic-hub,grunskis/nuimo-hub-backend
"""Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`.""" try: from subprocess import run except ImportError: from collections import namedtuple from subprocess import check_output def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None): stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, encoding=encoding, errors=errors, timeout=timeout) Output = namedtuple('Output', ['stdout']) return Output(stdout=stdout_bytes) Remove parameters that are not supported Apparently Python 3.4 doesn't have encoding and errors parameters
"""Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`.""" try: from subprocess import run except ImportError: from collections import namedtuple from subprocess import check_output def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None): stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, timeout=timeout) Output = namedtuple('Output', ['stdout']) return Output(stdout=stdout_bytes)
<commit_before>"""Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`.""" try: from subprocess import run except ImportError: from collections import namedtuple from subprocess import check_output def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None): stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, encoding=encoding, errors=errors, timeout=timeout) Output = namedtuple('Output', ['stdout']) return Output(stdout=stdout_bytes) <commit_msg>Remove parameters that are not supported Apparently Python 3.4 doesn't have encoding and errors parameters<commit_after>
"""Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`.""" try: from subprocess import run except ImportError: from collections import namedtuple from subprocess import check_output def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None): stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, timeout=timeout) Output = namedtuple('Output', ['stdout']) return Output(stdout=stdout_bytes)
"""Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`.""" try: from subprocess import run except ImportError: from collections import namedtuple from subprocess import check_output def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None): stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, encoding=encoding, errors=errors, timeout=timeout) Output = namedtuple('Output', ['stdout']) return Output(stdout=stdout_bytes) Remove parameters that are not supported Apparently Python 3.4 doesn't have encoding and errors parameters"""Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`.""" try: from subprocess import run except ImportError: from collections import namedtuple from subprocess import check_output def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None): stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, timeout=timeout) Output = namedtuple('Output', ['stdout']) return Output(stdout=stdout_bytes)
<commit_before>"""Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`.""" try: from subprocess import run except ImportError: from collections import namedtuple from subprocess import check_output def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None): stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, encoding=encoding, errors=errors, timeout=timeout) Output = namedtuple('Output', ['stdout']) return Output(stdout=stdout_bytes) <commit_msg>Remove parameters that are not supported Apparently Python 3.4 doesn't have encoding and errors parameters<commit_after>"""Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`.""" try: from subprocess import run except ImportError: from collections import namedtuple from subprocess import check_output def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None): stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, timeout=timeout) Output = namedtuple('Output', ['stdout']) return Output(stdout=stdout_bytes)
7894b066cde13eca75479921531e9d005970e9c3
go/billing/views.py
go/billing/views.py
# Create your views here.
import os from django.conf import settings from django.contrib.auth.decorators import login_required from django.http import HttpResponse from django.shortcuts import get_object_or_404 from django.template import RequestContext, loader from xhtml2pdf import pisa from go.billing.models import Statement @login_required def statement_view(request, statement_id=None): """Send a PDF version of the statement with the given ``statement_id`` to the user's browser. """ statement = get_object_or_404( Statement, pk=statement_id, account__user=request.user) response = HttpResponse(mimetype='application/pdf') filename = "%s (%s).pdf" % (statement.title, statement.from_date.strftime('%B %Y')) response['Content-Disposition'] = 'attachment; filename=%s' % (filename,) template = loader.get_template('billing/invoice.html') line_item_list = statement.lineitem_set.all() context = RequestContext(request, {'item_list': line_item_list}) html_result = template.render(context) pisa.CreatePDF(html_result, dest=response, link_callback=link_callback) return response # Convert HTML URIs to absolute system paths so xhtml2pdf can access those # resources def link_callback(uri, rel): # use short variable names sUrl = settings.STATIC_URL # Typically /static/ sRoot = settings.STATIC_ROOT # Typically /home/userX/project_static/ mUrl = settings.MEDIA_URL # Typically /static/media/ mRoot = settings.MEDIA_ROOT # Typically /home/userX/project_static/media/ # convert URIs to absolute system paths if uri.startswith(mUrl): path = os.path.join(mRoot, uri.replace(mUrl, "")) elif uri.startswith(sUrl): path = os.path.join(sRoot, uri.replace(sUrl, "")) # make sure that file exists if not os.path.isfile(path): raise Exception( 'media URI must start with %s or %s' % (sUrl, mUrl)) return path
Add basic PDF view for billing
Add basic PDF view for billing
Python
bsd-3-clause
praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go
# Create your views here. Add basic PDF view for billing
import os from django.conf import settings from django.contrib.auth.decorators import login_required from django.http import HttpResponse from django.shortcuts import get_object_or_404 from django.template import RequestContext, loader from xhtml2pdf import pisa from go.billing.models import Statement @login_required def statement_view(request, statement_id=None): """Send a PDF version of the statement with the given ``statement_id`` to the user's browser. """ statement = get_object_or_404( Statement, pk=statement_id, account__user=request.user) response = HttpResponse(mimetype='application/pdf') filename = "%s (%s).pdf" % (statement.title, statement.from_date.strftime('%B %Y')) response['Content-Disposition'] = 'attachment; filename=%s' % (filename,) template = loader.get_template('billing/invoice.html') line_item_list = statement.lineitem_set.all() context = RequestContext(request, {'item_list': line_item_list}) html_result = template.render(context) pisa.CreatePDF(html_result, dest=response, link_callback=link_callback) return response # Convert HTML URIs to absolute system paths so xhtml2pdf can access those # resources def link_callback(uri, rel): # use short variable names sUrl = settings.STATIC_URL # Typically /static/ sRoot = settings.STATIC_ROOT # Typically /home/userX/project_static/ mUrl = settings.MEDIA_URL # Typically /static/media/ mRoot = settings.MEDIA_ROOT # Typically /home/userX/project_static/media/ # convert URIs to absolute system paths if uri.startswith(mUrl): path = os.path.join(mRoot, uri.replace(mUrl, "")) elif uri.startswith(sUrl): path = os.path.join(sRoot, uri.replace(sUrl, "")) # make sure that file exists if not os.path.isfile(path): raise Exception( 'media URI must start with %s or %s' % (sUrl, mUrl)) return path
<commit_before># Create your views here. <commit_msg>Add basic PDF view for billing<commit_after>
import os from django.conf import settings from django.contrib.auth.decorators import login_required from django.http import HttpResponse from django.shortcuts import get_object_or_404 from django.template import RequestContext, loader from xhtml2pdf import pisa from go.billing.models import Statement @login_required def statement_view(request, statement_id=None): """Send a PDF version of the statement with the given ``statement_id`` to the user's browser. """ statement = get_object_or_404( Statement, pk=statement_id, account__user=request.user) response = HttpResponse(mimetype='application/pdf') filename = "%s (%s).pdf" % (statement.title, statement.from_date.strftime('%B %Y')) response['Content-Disposition'] = 'attachment; filename=%s' % (filename,) template = loader.get_template('billing/invoice.html') line_item_list = statement.lineitem_set.all() context = RequestContext(request, {'item_list': line_item_list}) html_result = template.render(context) pisa.CreatePDF(html_result, dest=response, link_callback=link_callback) return response # Convert HTML URIs to absolute system paths so xhtml2pdf can access those # resources def link_callback(uri, rel): # use short variable names sUrl = settings.STATIC_URL # Typically /static/ sRoot = settings.STATIC_ROOT # Typically /home/userX/project_static/ mUrl = settings.MEDIA_URL # Typically /static/media/ mRoot = settings.MEDIA_ROOT # Typically /home/userX/project_static/media/ # convert URIs to absolute system paths if uri.startswith(mUrl): path = os.path.join(mRoot, uri.replace(mUrl, "")) elif uri.startswith(sUrl): path = os.path.join(sRoot, uri.replace(sUrl, "")) # make sure that file exists if not os.path.isfile(path): raise Exception( 'media URI must start with %s or %s' % (sUrl, mUrl)) return path
# Create your views here. Add basic PDF view for billingimport os from django.conf import settings from django.contrib.auth.decorators import login_required from django.http import HttpResponse from django.shortcuts import get_object_or_404 from django.template import RequestContext, loader from xhtml2pdf import pisa from go.billing.models import Statement @login_required def statement_view(request, statement_id=None): """Send a PDF version of the statement with the given ``statement_id`` to the user's browser. """ statement = get_object_or_404( Statement, pk=statement_id, account__user=request.user) response = HttpResponse(mimetype='application/pdf') filename = "%s (%s).pdf" % (statement.title, statement.from_date.strftime('%B %Y')) response['Content-Disposition'] = 'attachment; filename=%s' % (filename,) template = loader.get_template('billing/invoice.html') line_item_list = statement.lineitem_set.all() context = RequestContext(request, {'item_list': line_item_list}) html_result = template.render(context) pisa.CreatePDF(html_result, dest=response, link_callback=link_callback) return response # Convert HTML URIs to absolute system paths so xhtml2pdf can access those # resources def link_callback(uri, rel): # use short variable names sUrl = settings.STATIC_URL # Typically /static/ sRoot = settings.STATIC_ROOT # Typically /home/userX/project_static/ mUrl = settings.MEDIA_URL # Typically /static/media/ mRoot = settings.MEDIA_ROOT # Typically /home/userX/project_static/media/ # convert URIs to absolute system paths if uri.startswith(mUrl): path = os.path.join(mRoot, uri.replace(mUrl, "")) elif uri.startswith(sUrl): path = os.path.join(sRoot, uri.replace(sUrl, "")) # make sure that file exists if not os.path.isfile(path): raise Exception( 'media URI must start with %s or %s' % (sUrl, mUrl)) return path
<commit_before># Create your views here. <commit_msg>Add basic PDF view for billing<commit_after>import os from django.conf import settings from django.contrib.auth.decorators import login_required from django.http import HttpResponse from django.shortcuts import get_object_or_404 from django.template import RequestContext, loader from xhtml2pdf import pisa from go.billing.models import Statement @login_required def statement_view(request, statement_id=None): """Send a PDF version of the statement with the given ``statement_id`` to the user's browser. """ statement = get_object_or_404( Statement, pk=statement_id, account__user=request.user) response = HttpResponse(mimetype='application/pdf') filename = "%s (%s).pdf" % (statement.title, statement.from_date.strftime('%B %Y')) response['Content-Disposition'] = 'attachment; filename=%s' % (filename,) template = loader.get_template('billing/invoice.html') line_item_list = statement.lineitem_set.all() context = RequestContext(request, {'item_list': line_item_list}) html_result = template.render(context) pisa.CreatePDF(html_result, dest=response, link_callback=link_callback) return response # Convert HTML URIs to absolute system paths so xhtml2pdf can access those # resources def link_callback(uri, rel): # use short variable names sUrl = settings.STATIC_URL # Typically /static/ sRoot = settings.STATIC_ROOT # Typically /home/userX/project_static/ mUrl = settings.MEDIA_URL # Typically /static/media/ mRoot = settings.MEDIA_ROOT # Typically /home/userX/project_static/media/ # convert URIs to absolute system paths if uri.startswith(mUrl): path = os.path.join(mRoot, uri.replace(mUrl, "")) elif uri.startswith(sUrl): path = os.path.join(sRoot, uri.replace(sUrl, "")) # make sure that file exists if not os.path.isfile(path): raise Exception( 'media URI must start with %s or %s' % (sUrl, mUrl)) return path
a3b6aa273900c9d9c98a6ad9c9735a323086c64f
user_clipboard/urls.py
user_clipboard/urls.py
from django.conf.urls import url from .views import ClipboardFileAPIView, ClipboardImageAPIView urlpatterns = [ url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"), url(r'^', ClipboardFileAPIView.as_view(), name="clipboard"), ]
from django.conf.urls import url from .views import ClipboardFileAPIView, ClipboardImageAPIView urlpatterns = [ url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"), url(r'^$', ClipboardFileAPIView.as_view(), name="clipboard"), ]
Fix clipboard URLs to not be greedy
Fix clipboard URLs to not be greedy
Python
mit
MagicSolutions/django-user-clipboard,IndustriaTech/django-user-clipboard,MagicSolutions/django-user-clipboard,IndustriaTech/django-user-clipboard
from django.conf.urls import url from .views import ClipboardFileAPIView, ClipboardImageAPIView urlpatterns = [ url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"), url(r'^', ClipboardFileAPIView.as_view(), name="clipboard"), ] Fix clipboard URLs to not be greedy
from django.conf.urls import url from .views import ClipboardFileAPIView, ClipboardImageAPIView urlpatterns = [ url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"), url(r'^$', ClipboardFileAPIView.as_view(), name="clipboard"), ]
<commit_before>from django.conf.urls import url from .views import ClipboardFileAPIView, ClipboardImageAPIView urlpatterns = [ url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"), url(r'^', ClipboardFileAPIView.as_view(), name="clipboard"), ] <commit_msg>Fix clipboard URLs to not be greedy<commit_after>
from django.conf.urls import url from .views import ClipboardFileAPIView, ClipboardImageAPIView urlpatterns = [ url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"), url(r'^$', ClipboardFileAPIView.as_view(), name="clipboard"), ]
from django.conf.urls import url from .views import ClipboardFileAPIView, ClipboardImageAPIView urlpatterns = [ url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"), url(r'^', ClipboardFileAPIView.as_view(), name="clipboard"), ] Fix clipboard URLs to not be greedyfrom django.conf.urls import url from .views import ClipboardFileAPIView, ClipboardImageAPIView urlpatterns = [ url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"), url(r'^$', ClipboardFileAPIView.as_view(), name="clipboard"), ]
<commit_before>from django.conf.urls import url from .views import ClipboardFileAPIView, ClipboardImageAPIView urlpatterns = [ url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"), url(r'^', ClipboardFileAPIView.as_view(), name="clipboard"), ] <commit_msg>Fix clipboard URLs to not be greedy<commit_after>from django.conf.urls import url from .views import ClipboardFileAPIView, ClipboardImageAPIView urlpatterns = [ url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"), url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"), url(r'^$', ClipboardFileAPIView.as_view(), name="clipboard"), ]
9a1a05c33258461c5d474b014654464892cd7b90
bake/bakedefaults.py
bake/bakedefaults.py
#!/usr/bin/env python LABEL_KEY = 'label' KEY_START = '@' KEY_END = '@' CFGFILE = 'bake.cfg'
#!/usr/bin/env python LABEL_KEY = 'label' KEY_START = '@' KEY_END = '@'
Remove mention of bake.cfg file
Remove mention of bake.cfg file
Python
mit
AlexSzatmary/bake
#!/usr/bin/env python LABEL_KEY = 'label' KEY_START = '@' KEY_END = '@' CFGFILE = 'bake.cfg' Remove mention of bake.cfg file
#!/usr/bin/env python LABEL_KEY = 'label' KEY_START = '@' KEY_END = '@'
<commit_before>#!/usr/bin/env python LABEL_KEY = 'label' KEY_START = '@' KEY_END = '@' CFGFILE = 'bake.cfg' <commit_msg>Remove mention of bake.cfg file<commit_after>
#!/usr/bin/env python LABEL_KEY = 'label' KEY_START = '@' KEY_END = '@'
#!/usr/bin/env python LABEL_KEY = 'label' KEY_START = '@' KEY_END = '@' CFGFILE = 'bake.cfg' Remove mention of bake.cfg file#!/usr/bin/env python LABEL_KEY = 'label' KEY_START = '@' KEY_END = '@'
<commit_before>#!/usr/bin/env python LABEL_KEY = 'label' KEY_START = '@' KEY_END = '@' CFGFILE = 'bake.cfg' <commit_msg>Remove mention of bake.cfg file<commit_after>#!/usr/bin/env python LABEL_KEY = 'label' KEY_START = '@' KEY_END = '@'
c859416d2d35aab83fc9e8f400e00f8f07c0b8a9
test/parser_test.py
test/parser_test.py
import socket client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) client_socket.connect(("localhost", 5002)) with open("resources/Matrix.java", "r") as java_file: source = java_file.read() + "\nEOS_BITSHIFT" client_socket.send("%d\n%s" % (len(source), source)); data = '' while True: data = client_socket.recv(10000) if data != '': client_socket.close() break; print data;
import socket, sys file_name = 'resources/<name>.c' server_socket_number = 5001 if __name__ == '__main__': if len(sys.argv) == 1: print "Please input a parser to test." elif len(sys.argv) > 2: print "Too many arguments." else: if sys.argv[1] == 'c': pass elif sys.argv[1] == 'java': file_name = "resources/Matrix.java" server_socket_number = 5002 elif sys.argv[1] == 'ruby': file_name = "resources/<name>.rb" server_socket_number = 5003 client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) client_socket.connect(("localhost", server_socket_number)) with open(file_name, "r") as source_file: source = source_file.read() client_socket.send("%d\n%s" % (len(source), source)); data = '' while True: data = client_socket.recv(10000) if data != '': client_socket.close() break; print data;
Change test file to support different parsers
Change test file to support different parsers
Python
mit
earwig/bitshift,earwig/bitshift,earwig/bitshift,earwig/bitshift,earwig/bitshift,earwig/bitshift
import socket client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) client_socket.connect(("localhost", 5002)) with open("resources/Matrix.java", "r") as java_file: source = java_file.read() + "\nEOS_BITSHIFT" client_socket.send("%d\n%s" % (len(source), source)); data = '' while True: data = client_socket.recv(10000) if data != '': client_socket.close() break; print data; Change test file to support different parsers
import socket, sys file_name = 'resources/<name>.c' server_socket_number = 5001 if __name__ == '__main__': if len(sys.argv) == 1: print "Please input a parser to test." elif len(sys.argv) > 2: print "Too many arguments." else: if sys.argv[1] == 'c': pass elif sys.argv[1] == 'java': file_name = "resources/Matrix.java" server_socket_number = 5002 elif sys.argv[1] == 'ruby': file_name = "resources/<name>.rb" server_socket_number = 5003 client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) client_socket.connect(("localhost", server_socket_number)) with open(file_name, "r") as source_file: source = source_file.read() client_socket.send("%d\n%s" % (len(source), source)); data = '' while True: data = client_socket.recv(10000) if data != '': client_socket.close() break; print data;
<commit_before>import socket client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) client_socket.connect(("localhost", 5002)) with open("resources/Matrix.java", "r") as java_file: source = java_file.read() + "\nEOS_BITSHIFT" client_socket.send("%d\n%s" % (len(source), source)); data = '' while True: data = client_socket.recv(10000) if data != '': client_socket.close() break; print data; <commit_msg>Change test file to support different parsers<commit_after>
import socket, sys file_name = 'resources/<name>.c' server_socket_number = 5001 if __name__ == '__main__': if len(sys.argv) == 1: print "Please input a parser to test." elif len(sys.argv) > 2: print "Too many arguments." else: if sys.argv[1] == 'c': pass elif sys.argv[1] == 'java': file_name = "resources/Matrix.java" server_socket_number = 5002 elif sys.argv[1] == 'ruby': file_name = "resources/<name>.rb" server_socket_number = 5003 client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) client_socket.connect(("localhost", server_socket_number)) with open(file_name, "r") as source_file: source = source_file.read() client_socket.send("%d\n%s" % (len(source), source)); data = '' while True: data = client_socket.recv(10000) if data != '': client_socket.close() break; print data;
import socket client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) client_socket.connect(("localhost", 5002)) with open("resources/Matrix.java", "r") as java_file: source = java_file.read() + "\nEOS_BITSHIFT" client_socket.send("%d\n%s" % (len(source), source)); data = '' while True: data = client_socket.recv(10000) if data != '': client_socket.close() break; print data; Change test file to support different parsersimport socket, sys file_name = 'resources/<name>.c' server_socket_number = 5001 if __name__ == '__main__': if len(sys.argv) == 1: print "Please input a parser to test." elif len(sys.argv) > 2: print "Too many arguments." else: if sys.argv[1] == 'c': pass elif sys.argv[1] == 'java': file_name = "resources/Matrix.java" server_socket_number = 5002 elif sys.argv[1] == 'ruby': file_name = "resources/<name>.rb" server_socket_number = 5003 client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) client_socket.connect(("localhost", server_socket_number)) with open(file_name, "r") as source_file: source = source_file.read() client_socket.send("%d\n%s" % (len(source), source)); data = '' while True: data = client_socket.recv(10000) if data != '': client_socket.close() break; print data;
<commit_before>import socket client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) client_socket.connect(("localhost", 5002)) with open("resources/Matrix.java", "r") as java_file: source = java_file.read() + "\nEOS_BITSHIFT" client_socket.send("%d\n%s" % (len(source), source)); data = '' while True: data = client_socket.recv(10000) if data != '': client_socket.close() break; print data; <commit_msg>Change test file to support different parsers<commit_after>import socket, sys file_name = 'resources/<name>.c' server_socket_number = 5001 if __name__ == '__main__': if len(sys.argv) == 1: print "Please input a parser to test." elif len(sys.argv) > 2: print "Too many arguments." else: if sys.argv[1] == 'c': pass elif sys.argv[1] == 'java': file_name = "resources/Matrix.java" server_socket_number = 5002 elif sys.argv[1] == 'ruby': file_name = "resources/<name>.rb" server_socket_number = 5003 client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) client_socket.connect(("localhost", server_socket_number)) with open(file_name, "r") as source_file: source = source_file.read() client_socket.send("%d\n%s" % (len(source), source)); data = '' while True: data = client_socket.recv(10000) if data != '': client_socket.close() break; print data;
6603d277eb84712a907008bad8bc8156ed9bdf40
setup.py
setup.py
"""A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f: long_description = f.read() setup( name='Graphistry', version='1.1.0.dev1', description='This is established as a Data Loader for Graphistry', long_description=long_description, url='https://github.com/graphistry/pygraphistry', author='Graphistry', author_email='xin@graphistry.com', license='MIT', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Graphistry User', 'Topic :: Data Visualization Development :: Load Tools', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='Python Data Loader', packages=find_packages(exclude=['contrib', 'docs', 'tests*']), )
"""A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f: long_description = f.read() setup( name='Graphistry', version='1.1.0.dev1', description='This is established as a Data Loader for Graphistry', long_description=long_description, url='https://github.com/graphistry/pygraphistry', author='Graphistry', author_email='xin@graphistry.com', license='MIT', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Graphistry User', 'Topic :: Data Visualization Development :: Load Tools', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='Python Data Loader', packages=find_packages(exclude=['contrib', 'docs', 'tests*']), )
Fix license MIT to BSD
Fix license MIT to BSD
Python
bsd-3-clause
graphistry/pygraphistry,kcompher/pygraphistry,graphistry/pygraphistry
"""A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f: long_description = f.read() setup( name='Graphistry', version='1.1.0.dev1', description='This is established as a Data Loader for Graphistry', long_description=long_description, url='https://github.com/graphistry/pygraphistry', author='Graphistry', author_email='xin@graphistry.com', license='MIT', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Graphistry User', 'Topic :: Data Visualization Development :: Load Tools', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='Python Data Loader', packages=find_packages(exclude=['contrib', 'docs', 'tests*']), ) Fix license MIT to BSD
"""A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f: long_description = f.read() setup( name='Graphistry', version='1.1.0.dev1', description='This is established as a Data Loader for Graphistry', long_description=long_description, url='https://github.com/graphistry/pygraphistry', author='Graphistry', author_email='xin@graphistry.com', license='MIT', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Graphistry User', 'Topic :: Data Visualization Development :: Load Tools', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='Python Data Loader', packages=find_packages(exclude=['contrib', 'docs', 'tests*']), )
<commit_before>"""A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f: long_description = f.read() setup( name='Graphistry', version='1.1.0.dev1', description='This is established as a Data Loader for Graphistry', long_description=long_description, url='https://github.com/graphistry/pygraphistry', author='Graphistry', author_email='xin@graphistry.com', license='MIT', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Graphistry User', 'Topic :: Data Visualization Development :: Load Tools', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='Python Data Loader', packages=find_packages(exclude=['contrib', 'docs', 'tests*']), ) <commit_msg>Fix license MIT to BSD<commit_after>
"""A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f: long_description = f.read() setup( name='Graphistry', version='1.1.0.dev1', description='This is established as a Data Loader for Graphistry', long_description=long_description, url='https://github.com/graphistry/pygraphistry', author='Graphistry', author_email='xin@graphistry.com', license='MIT', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Graphistry User', 'Topic :: Data Visualization Development :: Load Tools', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='Python Data Loader', packages=find_packages(exclude=['contrib', 'docs', 'tests*']), )
"""A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f: long_description = f.read() setup( name='Graphistry', version='1.1.0.dev1', description='This is established as a Data Loader for Graphistry', long_description=long_description, url='https://github.com/graphistry/pygraphistry', author='Graphistry', author_email='xin@graphistry.com', license='MIT', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Graphistry User', 'Topic :: Data Visualization Development :: Load Tools', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='Python Data Loader', packages=find_packages(exclude=['contrib', 'docs', 'tests*']), ) Fix license MIT to BSD"""A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f: long_description = f.read() setup( name='Graphistry', version='1.1.0.dev1', description='This is established as a Data Loader for Graphistry', long_description=long_description, url='https://github.com/graphistry/pygraphistry', author='Graphistry', author_email='xin@graphistry.com', license='MIT', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Graphistry User', 'Topic :: Data Visualization Development :: Load Tools', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='Python Data Loader', packages=find_packages(exclude=['contrib', 'docs', 'tests*']), )
<commit_before>"""A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f: long_description = f.read() setup( name='Graphistry', version='1.1.0.dev1', description='This is established as a Data Loader for Graphistry', long_description=long_description, url='https://github.com/graphistry/pygraphistry', author='Graphistry', author_email='xin@graphistry.com', license='MIT', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Graphistry User', 'Topic :: Data Visualization Development :: Load Tools', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='Python Data Loader', packages=find_packages(exclude=['contrib', 'docs', 'tests*']), ) <commit_msg>Fix license MIT to BSD<commit_after>"""A setuptools based setup module. See: https://packaging.python.org/en/latest/distributing.html https://github.com/pypa/sampleproject """ from setuptools import setup, find_packages from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f: long_description = f.read() setup( name='Graphistry', version='1.1.0.dev1', description='This is established as a Data Loader for Graphistry', long_description=long_description, url='https://github.com/graphistry/pygraphistry', author='Graphistry', author_email='xin@graphistry.com', license='MIT', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Graphistry User', 'Topic :: Data Visualization Development :: Load Tools', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], keywords='Python Data Loader', packages=find_packages(exclude=['contrib', 'docs', 'tests*']), )
0db9b203dddefa6576bbcf357453587f05403180
setup.py
setup.py
#!/usr/bin/env python # coding=utf-8 from setuptools import setup setup( name='alfred-workflow-packager', version='0.8b', description='A CLI utility for packaging and exporting Alfred workflows', url='https://github.com/caleb531/alfred-workflow-packager', author='Caleb Evans', author_email='caleb@calebevans.me', license='MIT', keywords='alfred workflow package export', packages=['awp'], install_requires=[ 'biplist >= 1, < 2', 'jsonschema >= 2, < 3' ], entry_points={ 'console_scripts': [ 'alfred-workflow-packager=awp.packager:main', 'workflow-packager=awp.packager:main' ] } )
#!/usr/bin/env python # coding=utf-8 from setuptools import setup setup( name='alfred-workflow-packager', version='0.9', description='A CLI utility for packaging and exporting Alfred workflows', url='https://github.com/caleb531/alfred-workflow-packager', author='Caleb Evans', author_email='caleb@calebevans.me', license='MIT', keywords='alfred workflow package export', packages=['awp'], install_requires=[ 'biplist >= 1, < 2', 'jsonschema >= 2, < 3' ], entry_points={ 'console_scripts': [ 'alfred-workflow-packager=awp.packager:main', 'workflow-packager=awp.packager:main' ] } )
Change initial version to 0.9
Change initial version to 0.9
Python
mit
caleb531/alfred-workflow-packager
#!/usr/bin/env python # coding=utf-8 from setuptools import setup setup( name='alfred-workflow-packager', version='0.8b', description='A CLI utility for packaging and exporting Alfred workflows', url='https://github.com/caleb531/alfred-workflow-packager', author='Caleb Evans', author_email='caleb@calebevans.me', license='MIT', keywords='alfred workflow package export', packages=['awp'], install_requires=[ 'biplist >= 1, < 2', 'jsonschema >= 2, < 3' ], entry_points={ 'console_scripts': [ 'alfred-workflow-packager=awp.packager:main', 'workflow-packager=awp.packager:main' ] } ) Change initial version to 0.9
#!/usr/bin/env python # coding=utf-8 from setuptools import setup setup( name='alfred-workflow-packager', version='0.9', description='A CLI utility for packaging and exporting Alfred workflows', url='https://github.com/caleb531/alfred-workflow-packager', author='Caleb Evans', author_email='caleb@calebevans.me', license='MIT', keywords='alfred workflow package export', packages=['awp'], install_requires=[ 'biplist >= 1, < 2', 'jsonschema >= 2, < 3' ], entry_points={ 'console_scripts': [ 'alfred-workflow-packager=awp.packager:main', 'workflow-packager=awp.packager:main' ] } )
<commit_before>#!/usr/bin/env python # coding=utf-8 from setuptools import setup setup( name='alfred-workflow-packager', version='0.8b', description='A CLI utility for packaging and exporting Alfred workflows', url='https://github.com/caleb531/alfred-workflow-packager', author='Caleb Evans', author_email='caleb@calebevans.me', license='MIT', keywords='alfred workflow package export', packages=['awp'], install_requires=[ 'biplist >= 1, < 2', 'jsonschema >= 2, < 3' ], entry_points={ 'console_scripts': [ 'alfred-workflow-packager=awp.packager:main', 'workflow-packager=awp.packager:main' ] } ) <commit_msg>Change initial version to 0.9<commit_after>
#!/usr/bin/env python # coding=utf-8 from setuptools import setup setup( name='alfred-workflow-packager', version='0.9', description='A CLI utility for packaging and exporting Alfred workflows', url='https://github.com/caleb531/alfred-workflow-packager', author='Caleb Evans', author_email='caleb@calebevans.me', license='MIT', keywords='alfred workflow package export', packages=['awp'], install_requires=[ 'biplist >= 1, < 2', 'jsonschema >= 2, < 3' ], entry_points={ 'console_scripts': [ 'alfred-workflow-packager=awp.packager:main', 'workflow-packager=awp.packager:main' ] } )
#!/usr/bin/env python # coding=utf-8 from setuptools import setup setup( name='alfred-workflow-packager', version='0.8b', description='A CLI utility for packaging and exporting Alfred workflows', url='https://github.com/caleb531/alfred-workflow-packager', author='Caleb Evans', author_email='caleb@calebevans.me', license='MIT', keywords='alfred workflow package export', packages=['awp'], install_requires=[ 'biplist >= 1, < 2', 'jsonschema >= 2, < 3' ], entry_points={ 'console_scripts': [ 'alfred-workflow-packager=awp.packager:main', 'workflow-packager=awp.packager:main' ] } ) Change initial version to 0.9#!/usr/bin/env python # coding=utf-8 from setuptools import setup setup( name='alfred-workflow-packager', version='0.9', description='A CLI utility for packaging and exporting Alfred workflows', url='https://github.com/caleb531/alfred-workflow-packager', author='Caleb Evans', author_email='caleb@calebevans.me', license='MIT', keywords='alfred workflow package export', packages=['awp'], install_requires=[ 'biplist >= 1, < 2', 'jsonschema >= 2, < 3' ], entry_points={ 'console_scripts': [ 'alfred-workflow-packager=awp.packager:main', 'workflow-packager=awp.packager:main' ] } )
<commit_before>#!/usr/bin/env python # coding=utf-8 from setuptools import setup setup( name='alfred-workflow-packager', version='0.8b', description='A CLI utility for packaging and exporting Alfred workflows', url='https://github.com/caleb531/alfred-workflow-packager', author='Caleb Evans', author_email='caleb@calebevans.me', license='MIT', keywords='alfred workflow package export', packages=['awp'], install_requires=[ 'biplist >= 1, < 2', 'jsonschema >= 2, < 3' ], entry_points={ 'console_scripts': [ 'alfred-workflow-packager=awp.packager:main', 'workflow-packager=awp.packager:main' ] } ) <commit_msg>Change initial version to 0.9<commit_after>#!/usr/bin/env python # coding=utf-8 from setuptools import setup setup( name='alfred-workflow-packager', version='0.9', description='A CLI utility for packaging and exporting Alfred workflows', url='https://github.com/caleb531/alfred-workflow-packager', author='Caleb Evans', author_email='caleb@calebevans.me', license='MIT', keywords='alfred workflow package export', packages=['awp'], install_requires=[ 'biplist >= 1, < 2', 'jsonschema >= 2, < 3' ], entry_points={ 'console_scripts': [ 'alfred-workflow-packager=awp.packager:main', 'workflow-packager=awp.packager:main' ] } )
41de24fa992939d8cf3daf762bd7c3235faeb5d0
setup.py
setup.py
import os from setuptools import setup # type: ignore VERSION = '4.5' setup( name='conllu', packages=["conllu"], python_requires=">=3.6", package_data={ "": ["py.typed"] }, version=VERSION, license='MIT License', description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), long_description_content_type="text/markdown", author=u'Emil Stenström', author_email="emil@emilstenstrom.se", url='https://github.com/EmilStenstrom/conllu/', keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Operating System :: OS Independent", ], )
import os from setuptools import setup # type: ignore VERSION = '4.5' with open(os.path.join(os.path.dirname(__file__), 'README.md')) as f: description = f.read() setup( name='conllu', packages=["conllu"], python_requires=">=3.6", package_data={ "": ["py.typed"] }, version=VERSION, license='MIT License', description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary', long_description=description, long_description_content_type="text/markdown", author=u'Emil Stenström', author_email="emil@emilstenstrom.se", url='https://github.com/EmilStenstrom/conllu/', keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Operating System :: OS Independent", ], )
Make sure file is closed properly.
Make sure file is closed properly.
Python
mit
EmilStenstrom/conllu
import os from setuptools import setup # type: ignore VERSION = '4.5' setup( name='conllu', packages=["conllu"], python_requires=">=3.6", package_data={ "": ["py.typed"] }, version=VERSION, license='MIT License', description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), long_description_content_type="text/markdown", author=u'Emil Stenström', author_email="emil@emilstenstrom.se", url='https://github.com/EmilStenstrom/conllu/', keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Operating System :: OS Independent", ], ) Make sure file is closed properly.
import os from setuptools import setup # type: ignore VERSION = '4.5' with open(os.path.join(os.path.dirname(__file__), 'README.md')) as f: description = f.read() setup( name='conllu', packages=["conllu"], python_requires=">=3.6", package_data={ "": ["py.typed"] }, version=VERSION, license='MIT License', description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary', long_description=description, long_description_content_type="text/markdown", author=u'Emil Stenström', author_email="emil@emilstenstrom.se", url='https://github.com/EmilStenstrom/conllu/', keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Operating System :: OS Independent", ], )
<commit_before>import os from setuptools import setup # type: ignore VERSION = '4.5' setup( name='conllu', packages=["conllu"], python_requires=">=3.6", package_data={ "": ["py.typed"] }, version=VERSION, license='MIT License', description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), long_description_content_type="text/markdown", author=u'Emil Stenström', author_email="emil@emilstenstrom.se", url='https://github.com/EmilStenstrom/conllu/', keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Operating System :: OS Independent", ], ) <commit_msg>Make sure file is closed properly.<commit_after>
import os from setuptools import setup # type: ignore VERSION = '4.5' with open(os.path.join(os.path.dirname(__file__), 'README.md')) as f: description = f.read() setup( name='conllu', packages=["conllu"], python_requires=">=3.6", package_data={ "": ["py.typed"] }, version=VERSION, license='MIT License', description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary', long_description=description, long_description_content_type="text/markdown", author=u'Emil Stenström', author_email="emil@emilstenstrom.se", url='https://github.com/EmilStenstrom/conllu/', keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Operating System :: OS Independent", ], )
import os from setuptools import setup # type: ignore VERSION = '4.5' setup( name='conllu', packages=["conllu"], python_requires=">=3.6", package_data={ "": ["py.typed"] }, version=VERSION, license='MIT License', description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), long_description_content_type="text/markdown", author=u'Emil Stenström', author_email="emil@emilstenstrom.se", url='https://github.com/EmilStenstrom/conllu/', keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Operating System :: OS Independent", ], ) Make sure file is closed properly.import os from setuptools import setup # type: ignore VERSION = '4.5' with open(os.path.join(os.path.dirname(__file__), 'README.md')) as f: description = f.read() setup( name='conllu', packages=["conllu"], python_requires=">=3.6", package_data={ "": ["py.typed"] }, version=VERSION, license='MIT License', description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary', long_description=description, long_description_content_type="text/markdown", author=u'Emil Stenström', author_email="emil@emilstenstrom.se", url='https://github.com/EmilStenstrom/conllu/', keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Operating System :: OS Independent", ], )
<commit_before>import os from setuptools import setup # type: ignore VERSION = '4.5' setup( name='conllu', packages=["conllu"], python_requires=">=3.6", package_data={ "": ["py.typed"] }, version=VERSION, license='MIT License', description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), long_description_content_type="text/markdown", author=u'Emil Stenström', author_email="emil@emilstenstrom.se", url='https://github.com/EmilStenstrom/conllu/', keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Operating System :: OS Independent", ], ) <commit_msg>Make sure file is closed properly.<commit_after>import os from setuptools import setup # type: ignore VERSION = '4.5' with open(os.path.join(os.path.dirname(__file__), 'README.md')) as f: description = f.read() setup( name='conllu', packages=["conllu"], python_requires=">=3.6", package_data={ "": ["py.typed"] }, version=VERSION, license='MIT License', description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary', long_description=description, long_description_content_type="text/markdown", author=u'Emil Stenström', author_email="emil@emilstenstrom.se", url='https://github.com/EmilStenstrom/conllu/', keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Operating System :: OS Independent", ], )
857133125fe5a633f209887b1b739f6f82c80f43
setup.py
setup.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup with open('README.rst') as readme_file: readme = readme_file.read() with open('HISTORY.rst') as history_file: history = history_file.read() setup( name='learnregex', version='0.4.2', description='A pyschool story for learning regular expressions.', long_description=readme + '\n\n' + history, author='Sophilabs', author_email='hi@sophilabs.co', url='https://github.com/sophilabs/learnregex', packages=['learnregex'], entry_points={ 'console_scripts': [ 'learnregex=learnregex.story:Story.begin' ] }, include_package_data=True, install_requires=['story'], license='MIT license', zip_safe=False, keywords='learnregex', classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests', tests_require=[] )
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup with open('README.rst') as readme_file: readme = readme_file.read() with open('HISTORY.rst') as history_file: history = history_file.read() setup( name='learnregex', version='0.4.2', description='A pyschool story for learning regular expressions.', long_description=readme + '\n\n' + history, author='Sophilabs', author_email='hi@sophilabs.co', url='https://github.com/sophilabs/learnregex', packages=['learnregex'], entry_points={ 'console_scripts': [ 'learnregex=learnregex.story:Story.begin' ] }, include_package_data=True, install_requires=[ 'story>=1.1.2' ], license='MIT license', zip_safe=False, keywords='learnregex', classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests', tests_require=[] )
Update mininum story version required.
Update mininum story version required.
Python
mit
sophilabs/learnregex
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup with open('README.rst') as readme_file: readme = readme_file.read() with open('HISTORY.rst') as history_file: history = history_file.read() setup( name='learnregex', version='0.4.2', description='A pyschool story for learning regular expressions.', long_description=readme + '\n\n' + history, author='Sophilabs', author_email='hi@sophilabs.co', url='https://github.com/sophilabs/learnregex', packages=['learnregex'], entry_points={ 'console_scripts': [ 'learnregex=learnregex.story:Story.begin' ] }, include_package_data=True, install_requires=['story'], license='MIT license', zip_safe=False, keywords='learnregex', classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests', tests_require=[] ) Update mininum story version required.
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup with open('README.rst') as readme_file: readme = readme_file.read() with open('HISTORY.rst') as history_file: history = history_file.read() setup( name='learnregex', version='0.4.2', description='A pyschool story for learning regular expressions.', long_description=readme + '\n\n' + history, author='Sophilabs', author_email='hi@sophilabs.co', url='https://github.com/sophilabs/learnregex', packages=['learnregex'], entry_points={ 'console_scripts': [ 'learnregex=learnregex.story:Story.begin' ] }, include_package_data=True, install_requires=[ 'story>=1.1.2' ], license='MIT license', zip_safe=False, keywords='learnregex', classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests', tests_require=[] )
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup with open('README.rst') as readme_file: readme = readme_file.read() with open('HISTORY.rst') as history_file: history = history_file.read() setup( name='learnregex', version='0.4.2', description='A pyschool story for learning regular expressions.', long_description=readme + '\n\n' + history, author='Sophilabs', author_email='hi@sophilabs.co', url='https://github.com/sophilabs/learnregex', packages=['learnregex'], entry_points={ 'console_scripts': [ 'learnregex=learnregex.story:Story.begin' ] }, include_package_data=True, install_requires=['story'], license='MIT license', zip_safe=False, keywords='learnregex', classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests', tests_require=[] ) <commit_msg>Update mininum story version required.<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup with open('README.rst') as readme_file: readme = readme_file.read() with open('HISTORY.rst') as history_file: history = history_file.read() setup( name='learnregex', version='0.4.2', description='A pyschool story for learning regular expressions.', long_description=readme + '\n\n' + history, author='Sophilabs', author_email='hi@sophilabs.co', url='https://github.com/sophilabs/learnregex', packages=['learnregex'], entry_points={ 'console_scripts': [ 'learnregex=learnregex.story:Story.begin' ] }, include_package_data=True, install_requires=[ 'story>=1.1.2' ], license='MIT license', zip_safe=False, keywords='learnregex', classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests', tests_require=[] )
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup with open('README.rst') as readme_file: readme = readme_file.read() with open('HISTORY.rst') as history_file: history = history_file.read() setup( name='learnregex', version='0.4.2', description='A pyschool story for learning regular expressions.', long_description=readme + '\n\n' + history, author='Sophilabs', author_email='hi@sophilabs.co', url='https://github.com/sophilabs/learnregex', packages=['learnregex'], entry_points={ 'console_scripts': [ 'learnregex=learnregex.story:Story.begin' ] }, include_package_data=True, install_requires=['story'], license='MIT license', zip_safe=False, keywords='learnregex', classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests', tests_require=[] ) Update mininum story version required.#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup with open('README.rst') as readme_file: readme = readme_file.read() with open('HISTORY.rst') as history_file: history = history_file.read() setup( name='learnregex', version='0.4.2', description='A pyschool story for learning regular expressions.', long_description=readme + '\n\n' + history, author='Sophilabs', author_email='hi@sophilabs.co', url='https://github.com/sophilabs/learnregex', packages=['learnregex'], entry_points={ 'console_scripts': [ 'learnregex=learnregex.story:Story.begin' ] }, include_package_data=True, install_requires=[ 'story>=1.1.2' ], license='MIT license', zip_safe=False, keywords='learnregex', classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests', tests_require=[] )
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup with open('README.rst') as readme_file: readme = readme_file.read() with open('HISTORY.rst') as history_file: history = history_file.read() setup( name='learnregex', version='0.4.2', description='A pyschool story for learning regular expressions.', long_description=readme + '\n\n' + history, author='Sophilabs', author_email='hi@sophilabs.co', url='https://github.com/sophilabs/learnregex', packages=['learnregex'], entry_points={ 'console_scripts': [ 'learnregex=learnregex.story:Story.begin' ] }, include_package_data=True, install_requires=['story'], license='MIT license', zip_safe=False, keywords='learnregex', classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests', tests_require=[] ) <commit_msg>Update mininum story version required.<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup with open('README.rst') as readme_file: readme = readme_file.read() with open('HISTORY.rst') as history_file: history = history_file.read() setup( name='learnregex', version='0.4.2', description='A pyschool story for learning regular expressions.', long_description=readme + '\n\n' + history, author='Sophilabs', author_email='hi@sophilabs.co', url='https://github.com/sophilabs/learnregex', packages=['learnregex'], entry_points={ 'console_scripts': [ 'learnregex=learnregex.story:Story.begin' ] }, include_package_data=True, install_requires=[ 'story>=1.1.2' ], license='MIT license', zip_safe=False, keywords='learnregex', classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests', tests_require=[] )