commit stringlengths 40 40 | old_file stringlengths 4 150 | new_file stringlengths 4 150 | old_contents stringlengths 0 3.26k | new_contents stringlengths 1 4.43k | subject stringlengths 15 501 | message stringlengths 15 4.06k | lang stringclasses 4 values | license stringclasses 13 values | repos stringlengths 5 91.5k | diff stringlengths 0 4.35k |
|---|---|---|---|---|---|---|---|---|---|---|
6547d653491adb6ab46e4a3a5f8251129719d3f7 | login/middleware.py | login/middleware.py | from django.conf import settings
from django.http import HttpResponseRedirect
DETACH_PATH = '/user/detach'
ACTIVATE_PATH = '/user/activate'
class DetachMiddleware(object):
def process_request(self, request):
if not request.path == '/login/' \
and not request.path.startswith('/api'):
if not request.user.is_anonymous \
and not request.user.is_native \
and not request.path == DETACH_PATH:
return HttpResponseRedirect(DETACH_PATH)
if not request.user.is_anonymous \
and not request.user.is_mail_verified \
and not request.path.startswith(ACTIVATE_PATH):
return HttpResponseRedirect(ACTIVATE_PATH) | from django.conf import settings
from django.http import HttpResponseRedirect
DETACH_PATH = '/user/detach'
ACTIVATE_PATH = '/user/activate'
class DetachMiddleware(object):
def process_request(self, request):
if not request.path == '/login/' \
and not request.path.startswith('/api') \
and not request.user.is_anonymous:
if not request.user.is_native:
if not request.path == DETACH_PATH:
return HttpResponseRedirect(DETACH_PATH)
elif not request.user.is_mail_verified \
and not request.path.startswith(ACTIVATE_PATH):
return HttpResponseRedirect(ACTIVATE_PATH) | Remove infinite loop if user is neither native nor verified | Remove infinite loop if user is neither native nor verified
| Python | agpl-3.0 | openego/oeplatform,openego/oeplatform,openego/oeplatform,openego/oeplatform | ---
+++
@@ -7,14 +7,12 @@
class DetachMiddleware(object):
def process_request(self, request):
if not request.path == '/login/' \
- and not request.path.startswith('/api'):
+ and not request.path.startswith('/api') \
+ and not request.user.is_anonymous:
- if not request.user.is_anonymous \
- and not request.user.is_native \
- and not request.path == DETACH_PATH:
- return HttpResponseRedirect(DETACH_PATH)
-
- if not request.user.is_anonymous \
- and not request.user.is_mail_verified \
+ if not request.user.is_native:
+ if not request.path == DETACH_PATH:
+ return HttpResponseRedirect(DETACH_PATH)
+ elif not request.user.is_mail_verified \
and not request.path.startswith(ACTIVATE_PATH):
return HttpResponseRedirect(ACTIVATE_PATH) |
de6de7c43932224f2845cc540170eeaf46650491 | mkerefuse/refuse.py | mkerefuse/refuse.py | from .util import XPathObject
class RefusePickup(XPathObject):
"""Defines attribute to XPath specification matching"""
input_properties = {
'success_msg': '//*[@id="nConf"]/h1',
'route_garbage': '//*[@id="nConf"]/strong[1]',
'next_pickup_garbage': '//*[@id="nConf"]/strong[2]',
'route_recyle': '//*[@id="nConf"]/strong[3]',
'next_pickup_recycle_after': '//*[@id="nConf"]/strong[4]',
'next_pickup_recycle_before': '//*[@id="nConf"]/strong[5]',
}
"""Maps the key to an attr name & value to an XPath lookup"""
| Add RefusePickup class (for parsing) | Add RefusePickup class (for parsing)
| Python | unlicense | tomislacker/python-mke-trash-pickup,tomislacker/python-mke-trash-pickup | ---
+++
@@ -1 +1,15 @@
+from .util import XPathObject
+
+class RefusePickup(XPathObject):
+ """Defines attribute to XPath specification matching"""
+
+ input_properties = {
+ 'success_msg': '//*[@id="nConf"]/h1',
+ 'route_garbage': '//*[@id="nConf"]/strong[1]',
+ 'next_pickup_garbage': '//*[@id="nConf"]/strong[2]',
+ 'route_recyle': '//*[@id="nConf"]/strong[3]',
+ 'next_pickup_recycle_after': '//*[@id="nConf"]/strong[4]',
+ 'next_pickup_recycle_before': '//*[@id="nConf"]/strong[5]',
+ }
+ """Maps the key to an attr name & value to an XPath lookup""" | |
016f14304b6b86634c4608927d3345f993178682 | config.py | config.py | ###
# Copyright (c) 2012, spline
# All rights reserved.
#
#
###
import supybot.conf as conf
import supybot.registry as registry
from supybot.i18n import PluginInternationalization, internationalizeDocstring
_ = PluginInternationalization('Scores')
def configure(advanced):
# This will be called by supybot to configure this module. advanced is
# a bool that specifies whether the user identified himself as an advanced
# user or not. You should effect your configuration by manipulating the
# registry as appropriate.
from supybot.questions import expect, anything, something, yn
conf.registerPlugin('Scores', True)
Scores = conf.registerPlugin('Scores')
# This is where your configuration variables (if any) should go. For example:
# conf.registerGlobalValue(Scores, 'someConfigVariableName',
# registry.Boolean(False, _("""Help for someConfigVariableName.""")))
# vim:set shiftwidth=4 tabstop=4 expandtab textwidth=79:
| ###
# Copyright (c) 2012, spline
# All rights reserved.
#
#
###
import supybot.conf as conf
import supybot.registry as registry
from supybot.i18n import PluginInternationalization, internationalizeDocstring
_ = PluginInternationalization('Scores')
def configure(advanced):
# This will be called by supybot to configure this module. advanced is
# a bool that specifies whether the user identified himself as an advanced
# user or not. You should effect your configuration by manipulating the
# registry as appropriate.
from supybot.questions import expect, anything, something, yn
conf.registerPlugin('Scores', True)
Scores = conf.registerPlugin('Scores')
conf.registerChannelValue(Scores, 'disableANSI', registry.Boolean(False, """Do not display any ANSI (color/bold) in output."""))
# vim:set shiftwidth=4 tabstop=4 expandtab textwidth=250:
| Add in channel value for disabling ansi | Add in channel value for disabling ansi
| Python | mit | reticulatingspline/Scores,cottongin/Scores | ---
+++
@@ -21,9 +21,7 @@
Scores = conf.registerPlugin('Scores')
-# This is where your configuration variables (if any) should go. For example:
-# conf.registerGlobalValue(Scores, 'someConfigVariableName',
-# registry.Boolean(False, _("""Help for someConfigVariableName.""")))
+conf.registerChannelValue(Scores, 'disableANSI', registry.Boolean(False, """Do not display any ANSI (color/bold) in output."""))
-# vim:set shiftwidth=4 tabstop=4 expandtab textwidth=79:
+# vim:set shiftwidth=4 tabstop=4 expandtab textwidth=250: |
61fe55efba2c491da6a93421fa702f123615bc32 | spacy/lang/en/__init__.py | spacy/lang/en/__init__.py | # coding: utf8
from __future__ import unicode_literals
from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS
from .tag_map import TAG_MAP
from .stop_words import STOP_WORDS
from .lex_attrs import LEX_ATTRS
from .morph_rules import MORPH_RULES
from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC
from .syntax_iterators import SYNTAX_ITERATORS
from ..tokenizer_exceptions import BASE_EXCEPTIONS
from ...language import Language
from ...attrs import LANG
from ...util import update_exc
class English(Language):
lang = 'en'
class Defaults(Language.Defaults):
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
lex_attr_getters[LANG] = lambda text: 'en'
lex_attr_getters.update(LEX_ATTRS)
tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS)
tag_map = dict(TAG_MAP)
stop_words = set(STOP_WORDS)
morph_rules = dict(MORPH_RULES)
lemma_rules = dict(LEMMA_RULES)
lemma_index = dict(LEMMA_INDEX)
lemma_exc = dict(LEMMA_EXC)
sytax_iterators = dict(SYNTAX_ITERATORS)
__all__ = ['English']
| # coding: utf8
from __future__ import unicode_literals
from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS
from .tag_map import TAG_MAP
from .stop_words import STOP_WORDS
from .lex_attrs import LEX_ATTRS
from .morph_rules import MORPH_RULES
from .lemmatizer import LEMMA_RULES, LEMMA_INDEX, LEMMA_EXC
from .syntax_iterators import SYNTAX_ITERATORS
from ..tokenizer_exceptions import BASE_EXCEPTIONS
from ...language import Language
from ...attrs import LANG
from ...util import update_exc
class EnglishDefaults(Language.Defaults):
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
lex_attr_getters[LANG] = lambda text: 'en'
lex_attr_getters.update(LEX_ATTRS)
tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS)
tag_map = dict(TAG_MAP)
stop_words = set(STOP_WORDS)
morph_rules = dict(MORPH_RULES)
lemma_rules = dict(LEMMA_RULES)
lemma_index = dict(LEMMA_INDEX)
lemma_exc = dict(LEMMA_EXC)
sytax_iterators = dict(SYNTAX_ITERATORS)
class English(Language):
lang = 'en'
Defaults = EnglishDefaults
__all__ = ['English', 'EnglishDefaults']
| Move EnglishDefaults class out of English | Move EnglishDefaults class out of English
| Python | mit | honnibal/spaCy,aikramer2/spaCy,recognai/spaCy,recognai/spaCy,honnibal/spaCy,aikramer2/spaCy,explosion/spaCy,aikramer2/spaCy,recognai/spaCy,aikramer2/spaCy,spacy-io/spaCy,recognai/spaCy,spacy-io/spaCy,explosion/spaCy,explosion/spaCy,explosion/spaCy,honnibal/spaCy,spacy-io/spaCy,recognai/spaCy,aikramer2/spaCy,recognai/spaCy,spacy-io/spaCy,honnibal/spaCy,explosion/spaCy,aikramer2/spaCy,explosion/spaCy,spacy-io/spaCy,spacy-io/spaCy | ---
+++
@@ -15,22 +15,25 @@
from ...util import update_exc
+class EnglishDefaults(Language.Defaults):
+ lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
+ lex_attr_getters[LANG] = lambda text: 'en'
+ lex_attr_getters.update(LEX_ATTRS)
+
+ tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS)
+ tag_map = dict(TAG_MAP)
+ stop_words = set(STOP_WORDS)
+ morph_rules = dict(MORPH_RULES)
+ lemma_rules = dict(LEMMA_RULES)
+ lemma_index = dict(LEMMA_INDEX)
+ lemma_exc = dict(LEMMA_EXC)
+ sytax_iterators = dict(SYNTAX_ITERATORS)
+
+
class English(Language):
lang = 'en'
- class Defaults(Language.Defaults):
- lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
- lex_attr_getters[LANG] = lambda text: 'en'
- lex_attr_getters.update(LEX_ATTRS)
-
- tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS)
- tag_map = dict(TAG_MAP)
- stop_words = set(STOP_WORDS)
- morph_rules = dict(MORPH_RULES)
- lemma_rules = dict(LEMMA_RULES)
- lemma_index = dict(LEMMA_INDEX)
- lemma_exc = dict(LEMMA_EXC)
- sytax_iterators = dict(SYNTAX_ITERATORS)
+ Defaults = EnglishDefaults
-__all__ = ['English']
+__all__ = ['English', 'EnglishDefaults'] |
d6cd7c16e32f64c4fd3627953d751e8c8bc26f1c | premis_event_service/forms.py | premis_event_service/forms.py | from django import forms
import settings
OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES
EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES
class EventSearchForm(forms.Form):
event_outcome = forms.ChoiceField(
widget=forms.Select(
attrs={
'id': 'prependedInput',
'class': 'input-small',
}
),
choices=OUTCOME_CHOICES,
required=False)
event_type = forms.ChoiceField(
widget=forms.Select(
attrs={
'id': 'prependedInput',
'class': 'input-medium',
}
),
choices=EVENT_TYPE_CHOICES,
required=False)
start_date = forms.DateField(
widget=forms.DateInput(
attrs={
'id': 'startdatepicker',
'placeholder': 'Start Date',
'class': 'input-small',
}
),
required=False)
end_date = forms.DateField(
widget=forms.DateInput(
attrs={
'id': 'enddatepicker',
'placeholder': 'End Date',
'class': 'input-small',
}
),
required=False)
linked_object_id = forms.CharField(
widget=forms.TextInput(
attrs={
'placeholder': 'Linked Object ID',
'class': 'input-medium',
}
),
max_length=20,
required=False,
)
| from django import forms
import settings
OUTCOME_CHOICES = settings.EVENT_OUTCOME_CHOICES
EVENT_TYPE_CHOICES = settings.EVENT_TYPE_CHOICES
class EventSearchForm(forms.Form):
event_outcome = forms.ChoiceField(
widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-small'}),
choices=OUTCOME_CHOICES,
required=False)
event_type = forms.ChoiceField(
widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-medium'}),
choices=EVENT_TYPE_CHOICES,
required=False)
start_date = forms.DateField(
widget=forms.DateInput(attrs={'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small'}), # noqa
required=False)
end_date = forms.DateField(
widget=forms.DateInput(attrs={'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small'}), # noqa
required=False)
linked_object_id = forms.CharField(
widget=forms.TextInput(attrs={'placeholder': 'Linked Object ID', 'class': 'input-medium'}),
max_length=20,
required=False)
| Fix formatting for the EventSearchForm class. | Fix formatting for the EventSearchForm class.
| Python | bsd-3-clause | unt-libraries/django-premis-event-service,unt-libraries/django-premis-event-service,unt-libraries/django-premis-event-service | ---
+++
@@ -8,52 +8,25 @@
class EventSearchForm(forms.Form):
event_outcome = forms.ChoiceField(
- widget=forms.Select(
- attrs={
- 'id': 'prependedInput',
- 'class': 'input-small',
- }
- ),
+ widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-small'}),
choices=OUTCOME_CHOICES,
required=False)
event_type = forms.ChoiceField(
- widget=forms.Select(
- attrs={
- 'id': 'prependedInput',
- 'class': 'input-medium',
- }
- ),
+ widget=forms.Select(attrs={'id': 'prependedInput', 'class': 'input-medium'}),
choices=EVENT_TYPE_CHOICES,
required=False)
start_date = forms.DateField(
- widget=forms.DateInput(
- attrs={
- 'id': 'startdatepicker',
- 'placeholder': 'Start Date',
- 'class': 'input-small',
- }
- ),
+ widget=forms.DateInput(attrs={'id': 'startdatepicker', 'placeholder': 'Start Date', 'class': 'input-small'}), # noqa
required=False)
end_date = forms.DateField(
- widget=forms.DateInput(
- attrs={
- 'id': 'enddatepicker',
- 'placeholder': 'End Date',
- 'class': 'input-small',
- }
- ),
+ widget=forms.DateInput(attrs={'id': 'enddatepicker', 'placeholder': 'End Date', 'class': 'input-small'}), # noqa
required=False)
linked_object_id = forms.CharField(
- widget=forms.TextInput(
- attrs={
- 'placeholder': 'Linked Object ID',
- 'class': 'input-medium',
- }
- ),
+ widget=forms.TextInput(attrs={'placeholder': 'Linked Object ID', 'class': 'input-medium'}),
max_length=20,
- required=False,
- )
+ required=False)
+ |
585fe3142860a294a6b30edb9879f979b6ac9c49 | apex/lib/db.py | apex/lib/db.py | from pyramid.httpexceptions import HTTPNotFound
from pyramid.i18n import TranslationString as _
def get_or_create(session, model, **kw):
""" Django's get_or_create function
http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create
"""
obj = session.query(model).filter_by(**kw).first()
if obj:
return obj
else:
obj = model(**kw)
session.add(obj)
session.flush()
return obj
def get_object_or_404(session, model, **kw):
""" Django's get_object_or_404 function
"""
obj = session.query(model).filter_by(**kw).first()
if obj is None:
raise HTTPNotFound(detail=_('No %s matches the given query.') % model.__name__)
return obj
def merge_session_with_post(session, post):
""" Basic function to merge data into an sql object.
This function doesn't work with relations.
"""
for key, value in post:
setattr(session, key, value)
return session
| from pyramid.httpexceptions import HTTPNotFound
from pyramid.i18n import TranslationString as _
def get_or_create(session, model, **kw):
""" Django's get_or_create function
http://stackoverflow.com/questions/2546207/does-sqlalchemy-have-an-equivalent-of-djangos-get-or-create
"""
obj = session.query(model).filter_by(**kw).first()
if obj:
return obj
else:
obj = model(**kw)
session.add(obj)
session.flush()
return obj
def get_object_or_404(session, model, **kw):
""" Django's get_object_or_404 function
"""
obj = session.query(model).filter_by(**kw).first()
if obj is None:
raise HTTPNotFound(detail='No %s matches the given query.' % model.__name__)
return obj
def merge_session_with_post(session, post):
""" Basic function to merge data into an sql object.
This function doesn't work with relations.
"""
for key, value in post:
setattr(session, key, value)
return session
| Fix issue with wsgi and translations. | Fix issue with wsgi and translations. | Python | mit | Qwait/apex,Qwait/apex | ---
+++
@@ -20,7 +20,7 @@
"""
obj = session.query(model).filter_by(**kw).first()
if obj is None:
- raise HTTPNotFound(detail=_('No %s matches the given query.') % model.__name__)
+ raise HTTPNotFound(detail='No %s matches the given query.' % model.__name__)
return obj
def merge_session_with_post(session, post): |
2937af0fe2f28ed9381b6b43c337c4cca14e4e78 | apps/polls/admin.py | apps/polls/admin.py | from django.contrib import admin
from apps.polls.models import Poll, Choice
class ChoiceInline(admin.TabularInline):
model = Choice
extra = 3
class PollAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['question']}),
('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}),
]
inlines = [ChoiceInline]
list_display = ('question', 'pub_date', 'was_published_recently')
list_filter = ['pub_date']
admin.site.register(Poll, PollAdmin)
admin.site.register(Choice) | from django.contrib import admin
from apps.polls.models import Poll, Choice
class ChoiceInline(admin.TabularInline):
model = Choice
extra = 3
class PollAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['question']}),
('Date information', {'fields': ['pub_date'], 'classes': ['collapse']}),
]
inlines = [ChoiceInline]
list_display = ('question', 'pub_date', 'was_published_recently')
list_filter = ['pub_date']
search_fields = ['question']
admin.site.register(Poll, PollAdmin)
admin.site.register(Choice) | Add search_fields = ['question'] to PollAdmin | Add search_fields = ['question'] to PollAdmin
| Python | bsd-3-clause | datphan/teracy-tutorial,teracyhq/django-tutorial | ---
+++
@@ -14,6 +14,7 @@
inlines = [ChoiceInline]
list_display = ('question', 'pub_date', 'was_published_recently')
list_filter = ['pub_date']
+ search_fields = ['question']
admin.site.register(Poll, PollAdmin) |
f50589ec9b61efbd2bd56cca802ffc542f5b3336 | pyrene/constants.py | pyrene/constants.py | class REPO:
'''Repo attributes'''
TYPE = 'type'
DIRECTORY = 'directory'
VOLATILE = 'volatile'
SERVE_INTERFACE = 'serve_interface'
SERVE_PORT = 'serve_port'
SERVE_USERNAME = 'serve_username'
SERVE_PASSWORD = 'serve_password'
USERNAME = 'username'
PASSWORD = 'password'
DOWNLOAD_URL = 'download_url'
UPLOAD_URL = 'upload_url'
class REPOTYPE:
'''Values for REPO.TYPE'''
DIRECTORY = 'directory'
HTTP = 'http'
| class REPO:
'''Repo attributes'''
TYPE = 'type'
DIRECTORY = 'directory'
VOLATILE = 'volatile'
SERVE_INTERFACE = 'interface'
SERVE_PORT = 'port'
SERVE_USERNAME = 'username'
SERVE_PASSWORD = 'password'
USERNAME = 'username'
PASSWORD = 'password'
DOWNLOAD_URL = 'download_url'
UPLOAD_URL = 'upload_url'
class REPOTYPE:
'''Values for REPO.TYPE'''
DIRECTORY = 'directory'
HTTP = 'http'
| Revert "make REPO.SERVE_* attributes distinct from other attributes (username, password)" | Revert "make REPO.SERVE_* attributes distinct from other attributes (username, password)"
This reverts commit 1553f4bae5f315666fac5ad9f6600ba8b076a84b.
| Python | mit | krisztianfekete/pyrene | ---
+++
@@ -4,10 +4,10 @@
DIRECTORY = 'directory'
VOLATILE = 'volatile'
- SERVE_INTERFACE = 'serve_interface'
- SERVE_PORT = 'serve_port'
- SERVE_USERNAME = 'serve_username'
- SERVE_PASSWORD = 'serve_password'
+ SERVE_INTERFACE = 'interface'
+ SERVE_PORT = 'port'
+ SERVE_USERNAME = 'username'
+ SERVE_PASSWORD = 'password'
USERNAME = 'username'
PASSWORD = 'password' |
2973b664e8c9cf551d5d7277ab4995125be5fad0 | python/reference.py | python/reference.py | import os
# Current directory
# If you call this from the current directory without abspath,
# then it will not work since __file__ is a relative path
os.path.dirname(os.path.abspath(__file__))
| import os
# Current directory
# If you call this from the current directory without abspath,
# then it will not work since __file__ is a relative path
os.path.dirname(os.path.abspath(__file__))
# Get all files in a directory
# Never use os.walk again
def all_sub_files(root):
for path, subdirs, files in os.walk(root):
for name in files:
yield os.path.join(path, name)
| Add util for not using os.walk | Add util for not using os.walk
| Python | mit | brycepg/how-to | ---
+++
@@ -3,3 +3,10 @@
# If you call this from the current directory without abspath,
# then it will not work since __file__ is a relative path
os.path.dirname(os.path.abspath(__file__))
+
+# Get all files in a directory
+# Never use os.walk again
+def all_sub_files(root):
+ for path, subdirs, files in os.walk(root):
+ for name in files:
+ yield os.path.join(path, name) |
38dfe0948b063b93f5375e065129838335d5ca95 | raiden/constants.py | raiden/constants.py | # -*- coding: utf-8 -*-
UINT64_MAX = 2 ** 64 - 1
UINT64_MIN = 0
INT64_MAX = 2 ** 63 - 1
INT64_MIN = -(2 ** 63)
UINT256_MAX = 2 ** 256 - 1
# Deployed to Ropsten revival on 2017-06-19 from commit 2677298a69c1b1f35b9ab26beafe457acfdcc0ee
ROPSTEN_REGISTRY_ADDRESS = 'aff1f958c69a6820b08a02549ff9041629ae8257'
ROPSTEN_DISCOVERY_ADDRESS = 'cf56165f4f6e8ec38bb463854c1fe28a5d320f4f'
DISCOVERY_REGISTRATION_GAS = 500000
MINUTE_SEC = 60
MINUTE_MS = 60 * 1000
NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6
# TODO: add this as an attribute of the transport class
UDP_MAX_MESSAGE_SIZE = 1200
| # -*- coding: utf-8 -*-
UINT64_MAX = 2 ** 64 - 1
UINT64_MIN = 0
INT64_MAX = 2 ** 63 - 1
INT64_MIN = -(2 ** 63)
UINT256_MAX = 2 ** 256 - 1
# Deployed to Ropsten revival on 2017-08-03 from commit 17aa7671159779ceef22fe90001970bed0685c4d
ROPSTEN_REGISTRY_ADDRESS = '25926b6d29f56ba8466601d7ce7dd29985af1f14'
ROPSTEN_DISCOVERY_ADDRESS = 'a647336e0b4eaddcee11fe6c13e9163ad2700d6d'
DISCOVERY_REGISTRATION_GAS = 500000
MINUTE_SEC = 60
MINUTE_MS = 60 * 1000
NETTINGCHANNEL_SETTLE_TIMEOUT_MIN = 6
# TODO: add this as an attribute of the transport class
UDP_MAX_MESSAGE_SIZE = 1200
| Update default smart contract addresses | Update default smart contract addresses
| Python | mit | tomashaber/raiden,tomashaber/raiden,hackaugusto/raiden,tomashaber/raiden,hackaugusto/raiden,tomashaber/raiden,tomashaber/raiden | ---
+++
@@ -8,9 +8,9 @@
UINT256_MAX = 2 ** 256 - 1
-# Deployed to Ropsten revival on 2017-06-19 from commit 2677298a69c1b1f35b9ab26beafe457acfdcc0ee
-ROPSTEN_REGISTRY_ADDRESS = 'aff1f958c69a6820b08a02549ff9041629ae8257'
-ROPSTEN_DISCOVERY_ADDRESS = 'cf56165f4f6e8ec38bb463854c1fe28a5d320f4f'
+# Deployed to Ropsten revival on 2017-08-03 from commit 17aa7671159779ceef22fe90001970bed0685c4d
+ROPSTEN_REGISTRY_ADDRESS = '25926b6d29f56ba8466601d7ce7dd29985af1f14'
+ROPSTEN_DISCOVERY_ADDRESS = 'a647336e0b4eaddcee11fe6c13e9163ad2700d6d'
DISCOVERY_REGISTRATION_GAS = 500000
|
ab1893301736e2cf5cd87cc30671898884377661 | test/test_events.py | test/test_events.py | import os
import sys
import unittest
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
from smartbot import events
class TestEvents(unittest.TestCase):
def test_empty(self):
event = events.Event()
self.assertEqual(len(event.trigger()), 0)
def test_with_handlers(self):
event = events.Event()
event.register(lambda: None)
self.assertEqual(len(event.trigger()), 1)
def test_custom_comparator(self):
comparator = lambda *args, **kwargs: False
event = events.Event(default_comparator=comparator)
event.register(lambda: None)
self.assertEqual(len(event.trigger()), 0)
event = events.Event()
event.register(lambda: None)
self.assertEqual(len(event.trigger(comparator=comparator)), 0)
def test_default_comparator(self):
event = events.Event()
event.register(lambda *args, **kwargs: None, a=10)
self.assertEqual(len(event.trigger()), 0)
self.assertEqual(len(event.trigger(a=10)), 1)
def test_decorator(self):
event = events.Event()
event()(lambda: None)
self.assertEqual(len(event.trigger()), 1)
| import os
import sys
import unittest
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
from smartbot import events
class TestEvents(unittest.TestCase):
def test_empty(self):
event = events.Event()
self.assertEqual(len(event.trigger()), 0)
def test_with_handlers(self):
event = events.Event()
event.register(lambda: None)
self.assertEqual(len(event.trigger()), 1)
def test_custom_comparator(self):
comparator = lambda *args, **kwargs: False
event = events.Event(default_comparator=comparator)
event.register(lambda: None)
self.assertEqual(len(event.trigger()), 0)
event = events.Event()
event.register(lambda: None)
self.assertEqual(len(event.trigger(comparator=comparator)), 0)
def test_default_comparator(self):
event = events.Event()
event.register(lambda *args, **kwargs: None, a=10)
self.assertEqual(len(event.trigger()), 1)
self.assertEqual(len(event.trigger(a=10)), 1)
def test_decorator(self):
event = events.Event()
event()(lambda: None)
self.assertEqual(len(event.trigger()), 1)
| Update unit tests to match correct behaviour | Update unit tests to match correct behaviour
| Python | mit | Cyanogenoid/smartbot,tomleese/smartbot,Muzer/smartbot,thomasleese/smartbot-old | ---
+++
@@ -30,7 +30,7 @@
def test_default_comparator(self):
event = events.Event()
event.register(lambda *args, **kwargs: None, a=10)
- self.assertEqual(len(event.trigger()), 0)
+ self.assertEqual(len(event.trigger()), 1)
self.assertEqual(len(event.trigger(a=10)), 1)
def test_decorator(self): |
0540b78a5c83cf307c4d629bb814c8359edd8709 | comrade/core/context_processors.py | comrade/core/context_processors.py | from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def context_processor(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
| from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
| Rename the only context processor. | Rename the only context processor.
| Python | mit | bueda/django-comrade | ---
+++
@@ -2,7 +2,7 @@
from django.contrib.sites.models import Site
from settings import DeploymentType
-def context_processor(request):
+def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current() |
b7939c13622d1134364e874da1d1903bcea6cffe | tests/test_graph.py | tests/test_graph.py |
import os
from xml.etree import ElementTree
from utils import with_app, pretty_print_xml
#=============================================================================
# Tests
@with_app(buildername="xml", srcdir="graph", warningiserror=True)
def test_graph(app, status, warning):
app.build()
tree = ElementTree.parse(app.outdir / "index.xml")
pretty_print_xml(tree.getroot())
# Verify that 1 graphviz node is found.
assert len(tree.findall(".//graphviz")) == 1
|
import os
from xml.etree import ElementTree
from utils import with_app, pretty_print_xml
#=============================================================================
# Tests
@with_app(buildername="xml", srcdir="graph", warningiserror=True)
def test_graph(app, status, warning):
app.build()
tree = ElementTree.parse(app.outdir / "index.xml")
# Verify that 1 graphviz node is found.
assert len(tree.findall(".//graphviz")) == 1
| Remove debug printing of doctree during a test | Remove debug printing of doctree during a test
| Python | apache-2.0 | t4ngo/sphinxcontrib-traceables | ---
+++
@@ -11,7 +11,6 @@
def test_graph(app, status, warning):
app.build()
tree = ElementTree.parse(app.outdir / "index.xml")
- pretty_print_xml(tree.getroot())
# Verify that 1 graphviz node is found.
assert len(tree.findall(".//graphviz")) == 1 |
942bd1abf9d4e74b0f1dbeb8fc5acaaaeed1a2a6 | tests/test_nexus.py | tests/test_nexus.py | import os, sys
sys.path.append(os.path.abspath('..'))
import unittest
from mock import patch
from pynexus import api_client
class NexusTest(unittest.TestCase):
def test_constructor_appends_base(self):
n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd')
self.assertEquals(n.uri, 'http://test.com/nexus/service/local/')
@patch.object(api_client.requests, 'get')
def test_get_users_return_list_with_just_anonymous_user(self, mock_get):
mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \
'service/local/users/anonymous","userId":"anonymous",' \
'"firstName":"Nexus","lastName":"Anonymous User",' \
'"status":"active","email":"changeme2@yourcompany.com"' \
',"roles":["anonymous","repository-any-read"]}'
mock_get.return_value = mock_output
n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd')
result = n.get_users()
self.assertEqual(result, mock_output)
def main():
unittest.main()
if __name__ == '__main__':
main()
| import os, sys
sys.path.append(os.path.abspath('..'))
import unittest
from mock import patch
from pynexus import api_client
class NexusTest(unittest.TestCase):
def test_constructor_appends_base(self):
n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd')
self.assertEquals(n.uri, 'http://test.com/nexus/service/local')
@patch.object(api_client.requests, 'get')
def test_get_users_return_list_with_just_anonymous_user(self, mock_get):
mock_output = u'{"data":[{"resourceURI":"http://test.com/nexus/' \
'service/local/users/anonymous","userId":"anonymous",' \
'"firstName":"Nexus","lastName":"Anonymous User",' \
'"status":"active","email":"changeme2@yourcompany.com"' \
',"roles":["anonymous","repository-any-read"]}'
mock_get.return_value = mock_output
n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd')
result = n.get_users()
self.assertEqual(result, mock_output)
def main():
unittest.main()
if __name__ == '__main__':
main()
| Fix constructor test, since constructor does not append / at the end | Fix constructor test, since constructor does not append / at the end
| Python | apache-2.0 | rcarrillocruz/pynexus | ---
+++
@@ -9,7 +9,7 @@
def test_constructor_appends_base(self):
n = api_client.ApiClient('http://test.com', 'testuser', 'testpwd')
- self.assertEquals(n.uri, 'http://test.com/nexus/service/local/')
+ self.assertEquals(n.uri, 'http://test.com/nexus/service/local')
@patch.object(api_client.requests, 'get')
def test_get_users_return_list_with_just_anonymous_user(self, mock_get): |
c76e3accff36eb993ea44f4e38adad9466af1f54 | tests/wsgi_tests.py | tests/wsgi_tests.py | import mock
import unittest
from resto.wsgi import Middleware
class AppTestCase(unittest.TestCase):
def setUp(self):
self.environ = {
'wsgi.version': (1, 0),
'wsgi.multithread': False,
'wsgi.multiprocess': False,
'wsgi.run_once': True,
}
self.mock_start = mock.Mock()
def test_application_init(self):
result = Middleware(self.environ, self.mock_start)
content = ''
for data in result:
content += data
self.assertGreater(content, '')
if __name__ == '__main__':
unittest.main()
| import mock
import unittest
from resto.wsgi import Middleware
class AppTestCase(unittest.TestCase):
def setUp(self):
self.environ = {
'wsgi.version': (1, 0),
'wsgi.multithread': False,
'wsgi.multiprocess': False,
'wsgi.run_once': True,
}
self.mock_start = mock.Mock()
def test_application_init(self):
result = Middleware(self.environ, self.mock_start)
content = ''
for data in result:
content += data
self.assertGreater(content, '')
self.assertTrue(self.mock_start.called)
if __name__ == '__main__':
unittest.main()
| Test WSGI start_response is called | Test WSGI start_response is called
| Python | mit | rafaelpivato/resto | ---
+++
@@ -21,6 +21,7 @@
for data in result:
content += data
self.assertGreater(content, '')
+ self.assertTrue(self.mock_start.called)
if __name__ == '__main__': |
ff662e931cf4d57016c97698d490bc75368ecb67 | pmxbot/testing/fixtures.py | pmxbot/testing/fixtures.py | import pytest
@pytest.fixture(scope='session', autouse=True)
def init_config():
__import__('pmxbot').config = {}
@pytest.fixture(params=['mongodb', 'sqlite'])
def db_uri(request):
if request.param == 'mongodb':
return request.getfuncargvalue('mongodb_uri')
return 'sqlite:pmxbot.sqlite'
| import pytest
@pytest.fixture(scope='session', autouse=True)
def init_config():
__import__('pmxbot').config = {}
@pytest.fixture(params=['mongodb', 'sqlite'])
def db_uri(request):
if request.param == 'mongodb':
return request.getfixturevalue('mongodb_uri')
return 'sqlite:pmxbot.sqlite'
| Fix deprecation warning in fixture retrieval. | Fix deprecation warning in fixture retrieval.
| Python | mit | yougov/pmxbot,yougov/pmxbot,yougov/pmxbot | ---
+++
@@ -9,5 +9,5 @@
@pytest.fixture(params=['mongodb', 'sqlite'])
def db_uri(request):
if request.param == 'mongodb':
- return request.getfuncargvalue('mongodb_uri')
+ return request.getfixturevalue('mongodb_uri')
return 'sqlite:pmxbot.sqlite' |
32f7b016d0e06fa85bef95119223636dc74918ad | reclass/defaults.py | reclass/defaults.py | #
# -*- coding: utf-8 -*-
#
# This file is part of reclass (http://github.com/madduck/reclass)
#
# Copyright © 2007–14 martin f. krafft <madduck@madduck.net>
# Released under the terms of the Artistic Licence 2.0
#
import os, sys
from version import RECLASS_NAME
# defaults for the command-line options
OPT_STORAGE_TYPE = 'yaml_fs'
OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME)
OPT_NODES_URI = 'nodes'
OPT_CLASSES_URI = 'classes'
OPT_PRETTY_PRINT = True
OPT_OUTPUT = 'yaml'
CONFIG_FILE_SEARCH_PATH = [os.getcwd(),
os.path.expanduser('~'),
OPT_INVENTORY_BASE_URI,
os.path.dirname(sys.argv[0])
]
CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml'
PARAMETER_INTERPOLATION_SENTINELS = ('${', '}')
PARAMETER_INTERPOLATION_DELIMITER = ':'
| #
# -*- coding: utf-8 -*-
#
# This file is part of reclass (http://github.com/madduck/reclass)
#
# Copyright © 2007–14 martin f. krafft <madduck@madduck.net>
# Released under the terms of the Artistic Licence 2.0
#
import os, sys
from version import RECLASS_NAME
# defaults for the command-line options
OPT_STORAGE_TYPE = 'yaml_fs'
OPT_INVENTORY_BASE_URI = os.path.join('/etc', RECLASS_NAME)
OPT_NODES_URI = 'nodes'
OPT_CLASSES_URI = 'classes'
OPT_PRETTY_PRINT = True
OPT_OUTPUT = 'yaml'
CONFIG_FILE_SEARCH_PATH = [os.getcwd(),
os.path.expanduser('~'),
OPT_INVENTORY_BASE_URI,
os.path.dirname(sys.argv[0])
]
CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml'
PARAMETER_INTERPOLATION_SENTINELS = ('{{', '}}')
PARAMETER_INTERPOLATION_DELIMITER = ':'
| Change interpolation sentinels to prevent collissions with pillar data | Change interpolation sentinels to prevent collissions with pillar data
| Python | artistic-2.0 | jeroen92/reclass,jeroen92/reclass | ---
+++
@@ -24,5 +24,5 @@
]
CONFIG_FILE_NAME = RECLASS_NAME + '-config.yml'
-PARAMETER_INTERPOLATION_SENTINELS = ('${', '}')
+PARAMETER_INTERPOLATION_SENTINELS = ('{{', '}}')
PARAMETER_INTERPOLATION_DELIMITER = ':' |
019aa0d78fbf54dda405cf8df3aab92dfdaba188 | tests/grammar_atomic_tests.py | tests/grammar_atomic_tests.py | from unittest import TestCase
from pyparsing import ParseException
from regparser.grammar.atomic import *
class GrammarAtomicTests(TestCase):
def test_em_digit_p(self):
result = em_digit_p.parseString('(<E T="03">2</E>)')
self.assertEqual('2', result.p5)
def test_double_alpha(self):
# Match (aa), (bb), etc.
result = lower_p.parseString('(a)')
self.assertEqual('a', result.p1)
result = lower_p.parseString('(aa)')
self.assertEqual('aa', result.p1)
result = lower_p.parseString('(i)')
self.assertEqual('i', result.p1)
# Except for roman numerals
with self.assertRaises(ParseException):
result = lower_p.parseString('(ii)')
with self.assertRaises(ParseException):
result = lower_p.parseString('(iv)')
| from unittest import TestCase
from pyparsing import ParseException
from regparser.grammar.atomic import *
class GrammarAtomicTests(TestCase):
def test_em_digit_p(self):
result = em_digit_p.parseString('(<E T="03">2</E>)')
self.assertEqual('2', result.p5)
def test_double_alpha(self):
for text, p1 in [('(a)', 'a'),
('(aa)', 'aa'),
('(i)','i')]:
result = lower_p.parseString(text)
self.assertEqual(p1, result.p1)
for text in ['(ii)', '(iv)', '(vi)']:
try:
result = lower_p.parseString(text)
except ParseException:
pass
except e:
self.fail("Unexpected error:", e)
else:
self.fail("Didn't raise ParseException")
| Refactor test to eliminate assertRaises() error with Python 2.6 | Refactor test to eliminate assertRaises() error with Python 2.6
| Python | cc0-1.0 | adderall/regulations-parser,willbarton/regulations-parser,grapesmoker/regulations-parser | ---
+++
@@ -10,19 +10,19 @@
self.assertEqual('2', result.p5)
def test_double_alpha(self):
- # Match (aa), (bb), etc.
- result = lower_p.parseString('(a)')
- self.assertEqual('a', result.p1)
+ for text, p1 in [('(a)', 'a'),
+ ('(aa)', 'aa'),
+ ('(i)','i')]:
+ result = lower_p.parseString(text)
+ self.assertEqual(p1, result.p1)
- result = lower_p.parseString('(aa)')
- self.assertEqual('aa', result.p1)
+ for text in ['(ii)', '(iv)', '(vi)']:
+ try:
+ result = lower_p.parseString(text)
+ except ParseException:
+ pass
+ except e:
+ self.fail("Unexpected error:", e)
+ else:
+ self.fail("Didn't raise ParseException")
- result = lower_p.parseString('(i)')
- self.assertEqual('i', result.p1)
-
- # Except for roman numerals
- with self.assertRaises(ParseException):
- result = lower_p.parseString('(ii)')
- with self.assertRaises(ParseException):
- result = lower_p.parseString('(iv)')
- |
c34817c2740e860493692b630a11fdb7acab76aa | tests/test_simple_features.py | tests/test_simple_features.py | from wordgraph.points import Point
import wordgraph
EPOCH_START = 1407109280
def time_values(values, start=EPOCH_START, increment=1):
datapoints = []
for index, value in enumerate(values):
datapoints.append(Point(x=value, y=start + (increment * index)))
return datapoints
def test_monotonic_up_per_second():
datapoints = time_values(float(i) for i in range(POINTS))
features = wordgraph.describe(datapoints)
assert "" in features
| from wordgraph.points import Point
import wordgraph
EPOCH_START = 1407109280
def time_values(values, start=EPOCH_START, increment=1):
datapoints = []
for index, value in enumerate(values):
datapoints.append(Point(x=value, y=start + (increment * index)))
return datapoints
def test_monotonic_up_per_second():
datapoints = time_values(float(i) for i in range(10))
features = wordgraph.describe(datapoints)
assert "" in features
def test_monotonic_down_per_second():
datapoints = time_values(10.0 - i for i in range(10))
features = wordgraph.describe(datapoints)
assert "" in features
| Test case for monotonically decreasing graphs | Test case for monotonically decreasing graphs
Generate time series data for values that decrease monotonically over
time.
| Python | apache-2.0 | tleeuwenburg/wordgraph,tleeuwenburg/wordgraph | ---
+++
@@ -10,6 +10,11 @@
return datapoints
def test_monotonic_up_per_second():
- datapoints = time_values(float(i) for i in range(POINTS))
+ datapoints = time_values(float(i) for i in range(10))
features = wordgraph.describe(datapoints)
assert "" in features
+
+def test_monotonic_down_per_second():
+ datapoints = time_values(10.0 - i for i in range(10))
+ features = wordgraph.describe(datapoints)
+ assert "" in features |
e2991360bb35060e570e59bbea6a1e6d23a54aa7 | transport_tester.py | transport_tester.py | from gevent import monkey
monkey.patch_all() # noqa
import sys
import time
from raiden.network.transport import UDPTransport
from raiden.network.sockfactory import socket_factory
class DummyProtocol(object):
def __init__(self):
self.raiden = None
def receive(self, data):
print data
if __name__ == "__main__":
with socket_factory('0.0.0.0', 8885) as mapped_socket:
print mapped_socket
t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol())
while True:
time.sleep(1)
if len(sys.argv) > 1:
t.send(None, (sys.argv[1], 8885), b'hello')
| from gevent import monkey
monkey.patch_all() # noqa
import sys
import time
from ethereum import slogging
from raiden.network.transport import UDPTransport
from raiden.network.sockfactory import socket_factory
class DummyProtocol(object):
def __init__(self):
self.raiden = None
def receive(self, data):
print data
if __name__ == "__main__":
slogging.configure(':DEBUG')
with socket_factory('0.0.0.0', 8885) as mapped_socket:
print mapped_socket
t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol())
while True:
time.sleep(1)
if len(sys.argv) > 1:
t.send(None, (sys.argv[1], 8885), b'hello')
| Add logging to script test | Add logging to script test
| Python | mit | hackaugusto/raiden,tomashaber/raiden,tomashaber/raiden,tomashaber/raiden,hackaugusto/raiden,tomashaber/raiden,tomashaber/raiden,charles-cooper/raiden,charles-cooper/raiden | ---
+++
@@ -2,6 +2,8 @@
monkey.patch_all() # noqa
import sys
import time
+
+from ethereum import slogging
from raiden.network.transport import UDPTransport
from raiden.network.sockfactory import socket_factory
@@ -17,6 +19,7 @@
if __name__ == "__main__":
+ slogging.configure(':DEBUG')
with socket_factory('0.0.0.0', 8885) as mapped_socket:
print mapped_socket
t = UDPTransport(mapped_socket.socket, protocol=DummyProtocol()) |
156ebae630c3690db875b8925bfbdc5ded396fdd | src/Sensors/Factory.py | src/Sensors/Factory.py | from src.Sensors.BME280 import BME280
from src.Sensors.BME680 import BME680
from src.Sensors.DS18B20 import DS18B20
from src.Notification.Subscriber.LED.RGB import RGB
class Factory:
@staticmethod
def create_sensor(device, address):
if device == 'BME280':
return BME280(address=address)
elif device == 'BME680':
return BME680(address=address)
elif device == 'DS18B20':
return DS18B20(address=address)
@staticmethod
def create_led(device, configuration, notification_manager):
if device == 'rgb':
return RGB(configuration=configuration, notification_manager=notification_manager)
| from src.Sensors.BME280 import BME280
from src.Sensors.BME680 import BME680
from src.Sensors.DS18B20 import DS18B20
from src.Sensors.CCS811 import CCS811
from src.Notification.Subscriber.LED.RGB import RGB
class Factory:
@staticmethod
def create_sensor(device, address):
if device == 'BME280':
return BME280(address=address)
elif device == 'BME680':
return BME680(address=address)
elif device == 'DS18B20':
return DS18B20(address=address)
elif device == 'CCS811':
return CCS811(address=address)
@staticmethod
def create_led(device, configuration, notification_manager):
if device == 'rgb':
return RGB(configuration=configuration, notification_manager=notification_manager)
| Add CCS811 to device factory | Add CCS811 to device factory
| Python | mit | dashford/sentinel | ---
+++
@@ -1,6 +1,7 @@
from src.Sensors.BME280 import BME280
from src.Sensors.BME680 import BME680
from src.Sensors.DS18B20 import DS18B20
+from src.Sensors.CCS811 import CCS811
from src.Notification.Subscriber.LED.RGB import RGB
@@ -14,6 +15,8 @@
return BME680(address=address)
elif device == 'DS18B20':
return DS18B20(address=address)
+ elif device == 'CCS811':
+ return CCS811(address=address)
@staticmethod
def create_led(device, configuration, notification_manager): |
b05ba019143cc39ba0d02d822824172313e78591 | aubergine/celery.py | aubergine/celery.py | from __future__ import absolute_import
from celery import Celery
app = Celery('aubergine')
app.config_from_object('aubergine.settings.celeryconfig')
app.autodiscover_tasks(['aubergine'], related_name='tasks')
app.setup_security()
@app.task(bind=True)
def debug_task(self):
print('Request: {0!r}'.format(self.request))
@app.task()
def tadd(x, y):
return x + y
@app.task()
def tsum(numbers):
return sum(numbers)
| from __future__ import absolute_import
from celery import Celery
app = Celery('aubergine')
app.config_from_object('aubergine.settings.celeryconfig')
app.autodiscover_tasks(['aubergine'], related_name='tasks')
# For message signing
# app.setup_security()
@app.task(bind=True)
def debug_task(self):
print('Request: {0!r}'.format(self.request))
@app.task()
def tadd(x, y):
return x + y
@app.task()
def tsum(numbers):
return sum(numbers)
| Disable message signing for now | Disable message signing for now
| Python | bsd-3-clause | michaelwisely/aubergine | ---
+++
@@ -6,7 +6,9 @@
app = Celery('aubergine')
app.config_from_object('aubergine.settings.celeryconfig')
app.autodiscover_tasks(['aubergine'], related_name='tasks')
-app.setup_security()
+
+# For message signing
+# app.setup_security()
@app.task(bind=True) |
f1af7dad41992b53e90a5f8dd20e1635f11a7ce1 | pstats_print2list/__init__.py | pstats_print2list/__init__.py | # -*- coding: utf-8 -*-
__author__ = 'Vauxoo'
__email__ = 'info@vauxoo.com'
__version__ = '0.1.0'
| # -*- coding: utf-8 -*-
from pstats_print2list import print_stats
__author__ = 'Vauxoo'
__email__ = 'info@vauxoo.com'
__version__ = '0.1.0'
| Add print_stats to init file | [REF] pstats_print2list: Add print_stats to init file
| Python | isc | Vauxoo/pstats-print2list | ---
+++
@@ -1,4 +1,6 @@
# -*- coding: utf-8 -*-
+
+from pstats_print2list import print_stats
__author__ = 'Vauxoo'
__email__ = 'info@vauxoo.com' |
875e25231dec76760bde97107d52036560347914 | salt/modules/win_shadow.py | salt/modules/win_shadow.py | '''
Manage the shadow file
'''
import salt.utils
def __virtual__():
'''
Only works on Windows systems
'''
if salt.utils.is_windows():
return 'shadow'
return False
def info(name):
'''
Return information for the specified user
This is just returns dummy data so that salt states can work.
CLI Example::
salt '*' shadow.info root
'''
ret = {
'name': name,
'pwd': '',
'lstchg': '',
'min': '',
'max': '',
'warn': '',
'inact': '',
'expire': ''}
return ret
def set_password(name, password):
'''
Set the password for a named user.
CLI Example::
salt '*' shadow.set_password root mysecretpassword
'''
cmd = 'net user {0} {1}'.format(name, password)
ret = __salt__['cmd.run_all'](cmd)
return not ret['retcode']
| '''
Manage the shadow file
'''
import salt.utils
def __virtual__():
'''
Only works on Windows systems
'''
if salt.utils.is_windows():
return 'shadow'
return False
def info(name):
'''
Return information for the specified user
This is just returns dummy data so that salt states can work.
CLI Example::
salt '*' shadow.info root
'''
ret = {
'name': name,
'passwd': '',
'lstchg': '',
'min': '',
'max': '',
'warn': '',
'inact': '',
'expire': ''}
return ret
def set_password(name, password):
'''
Set the password for a named user.
CLI Example::
salt '*' shadow.set_password root mysecretpassword
'''
cmd = 'net user {0} {1}'.format(name, password)
ret = __salt__['cmd.run_all'](cmd)
return not ret['retcode']
| Rename 'pwd' field in windows shadow.info output | Rename 'pwd' field in windows shadow.info output
This makes the field name consistent with the other shadow modules. Note
that the passwd field is not used at all in Windows user management, so
this is merely a cosmetic change.
| Python | apache-2.0 | saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt | ---
+++
@@ -25,7 +25,7 @@
'''
ret = {
'name': name,
- 'pwd': '',
+ 'passwd': '',
'lstchg': '',
'min': '',
'max': '', |
d279572c255a302dd5f191b0f047c46c9184ec2a | tests/test_write.py | tests/test_write.py | from __future__ import absolute_import
from ofxparse import OfxParser as op, OfxPrinter
from unittest import TestCase
from os import close, remove
from tempfile import mkstemp
import sys
sys.path.append('..')
from .support import open_file
class TestOfxWrite(TestCase):
def test_write(self):
test_file = open_file('fidelity.ofx')
ofx_doc = op.parse(test_file)
self.assertEqual(str(ofx_doc), "")
def test_using_ofx_printer(self):
test_file = open_file('checking.ofx')
ofx_doc = op.parse(test_file)
fd, name = mkstemp()
close(fd)
printer = OfxPrinter(ofx=ofx_doc, filename=name)
printer.write(tabs=1)
if __name__ == "__main__":
import unittest
unittest.main()
| from __future__ import absolute_import
from ofxparse import OfxParser, OfxPrinter
from unittest import TestCase
from os import close, remove
from tempfile import mkstemp
import sys
sys.path.append('..')
from .support import open_file
class TestOfxWrite(TestCase):
def test_write(self):
with open_file('fidelity.ofx') as f:
ofx = OfxParser.parse(f)
self.assertEqual(str(ofx), "")
def test_using_ofx_printer(self):
with open_file('checking.ofx') as f:
ofx = OfxParser.parse(f)
fd, name = mkstemp()
close(fd)
printer = OfxPrinter(ofx=ofx, filename=name)
printer.write(tabs=1)
if __name__ == "__main__":
import unittest
unittest.main()
| Fix warnings under Python 3 | Fix warnings under Python 3
| Python | mit | rdsteed/ofxparse,jaraco/ofxparse,jseutter/ofxparse,udibr/ofxparse | ---
+++
@@ -1,6 +1,6 @@
from __future__ import absolute_import
-from ofxparse import OfxParser as op, OfxPrinter
+from ofxparse import OfxParser, OfxPrinter
from unittest import TestCase
from os import close, remove
from tempfile import mkstemp
@@ -11,16 +11,16 @@
class TestOfxWrite(TestCase):
def test_write(self):
- test_file = open_file('fidelity.ofx')
- ofx_doc = op.parse(test_file)
- self.assertEqual(str(ofx_doc), "")
+ with open_file('fidelity.ofx') as f:
+ ofx = OfxParser.parse(f)
+ self.assertEqual(str(ofx), "")
def test_using_ofx_printer(self):
- test_file = open_file('checking.ofx')
- ofx_doc = op.parse(test_file)
+ with open_file('checking.ofx') as f:
+ ofx = OfxParser.parse(f)
fd, name = mkstemp()
close(fd)
- printer = OfxPrinter(ofx=ofx_doc, filename=name)
+ printer = OfxPrinter(ofx=ofx, filename=name)
printer.write(tabs=1)
if __name__ == "__main__": |
491613b34cb3c89e8d49670457a46b924a109529 | pypinksign/__init__.py | pypinksign/__init__.py | """
Basic Template system for project pinksign,
similar to the template part of PasteScript but without any dependencies.
"""
from .pypinksign import (
PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator,
bit2string, separate_p12_into_npki
)
| """
Basic Template system for project pinksign,
similar to the template part of PasteScript but without any dependencies.
"""
from .pypinksign import (
PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator,
bit2string, separate_p12_into_npki, encrypt_decrypted_prikey
)
| Add new function in init | Add new function in init
| Python | mit | bandoche/PyPinkSign | ---
+++
@@ -6,5 +6,5 @@
from .pypinksign import (
PinkSign, get_npki_path, url_encode, paramize, choose_cert, seed_cbc_128_encrypt, seed_cbc_128_decrypt, seed_generator,
- bit2string, separate_p12_into_npki
+ bit2string, separate_p12_into_npki, encrypt_decrypted_prikey
) |
d7534dc3536ebe035abf063d83aa8d471cdadb16 | python/pyqt_version.py | python/pyqt_version.py | import PySide2.QtCore
# Prints PySide2 version
# e.g. 5.11.1a1
print(PySide2.__version__)
# Gets a tuple with each version component
# e.g. (5, 11, 1, 'a', 1)
print(PySide2.__version_info__)
# Prints the Qt version used to compile PySide2
# e.g. "5.11.2"
print(PySide2.QtCore.__version__)
# Gets a tuple with each version components of Qt used to compile PySide2
# e.g. (5, 11, 2)
print(PySide2.QtCore.__version_info__)
print(PySide2.QtCore.qVersion())
| #!/usr/bin/env python3
# coding: utf-8
'''
PySide2
'''
import sys
try:
import PySide2.QtCore
except ImportError:
print('cannot load module: PySide2.QtCore')
sys.exit(1)
# Prints PySide2 version
# e.g. 5.11.1a1
print(PySide2.__version__)
# Gets a tuple with each version component
# e.g. (5, 11, 1, 'a', 1)
print(PySide2.__version_info__)
# Prints the Qt version used to compile PySide2
# e.g. "5.11.2"
print(PySide2.QtCore.__version__)
# Gets a tuple with each version components of Qt used to compile PySide2
# e.g. (5, 11, 2)
print(PySide2.QtCore.__version_info__)
print(PySide2.QtCore.qVersion())
| ADD try-except to handle import error | ADD try-except to handle import error
| Python | mit | ericosur/myqt,ericosur/myqt,ericosur/myqt,ericosur/myqt,ericosur/myqt,ericosur/myqt | ---
+++
@@ -1,4 +1,16 @@
-import PySide2.QtCore
+#!/usr/bin/env python3
+# coding: utf-8
+
+'''
+PySide2
+'''
+
+import sys
+try:
+ import PySide2.QtCore
+except ImportError:
+ print('cannot load module: PySide2.QtCore')
+ sys.exit(1)
# Prints PySide2 version
# e.g. 5.11.1a1 |
da67ce3f25a708b99cb81f17703e74965dbea960 | rtrss/filestorage/httputil.py | rtrss/filestorage/httputil.py | import logging
import time
import requests
from googleapiclient.errors import HttpError
# Number of retries in case of API errors
NUM_RETRIES = 3
# Delay between retry attempts, seconds
RETRY_DELAY = 1
_logger = logging.getLogger(__name__)
def is_retryable(exc):
retryable_codes = [500, 502, 503, 504]
"""Returns True if exception is "retryable", eg. HTTP 503"""
if issubclass(exc, requests.exceptions.RequestException):
code = exc.response.status_code
elif issubclass(exc, HttpError):
code = exc.resp.status
else:
return False
return code in retryable_codes
def retry_on_exception(
exceptions=(HttpError, requests.exceptions.RequestException),
retryable=is_retryable,
tries=NUM_RETRIES,
delay=RETRY_DELAY):
"""Retry call if function raises retryable exception"""
def wrap(f):
def wrapped_f(*args, **kwargs):
mtries = tries
while mtries > 1:
try:
return f(*args, **kwargs)
except exceptions as err:
# Reraise if non-retryable error
if not retryable(err):
raise
_logger.warn("Retrying in %.2f seconds ...", delay)
time.sleep(delay)
mtries -= 1
# Only one last try left
return f(*args, **kwargs)
return wrapped_f
return wrap
| import logging
import time
import requests
from googleapiclient.errors import HttpError
# Number of retries in case of API errors
NUM_RETRIES = 3
# Delay between retry attempts, seconds
RETRY_DELAY = 1
_logger = logging.getLogger(__name__)
def is_retryable(exc):
retryable_codes = [500, 502, 503, 504]
"""Returns True if exception is "retryable", eg. HTTP 503"""
if isinstance(exc, requests.exceptions.RequestException):
code = exc.response.status_code
elif isinstance(exc, HttpError):
code = exc.resp.status
else:
return False
return code in retryable_codes
def retry_on_exception(
retryable=is_retryable,
tries=NUM_RETRIES,
delay=RETRY_DELAY):
"""Retry call if function raises retryable exception"""
def wrap(f):
def wrapped_f(*args, **kwargs):
mtries = tries
while mtries > 1:
try:
return f(*args, **kwargs)
except Exception as err:
# Re-raise if non-retryable error
if not retryable(err):
raise
_logger.warn("Retrying in %.2f seconds ...", delay)
time.sleep(delay)
mtries -= 1
# Only one last try left
return f(*args, **kwargs)
return wrapped_f
return wrap
| Remove unnecessary parameter, fix type detection bug | Remove unnecessary parameter, fix type detection bug
| Python | apache-2.0 | notapresent/rtrss,notapresent/rtrss,notapresent/rtrss,notapresent/rtrss | ---
+++
@@ -3,6 +3,7 @@
import requests
from googleapiclient.errors import HttpError
+
# Number of retries in case of API errors
@@ -17,9 +18,9 @@
def is_retryable(exc):
retryable_codes = [500, 502, 503, 504]
"""Returns True if exception is "retryable", eg. HTTP 503"""
- if issubclass(exc, requests.exceptions.RequestException):
+ if isinstance(exc, requests.exceptions.RequestException):
code = exc.response.status_code
- elif issubclass(exc, HttpError):
+ elif isinstance(exc, HttpError):
code = exc.resp.status
else:
return False
@@ -27,7 +28,6 @@
def retry_on_exception(
- exceptions=(HttpError, requests.exceptions.RequestException),
retryable=is_retryable,
tries=NUM_RETRIES,
delay=RETRY_DELAY):
@@ -38,8 +38,8 @@
while mtries > 1:
try:
return f(*args, **kwargs)
- except exceptions as err:
- # Reraise if non-retryable error
+ except Exception as err:
+ # Re-raise if non-retryable error
if not retryable(err):
raise
|
5a3cdba13cb4898b76d83c949fc3ab6895c267ff | scripts/cpuflags.py | scripts/cpuflags.py | import errno
class CPUFlags:
def __init__(self):
self.flags = set()
try:
self.flags = self.__parse_cpuinfo()
except IOError as e:
if e.errno == errno.ENOENT:
return
raise
def __contains__(self, name):
return name in self.flags
def __parse_cpuinfo(self):
def get_flags():
with open('/proc/cpuinfo', 'r') as f:
for line in f:
if line.startswith('flags'):
return line
line = get_flags().split()
del line[:2] # remove tokens "flags", ":"
return set(line)
def main():
import sys
flags = CPUFlags()
if len(sys.argv) == 2:
if sys.argv[1] in flags:
print "present"
if __name__ == '__main__':
main()
| import errno
import sys
class CPUFlags:
def __init__(self):
self.flags = set()
try:
self.flags = self.__parse_cpuinfo()
except IOError as e:
if e.errno == errno.ENOENT:
return
raise
def __contains__(self, name):
return name in self.flags
def __parse_cpuinfo(self):
def get_flags():
with open('/proc/cpuinfo', 'r') as f:
for line in f:
if line.startswith('flags'):
return line
line = get_flags().split()
del line[:2] # remove tokens "flags", ":"
return set(line)
def main():
import sys
flags = CPUFlags()
if len(sys.argv) == 2:
if sys.argv[1] in flags:
print "present"
return 0
return 1
if __name__ == '__main__':
sys.exit(main())
| Check script returns an exit code | Check script returns an exit code
| Python | bsd-2-clause | WojciechMula/toys,WojciechMula/toys,WojciechMula/toys,WojciechMula/toys,WojciechMula/toys | ---
+++
@@ -1,4 +1,5 @@
import errno
+import sys
class CPUFlags:
def __init__(self):
@@ -40,7 +41,9 @@
if len(sys.argv) == 2:
if sys.argv[1] in flags:
print "present"
-
+ return 0
+
+ return 1
if __name__ == '__main__':
- main()
+ sys.exit(main()) |
ea62a1cd9642dbff69cbfae3f8b540604a8a8fca | mine/__init__.py | mine/__init__.py | #!/usr/bin/env python
"""Package for mine."""
import sys
__project__ = 'mine'
__version__ = '0.1'
CLI = 'mine'
VERSION = __project__ + '-' + __version__
DESCRIPTION = "Manages running applications across multiple computers."
PYTHON_VERSION = 3, 3
if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test)
exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
| #!/usr/bin/env python
"""Package for mine."""
import sys
__project__ = 'mine'
__version__ = '0.1'
CLI = 'mine'
VERSION = __project__ + '-' + __version__
DESCRIPTION = "For applications that haven't learned to share."
PYTHON_VERSION = 3, 3
if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test)
exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
| Update headline to match GitHub | Update headline to match GitHub
| Python | mit | jacebrowning/mine | ---
+++
@@ -9,7 +9,7 @@
CLI = 'mine'
VERSION = __project__ + '-' + __version__
-DESCRIPTION = "Manages running applications across multiple computers."
+DESCRIPTION = "For applications that haven't learned to share."
PYTHON_VERSION = 3, 3
|
f95460070e80e1d83394fa6ed08bad9dad34802c | ovp_users/emails.py | ovp_users/emails.py | from ovp_core.emails import BaseMail
class UserMail(BaseMail):
"""
This class is responsible for firing emails for Users
"""
def __init__(self, user, async_mail=None):
super(UserMail, self).__init__(user.email, async_mail)
def sendWelcome(self, context={}):
"""
Sent when user registers
"""
return self.sendEmail('welcome', 'Welcome', context)
def sendRecoveryToken(self, context):
"""
Sent when volunteer requests recovery token
"""
context.update({
'user_email': self.email_address
})
return self.sendEmail('recoveryToken', 'Password recovery', context)
| from ovp_core.emails import BaseMail
class UserMail(BaseMail):
"""
This class is responsible for firing emails for Users
"""
def __init__(self, user, async_mail=None):
super(UserMail, self).__init__(user.email, async_mail)
def sendWelcome(self, context={}):
"""
Sent when user registers
"""
return self.sendEmail('welcome', 'Welcome', context)
def sendRecoveryToken(self, context):
"""
Sent when volunteer requests recovery token
"""
context.update({
'user_email': user.email
})
return self.sendEmail('recoveryToken', 'Password recovery', context)
| Revert "fix getting user email for recoveryToken" | Revert "fix getting user email for recoveryToken"
This reverts commit a47b098e4d644391213958f9e05c179a7410208d.
| Python | agpl-3.0 | OpenVolunteeringPlatform/django-ovp-users,OpenVolunteeringPlatform/django-ovp-users | ---
+++
@@ -19,7 +19,7 @@
Sent when volunteer requests recovery token
"""
context.update({
- 'user_email': self.email_address
+ 'user_email': user.email
})
return self.sendEmail('recoveryToken', 'Password recovery', context)
|
839a0cafca1d172f7a061dcec5f6a4eca6d725c8 | superlists/lists/tests.py | superlists/lists/tests.py | from django.test import TestCase
# Create your tests here.
| from django.test import TestCase
class SmokeTest(TestCase):
def test_bad_maths(self):
self.assertEqual(1 + 1, 3)
| Add app for lists, with deliberately failing unit test | Add app for lists, with deliberately failing unit test
| Python | apache-2.0 | Alfawuhn/test-driven-python | ---
+++
@@ -1,3 +1,5 @@
from django.test import TestCase
-# Create your tests here.
+class SmokeTest(TestCase):
+ def test_bad_maths(self):
+ self.assertEqual(1 + 1, 3) |
13d9cf933e49849a3c5343e7bdbf887b9aee6097 | busbus/entity.py | busbus/entity.py | from busbus import util
class LazyEntityProperty(object):
def __init__(self, f, *args, **kwargs):
self.f = f
self.args = args
self.kwargs = kwargs
def __call__(self):
return self.f(*self.args, **self.kwargs)
class BaseEntity(object):
def __init__(self, provider, **kwargs):
self._provider = provider
self._lazy_properties = {}
for attr in getattr(self, '__attrs__', []):
if isinstance(kwargs.get(attr, None), LazyEntityProperty):
self._lazy_properties[attr] = kwargs[attr]
else:
setattr(self, attr, kwargs.get(attr, None))
provider._new_entity(self)
def __repr__(self, args=['id']):
return u'<{0}({1})>'.format(
util.clsname(self),
','.join('{0}={1!r}'.format(i, getattr(self, i)) for i in args))
def __getattr__(self, name):
if name in self._lazy_properties:
value = self._lazy_properties[name]()
del self._lazy_properties[name]
setattr(self, name, value)
return value
else:
raise AttributeError(name)
def to_dict(self):
return dict((attr, getattr(self, attr)) for attr in self.__attrs__
if getattr(self, attr))
| from busbus import util
class LazyEntityProperty(object):
def __init__(self, f, *args, **kwargs):
self.f = f
self.args = args
self.kwargs = kwargs
def __call__(self):
return self.f(*self.args, **self.kwargs)
class BaseEntity(object):
__repr_attrs__ = ('id',)
def __init__(self, provider, **kwargs):
self._provider = provider
self._lazy_properties = {}
for attr in getattr(self, '__attrs__', []):
if isinstance(kwargs.get(attr, None), LazyEntityProperty):
self._lazy_properties[attr] = kwargs[attr]
else:
setattr(self, attr, kwargs.get(attr, None))
provider._new_entity(self)
def __repr__(self):
return u'<{0}({1})>'.format(
util.clsname(self), ','.join(
'{0}={1!r}'.format(i, getattr(self, i))
for i in self.__repr_attrs__))
def __getattr__(self, name):
if name in self._lazy_properties:
value = self._lazy_properties[name]()
del self._lazy_properties[name]
setattr(self, name, value)
return value
else:
raise AttributeError(name)
def to_dict(self):
return dict((attr, getattr(self, attr)) for attr in self.__attrs__
if getattr(self, attr))
| Use an instance variable instead of a non-standard argument to __repr__ | Use an instance variable instead of a non-standard argument to __repr__
| Python | mit | spaceboats/busbus | ---
+++
@@ -13,6 +13,7 @@
class BaseEntity(object):
+ __repr_attrs__ = ('id',)
def __init__(self, provider, **kwargs):
self._provider = provider
@@ -26,10 +27,11 @@
provider._new_entity(self)
- def __repr__(self, args=['id']):
+ def __repr__(self):
return u'<{0}({1})>'.format(
- util.clsname(self),
- ','.join('{0}={1!r}'.format(i, getattr(self, i)) for i in args))
+ util.clsname(self), ','.join(
+ '{0}={1!r}'.format(i, getattr(self, i))
+ for i in self.__repr_attrs__))
def __getattr__(self, name):
if name in self._lazy_properties: |
5b6823ec19185ed5b413d1c01d3afeb5b1716778 | taca/server_status/cli.py | taca/server_status/cli.py | import click
import logging
from taca.server_status import server_status as status
from taca.utils.config import CONFIG
from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception
@click.group(name='server_status')
def server_status():
""" Monitor server status """
if not CONFIG.get('server_status', ''):
logging.warning("Configuration missing required entries: server_status")
# server status subcommands
@server_status.command()
@click.option('--statusdb', is_flag=True, help="Update the statusdb")
def nases(statusdb):
""" Checks the available space on all the nases
"""
disk_space = status.get_nases_disk_space()
if statusdb:
status.update_status_db(disk_space, server_type='nas')
@server_status.command()
def cronjobs():
""" Monitors cronjobs and updates statusdb
"""
cj.update_cronjob_db()
| import click
import logging
from taca.server_status import server_status as status
from taca.utils.config import CONFIG
from taca.server_status import cronjobs as cj # to avoid similar names with command, otherwise exception
@click.group(name='server_status')
def server_status():
""" Monitor server status """
# server status subcommands
@server_status.command()
@click.option('--statusdb', is_flag=True, help="Update the statusdb")
def nases(statusdb):
""" Checks the available space on all the nases
"""
if not CONFIG.get('server_status', ''):
logging.warning("Configuration missing required entries: server_status")
disk_space = status.get_nases_disk_space()
if statusdb:
status.update_status_db(disk_space, server_type='nas')
@server_status.command()
def cronjobs():
""" Monitors cronjobs and updates statusdb
"""
cj.update_cronjob_db()
| Move warning about missing config entry to relevant subcommand | Move warning about missing config entry to relevant subcommand
| Python | mit | SciLifeLab/TACA,SciLifeLab/TACA,SciLifeLab/TACA | ---
+++
@@ -9,8 +9,6 @@
@click.group(name='server_status')
def server_status():
""" Monitor server status """
- if not CONFIG.get('server_status', ''):
- logging.warning("Configuration missing required entries: server_status")
# server status subcommands
@server_status.command()
@@ -18,6 +16,8 @@
def nases(statusdb):
""" Checks the available space on all the nases
"""
+ if not CONFIG.get('server_status', ''):
+ logging.warning("Configuration missing required entries: server_status")
disk_space = status.get_nases_disk_space()
if statusdb:
status.update_status_db(disk_space, server_type='nas') |
d20347f4a57bb195291ebc79fc1ca0858b3f1d65 | PyLunch/pylunch/specials/models.py | PyLunch/pylunch/specials/models.py | from django.db import models
MAX_PRICE_FORMAT = {
'max_digits': 5,
'decimal_places': 2
}
SPECIAL_TYPES = (
('LU', 'Lunch'),
('BR', 'Breakfast'),
('DI', 'Dinner'),
)
MAX_RESTAURANT_NAME_LENGTH = 50
MAX_DESCRIPTION_LENGTH = 500
class Restaurant(models.Model):
name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH)
description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH)
class Special(models.Model):
restaurant = models.ForeignKey(Restaurant)
description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH)
special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES)
special_price = models.DecimalField(**MAX_PRICE_FORMAT)
normal_price = models.DecimalField(**MAX_PRICE_FORMAT) | from django.db import models
MAX_PRICE_FORMAT = {
'max_digits': 5,
'decimal_places': 2
}
SPECIAL_TYPES = (
('LU', 'Lunch'),
('BR', 'Breakfast'),
('DI', 'Dinner'),
)
MAX_RESTAURANT_NAME_LENGTH = 50
MAX_DESCRIPTION_LENGTH = 500
class Restaurant(models.Model):
name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH)
description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH)
def __unicode__(self):
return self.name
class Special(models.Model):
restaurant = models.ForeignKey(Restaurant)
description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH)
special_type = models.CharField(max_length=2, choices=SPECIAL_TYPES)
special_price = models.DecimalField(**MAX_PRICE_FORMAT)
normal_price = models.DecimalField(**MAX_PRICE_FORMAT)
valid_from = models.DateField()
valid_until = models.DateField()
def __unicode__(self):
return "%s: %s" % (self.restaurant.name, self.description) | Add fields to Special model | Add fields to Special model
| Python | unlicense | wiehan-a/pylunch | ---
+++
@@ -18,6 +18,9 @@
name = models.CharField(max_length=MAX_RESTAURANT_NAME_LENGTH)
description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH)
+ def __unicode__(self):
+ return self.name
+
class Special(models.Model):
restaurant = models.ForeignKey(Restaurant)
description = models.CharField(max_length=MAX_DESCRIPTION_LENGTH)
@@ -26,3 +29,9 @@
special_price = models.DecimalField(**MAX_PRICE_FORMAT)
normal_price = models.DecimalField(**MAX_PRICE_FORMAT)
+
+ valid_from = models.DateField()
+ valid_until = models.DateField()
+
+ def __unicode__(self):
+ return "%s: %s" % (self.restaurant.name, self.description) |
d72e34de631e3f6984a1810cbd8ec2b128a196de | sigma_core/tests/factories.py | sigma_core/tests/factories.py | import factory
from django.utils.text import slugify
from sigma_core.models.user import User
from sigma_core.models.group import Group
class UserFactory(factory.django.DjangoModelFactory):
class Meta:
model = User
lastname = factory.Faker('last_name')
firstname = factory.Faker('first_name')
email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname)))
class AdminUserFactory(UserFactory):
is_staff = True
class GroupFactory(factory.django.DjangoModelFactory):
class Meta:
model = Group
name = factory.Sequence(lambda n: 'Group %d' % n)
| import factory
from django.utils.text import slugify
from sigma_core.models.user import User
from sigma_core.models.group import Group
from sigma_core.models.user_group import UserGroup
class UserFactory(factory.django.DjangoModelFactory):
class Meta:
model = User
lastname = factory.Faker('last_name')
firstname = factory.Faker('first_name')
email = factory.LazyAttribute(lambda obj: '%s.%s@school.edu' % (slugify(obj.firstname), slugify(obj.lastname)))
class AdminUserFactory(UserFactory):
is_staff = True
class GroupFactory(factory.django.DjangoModelFactory):
class Meta:
model = Group
name = factory.Sequence(lambda n: 'Group %d' % n)
class UserGroupFactory(factory.django.DjangoModelFactory):
class Meta:
model = UserGroup
user = factory.SubFactory(UserFactory)
group = factory.SubFactory(GroupFactory)
join_date = factory.Faker('date')
| Add UserGroupFactory for future tests | Add UserGroupFactory for future tests
| Python | agpl-3.0 | ProjetSigma/backend,ProjetSigma/backend | ---
+++
@@ -4,6 +4,8 @@
from sigma_core.models.user import User
from sigma_core.models.group import Group
+from sigma_core.models.user_group import UserGroup
+
class UserFactory(factory.django.DjangoModelFactory):
class Meta:
@@ -23,4 +25,12 @@
model = Group
name = factory.Sequence(lambda n: 'Group %d' % n)
-
+
+
+class UserGroupFactory(factory.django.DjangoModelFactory):
+ class Meta:
+ model = UserGroup
+
+ user = factory.SubFactory(UserFactory)
+ group = factory.SubFactory(GroupFactory)
+ join_date = factory.Faker('date') |
70f69f7b801404f7091e91b6ed997602709f9f42 | commands/globaladd.py | commands/globaladd.py | from devbot import chat
def call(message: str, name, protocol, cfg, commands):
if message is '':
chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd')))
return
if ' ' in message:
chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name))
return
chat.say('/msg {} Invited {} to GlobalChat'.format(name, message))
chat.say_wrap('/msg {}'.format(message),
'You have been added to global chat. Use /g GlobalChat to speak in the group, and /e to exit.')
chat.say('/nlip GlobalChat {}'.format(message))
| from devbot import chat
def call(message: str, name, protocol, cfg, commands):
if message is '':
chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd')))
return
if ' ' in message:
chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name))
return
chat.say('/msg {} Invited {} to GlobalChat'.format(name, message))
chat.say(
'/msg {} You have been invited to global chat. Use /g GlobalChat to enter, and /e to exit.'.format(
message))
chat.say('/nlip GlobalChat {}'.format(message))
| Fix gadd not sending tutorial | Fix gadd not sending tutorial
| Python | mit | Ameliorate/DevotedBot,Ameliorate/DevotedBot | ---
+++
@@ -4,11 +4,12 @@
def call(message: str, name, protocol, cfg, commands):
if message is '':
chat.say('/msg {} {}'.format(name, commands['help']['globaladd'].format('globaladd')))
- return
+ return
if ' ' in message:
chat.say('/msg {} Sorry, that was not a valid player name: It contains spaces.'.format(name))
return
chat.say('/msg {} Invited {} to GlobalChat'.format(name, message))
- chat.say_wrap('/msg {}'.format(message),
- 'You have been added to global chat. Use /g GlobalChat to speak in the group, and /e to exit.')
+ chat.say(
+ '/msg {} You have been invited to global chat. Use /g GlobalChat to enter, and /e to exit.'.format(
+ message))
chat.say('/nlip GlobalChat {}'.format(message)) |
559f3c18a7e27e4bb1147b03a27ec083a66749d0 | was/photo/models.py | was/photo/models.py | from django.db import models
class Photo(models.Model):
artist = models.ForeignKey('artists.Artists')
picture = models.ImageField(null=True, blank=True, upload_to="art_picture/")
comment = models.TextField(max_length=500) | from django.db import models
class Photo(models.Model):
artist = models.ForeignKey('artists.Artists')
picture = models.ImageField(null=True, blank=True, upload_to="art_picture/")
comment = models.TextField(max_length=500)
def __str__(self):
return '{}'.format(self.picture) | Define a '__str__' method for photo model | Define a '__str__' method for photo model
| Python | mit | KeserOner/where-artists-share,KeserOner/where-artists-share | ---
+++
@@ -5,3 +5,6 @@
artist = models.ForeignKey('artists.Artists')
picture = models.ImageField(null=True, blank=True, upload_to="art_picture/")
comment = models.TextField(max_length=500)
+
+ def __str__(self):
+ return '{}'.format(self.picture) |
c967776872e9a5c6fa840d202f25a067020c148f | cyder/cydhcp/range/forms.py | cyder/cydhcp/range/forms.py | from django import forms
from cyder.base.eav.forms import get_eav_form
from cyder.base.mixins import UsabilityFormMixin
from cyder.cydhcp.range.models import Range, RangeAV
from cyder.cydns.forms import ViewChoiceForm
class RangeForm(ViewChoiceForm, UsabilityFormMixin):
class Meta:
model = Range
exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower')
fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str',
'domain', 'is_reserved', 'allow', 'views',
'dhcpd_raw_include', 'dhcp_enabled')
widgets = {'views': forms.CheckboxSelectMultiple,
'range_type': forms.RadioSelect,
'ip_type': forms.RadioSelect}
exclude = 'range_usage'
def __init__(self, *args, **kwargs):
super(RangeForm, self).__init__(*args, **kwargs)
self.fields['dhcpd_raw_include'].label = "DHCP Config Extras"
self.fields['dhcpd_raw_include'].widget.attrs.update(
{'cols': '80',
'style': 'display: none;width: 680px'})
RangeAVForm = get_eav_form(RangeAV, Range)
| from django import forms
from cyder.base.eav.forms import get_eav_form
from cyder.base.mixins import UsabilityFormMixin
from cyder.cydhcp.range.models import Range, RangeAV
from cyder.cydns.forms import ViewChoiceForm
class RangeForm(ViewChoiceForm, UsabilityFormMixin):
class Meta:
model = Range
exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower')
fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str',
'domain', 'is_reserved', 'allow', 'views',
'dhcpd_raw_include', 'dhcp_enabled', 'name')
widgets = {'views': forms.CheckboxSelectMultiple,
'range_type': forms.RadioSelect,
'ip_type': forms.RadioSelect}
exclude = 'range_usage'
def __init__(self, *args, **kwargs):
super(RangeForm, self).__init__(*args, **kwargs)
self.fields['dhcpd_raw_include'].label = "DHCP Config Extras"
self.fields['dhcpd_raw_include'].widget.attrs.update(
{'cols': '80',
'style': 'display: none;width: 680px'})
RangeAVForm = get_eav_form(RangeAV, Range)
| Include name field in range form | Include name field in range form
| Python | bsd-3-clause | OSU-Net/cyder,zeeman/cyder,drkitty/cyder,OSU-Net/cyder,akeym/cyder,murrown/cyder,murrown/cyder,akeym/cyder,zeeman/cyder,akeym/cyder,murrown/cyder,murrown/cyder,drkitty/cyder,zeeman/cyder,akeym/cyder,zeeman/cyder,OSU-Net/cyder,drkitty/cyder,drkitty/cyder,OSU-Net/cyder | ---
+++
@@ -12,7 +12,7 @@
exclude = ('start_upper', 'start_lower', 'end_upper', 'end_lower')
fields = ('network', 'ip_type', 'range_type', 'start_str', 'end_str',
'domain', 'is_reserved', 'allow', 'views',
- 'dhcpd_raw_include', 'dhcp_enabled')
+ 'dhcpd_raw_include', 'dhcp_enabled', 'name')
widgets = {'views': forms.CheckboxSelectMultiple,
'range_type': forms.RadioSelect,
'ip_type': forms.RadioSelect} |
13ee0e2084765dcf958f4dbc844da54750878242 | snapshottest/django.py | snapshottest/django.py | from __future__ import absolute_import
from django.test import TestCase as dTestCase
from django.test.runner import DiscoverRunner
from snapshottest.reporting import reporting_lines
from .unittest import TestCase as uTestCase
class TestRunner(DiscoverRunner):
separator1 = "=" * 70
separator2 = "-" * 70
def __init__(self, snapshot_update=False, **kwargs):
super(TestRunner, self).__init__(**kwargs)
TestCase.snapshot_should_update = snapshot_update
@classmethod
def add_arguments(cls, parser):
super(TestRunner, cls).add_arguments(parser)
parser.add_argument(
'--snapshot-update', default=False, action='store_true',
dest='snapshot_update', help='Update the snapshots automatically.',
)
def run_tests(self, test_labels, extra_tests=None, **kwargs):
result = super(TestRunner, self).run_tests(
test_labels=test_labels,
extra_tests=extra_tests,
**kwargs
)
self.print_report()
return result
def print_report(self):
print("\n" + self.separator1)
print('SnapshotTest summary')
print(self.separator2)
for line in reporting_lines('python manage.py test'):
print(line)
print(self.separator1)
class TestCase(uTestCase, dTestCase):
pass
| from __future__ import absolute_import
from django.test import TestCase as dTestCase
from django.test import SimpleTestCase as dSimpleTestCase
from django.test.runner import DiscoverRunner
from snapshottest.reporting import reporting_lines
from .unittest import TestCase as uTestCase
class TestRunner(DiscoverRunner):
separator1 = "=" * 70
separator2 = "-" * 70
def __init__(self, snapshot_update=False, **kwargs):
super(TestRunner, self).__init__(**kwargs)
uTestCase.snapshot_should_update = snapshot_update
@classmethod
def add_arguments(cls, parser):
super(TestRunner, cls).add_arguments(parser)
parser.add_argument(
'--snapshot-update', default=False, action='store_true',
dest='snapshot_update', help='Update the snapshots automatically.',
)
def run_tests(self, test_labels, extra_tests=None, **kwargs):
result = super(TestRunner, self).run_tests(
test_labels=test_labels,
extra_tests=extra_tests,
**kwargs
)
self.print_report()
return result
def print_report(self):
print("\n" + self.separator1)
print('SnapshotTest summary')
print(self.separator2)
for line in reporting_lines('python manage.py test'):
print(line)
print(self.separator1)
class TestCase(uTestCase, dTestCase):
pass
class SimpleTestCase(uTestCase, dSimpleTestCase):
pass
| Allow use of alternate Django test cases | Allow use of alternate Django test cases
| Python | mit | syrusakbary/snapshottest | ---
+++
@@ -1,5 +1,6 @@
from __future__ import absolute_import
from django.test import TestCase as dTestCase
+from django.test import SimpleTestCase as dSimpleTestCase
from django.test.runner import DiscoverRunner
from snapshottest.reporting import reporting_lines
@@ -13,7 +14,7 @@
def __init__(self, snapshot_update=False, **kwargs):
super(TestRunner, self).__init__(**kwargs)
- TestCase.snapshot_should_update = snapshot_update
+ uTestCase.snapshot_should_update = snapshot_update
@classmethod
def add_arguments(cls, parser):
@@ -43,3 +44,6 @@
class TestCase(uTestCase, dTestCase):
pass
+
+class SimpleTestCase(uTestCase, dSimpleTestCase):
+ pass |
7c591a38bc89350ea2586fb83a6880cdf71b4a9a | passwd_change.py | passwd_change.py | #!/usr/bin/env python3
import sys
_args = sys.argv
if __name__ == "__main__":
if len(_args) == 4:
keys_file = _args[1]
target_file = _args[2]
result_file = _args[3]
with open(keys_file, 'r') as k:
keys = k.readlines()
keys = [key.strip() for key in keys]
keys = [key for key in keys if key != '']
with open(target_file, 'r') as t:
target_lines = t.readlines()
with open(result_file, 'w') as r:
for line in target_lines:
if line.split(':')[0] in keys:
r.write(line)
else:
print('./passwd_change.py keys_file.txt passwd_file result_file')
| #!/usr/bin/env python3
import sys
_args = sys.argv
if __name__ == "__main__":
if len(_args) == 4:
keys_file = _args[1]
target_file = _args[2]
result_file = _args[3]
try:
with open(keys_file, 'r') as k:
keys = k.readlines()
keys = [key.strip().split('@')[0] for key in keys]
keys = [key for key in keys if key != '']
with open(target_file, 'r') as t:
target_lines = t.readlines()
with open(result_file, 'w') as r:
for line in target_lines:
if line.split(':')[0] in keys or line.split(':')[3] != '12':
r.write(line)
except Exception as e:
print(str(e))
sys.exit()
else:
print('./passwd_change.py keys_file.txt passwd_file result_file')
| Add Exception to all with's block. | Add Exception to all with's block.
| Python | mit | maxsocl/oldmailer | ---
+++
@@ -1,7 +1,6 @@
#!/usr/bin/env python3
import sys
-
_args = sys.argv
@@ -11,17 +10,21 @@
target_file = _args[2]
result_file = _args[3]
- with open(keys_file, 'r') as k:
- keys = k.readlines()
- keys = [key.strip() for key in keys]
- keys = [key for key in keys if key != '']
- with open(target_file, 'r') as t:
- target_lines = t.readlines()
+ try:
+ with open(keys_file, 'r') as k:
+ keys = k.readlines()
+ keys = [key.strip().split('@')[0] for key in keys]
+ keys = [key for key in keys if key != '']
+ with open(target_file, 'r') as t:
+ target_lines = t.readlines()
- with open(result_file, 'w') as r:
- for line in target_lines:
- if line.split(':')[0] in keys:
- r.write(line)
+ with open(result_file, 'w') as r:
+ for line in target_lines:
+ if line.split(':')[0] in keys or line.split(':')[3] != '12':
+ r.write(line)
+
+ except Exception as e:
+ print(str(e))
+ sys.exit()
else:
print('./passwd_change.py keys_file.txt passwd_file result_file')
- |
c79c573c93a96bf5b631472c5e7efccc60102813 | yatsm/log_yatsm.py | yatsm/log_yatsm.py | import logging
FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s'
logging.basicConfig(format=FORMAT, level=logging.INFO, datefmt='%H:%M:%S')
logger = logging.getLogger('yatsm')
| import logging
_FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s'
_formatter = logging.Formatter(_FORMAT)
_handler = logging.StreamHandler()
_handler.setFormatter(_formatter)
logger = logging.getLogger('yatsm')
logger.addHandler(_handler)
logger.setLevel(logging.INFO)
| Change logger to be more friendly and play nice | Change logger to be more friendly and play nice
| Python | mit | ceholden/yatsm,ceholden/yatsm,valpasq/yatsm,c11/yatsm,valpasq/yatsm,c11/yatsm | ---
+++
@@ -1,5 +1,10 @@
import logging
-FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s'
-logging.basicConfig(format=FORMAT, level=logging.INFO, datefmt='%H:%M:%S')
+_FORMAT = '%(asctime)s:%(levelname)s:%(module)s.%(funcName)s:%(message)s'
+_formatter = logging.Formatter(_FORMAT)
+_handler = logging.StreamHandler()
+_handler.setFormatter(_formatter)
+
logger = logging.getLogger('yatsm')
+logger.addHandler(_handler)
+logger.setLevel(logging.INFO) |
03ebfe0518a7ac39f9414b3e8d8638c9dcba917c | tests/auth/test_models.py | tests/auth/test_models.py | # -*- coding: utf-8 -*-
from django.core.urlresolvers import reverse
from django.test import TestCase
from django.utils import unittest
from bakery.auth.models import BakeryUser
class TestBakeryUserModel(TestCase):
@unittest.skip('Not yet implemented')
def test_get_absolute_url(self):
user = BakeryUser.objects.create_user('user', 'password')
user.name = 'John Doe'
self.assertEqual(user.get_absolute_url(), reverse('user-detail-view'))
def test_get_full_name(self):
user = BakeryUser.objects.create_user('user', 'password')
user.name = 'John Doe'
self.assertEqual(user.get_full_name(), 'John Doe')
def test_get_short_name(self):
user = BakeryUser.objects.create_user('user', 'password')
user.name = 'John Doe'
self.assertEqual(user.get_short_name(), 'John Doe')
| # -*- coding: utf-8 -*-
from django.test import TestCase
from bakery.auth.models import BakeryUser
class TestBakeryUserModel(TestCase):
def test_get_absolute_url(self):
user = BakeryUser.objects.create_user('user', 'password')
user.name = 'John Doe'
self.assertEqual(user.get_absolute_url(), '/profile/user/')
def test_get_full_name(self):
user = BakeryUser.objects.create_user('user', 'password')
user.name = 'John Doe'
self.assertEqual(user.get_full_name(), 'John Doe')
def test_get_short_name(self):
user = BakeryUser.objects.create_user('user', 'password')
user.name = 'John Doe'
self.assertEqual(user.get_short_name(), 'John Doe')
| Adjust test (refers prev commit) | Adjust test (refers prev commit)
| Python | bsd-3-clause | muffins-on-dope/bakery,muffins-on-dope/bakery,muffins-on-dope/bakery | ---
+++
@@ -1,19 +1,16 @@
# -*- coding: utf-8 -*-
-from django.core.urlresolvers import reverse
from django.test import TestCase
-from django.utils import unittest
from bakery.auth.models import BakeryUser
class TestBakeryUserModel(TestCase):
- @unittest.skip('Not yet implemented')
def test_get_absolute_url(self):
user = BakeryUser.objects.create_user('user', 'password')
user.name = 'John Doe'
- self.assertEqual(user.get_absolute_url(), reverse('user-detail-view'))
+ self.assertEqual(user.get_absolute_url(), '/profile/user/')
def test_get_full_name(self):
user = BakeryUser.objects.create_user('user', 'password') |
853eb4896315c7fc60b1cbd7c87be9f7674f01ba | urls.py | urls.py | from django.conf.urls.defaults import *
from django.contrib import admin
from django.contrib.auth.decorators import login_required
from django.views.generic.simple import direct_to_template
import settings
admin.autodiscover()
urlpatterns = patterns('',
(r'^admin/', include(admin.site.urls)),
(r'^idp/', include('authentic.idp.urls')),
(r'^accounts/', include('registration.urls')),
(r'^$', login_required(direct_to_template),
{ 'template': 'index.html' }, 'index'),
)
if settings.AUTH_OPENID:
urlpatterns += patterns('',
(r'^openid/', include('django_authopenid.urls')),
)
if settings.AUTH_SSL:
urlpatterns += patterns('',
url(r'^sslauth/$',
'authentic.sslauth.login_ssl.process_request',
name='user_signin_ssl'),
url(r'^error_ssl/$', direct_to_template,
{'template': 'error_ssl.html'}, 'error_ssl'),
)
if settings.STATIC_SERVE:
urlpatterns += patterns('',
url(
regex = r'^media/(?P<path>.*)$',
view = 'django.views.static.serve',
kwargs = {'document_root': settings.MEDIA_ROOT}),
)
| from django.conf.urls.defaults import *
from django.contrib import admin
from django.contrib.auth.decorators import login_required
from django.views.generic.simple import direct_to_template
import settings
admin.autodiscover()
urlpatterns = patterns('',
(r'^admin/', include(admin.site.urls)),
(r'^idp/', include('authentic.idp.urls')),
(r'^$', login_required(direct_to_template),
{ 'template': 'index.html' }, 'index'),
)
if settings.AUTH_OPENID:
urlpatterns += patterns('',
(r'^accounts/openid/', include('django_authopenid.urls')),
)
urlpatterns += patterns('',
(r'^accounts/', include('registration.urls')),
)
if settings.AUTH_SSL:
urlpatterns += patterns('',
url(r'^sslauth/$',
'authentic.sslauth.login_ssl.process_request',
name='user_signin_ssl'),
url(r'^error_ssl/$', direct_to_template,
{'template': 'error_ssl.html'}, 'error_ssl'),
)
if settings.STATIC_SERVE:
urlpatterns += patterns('',
url(
regex = r'^media/(?P<path>.*)$',
view = 'django.views.static.serve',
kwargs = {'document_root': settings.MEDIA_ROOT}),
)
| Move OpenID stuff under /accounts/openid/ | Move OpenID stuff under /accounts/openid/
| Python | agpl-3.0 | adieu/authentic2,adieu/authentic2,pu239ppy/authentic2,incuna/authentic,incuna/authentic,BryceLohr/authentic,BryceLohr/authentic,pu239ppy/authentic2,incuna/authentic,adieu/authentic2,adieu/authentic2,BryceLohr/authentic,BryceLohr/authentic,incuna/authentic,incuna/authentic,pu239ppy/authentic2,pu239ppy/authentic2 | ---
+++
@@ -10,15 +10,18 @@
urlpatterns = patterns('',
(r'^admin/', include(admin.site.urls)),
(r'^idp/', include('authentic.idp.urls')),
- (r'^accounts/', include('registration.urls')),
(r'^$', login_required(direct_to_template),
{ 'template': 'index.html' }, 'index'),
)
if settings.AUTH_OPENID:
urlpatterns += patterns('',
- (r'^openid/', include('django_authopenid.urls')),
+ (r'^accounts/openid/', include('django_authopenid.urls')),
)
+
+urlpatterns += patterns('',
+ (r'^accounts/', include('registration.urls')),
+)
if settings.AUTH_SSL:
urlpatterns += patterns('', |
73af0eed3ce746154b957af5c05137f9e432c7a3 | tests/test_pkgmanifest.py | tests/test_pkgmanifest.py | # Copyright (C) Ivan Kravets <me@ikravets.com>
# See LICENSE for details.
import requests
from platformio.util import get_api_result
def pytest_generate_tests(metafunc):
if "package_data" not in metafunc.fixturenames:
return
pkgs_manifest = get_api_result("/packages")
assert isinstance(pkgs_manifest, dict)
packages = []
for _, variants in pkgs_manifest.iteritems():
for item in variants:
packages.append(item)
metafunc.parametrize("package_data", packages)
def validate_response(req):
assert req.status_code == 200
assert int(req.headers['Content-Length']) > 0
def validate_package(url):
r = requests.head(url, allow_redirects=True)
validate_response(r)
assert r.headers['Content-Type'] == "application/x-gzip"
def test_package(package_data):
assert package_data['url'].endswith("%d.tar.gz" % package_data['version'])
validate_package(package_data['url'])
| # Copyright (C) Ivan Kravets <me@ikravets.com>
# See LICENSE for details.
import requests
from platformio.util import get_api_result
def pytest_generate_tests(metafunc):
if "package_data" not in metafunc.fixturenames:
return
pkgs_manifest = get_api_result("/packages")
assert isinstance(pkgs_manifest, dict)
packages = []
for _, variants in pkgs_manifest.iteritems():
for item in variants:
packages.append(item)
metafunc.parametrize("package_data", packages)
def validate_response(req):
assert req.status_code == 200
assert int(req.headers['Content-Length']) > 0
def validate_package(url):
r = requests.head(url, allow_redirects=True)
validate_response(r)
assert r.headers['Content-Type'] in ("application/x-gzip",
"application/octet-stream")
def test_package(package_data):
assert package_data['url'].endswith("%d.tar.gz" % package_data['version'])
validate_package(package_data['url'])
| Add "application/octet-stream" mime type for package | Add "application/octet-stream" mime type for package
| Python | apache-2.0 | bkudria/platformio,awong1900/platformio,bkudria/platformio,platformio/platformio,jrobeson/platformio,platformio/platformio-core,TimJay/platformio,jrobeson/platformio,eiginn/platformio,platformio/platformio-core,bkudria/platformio,TimJay/platformio,TimJay/platformio,awong1900/platformio,valeros/platformio,ZachMassia/platformio,TimJay/platformio,mseroczynski/platformio,awong1900/platformio,jrobeson/platformio,mcanthony/platformio,dkuku/platformio,mplewis/platformio,TimJay/platformio,bkudria/platformio,jrobeson/platformio,atyenoria/platformio | ---
+++
@@ -25,7 +25,8 @@
def validate_package(url):
r = requests.head(url, allow_redirects=True)
validate_response(r)
- assert r.headers['Content-Type'] == "application/x-gzip"
+ assert r.headers['Content-Type'] in ("application/x-gzip",
+ "application/octet-stream")
def test_package(package_data): |
76dbf84facac62345ba02f17b33844a41d326d1f | tg/tests/test_testutil.py | tg/tests/test_testutil.py | """Tests for the testutil module"""
from unittest import TestCase
from sqlalchemy import create_engine
from sqlalchemy.exc import DBAPIError
import transaction
from tg.testutil import DBTest
from tg.tests.fixtures import model
# Ideally, we would have defined several different descendants of DBTest,
# in order to test its behavior in different situations, but there seem to be
# a problem in unittests and grand-grandchildren of TestCase won't work. You
# may try this code if you want: http://paste.turbogears.org/paste/4721
# or http://paste.turbogears.org/paste/4724
class BaseModelTest(DBTest):
database = create_engine("sqlite:///:memory:")
model = model
class TestGroup(BaseModelTest):
"""Test case for the Group model.
This should tell us whether the setUp() and tearDown() of DBTest work as
expected.
"""
def test_group_creation(self):
group = model.Group()
group.group_name = u"turbogears"
group.display_name = u"The TurboGears Team"
model.DBSession.save(group)
model.DBSession.flush()
transaction.commit()
def test_this_group_was_already_removed(self):
group = model.Group()
group.group_name = u"turbogears"
group.display_name = u"The TurboGears Team"
model.DBSession.save(group)
model.DBSession.flush()
transaction.commit()
| """Tests for the testutil module"""
from unittest import TestCase
from sqlalchemy import create_engine
from sqlalchemy.exc import DBAPIError
import transaction
from tg.testutil import DBTest
from tg.tests.fixtures import model
# Ideally, we would have defined several different descendants of DBTest,
# in order to test its behavior in different situations, but there seem to be
# a problem in unittests and grand-grandchildren of TestCase won't work. You
# may try this code if you want: http://paste.turbogears.org/paste/4721
# or http://paste.turbogears.org/paste/4724
class BaseModelTest(DBTest):
database = create_engine("sqlite:///:memory:")
model = model
class TestGroup(BaseModelTest):
"""Test case for the Group model.
This should tell us whether the setUp() and tearDown() of DBTest work as
expected.
"""
def test_group_creation(self):
group = model.Group()
group.group_name = u"turbogears"
group.display_name = u"The TurboGears Team"
model.DBSession.add(group)
model.DBSession.flush()
transaction.commit()
def test_this_group_was_already_removed(self):
group = model.Group()
group.group_name = u"turbogears"
group.display_name = u"The TurboGears Team"
model.DBSession.add(group)
model.DBSession.flush()
transaction.commit()
| Fix from cdevienne for SA Deprecation warnings | Fix from cdevienne for SA Deprecation warnings
--HG--
extra : convert_revision : svn%3A77541ad4-5f01-0410-9ede-a1b63cd9a898/trunk%405878
| Python | mit | lucius-feng/tg2,lucius-feng/tg2 | ---
+++
@@ -33,7 +33,7 @@
group = model.Group()
group.group_name = u"turbogears"
group.display_name = u"The TurboGears Team"
- model.DBSession.save(group)
+ model.DBSession.add(group)
model.DBSession.flush()
transaction.commit()
@@ -41,6 +41,6 @@
group = model.Group()
group.group_name = u"turbogears"
group.display_name = u"The TurboGears Team"
- model.DBSession.save(group)
+ model.DBSession.add(group)
model.DBSession.flush()
transaction.commit() |
daa4021011778f7511ad2c97648155bb17539d98 | tests/func/test_examples.py | tests/func/test_examples.py | import pytest # noqa
import os
import sys
import glob
import imp
def test_examples():
examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)),
'../../examples/*/*.py')
# Filter out __init__.py
examples = [f for f in glob.glob(examples_pat)
if not any([x in f for x in ['__init__.py',
'molecular',
'custom_table_caching']])]
for e in examples:
example_dir = os.path.dirname(e)
sys.path.insert(0, example_dir)
(module_name, _) = os.path.splitext(os.path.basename(e))
(module_file, module_path, desc) = \
imp.find_module(module_name, [example_dir])
m = imp.load_module(module_name, module_file, module_path, desc)
if hasattr(m, 'main'):
m.main(debug=False)
| import pytest # noqa
import os
import sys
import glob
import importlib
def test_examples():
examples_pat = os.path.join(os.path.abspath(os.path.dirname(__file__)),
'../../examples/*/*.py')
# Filter out __init__.py
examples = [f for f in glob.glob(examples_pat)
if not any([x in f for x in ['__init__.py',
'molecular',
'custom_table_caching']])]
for e in examples:
example_dir = os.path.dirname(e)
sys.path.insert(0, example_dir)
(module_name, _) = os.path.splitext(os.path.basename(e))
m = importlib.import_module(module_name)
if hasattr(m, 'main'):
m.main(debug=False)
| Replace deprecated imp with importlib | Replace deprecated imp with importlib
| Python | mit | igordejanovic/parglare,igordejanovic/parglare | ---
+++
@@ -2,7 +2,7 @@
import os
import sys
import glob
-import imp
+import importlib
def test_examples():
@@ -19,10 +19,7 @@
example_dir = os.path.dirname(e)
sys.path.insert(0, example_dir)
(module_name, _) = os.path.splitext(os.path.basename(e))
- (module_file, module_path, desc) = \
- imp.find_module(module_name, [example_dir])
-
- m = imp.load_module(module_name, module_file, module_path, desc)
+ m = importlib.import_module(module_name)
if hasattr(m, 'main'):
m.main(debug=False) |
fe98a627943c235ba24fc6de781deec69e7fd02e | relayer/__init__.py | relayer/__init__.py | from kafka import KafkaProducer
from .event_emitter import EventEmitter
from .exceptions import ConfigurationError
__version__ = '0.1.3'
class Relayer(object):
def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''):
self.logging_topic = logging_topic
if not kafka_hosts:
raise ConfigurationError()
if source == '':
self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix)
else:
self.source = source
producer = KafkaProducer(bootstrap_servers=kafka_hosts)
emitter = EventEmitter(producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix)
self.context = context_handler_class(emitter)
def emit(self, event_type, event_subtype, payload, partition_key=None):
payload = {
'source': self.source,
'event_type': event_type,
'event_subtype': event_subtype,
'payload': payload
}
self.context.emit(event_type, payload, partition_key)
def emit_raw(self, topic, message, partition_key=None):
self.context.emit(topic, message, partition_key)
def log(self, log_level, payload):
message = {
'log_level': log_level,
'payload': payload
}
self.context.log(message)
def flush(self):
self.emitter.flush()
| from kafka import KafkaProducer
from .event_emitter import EventEmitter
from .exceptions import ConfigurationError
__version__ = '0.1.3'
class Relayer(object):
def __init__(self, logging_topic, context_handler_class, kafka_hosts=None, topic_prefix='', topic_suffix='', source=''):
self.logging_topic = logging_topic
if not kafka_hosts:
raise ConfigurationError()
if source == '':
self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix)
else:
self.source = source
self._producer = KafkaProducer(bootstrap_servers=kafka_hosts)
self._emitter = EventEmitter(self._producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix)
self.context = context_handler_class(self._emitter)
def emit(self, event_type, event_subtype, payload, partition_key=None):
payload = {
'source': self.source,
'event_type': event_type,
'event_subtype': event_subtype,
'payload': payload
}
self.context.emit(event_type, payload, partition_key)
def emit_raw(self, topic, message, partition_key=None):
self.context.emit(topic, message, partition_key)
def log(self, log_level, payload):
message = {
'log_level': log_level,
'payload': payload
}
self.context.log(message)
def flush(self):
self._emitter.flush()
| Save event emitter y producer reference in relayer instance | Save event emitter y producer reference in relayer instance
| Python | mit | wizeline/relayer | ---
+++
@@ -16,9 +16,9 @@
self.source = '{0}{1}{2}'.format(topic_prefix, logging_topic, topic_suffix)
else:
self.source = source
- producer = KafkaProducer(bootstrap_servers=kafka_hosts)
- emitter = EventEmitter(producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix)
- self.context = context_handler_class(emitter)
+ self._producer = KafkaProducer(bootstrap_servers=kafka_hosts)
+ self._emitter = EventEmitter(self._producer, topic_prefix=topic_prefix, topic_suffix=topic_suffix)
+ self.context = context_handler_class(self._emitter)
def emit(self, event_type, event_subtype, payload, partition_key=None):
payload = {
@@ -40,4 +40,4 @@
self.context.log(message)
def flush(self):
- self.emitter.flush()
+ self._emitter.flush() |
64ac175029393c75a6b308bb36287081ab1bde8f | alignak_backend_import/__init__.py | alignak_backend_import/__init__.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
Alignak backend import
This module contains utility tools to import Nagios-like flat files configuration into
an Alignak REST backend.
"""
# Application version and manifest
VERSION = (0, 4, 3)
__application__ = u"Alignak backend import"
__short_version__ = '.'.join((str(each) for each in VERSION[:2]))
__version__ = '.'.join((str(each) for each in VERSION[:4]))
__author__ = u"Alignak team"
__copyright__ = u"(c) 2015-2016, %s" % __author__
__license__ = u"GNU Affero General Public License, version 3"
__description__ = u"Alignak backend import tools"
__releasenotes__ = u"""Alignak Backend import tools"""
__doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import"
# Application manifest
manifest = {
'name': __application__,
'version': __version__,
'author': __author__,
'description': __description__,
'copyright': __copyright__,
'license': __license__,
'release': __releasenotes__,
'doc': __doc_url__
}
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
Alignak backend import
This module contains utility tools to import Nagios-like flat files configuration into
an Alignak REST backend.
"""
# Application version and manifest
VERSION = (0, 4, 3)
__application__ = u"Alignak backend import"
__short_version__ = '.'.join((str(each) for each in VERSION[:2]))
__version__ = '.'.join((str(each) for each in VERSION[:4]))
__author__ = u"Alignak team"
__copyright__ = u"(c) 2015-2016, %s" % __author__
__license__ = u"GNU Affero General Public License, version 3"
__description__ = u"Alignak backend import tools"
__releasenotes__ = u"""Alignak Backend import tools"""
__doc_url__ = "https://github.com/Alignak-monitoring-contrib/alignak-backend-import"
# Application manifest
manifest = {
'name': __application__,
'version': __version__,
'author': __author__,
'description': __description__,
'copyright': __copyright__,
'license': __license__,
'release': __releasenotes__,
'doc': __doc_url__
}
| Fix bad indentation that broke PEP8 ! | Fix bad indentation that broke PEP8 !
| Python | agpl-3.0 | Alignak-monitoring-contrib/alignak-backend-import,Alignak-monitoring-contrib/alignak-backend-import | ---
+++
@@ -4,9 +4,9 @@
"""
Alignak backend import
-
- This module contains utility tools to import Nagios-like flat files configuration into
- an Alignak REST backend.
+
+ This module contains utility tools to import Nagios-like flat files configuration into
+ an Alignak REST backend.
"""
# Application version and manifest
VERSION = (0, 4, 3) |
da447a06c65a421b8d2fe122cb7b6bee5d11614e | app.py | app.py | import sys
from upload_s3 import set_metadata
from flask import Flask, render_template
from flask_frozen import Freezer
app = Flask(__name__)
freezer = Freezer(app)
app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html'
app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True
app.config['FREEZER_BASE_URL'] = 'http://www.vpr.net/apps/test'
# If project doesn't have it's own domain/subdomain, use BASE_URL
# app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base'
# If Flask is needed to generate URLs, use freezer.register_generator
# see: http://pythonhosted.org/Frozen-Flask/#url-generators
@app.route('/')
def index():
return render_template('content.html')
@app.route('/test/work')
def work():
return render_template('content.html')
if __name__ == '__main__':
if len(sys.argv) > 1 and sys.argv[1] == 'build':
freezer.freeze()
set_metadata()
else:
app.run(debug=True)
| import sys
from upload_s3 import set_metadata
from flask import Flask, render_template
from flask_frozen import Freezer
app = Flask(__name__)
freezer = Freezer(app)
app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html'
app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True
# If project doesn't have it's own domain/subdomain, use BASE_URL
# app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base'
# If Flask is needed to generate URLs, use freezer.register_generator
# see: http://pythonhosted.org/Frozen-Flask/#url-generators
@app.route('/')
def index():
return render_template('content.html')
if __name__ == '__main__':
if len(sys.argv) > 1 and sys.argv[1] == 'build':
freezer.freeze()
set_metadata()
else:
app.run(debug=True)
| Comment out optional base url configuration | Comment out optional base url configuration
| Python | apache-2.0 | vprnet/interactive-transcript-gov-peter-shumlins-2015-budget-speech,vprnet/google-s3-json,vprnet/app-template,vprnet/EOTS-iframe-widget,vprnet/soundcloud-podcast,vprnet/interactive-transcript-gov-peter-shumlins-2015-budget-speech,vprnet/timeline-dcf-systemic-failure,vprnet/live-from-the-fort,vprnet/live-from-the-fort,vprnet/EOTS-iframe-widget,vprnet/timeline-dcf-systemic-failure,vprnet/app-template,vprnet/old-app-template,vprnet/interactive-transcript-gov-peter-shumlins-third-inaugural-address,vprnet/live-from-the-fort,vprnet/EOTS-iframe-widget,vprnet/old-app-template,vprnet/app-template,vprnet/interactive-transcript-gov-peter-shumlins-third-inaugural-address | ---
+++
@@ -9,7 +9,6 @@
app.config['FREEZER_DEFAULT_MIMETYPE'] = 'text/html'
app.config['FREEZER_IGNORE_MIMETYPE_WARNINGS'] = True
-app.config['FREEZER_BASE_URL'] = 'http://www.vpr.net/apps/test'
# If project doesn't have it's own domain/subdomain, use BASE_URL
# app.config['FREEZER_BASE_URL'] = 'http://www.example.com/not_base'
@@ -22,12 +21,6 @@
def index():
return render_template('content.html')
-
-@app.route('/test/work')
-def work():
- return render_template('content.html')
-
-
if __name__ == '__main__':
if len(sys.argv) > 1 and sys.argv[1] == 'build':
freezer.freeze() |
07b22d0e4912678e3e2e5507e8a7fe2488ce2b3c | djangocms_installer/config/urls.py | djangocms_installer/config/urls.py | # -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function, unicode_literals
from cms.sitemaps import CMSSitemap
from django.conf import settings
from django.conf.urls import include, patterns, url
from django.conf.urls.i18n import i18n_patterns
from django.contrib import admin
from django.contrib.staticfiles.urls import staticfiles_urlpatterns
admin.autodiscover()
urlpatterns = i18n_patterns('',
url(r'^admin/', include(admin.site.urls)), # NOQA
url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap',
{'sitemaps': {'cmspages': CMSSitemap}}),
url(r'^select2/', include('django_select2.urls')),
url(r'^', include('cms.urls')),
)
# This is only needed when using runserver.
if settings.DEBUG:
urlpatterns = patterns('',
url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA
{'document_root': settings.MEDIA_ROOT, 'show_indexes': True}),
) + staticfiles_urlpatterns() + urlpatterns # NOQA
| # -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function, unicode_literals
from cms.sitemaps import CMSSitemap
from django.conf import settings
from django.conf.urls import include, patterns, url
from django.conf.urls.i18n import i18n_patterns
from django.contrib import admin
from django.contrib.staticfiles.urls import staticfiles_urlpatterns
admin.autodiscover()
urlpatterns = [
url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap',
{'sitemaps': {'cmspages': CMSSitemap}}),
url(r'^select2/', include('django_select2.urls')),
]
urlpatterns += i18n_patterns('',
url(r'^admin/', include(admin.site.urls)), # NOQA
url(r'^', include('cms.urls')),
)
# This is only needed when using runserver.
if settings.DEBUG:
urlpatterns = patterns('',
url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA
{'document_root': settings.MEDIA_ROOT, 'show_indexes': True}),
) + staticfiles_urlpatterns() + urlpatterns # NOQA
| Move sitemaps to non-language prefix url | Move sitemaps to non-language prefix url
| Python | bsd-3-clause | nephila/djangocms-installer,nephila/djangocms-installer | ---
+++
@@ -10,11 +10,14 @@
admin.autodiscover()
-urlpatterns = i18n_patterns('',
- url(r'^admin/', include(admin.site.urls)), # NOQA
+urlpatterns = [
url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap',
{'sitemaps': {'cmspages': CMSSitemap}}),
url(r'^select2/', include('django_select2.urls')),
+]
+
+urlpatterns += i18n_patterns('',
+ url(r'^admin/', include(admin.site.urls)), # NOQA
url(r'^', include('cms.urls')),
)
|
c2a2776f6bbff866078ed2ecaa30f9756af1a5dc | flow_workflow/historian/handler.py | flow_workflow/historian/handler.py | import logging
import sys
from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError
from flow_workflow.historian.messages import UpdateMessage
LOG = logging.getLogger(__name__)
class WorkflowHistorianMessageHandler(object):
message_class = UpdateMessage
def __init__(self, broker=None, storage=None, queue_name=None):
self.broker = broker
self.storage = storage
self.queue_name = queue_name
def __call__(self, message):
message_dict = message.to_dict()
LOG.info("Updating [net_key='%s', operation_id='%s']: %r",
message.net_key, message.operation_id, message_dict)
try:
self.storage.update(message_dict)
except (ResourceClosedError, TimeoutError, DisconnectionError):
LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.")
sys._exit()
| import logging
import os
from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError
from flow_workflow.historian.messages import UpdateMessage
LOG = logging.getLogger(__name__)
class WorkflowHistorianMessageHandler(object):
message_class = UpdateMessage
def __init__(self, broker=None, storage=None, queue_name=None):
self.broker = broker
self.storage = storage
self.queue_name = queue_name
def __call__(self, message):
message_dict = message.to_dict()
LOG.info("Updating [net_key='%s', operation_id='%s']: %r",
message.net_key, message.operation_id, message_dict)
try:
self.storage.update(message_dict)
except (ResourceClosedError, TimeoutError, DisconnectionError):
LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.")
os._exit(1)
| Use the more betterer os._exit | Use the more betterer os._exit
| Python | agpl-3.0 | genome/flow-workflow,genome/flow-workflow,genome/flow-workflow | ---
+++
@@ -1,5 +1,5 @@
import logging
-import sys
+import os
from sqlalchemy.exc import ResourceClosedError, TimeoutError, DisconnectionError
@@ -23,5 +23,5 @@
self.storage.update(message_dict)
except (ResourceClosedError, TimeoutError, DisconnectionError):
LOG.exception("This historian cannot handle messages anymore because it lost access to Oracle... exiting.")
- sys._exit()
+ os._exit(1)
|
b698c7ab1c13353d8e9538bb42797344049812c1 | astro.py | astro.py | import ephem
from datetime import datetime
def const(planet_name): # function name and parameters
planet_class = getattr(ephem, planet_name) # sets ephem object class
date_class = datetime.now()
planet = planet_class() # sets planet variable
south_bend = ephem.Observer() # Creates the Observer object
south_bend.lat = '41.40' # latitude
south_bend.lon = '-86.15'
south_bend.date = date_class # sets date parameter
planet.compute(south_bend) # calculates the location data
print date_class
print planet.ra, planet.dec
return ephem.constellation((planet.ra, planet.dec))
print const(raw_input('Planet: '))
| import ephem
from datetime import datetime
def const(planet_name): # function name and parameters
planet_class = getattr(ephem, planet_name) # sets ephem object class
date_class = datetime.now()
planet = planet_class() # sets planet variable
south_bend = ephem.Observer() # Creates the Observer object
south_bend.lat = '41.40' # latitude
south_bend.lon = '-86.15'
south_bend.date = date_class # sets date parameter
planet.compute(south_bend) # calculates the location data
print date_class
print planet.ra, planet.dec
print planet.alt, planet.az
return ephem.constellation((planet.ra, planet.dec))
print const(raw_input('Planet: '))
| Add to line 15 for testing. | Add to line 15 for testing.
| Python | mit | bennettscience/PySky | ---
+++
@@ -12,6 +12,7 @@
planet.compute(south_bend) # calculates the location data
print date_class
print planet.ra, planet.dec
+ print planet.alt, planet.az
return ephem.constellation((planet.ra, planet.dec))
print const(raw_input('Planet: ')) |
7816fc20c2e46a9d1a1e7c11a061862163018069 | tests/test_vfg_path.py | tests/test_vfg_path.py | import angr
import logging
import os
l = logging.getLogger("angr_tests")
test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)),
'../../binaries/tests'))
def test_vfg_paths():
p = angr.Project(os.path.join(test_location, "x86_64/track_user_input"))
main_addr = p.loader.main_bin.get_symbol("main").addr
printf_addr = 0x4005e1 # actually where it returns
vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=4)
paths = vfg.get_paths(main_addr, printf_addr)
if __name__ == '__main__':
test_vfg_paths()
| import angr
import logging
import os
l = logging.getLogger("angr_tests")
test_location = str(os.path.join(os.path.dirname(os.path.realpath(__file__)),
'../../binaries/tests'))
def test_vfg_paths():
p = angr.Project(os.path.join(test_location, "x86_64/track_user_input"))
main_addr = p.loader.main_bin.get_symbol("main").addr
printf_addr = 0x4005e1 # actually where it returns
vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=5)
paths = vfg.get_paths(main_addr, printf_addr)
if __name__ == '__main__':
test_vfg_paths()
| Fix the VFG path test. | Fix the VFG path test.
| Python | bsd-2-clause | tyb0807/angr,schieb/angr,schieb/angr,tyb0807/angr,chubbymaggie/angr,schieb/angr,chubbymaggie/angr,f-prettyland/angr,chubbymaggie/angr,angr/angr,angr/angr,axt/angr,axt/angr,axt/angr,f-prettyland/angr,angr/angr,tyb0807/angr,f-prettyland/angr,iamahuman/angr,iamahuman/angr,iamahuman/angr | ---
+++
@@ -11,7 +11,7 @@
main_addr = p.loader.main_bin.get_symbol("main").addr
printf_addr = 0x4005e1 # actually where it returns
- vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=4)
+ vfg = p.analyses.VFG(context_sensitivity_level=1, interfunction_level=5)
paths = vfg.get_paths(main_addr, printf_addr)
if __name__ == '__main__': |
d0018748cae3f0af4818106643926f7e8effe3c6 | monasca_log_api_tempest/clients.py | monasca_log_api_tempest/clients.py | # Copyright 2015 FUJITSU LIMITED
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from tempest import clients
from monasca_log_api_tempest.services import log_api_v2_client
from monasca_log_api_tempest.services import log_api_v3_client
from monasca_log_api_tempest.services import log_search_client
class Manager(clients.Manager):
def __init__(self, credentials=None, service=None):
super(Manager, self).__init__(credentials, service)
self.log_api_clients = {
"v2": log_api_v2_client.LogApiV2Client(
self.auth_provider,
'logs_v2',
None
),
"v3": log_api_v3_client.LogApiV3Client(
self.auth_provider,
'logs',
None
)
}
self.log_search_client = log_search_client.LogsSearchClient(
self.auth_provider,
'logs-search',
None
)
| # Copyright 2015-2016 FUJITSU LIMITED
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from tempest import clients
from monasca_log_api_tempest.services import log_api_v2_client
from monasca_log_api_tempest.services import log_api_v3_client
from monasca_log_api_tempest.services import log_search_client
class Manager(clients.Manager):
def __init__(self, credentials=None):
super(Manager, self).__init__(credentials)
self.log_api_clients = {
"v2": log_api_v2_client.LogApiV2Client(
self.auth_provider,
'logs_v2',
None
),
"v3": log_api_v3_client.LogApiV3Client(
self.auth_provider,
'logs',
None
)
}
self.log_search_client = log_search_client.LogsSearchClient(
self.auth_provider,
'logs-search',
None
)
| Fix the Monasca Log API tempest tests | Fix the Monasca Log API tempest tests
The Tempest Manager class must have changed and the service
argument apparently no longer exists. Instead, it was being set
as the scope which caused the catalog to not be retrieved
See-also: If934bac4e2cd833fe4e381c373218383354969ec
Change-Id: I43c023e91eb93e2c19096b0de812eabf7b2db62c
| Python | apache-2.0 | openstack/monasca-log-api,stackforge/monasca-log-api,stackforge/monasca-log-api,stackforge/monasca-log-api,openstack/monasca-log-api,openstack/monasca-log-api | ---
+++
@@ -1,4 +1,4 @@
-# Copyright 2015 FUJITSU LIMITED
+# Copyright 2015-2016 FUJITSU LIMITED
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
@@ -20,8 +20,8 @@
class Manager(clients.Manager):
- def __init__(self, credentials=None, service=None):
- super(Manager, self).__init__(credentials, service)
+ def __init__(self, credentials=None):
+ super(Manager, self).__init__(credentials)
self.log_api_clients = {
"v2": log_api_v2_client.LogApiV2Client( |
66fe6f98c079490d2d5de4c161da1d8b3801cda4 | monasca_persister/conf/influxdb.py | monasca_persister/conf/influxdb.py | # (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP
# Copyright 2017 FUJITSU LIMITED
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo_config import cfg
influxdb_opts = [
cfg.StrOpt('database_name',
help='database name where metrics are stored',
default='mon'),
cfg.IPOpt('ip_address',
help='ip address to influxdb'),
cfg.PortOpt('port',
help='port to influxdb',
default=8086),
cfg.StrOpt('user',
help='influxdb user ',
default='mon_persister'),
cfg.StrOpt('password',
secret=True,
help='influxdb password')]
influxdb_group = cfg.OptGroup(name='influxdb',
title='influxdb')
def register_opts(conf):
conf.register_group(influxdb_group)
conf.register_opts(influxdb_opts, influxdb_group)
def list_opts():
return influxdb_group, influxdb_opts
| # (C) Copyright 2016-2017 Hewlett Packard Enterprise Development LP
# Copyright 2017 FUJITSU LIMITED
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo_config import cfg
influxdb_opts = [
cfg.StrOpt('database_name',
help='database name where metrics are stored',
default='mon'),
cfg.HostAddressOpt('ip_address',
help='Valid IP address or hostname '
'to InfluxDB instance'),
cfg.PortOpt('port',
help='port to influxdb',
default=8086),
cfg.StrOpt('user',
help='influxdb user ',
default='mon_persister'),
cfg.StrOpt('password',
secret=True,
help='influxdb password')]
influxdb_group = cfg.OptGroup(name='influxdb',
title='influxdb')
def register_opts(conf):
conf.register_group(influxdb_group)
conf.register_opts(influxdb_opts, influxdb_group)
def list_opts():
return influxdb_group, influxdb_opts
| Allow hostnames to be used as ip_address | Allow hostnames to be used as ip_address
Previously introduced change for monasca-persister
had enforced the IPAddress as the only type one can
configure influxdb.ip_address property with.
Following change makes it possible to use also hostname.
Using IPAdress is still possible.
Change-Id: Ib0d7f19b3ac2dcb7c84923872d94f180cda58b2b
| Python | apache-2.0 | stackforge/monasca-persister,openstack/monasca-persister,stackforge/monasca-persister,stackforge/monasca-persister,openstack/monasca-persister,openstack/monasca-persister | ---
+++
@@ -20,8 +20,9 @@
cfg.StrOpt('database_name',
help='database name where metrics are stored',
default='mon'),
- cfg.IPOpt('ip_address',
- help='ip address to influxdb'),
+ cfg.HostAddressOpt('ip_address',
+ help='Valid IP address or hostname '
+ 'to InfluxDB instance'),
cfg.PortOpt('port',
help='port to influxdb',
default=8086), |
3ecca9dfa3f79a4c42a386e0bfe27cdca7e46a69 | tests/21-ct-clean-up-nc.py | tests/21-ct-clean-up-nc.py | import socket, sys
if len(sys.argv) != 6:
print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>')
localport = int(sys.argv[1])
timeout = int(sys.argv[2])
serverAddr = sys.argv[3]
serverPort = int(sys.argv[4])
httpPath = sys.argv[5]
if ":" not in serverAddr:
clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
host = serverAddr
else:
clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM)
host = "["+serverAddr+"]"
clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
clientsocket.bind(('', localport))
clientsocket.settimeout(timeout)
clientsocket.connect((serverAddr, serverPort))
clientsocket.send('GET '+httpPath+' HTTP/1.0\r\nHost: '+host+
'\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n')
data = clientsocket.recv(4096)
print(data)
| import socket, sys
if len(sys.argv) != 6:
print('Wrong number of arguments. Usage: ./21-ct-clean-up-nc.py <localport> <timeout> <remote-address> <remote-port> <HTTP path>')
localport = int(sys.argv[1])
timeout = int(sys.argv[2])
serverAddr = sys.argv[3]
serverPort = int(sys.argv[4])
httpPath = sys.argv[5]
if ":" not in serverAddr:
clientsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
host = serverAddr
else:
clientsocket = socket.socket(socket.AF_INET6, socket.SOCK_STREAM)
host = "["+serverAddr+"]"
clientsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
clientsocket.bind(('', localport))
clientsocket.settimeout(timeout)
clientsocket.connect((serverAddr, serverPort))
clientsocket.send('GET '+httpPath+' HTTP/1.1\r\nHost: '+host+
'\r\nConnection: close\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n')
data = clientsocket.recv(4096)
print(data)
| Use HTTP 1.1 instead of HTTP 1.0 | tests: Use HTTP 1.1 instead of HTTP 1.0
Envoy does not support HTTP 1.0, so use HTTP 1.1 instead.
Signed-off-by: Jarno Rajahalme <0f1ab0ac7dffd9db21aa539af2fd4bb04abc3ad4@covalent.io>
| Python | apache-2.0 | cilium/cilium,cilium/cilium,tgraf/cilium,michi-covalent/cilium,cilium-team/cilium,eloycoto/cilium,tgraf/cilium,eloycoto/cilium,cilium/cilium,eloycoto/cilium,scanf/cilium,cilium-team/cilium,scanf/cilium,tklauser/cilium,scanf/cilium,cilium/cilium,scanf/cilium,cilium/cilium,tgraf/cilium,tgraf/cilium,tklauser/cilium,eloycoto/cilium,eloycoto/cilium,tgraf/cilium,tklauser/cilium,tklauser/cilium,eloycoto/cilium,eloycoto/cilium,scanf/cilium,scanf/cilium,michi-covalent/cilium,scanf/cilium,michi-covalent/cilium,michi-covalent/cilium,tklauser/cilium,tgraf/cilium,michi-covalent/cilium | ---
+++
@@ -20,8 +20,8 @@
clientsocket.bind(('', localport))
clientsocket.settimeout(timeout)
clientsocket.connect((serverAddr, serverPort))
-clientsocket.send('GET '+httpPath+' HTTP/1.0\r\nHost: '+host+
- '\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n')
+clientsocket.send('GET '+httpPath+' HTTP/1.1\r\nHost: '+host+
+ '\r\nConnection: close\r\nUser-Agent: curl/7.38.0\r\nAccept: */*\r\n\r\n')
data = clientsocket.recv(4096)
print(data) |
5dd5d4e6ac93ad3867f98bd35412de11dd6b1dc2 | tests/test_generalwords.py | tests/test_generalwords.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
test_generalwords
----------------------------------
All the tests for the generalword module. Simple module, simple tests.
"""
import unittest
from generalwords import get_word
class TestGeneralwords(unittest.TestCase):
def setUp(self):
pass
def test_get_word(self):
self.assertIsNotNone(get_word)
def test_get_word_is_somewhat_random(self):
sample_size = 100
words = {get_word() for i in range(sample_size)}
self.assertAlmostEqual(len(words), sample_size,
delta=int((sample_size * 0.1)))
def tearDown(self):
pass
if __name__ == '__main__':
unittest.main()
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
test_generalwords
----------------------------------
All the tests for the generalword module. Simple module, simple tests.
"""
import unittest
from generalwords import get_word
class TestGeneralwords(unittest.TestCase):
def setUp(self):
pass
def test_get_word(self):
self.assertIsNotNone(get_word)
def test_get_word_is_somewhat_random(self):
sample_size = 100
words = set(get_word() for i in range(sample_size))
self.assertAlmostEqual(len(words), sample_size,
delta=int((sample_size * 0.1)))
def tearDown(self):
pass
if __name__ == '__main__':
unittest.main()
| Remove set-comprehensions so that tests will pass on 2.6 | Remove set-comprehensions so that tests will pass on 2.6
| Python | bsd-3-clause | petrilli/generalwords | ---
+++
@@ -22,7 +22,7 @@
def test_get_word_is_somewhat_random(self):
sample_size = 100
- words = {get_word() for i in range(sample_size)}
+ words = set(get_word() for i in range(sample_size))
self.assertAlmostEqual(len(words), sample_size,
delta=int((sample_size * 0.1)))
|
fd5634902f079c000c870192bb28fa4c35b956ec | snippets/urls.py | snippets/urls.py | from django.conf import settings
from django.conf.urls import patterns, include, url
from django.contrib.staticfiles.urls import staticfiles_urlpatterns
from django.http import HttpResponse
from funfactory.monkeypatches import patch
# Apply funfactory monkeypatches.
patch()
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
def robots_txt(request):
permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'
return HttpResponse('User-agent: *\n{0}: /'.format(permission),
mimetype='text/plain')
urlpatterns = patterns('',
url(r'', include('snippets.base.urls')),
url(r'^admin/', include('smuggler.urls')),
url(r'^admin/', include(admin.site.urls)),
url(r'^robots\.txt$', robots_txt)
)
## In DEBUG mode, serve media files through Django.
if settings.DEBUG:
urlpatterns += patterns('',
url(r'^media/(?P<path>.*)$', 'django.views.static.serve', {
'document_root': settings.MEDIA_ROOT,
}),
) + staticfiles_urlpatterns()
| from django.conf import settings
from django.conf.urls import patterns, include, url
from django.contrib.staticfiles.urls import staticfiles_urlpatterns
from django.http import HttpResponse
from django.views.static import serve as static_serve
from funfactory.monkeypatches import patch
# Apply funfactory monkeypatches.
patch()
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
def robots_txt(request):
permission = 'Allow' if settings.ENGAGE_ROBOTS else 'Disallow'
return HttpResponse('User-agent: *\n{0}: /'.format(permission),
mimetype='text/plain')
urlpatterns = patterns('',
url(r'', include('snippets.base.urls')),
url(r'^admin/', include('smuggler.urls')),
url(r'^admin/', include(admin.site.urls)),
url(r'^robots\.txt$', robots_txt)
)
## In DEBUG mode, serve media files through Django.
if settings.DEBUG:
# Use custom serve function that adds necessary headers.
def serve_media(*args, **kwargs):
response = static_serve(*args, **kwargs)
response['Access-Control-Allow-Origin'] = '*'
return response
urlpatterns += patterns('',
url(r'^media/(?P<path>.*)$', serve_media, {
'document_root': settings.MEDIA_ROOT,
}),
) + staticfiles_urlpatterns()
| Add CORS headers to dev server media. | Add CORS headers to dev server media. | Python | mpl-2.0 | mozilla/snippets-service,glogiotatidis/snippets-service,glogiotatidis/snippets-service,mozmar/snippets-service,schalkneethling/snippets-service,Osmose/snippets-service,bensternthal/snippets-service,bensternthal/snippets-service,Osmose/snippets-service,akatsoulas/snippets-service,schalkneethling/snippets-service,glogiotatidis/snippets-service,mozmar/snippets-service,akatsoulas/snippets-service,mozilla/snippets-service,mozilla/snippets-service,schalkneethling/snippets-service,schalkneethling/snippets-service,akatsoulas/snippets-service,bensternthal/snippets-service,Osmose/snippets-service,mozmar/snippets-service,bensternthal/snippets-service,mozilla/snippets-service,Osmose/snippets-service,glogiotatidis/snippets-service,mozmar/snippets-service,akatsoulas/snippets-service | ---
+++
@@ -2,6 +2,7 @@
from django.conf.urls import patterns, include, url
from django.contrib.staticfiles.urls import staticfiles_urlpatterns
from django.http import HttpResponse
+from django.views.static import serve as static_serve
from funfactory.monkeypatches import patch
@@ -29,8 +30,14 @@
## In DEBUG mode, serve media files through Django.
if settings.DEBUG:
+ # Use custom serve function that adds necessary headers.
+ def serve_media(*args, **kwargs):
+ response = static_serve(*args, **kwargs)
+ response['Access-Control-Allow-Origin'] = '*'
+ return response
+
urlpatterns += patterns('',
- url(r'^media/(?P<path>.*)$', 'django.views.static.serve', {
+ url(r'^media/(?P<path>.*)$', serve_media, {
'document_root': settings.MEDIA_ROOT,
}),
) + staticfiles_urlpatterns() |
8d0d4704f62b223128bab193cd5f5cda8e978c19 | polling_stations/apps/pollingstations/tests/test_urls.py | polling_stations/apps/pollingstations/tests/test_urls.py | import json
from django.test import TestCase
from django_extensions.management.commands.show_urls import Command
class UrlTests(TestCase):
def is_exception(self, url):
exceptions = [".txt", ".ics", ".geojson"]
for exception in exceptions:
if exception in url:
return True
return False
def test_trailing_slashes(self):
c = Command()
data = json.loads(
c.handle(
**{
"unsorted": False,
"language": None,
"decorator": [],
"format_style": "json",
"urlconf": "ROOT_URLCONF",
"no_color": True,
}
)
)
urls = [rec["url"] for rec in data]
for url in urls:
if self.is_exception(url):
continue
assert url[-1] == "/" or ">", url + " does not end with /"
| import json
from django.test import TestCase
from django_extensions.management.commands.show_urls import Command
class UrlTests(TestCase):
def is_exception(self, url):
exceptions = [".txt", ".ics", ".geojson"]
for exception in exceptions:
if exception in url:
return True
return False
def test_trailing_slashes(self):
c = Command()
data = json.loads(
c.handle(
**{
"unsorted": False,
"language": None,
"decorator": [],
"format_style": "json",
"urlconf": "ROOT_URLCONF",
"no_color": True,
}
)
)
urls = [rec["url"] for rec in data]
urls.remove("/admin/<url>")
for url in urls:
if self.is_exception(url):
continue
assert url[-1] == "/", url + " does not end with /"
| Refactor test for new view added with upgrade | Refactor test for new view added with upgrade
| Python | bsd-3-clause | DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations | ---
+++
@@ -26,8 +26,9 @@
)
)
urls = [rec["url"] for rec in data]
+ urls.remove("/admin/<url>")
for url in urls:
if self.is_exception(url):
continue
- assert url[-1] == "/" or ">", url + " does not end with /"
+ assert url[-1] == "/", url + " does not end with /" |
fb675239ae79adcdc5f050bcf8403effb067a59b | smsgateway/utils.py | smsgateway/utils.py | import logging
logger = logging.getLogger(__name__)
def strspn(source, allowed):
newchrs = []
for c in source:
if c in allowed:
newchrs.append(c)
return u''.join(newchrs)
def check_cell_phone_number(number):
cleaned_number = strspn(number, u'+0123456789')
if not u'+' in cleaned_number[:1]:
cleaned_number = u'+%s' % cleaned_number
return cleaned_number
def truncate_sms(text, max_length=160):
if len(text) <= max_length:
if len(text) > 140:
logger.warning("SMS possibly too long (>140 chars): %s", text)
return text
else:
logger.error("SMS is too long (>160 chars): %s", text)
return text[:max_length-3] + '...'
def parse_sms(content):
content = content.upper().strip()
from smsgateway.backends.base import hook
for keyword, subkeywords in hook.iteritems():
if content[:len(keyword)] == unicode(keyword):
remainder = content[len(keyword):].strip()
if '*' in subkeywords:
parts = remainder.split(u' ')
subkeyword = parts[0].strip()
if subkeyword in subkeywords:
return [keyword] + parts
return keyword, remainder
else:
for subkeyword in subkeywords:
if remainder[:len(subkeyword)] == unicode(subkeyword):
subremainder = remainder[len(subkeyword):].strip()
return [keyword, subkeyword] + subremainder.split()
return None
| import logging
logger = logging.getLogger(__name__)
def strspn(source, allowed):
newchrs = []
for c in source:
if c in allowed:
newchrs.append(c)
return u''.join(newchrs)
def check_cell_phone_number(number):
cleaned_number = strspn(number, u'+0123456789')
if not u'+' in cleaned_number[:1]:
cleaned_number = u'+%s' % cleaned_number
return cleaned_number
def truncate_sms(text, max_length=160):
if len(text) <= max_length:
return text
else:
logger.error("Trying to send an SMS that is too long: %s", text)
return text[:max_length-3] + '...'
def parse_sms(content):
content = content.upper().strip()
from smsgateway.backends.base import hook
for keyword, subkeywords in hook.iteritems():
if content[:len(keyword)] == unicode(keyword):
remainder = content[len(keyword):].strip()
if '*' in subkeywords:
parts = remainder.split(u' ')
subkeyword = parts[0].strip()
if subkeyword in subkeywords:
return [keyword] + parts
return keyword, remainder
else:
for subkeyword in subkeywords:
if remainder[:len(subkeyword)] == unicode(subkeyword):
subremainder = remainder[len(subkeyword):].strip()
return [keyword, subkeyword] + subremainder.split()
return None
| Remove warning for long messages | Remove warning for long messages
| Python | bsd-3-clause | mvpoland/django-smsgateway,peterayeni/django-smsgateway,peterayeni/django-smsgateway,peterayeni/django-smsgateway,mvpoland/django-smsgateway,peterayeni/django-smsgateway,mvpoland/django-smsgateway | ---
+++
@@ -18,11 +18,9 @@
def truncate_sms(text, max_length=160):
if len(text) <= max_length:
- if len(text) > 140:
- logger.warning("SMS possibly too long (>140 chars): %s", text)
return text
else:
- logger.error("SMS is too long (>160 chars): %s", text)
+ logger.error("Trying to send an SMS that is too long: %s", text)
return text[:max_length-3] + '...'
def parse_sms(content): |
e2831c6241e18d6f611eaaf51854b75b970867d7 | examples/demo.py | examples/demo.py | #-------------------------------------------------------------------------------
#
# Copyright (c) 2009, Enthought, Inc.
# All rights reserved.
#
# This software is provided without warranty under the terms of the BSD
# license included in enthought/LICENSE.txt and may be redistributed only
# under the conditions described in the aforementioned license. The license
# is also available online at http://www.enthought.com/licenses/BSD.txt
#
# Thanks for using Enthought open source!
#
# Author: Vibha Srinivasan
# Date: 02/03/2009
#
#-------------------------------------------------------------------------------
""" Run the Chaco demo.
"""
from enthought.traits.ui.extras.demo import demo
# Uncomment the config_filename portion to see a tree editor based on the
# examples.cfg file.
demo(use_files=True,
# config_filename='examples.cfg'
)
| #-------------------------------------------------------------------------------
#
# Copyright (c) 2009, Enthought, Inc.
# All rights reserved.
#
# This software is provided without warranty under the terms of the BSD
# license included in enthought/LICENSE.txt and may be redistributed only
# under the conditions described in the aforementioned license. The license
# is also available online at http://www.enthought.com/licenses/BSD.txt
#
# Thanks for using Enthought open source!
#
# Author: Vibha Srinivasan
# Date: 02/03/2009
#
#-------------------------------------------------------------------------------
""" Run the Chaco demo.
"""
from enthought.traits.ui.extras.demo import demo
# Uncomment the config_filename portion to see a tree editor based on the
# examples.cfg file.
demo(use_files=True,
config_filename='examples.cfg'
)
| Use config files to construct the tree of examples. | Use config files to construct the tree of examples.
| Python | bsd-3-clause | ContinuumIO/chaco,ContinuumIO/chaco,tommy-u/chaco,ContinuumIO/chaco,burnpanck/chaco,tommy-u/chaco,tommy-u/chaco,burnpanck/chaco,burnpanck/chaco,ContinuumIO/chaco | ---
+++
@@ -23,6 +23,6 @@
# Uncomment the config_filename portion to see a tree editor based on the
# examples.cfg file.
demo(use_files=True,
- # config_filename='examples.cfg'
+ config_filename='examples.cfg'
)
|
f9c9cd4505e9055a2905a87f91cdaab399352b27 | dthm4kaiako/config/__init__.py | dthm4kaiako/config/__init__.py | """Configuration for Django system."""
__version__ = "0.15.2"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
| """Configuration for Django system."""
__version__ = "0.16.0"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
| Increment version number to 0.16.0 | Increment version number to 0.16.0
| Python | mit | uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers | ---
+++
@@ -1,6 +1,6 @@
"""Configuration for Django system."""
-__version__ = "0.15.2"
+__version__ = "0.16.0"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num |
e93ec5939d5a47e20501b4812e02a6cdac4a3ba6 | rplugin/python3/deoplete/sources/LanguageClientSource.py | rplugin/python3/deoplete/sources/LanguageClientSource.py | from .base import Base
import re
CompleteOutputs = "g:LanguageClient_omniCompleteResults"
class Source(Base):
def __init__(self, vim):
super().__init__(vim)
self.name = "LanguageClient"
self.mark = "[LC]"
self.rank = 1000
self.min_pattern_length = 1
self.filetypes = vim.eval(
"get(g:, 'LanguageClient_serverCommands', {})").keys()
self.input_pattern += r'(\.|::|->)\w*$'
self.complete_pos = re.compile(r"\w*$")
def get_complete_position(self, context):
m = self.complete_pos.search(context['input'])
return m.start() if m else -1
def gather_candidates(self, context):
if context["is_async"]:
outputs = self.vim.eval(CompleteOutputs)
if len(outputs) != 0:
context["is_async"] = False
# TODO: error handling.
candidates = outputs[0].get("result", [])
# log(str(candidates))
return candidates
else:
context["is_async"] = True
self.vim.command("let {} = []".format(CompleteOutputs))
self.vim.funcs.LanguageClient_omniComplete({
"character": context["complete_position"],
})
return []
# f = open("/tmp/deoplete.log", "w")
# def log(message):
# f.writelines([message])
# f.flush()
| from .base import Base
import re
CompleteOutputs = "g:LanguageClient_omniCompleteResults"
class Source(Base):
def __init__(self, vim):
super().__init__(vim)
self.name = "LanguageClient"
self.mark = "[LC]"
self.rank = 1000
self.min_pattern_length = 0
self.filetypes = vim.eval(
"get(g:, 'LanguageClient_serverCommands', {})").keys()
self.input_pattern += r'(\.|::|->)\w*$'
self.complete_pos = re.compile(r"\w*$")
def get_complete_position(self, context):
m = self.complete_pos.search(context['input'])
return m.start() if m else -1
def gather_candidates(self, context):
if context["is_async"]:
outputs = self.vim.eval(CompleteOutputs)
if len(outputs) != 0:
context["is_async"] = False
# TODO: error handling.
candidates = outputs[0].get("result", [])
# log(str(candidates))
return candidates
else:
context["is_async"] = True
self.vim.command("let {} = []".format(CompleteOutputs))
self.vim.funcs.LanguageClient_omniComplete({
"character": context["complete_position"] + len(context["complete_str"]),
})
return []
# f = open("/tmp/deoplete.log", "w")
# def log(message):
# f.writelines([message])
# f.flush()
| Fix deoplete source compete position sent to language servers. | Fix deoplete source compete position sent to language servers.
| Python | mit | autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim,autozimu/LanguageClient-neovim | ---
+++
@@ -12,7 +12,7 @@
self.name = "LanguageClient"
self.mark = "[LC]"
self.rank = 1000
- self.min_pattern_length = 1
+ self.min_pattern_length = 0
self.filetypes = vim.eval(
"get(g:, 'LanguageClient_serverCommands', {})").keys()
self.input_pattern += r'(\.|::|->)\w*$'
@@ -35,7 +35,7 @@
context["is_async"] = True
self.vim.command("let {} = []".format(CompleteOutputs))
self.vim.funcs.LanguageClient_omniComplete({
- "character": context["complete_position"],
+ "character": context["complete_position"] + len(context["complete_str"]),
})
return []
|
e25f085025f881ccf0a0da2e620b09787819507a | sub.py | sub.py | import csv
import sys
import threading
from time import sleep
from datetime import datetime
import msgpack
import zmq
context = zmq.Context()
socket = context.socket(zmq.SUB)
socket.setsockopt(zmq.SUBSCRIBE, b'')
socket.bind("tcp://*:4200")
terminate = threading.Event()
def go():
global terminate
writer = None
firsttime = True
with open('ani.csv', 'w', newline='') as csvfile:
while not terminate.is_set():
try:
msg = socket.recv(flags=zmq.NOBLOCK)
except zmq.Again as e:
# No message received
continue
orig, msgpackdata = msg.split(b' ', 1)
unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8')
if not isinstance(unpacked, dict):
print("Message garbled: {}", unpacked)
continue
unpacked.update({'datetime': str(datetime.now())})
if firsttime:
writer = csv.DictWriter(csvfile, fieldnames=list(unpacked.keys()))
writer.writeheader()
firsttime = False
writer.writerow(unpacked)
print(msgpackdata, unpacked)
anithread = threading.Thread(target=go)
anithread.start()
while True:
try:
sleep(1)
except KeyboardInterrupt:
terminate.set()
anithread.join()
break
| import csv
import sys
import threading
from time import sleep
from datetime import datetime
import msgpack
import zmq
context = zmq.Context()
socket = context.socket(zmq.SUB)
socket.setsockopt(zmq.SUBSCRIBE, b'')
socket.bind("tcp://*:4200")
terminate = threading.Event()
def go():
global terminate
writer = None
firsttime = True
with open('ani.csv', 'w', newline='') as csvfile:
while not terminate.is_set():
try:
msg = socket.recv(flags=zmq.NOBLOCK)
except zmq.Again as e:
# No message received
continue
orig, msgpackdata = msg.split(b' ', 1)
unpacked = msgpack.unpackb(msgpackdata, encoding='utf-8')
if not isinstance(unpacked, dict):
print("Message garbled: {}", unpacked)
continue
if firsttime:
headers = ['datetime'] + list(unpacked.keys())
writer = csv.DictWriter(csvfile, fieldnames=headers)
writer.writeheader()
firsttime = False
unpacked.update({'datetime': str(datetime.now())})
writer.writerow(unpacked)
print(msgpackdata, unpacked)
anithread = threading.Thread(target=go)
anithread.start()
while True:
try:
sleep(1)
except KeyboardInterrupt:
terminate.set()
anithread.join()
break
| Move date/time to the first csv column. | Move date/time to the first csv column.
| Python | isc | jaj42/hsmedstream,jaj42/phystream | ---
+++
@@ -31,11 +31,12 @@
if not isinstance(unpacked, dict):
print("Message garbled: {}", unpacked)
continue
- unpacked.update({'datetime': str(datetime.now())})
if firsttime:
- writer = csv.DictWriter(csvfile, fieldnames=list(unpacked.keys()))
+ headers = ['datetime'] + list(unpacked.keys())
+ writer = csv.DictWriter(csvfile, fieldnames=headers)
writer.writeheader()
firsttime = False
+ unpacked.update({'datetime': str(datetime.now())})
writer.writerow(unpacked)
print(msgpackdata, unpacked)
|
866026a5d2f89a8ac76a726720e4fbe812c94eb4 | ds/providers/shell.py | ds/providers/shell.py | from __future__ import absolute_import, unicode_literals
__all__ = ['ShellProvider']
from .base import Provider
class ShellProvider(Provider):
def get_options(self):
return {
'command': {'required': True},
}
def execute(self, workspace, task):
command = task.provider_config['command'].format({
'environment': task.environment,
'sha': task.sha,
'ref': task.ref,
'task': task.name,
})
return workspace.run(command)
| from __future__ import absolute_import, unicode_literals
__all__ = ['ShellProvider']
from .base import Provider
class ShellProvider(Provider):
def get_options(self):
return {
'command': {'required': True},
}
def execute(self, workspace, task):
command = task.provider_config['command'].format(
environment=task.environment,
sha=task.sha,
ref=task.ref,
task=task.name,
)
return workspace.run(command)
| Fix arg passing to command | Fix arg passing to command
| Python | apache-2.0 | rshk/freight,klynton/freight,jkimbo/freight,getsentry/freight,getsentry/freight,klynton/freight,getsentry/freight,klynton/freight,rshk/freight,klynton/freight,jkimbo/freight,getsentry/freight,rshk/freight,jkimbo/freight,jkimbo/freight,rshk/freight,getsentry/freight | ---
+++
@@ -12,11 +12,11 @@
}
def execute(self, workspace, task):
- command = task.provider_config['command'].format({
- 'environment': task.environment,
- 'sha': task.sha,
- 'ref': task.ref,
- 'task': task.name,
- })
+ command = task.provider_config['command'].format(
+ environment=task.environment,
+ sha=task.sha,
+ ref=task.ref,
+ task=task.name,
+ )
return workspace.run(command) |
b05d2666c834a9c4d151d0340612010851bd4610 | eniric/Qcalculator.py | eniric/Qcalculator.py | """
Created on Mon Dec 29 00:14:56 2014
@author: pfigueira
Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric.
"""
# from eniric.IOmodule import read_2col
import numpy as np
import pandas as pd
c = 299792458 # m/s
def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"):
"""
function that claculates the RV precision achievable on a spectrum
"""
data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64)
wav, flux = data["wavelength"].values, data["flux"].values
return [c / SqrtSumWis(wavelength, flux)]
def SqrtSumWis(wavelength, flux):
"""
Calculation of the SquareRoot of the sum of the Wis for a spectrum
"""
delta_F = (np.array(flux[1:]) - np.array(flux[:-1]))
delta_l = np.array(wavelength[:-1])
derivF_over_lambda = delta_F/delta_l
return np.sqrt(np.sum(np.array(wavelength[:-1])**2.0 * derivF_over_lambda**2.0 / np.array(flux[:-1])))
| """
Created on Mon Dec 29 00:14:56 2014
@author: pfigueira
Editied Thur Dec 15 13:00 2016 by Jason Neal for eniric.
"""
# from eniric.IOmodule import read_2col
import numpy as np
import pandas as pd
c = 299792458 # m/s
def RVprec_calc(spectrum_file="resampled/Spectrum_M0-PHOENIX-ACES_Hband_vsini1.0_R60k_res3.txt"):
"""
function that claculates the RV precision achievable on a spectrum
"""
data = pd.read_table(spectrum_file, comment='#', names=["wavelength", "flux"], dtype=np.float64)
wav, flux = data["wavelength"].values, data["flux"].values
return [c / SqrtSumWis(wavelength, flux)]
def SqrtSumWis(wavelength, flux):
"""
Calculation of the SquareRoot of the sum of the Wis for a spectrum
"""
delta_F = np.diff(flux)
delta_l = wavelength[:-1]
derivF_over_lambda = delta_F/delta_l
return np.sqrt(np.sum(wavelength[:-1]**2.0 * derivF_over_lambda**2.0 / flux[:-1]))
| Update RVprec_calculation use numpy.diff() and remove unnecessary array calls. | Update RVprec_calculation
use numpy.diff() and remove unnecessary array calls.
Former-commit-id: 646ff0cea061feb87c08b819a47d8e9f3dd55b55 | Python | mit | jason-neal/eniric,jason-neal/eniric | ---
+++
@@ -28,9 +28,9 @@
Calculation of the SquareRoot of the sum of the Wis for a spectrum
"""
- delta_F = (np.array(flux[1:]) - np.array(flux[:-1]))
- delta_l = np.array(wavelength[:-1])
+ delta_F = np.diff(flux)
+ delta_l = wavelength[:-1]
derivF_over_lambda = delta_F/delta_l
- return np.sqrt(np.sum(np.array(wavelength[:-1])**2.0 * derivF_over_lambda**2.0 / np.array(flux[:-1])))
+ return np.sqrt(np.sum(wavelength[:-1]**2.0 * derivF_over_lambda**2.0 / flux[:-1])) |
7d0f3ba1aa82c2ea5a4a2eca2bbe842b63a82c72 | wafer/talks/serializers.py | wafer/talks/serializers.py | from rest_framework import serializers
from reversion import revisions
from wafer.talks.models import Talk
class TalkSerializer(serializers.ModelSerializer):
class Meta:
model = Talk
exclude = ('_abstract_rendered', )
@revisions.create_revision()
def create(self, validated_data):
revisions.set_comment("Created via REST api")
return super(TalkSerializer, self).create(validated_data)
@revisions.create_revision()
def update(self, talk, validated_data):
revisions.set_comment("Changed via REST api")
talk.abstract = validated_data['abstract']
talk.title = validated_data['title']
talk.status = validated_data['status']
talk.talk_type = validated_data['talk_type']
talk.notes = validated_data['notes']
talk.private_notes = validated_data['private_notes']
talk.save()
return talk
| from rest_framework import serializers
from reversion import revisions
from wafer.talks.models import Talk
class TalkSerializer(serializers.ModelSerializer):
class Meta:
model = Talk
# private_notes should possibly be accessible to
# talk reviewers by the API, but certainly
# not to the other users.
# Similar considerations apply to notes, which should
# not be generally accessible
exclude = ('_abstract_rendered', 'private_notes', 'notes')
@revisions.create_revision()
def create(self, validated_data):
revisions.set_comment("Created via REST api")
return super(TalkSerializer, self).create(validated_data)
@revisions.create_revision()
def update(self, talk, validated_data):
revisions.set_comment("Changed via REST api")
talk.abstract = validated_data['abstract']
talk.title = validated_data['title']
talk.status = validated_data['status']
talk.talk_type = validated_data['talk_type']
talk.notes = validated_data['notes']
talk.private_notes = validated_data['private_notes']
talk.save()
return talk
| Exclude notes and private_notes from api for now | Exclude notes and private_notes from api for now
| Python | isc | CTPUG/wafer,CTPUG/wafer,CTPUG/wafer,CTPUG/wafer | ---
+++
@@ -8,7 +8,12 @@
class Meta:
model = Talk
- exclude = ('_abstract_rendered', )
+ # private_notes should possibly be accessible to
+ # talk reviewers by the API, but certainly
+ # not to the other users.
+ # Similar considerations apply to notes, which should
+ # not be generally accessible
+ exclude = ('_abstract_rendered', 'private_notes', 'notes')
@revisions.create_revision() |
0600f3a4f9e13ac0a2a2b4d542db949f90e4185c | challenge_2/python/ning/challenge_2.py | challenge_2/python/ning/challenge_2.py | def find_unique(sequence):
item_counter = dict()
uniques = list()
for item in sequence:
if item not in item_counter:
item_counter[item] = 1
else:
item_counter[item] += 1
for item, item_count in item_counter.items():
if item_count == 1:
uniques.append(item)
return uniques
test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7]
print(find_unique(test_sequence_list))
| import collections
def find_unique(sequence):
counter_dict = collections.defaultdict(int)
uniques = []
for item in sequence:
counter_dict[item] += 1
for item, count in counter_dict.items():
if count == 1:
uniques.append(item)
return uniques
test_sequence_list = [2,'a','l',3,'l',4,'k',2,3,4,'a',6,'c',4,'m',6,'m','k',9,10,9,8,7,8,10,7]
print(find_unique(test_sequence_list))
| Change logic to use defaultdict, lists initiate now with | Change logic to use defaultdict, lists initiate now with []
| Python | mit | erocs/2017Challenges,popcornanachronism/2017Challenges,erocs/2017Challenges,erocs/2017Challenges,popcornanachronism/2017Challenges,DakRomo/2017Challenges,popcornanachronism/2017Challenges,mindm/2017Challenges,erocs/2017Challenges,DakRomo/2017Challenges,popcornanachronism/2017Challenges,mindm/2017Challenges,DakRomo/2017Challenges,popcornanachronism/2017Challenges,popcornanachronism/2017Challenges,erocs/2017Challenges,popcornanachronism/2017Challenges,erocs/2017Challenges,erocs/2017Challenges,mindm/2017Challenges,DakRomo/2017Challenges,erocs/2017Challenges,mindm/2017Challenges,DakRomo/2017Challenges,DakRomo/2017Challenges,DakRomo/2017Challenges,DakRomo/2017Challenges,mindm/2017Challenges,DakRomo/2017Challenges,popcornanachronism/2017Challenges,DakRomo/2017Challenges,mindm/2017Challenges,popcornanachronism/2017Challenges,mindm/2017Challenges,mindm/2017Challenges,erocs/2017Challenges,erocs/2017Challenges,mindm/2017Challenges,popcornanachronism/2017Challenges,popcornanachronism/2017Challenges,mindm/2017Challenges,mindm/2017Challenges,mindm/2017Challenges,mindm/2017Challenges,erocs/2017Challenges,popcornanachronism/2017Challenges,DakRomo/2017Challenges,DakRomo/2017Challenges,DakRomo/2017Challenges,popcornanachronism/2017Challenges,erocs/2017Challenges,DakRomo/2017Challenges,mindm/2017Challenges,erocs/2017Challenges,erocs/2017Challenges,popcornanachronism/2017Challenges | ---
+++
@@ -1,15 +1,14 @@
+import collections
+
def find_unique(sequence):
- item_counter = dict()
- uniques = list()
+ counter_dict = collections.defaultdict(int)
+ uniques = []
for item in sequence:
- if item not in item_counter:
- item_counter[item] = 1
- else:
- item_counter[item] += 1
+ counter_dict[item] += 1
- for item, item_count in item_counter.items():
- if item_count == 1:
+ for item, count in counter_dict.items():
+ if count == 1:
uniques.append(item)
return uniques |
415c9bd14016453d474901fe16f2da8246b44e1b | dojango/__init__.py | dojango/__init__.py | VERSION = (0, 5, 3, 'final', 0)
def get_version():
version = '%s.%s' % (VERSION[0], VERSION[1])
if VERSION[2]:
version = '%s.%s' % (version, VERSION[2])
if VERSION[3:] == ('alpha', 0):
version = '%s pre-alpha' % version
else:
if VERSION[3] != 'final':
version = '%s %s %s' % (version, VERSION[3], VERSION[4])
#from django.utils.version import get_svn_revision
#svn_rev = get_svn_revision()
#if svn_rev != u'SVN-unknown':
# version = "%s %s" % (version, svn_rev)
return version
| VERSION = (0, 5, 4, 'alpha', 0)
def get_version():
version = '%s.%s' % (VERSION[0], VERSION[1])
if VERSION[2]:
version = '%s.%s' % (version, VERSION[2])
if VERSION[3:] == ('alpha', 0):
version = '%s pre-alpha' % version
else:
if VERSION[3] != 'final':
version = '%s %s %s' % (version, VERSION[3], VERSION[4])
#from django.utils.version import get_svn_revision
#svn_rev = get_svn_revision()
#if svn_rev != u'SVN-unknown':
# version = "%s %s" % (version, svn_rev)
return version
| Update to 0.5.4 pre-alpha (preparing next build version) | Update to 0.5.4 pre-alpha (preparing next build version)
| Python | bsd-3-clause | ofirr/dojango,ofirr/dojango,william-gr/dojango,ofirr/dojango,william-gr/dojango,ricard33/dojango,ricard33/dojango,klipstein/dojango,ricard33/dojango,william-gr/dojango,klipstein/dojango | ---
+++
@@ -1,4 +1,4 @@
-VERSION = (0, 5, 3, 'final', 0)
+VERSION = (0, 5, 4, 'alpha', 0)
def get_version():
version = '%s.%s' % (VERSION[0], VERSION[1]) |
bd850668e2c599abcf110c5e3ebb65b79ee91e22 | xpserver_web/models.py | xpserver_web/models.py | from django.db import models
from django.contrib.auth.models import User
class Profile(models.Model):
user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True)
activation_code = models.CharField(max_length=255, default="0000")
fcm_registration_id = models.CharField(max_length=255, blank=True)
def __str__(self):
if self.user:
if self.fcm_registration_id:
return '%s - profile with activated mobile app.' % self.user.username
else:
return '%s - profile.' % self.user.username
else:
return 'Profile unknown user'
| from django.db import models
from django.contrib.auth.models import User
class Profile(models.Model):
user = models.OneToOneField(User, on_delete=models.CASCADE, null=True, blank=True)
activation_code = models.CharField(max_length=255, default="0000")
fcm_registration_id = models.CharField(max_length=255, blank=True)
def __str__(self):
if self.user:
if self.fcm_registration_id:
return '%s - profile with activated mobile app.' % self.user.username
else:
return '%s - profile.' % self.user.username
else:
return 'Profile with unknown user'
| Change str method for unknow user | Change str method for unknow user
| Python | mit | xp2017-hackergarden/server,xp2017-hackergarden/server,xp2017-hackergarden/server,xp2017-hackergarden/server | ---
+++
@@ -14,4 +14,4 @@
else:
return '%s - profile.' % self.user.username
else:
- return 'Profile unknown user'
+ return 'Profile with unknown user' |
49f3d4d168fa62b545aa308684667c355db5c9d8 | examples/write_avi.py | examples/write_avi.py | #!/usr/bin/env python
# Here we use the camhd module to write a single-frame AVI file, only
# downloading data from the moov atom and one frame. This AVI file can
# easily be converted to a PNG or losslessly to a raw YUV using ffmpeg.
#
# Timothy Crone (tjcrone@gmail.com)
import camhd
# remote file
filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov'
# write single frame to avi file
frame_number = 4976
write_frame(filename, frame_number)
| #!/usr/bin/env python
# Here we use the camhd module to write a single-frame AVI file, only
# downloading data from the moov atom and one frame. This AVI file can
# easily be converted to a PNG or losslessly to a raw YUV using ffmpeg.
#
# Timothy Crone (tjcrone@gmail.com)
import camhd
# remote file
filename = 'https://rawdata.oceanobservatories.org/files/RS03ASHS/PN03B/06-CAMHDA301/2016/11/13/CAMHDA301-20161113T000000Z.mov'
# write single frame to avi file
frame_number = 4976
camhd.write_frame(filename, frame_number)
| Fix minor bug in example program | Fix minor bug in example program
| Python | mit | tjcrone/pycamhd | ---
+++
@@ -12,4 +12,4 @@
# write single frame to avi file
frame_number = 4976
-write_frame(filename, frame_number)
+camhd.write_frame(filename, frame_number) |
4f70897d5a85f1822a93df9bc91979ea79594901 | nose2/tests/unit/test_generators_plugin.py | nose2/tests/unit/test_generators_plugin.py | from nose2.plugins.loader import functions
from nose2.tests._common import TestCase
class TestGeneratorUnpack(TestCase):
tags = ['unit']
def setUp(self):
self.expect = [(0, ('call', (0, 1))),
(1, ('call', (1, 2))),
(2, ('call', (2, 3))),]
def test_unpack_handles_nose_style_generators(self):
def gen():
for i in range(0, 3):
yield 'call', i, i + 1
out = list(functions.Functions().unpack(gen()))
self.assertEqual(out, self.expect)
def test_unpack_handles_unittest2_style_generators(self):
def gen():
for i in range(0, 3):
yield 'call', (i, i + 1)
out = list(functions.Functions().unpack(gen()))
self.assertEqual(out, self.expect)
| from nose2 import events, loader, session
from nose2.plugins.loader import generators
from nose2.tests._common import TestCase
class TestGeneratorUnpack(TestCase):
tags = ['unit']
def setUp(self):
self.session = session.Session()
self.loader = loader.PluggableTestLoader(self.session)
self.expect = [(0, ('call', (0, 1))),
(1, ('call', (1, 2))),
(2, ('call', (2, 3))),]
self.plugin = generators.Generators(session=self.session)
def test_unpack_handles_nose_style_generators(self):
def gen():
for i in range(0, 3):
yield 'call', i, i + 1
out = list(self.plugin.unpack(gen()))
self.assertEqual(out, self.expect)
def test_unpack_handles_unittest2_style_generators(self):
def gen():
for i in range(0, 3):
yield 'call', (i, i + 1)
out = list(self.plugin.unpack(gen()))
self.assertEqual(out, self.expect)
def test_ignores_ordinary_functions(self):
class Mod(object):
pass
def test():
pass
m = Mod()
m.test = test
event = events.LoadFromModuleEvent(self.loader, m)
self.session.hooks.loadTestsFromModule(event)
self.assertEqual(len(event.extraTests), 0)
def test_can_load_tests_from_generator_functions(self):
class Mod(object):
pass
def check(x):
assert x == 1
def test():
yield check, 1
yield check, 2
m = Mod()
m.test = test
event = events.LoadFromModuleEvent(self.loader, m)
self.session.hooks.loadTestsFromModule(event)
self.assertEqual(len(event.extraTests), 2)
| Add initial tests for generators plugin | Add initial tests for generators plugin
| Python | bsd-2-clause | ojengwa/nose2,ojengwa/nose2,leth/nose2,little-dude/nose2,ptthiem/nose2,ezigman/nose2,ezigman/nose2,ptthiem/nose2,leth/nose2,little-dude/nose2 | ---
+++
@@ -1,4 +1,5 @@
-from nose2.plugins.loader import functions
+from nose2 import events, loader, session
+from nose2.plugins.loader import generators
from nose2.tests._common import TestCase
@@ -6,20 +7,48 @@
tags = ['unit']
def setUp(self):
+ self.session = session.Session()
+ self.loader = loader.PluggableTestLoader(self.session)
self.expect = [(0, ('call', (0, 1))),
(1, ('call', (1, 2))),
(2, ('call', (2, 3))),]
+ self.plugin = generators.Generators(session=self.session)
def test_unpack_handles_nose_style_generators(self):
def gen():
for i in range(0, 3):
yield 'call', i, i + 1
- out = list(functions.Functions().unpack(gen()))
+ out = list(self.plugin.unpack(gen()))
self.assertEqual(out, self.expect)
def test_unpack_handles_unittest2_style_generators(self):
def gen():
for i in range(0, 3):
yield 'call', (i, i + 1)
- out = list(functions.Functions().unpack(gen()))
+ out = list(self.plugin.unpack(gen()))
self.assertEqual(out, self.expect)
+
+ def test_ignores_ordinary_functions(self):
+ class Mod(object):
+ pass
+ def test():
+ pass
+ m = Mod()
+ m.test = test
+ event = events.LoadFromModuleEvent(self.loader, m)
+ self.session.hooks.loadTestsFromModule(event)
+ self.assertEqual(len(event.extraTests), 0)
+
+ def test_can_load_tests_from_generator_functions(self):
+ class Mod(object):
+ pass
+ def check(x):
+ assert x == 1
+ def test():
+ yield check, 1
+ yield check, 2
+ m = Mod()
+ m.test = test
+ event = events.LoadFromModuleEvent(self.loader, m)
+ self.session.hooks.loadTestsFromModule(event)
+ self.assertEqual(len(event.extraTests), 2) |
8aa02a12883d058b498ac56673079c1e9b53fdf7 | src/libcask/network.py | src/libcask/network.py | import subprocess
class SetupNetworkMixin(object):
def _setup_hostname(self):
with self.get_attachment().attach():
subprocess.check_call(['hostname', self.hostname])
def _setup_virtual_ethernet(self):
# Setup virtual ethernet interface on the host
# TODO - Need to allocate virtual interface names to containers!
subprocess.check_call([
'ip', 'link', 'add',
'name', 'veth0', 'type', 'veth',
'peer', 'name', 'veth1', 'netns', str(self.pid())
])
subprocess.check_call([
'ifconfig', 'veth0', self.ipaddr_host, 'up',
])
# Set up virtual ethernet interface inside the container
# TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're
# not relying on the container having ifconfig.
with self.get_attachment().attach():
subprocess.check_call([
'ifconfig', 'veth1', self.ipaddr, 'up',
])
def setup_network(self):
self._setup_hostname()
self._setup_virtual_ethernet()
| import subprocess
class SetupNetworkMixin(object):
def _setup_hostname(self):
with self.get_attachment().attach():
subprocess.check_call(['hostname', self.hostname])
def _setup_virtual_ethernet(self):
veth_name = 'veth-{hostname}'.format(hostname=self.hostname)
veth_host_name = 'hveth-{hostname}'.format(hostname=self.hostname)
# Create virtual ethernet pair
subprocess.check_call([
'ip', 'link', 'add',
'name', veth_host_name, 'type', 'veth',
'peer', 'name', veth_name, 'netns', str(self.pid())
])
# Add the container's host IP address and bring the interface up
subprocess.check_call(['ip', 'addr', 'add', self.ipaddr_host, 'dev', veth_host_name])
subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'up'])
# Add the host interface to the bridge
# Assuming here that `cask0` bridge interface exists. It should
# be created and initialized by the Makefile.
subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'master', 'cask0'])
# Set up virtual ethernet interface inside the container
# TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're
# not relying on the container having ifconfig.
with self.get_attachment().attach():
subprocess.check_call([
'ifconfig', veth_name, self.ipaddr, 'up',
])
def setup_network(self):
self._setup_hostname()
self._setup_virtual_ethernet()
| Set up separate veth interfaces for each container with a global bridge | Set up separate veth interfaces for each container with a global bridge
| Python | mit | ianpreston/cask,ianpreston/cask | ---
+++
@@ -7,23 +7,31 @@
subprocess.check_call(['hostname', self.hostname])
def _setup_virtual_ethernet(self):
- # Setup virtual ethernet interface on the host
- # TODO - Need to allocate virtual interface names to containers!
+ veth_name = 'veth-{hostname}'.format(hostname=self.hostname)
+ veth_host_name = 'hveth-{hostname}'.format(hostname=self.hostname)
+
+ # Create virtual ethernet pair
subprocess.check_call([
'ip', 'link', 'add',
- 'name', 'veth0', 'type', 'veth',
- 'peer', 'name', 'veth1', 'netns', str(self.pid())
+ 'name', veth_host_name, 'type', 'veth',
+ 'peer', 'name', veth_name, 'netns', str(self.pid())
])
- subprocess.check_call([
- 'ifconfig', 'veth0', self.ipaddr_host, 'up',
- ])
+
+ # Add the container's host IP address and bring the interface up
+ subprocess.check_call(['ip', 'addr', 'add', self.ipaddr_host, 'dev', veth_host_name])
+ subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'up'])
+
+ # Add the host interface to the bridge
+ # Assuming here that `cask0` bridge interface exists. It should
+ # be created and initialized by the Makefile.
+ subprocess.check_call(['ip', 'link', 'set', veth_host_name, 'master', 'cask0'])
# Set up virtual ethernet interface inside the container
# TODO - Only attach CLONE_NEWNET and use the host's ifconfig, so we're
# not relying on the container having ifconfig.
with self.get_attachment().attach():
subprocess.check_call([
- 'ifconfig', 'veth1', self.ipaddr, 'up',
+ 'ifconfig', veth_name, self.ipaddr, 'up',
])
def setup_network(self): |
afdab20403a360508bced14f4750dd6ef4e6aa57 | flask_apidoc/utils.py | flask_apidoc/utils.py | """
Helpers.
"""
def cached(f):
"""
Cache decorator for functions taking one or more arguments.
:param f: The function to be cached.
:return: The cached value.
"""
class CachedDict(dict):
def __init__(self, f):
self.f = f
def __call__(self, *args):
return self[args]
def __missing__(self, key):
ret = self[key] = self.f(*key)
return ret
return CachedDict(f)
| """
Helpers.
"""
import functools
def cached(f):
"""
Cache decorator for functions taking one or more arguments.
:param f: The function to be cached.
:return: The cached value.
"""
cache = f.cache = {}
@functools.wraps(f)
def decorator(*args, **kwargs):
key = str(args) + str(kwargs)
if key not in cache:
cache[key] = f(*args, **kwargs)
return cache[key]
return decorator
| Improve cached decorator to support class based methods | Improve cached decorator to support class based methods
| Python | mit | viniciuschiele/flask-apidoc | ---
+++
@@ -1,6 +1,8 @@
"""
Helpers.
"""
+
+import functools
def cached(f):
@@ -9,15 +11,13 @@
:param f: The function to be cached.
:return: The cached value.
"""
- class CachedDict(dict):
- def __init__(self, f):
- self.f = f
- def __call__(self, *args):
- return self[args]
+ cache = f.cache = {}
- def __missing__(self, key):
- ret = self[key] = self.f(*key)
- return ret
-
- return CachedDict(f)
+ @functools.wraps(f)
+ def decorator(*args, **kwargs):
+ key = str(args) + str(kwargs)
+ if key not in cache:
+ cache[key] = f(*args, **kwargs)
+ return cache[key]
+ return decorator |
540b7cbcf51745e7a7eddb5f0030bc9f52dab35a | Demo/scripts/makedir.py | Demo/scripts/makedir.py | #! /usr/bin/env python
# Like mkdir, but also make intermediate directories if necessary.
# It is not an error if the given directory already exists (as long
# as it is a directory).
# Errors are not treated specially -- you just get a Python exception.
import sys, os
def main():
for p in sys.argv[1:]:
makedirs(p)
def makedirs(p):
if not os.path.isdir(p):
head, tail = os.path.split(p)
makedirs(head)
os.mkdir(p, 0777)
main()
| #! /usr/bin/env python
# Like mkdir, but also make intermediate directories if necessary.
# It is not an error if the given directory already exists (as long
# as it is a directory).
# Errors are not treated specially -- you just get a Python exception.
import sys, os
def main():
for p in sys.argv[1:]:
makedirs(p)
def makedirs(p):
if p and not os.path.isdir(p):
head, tail = os.path.split(p)
makedirs(head)
os.mkdir(p, 0777)
main()
| Fix bug discovered by Klaus-Juergen Wolf -- it runs into infinite recursion! | Fix bug discovered by Klaus-Juergen Wolf -- it runs into infinite recursion!
| Python | mit | sk-/python2.7-type-annotator,sk-/python2.7-type-annotator,sk-/python2.7-type-annotator | ---
+++
@@ -12,7 +12,7 @@
makedirs(p)
def makedirs(p):
- if not os.path.isdir(p):
+ if p and not os.path.isdir(p):
head, tail = os.path.split(p)
makedirs(head)
os.mkdir(p, 0777) |
d66e44fa9fd9b8e8944907b2490d32102c3fba82 | keystoneclient/hacking/checks.py | keystoneclient/hacking/checks.py | # Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
"""python-keystoneclient's pep8 extensions.
In order to make the review process faster and easier for core devs we are
adding some python-keystoneclient specific pep8 checks. This will catch common
errors so that core devs don't have to.
"""
import re
def check_oslo_namespace_imports(logical_line, blank_before, filename):
oslo_namespace_imports = re.compile(
r"(((from)|(import))\s+oslo\."
"((config)|(serialization)|(utils)|(i18n)))|"
"(from\s+oslo\s+import\s+((config)|(serialization)|(utils)|(i18n)))")
if re.match(oslo_namespace_imports, logical_line):
msg = ("K333: '%s' must be used instead of '%s'.") % (
logical_line.replace('oslo.', 'oslo_'),
logical_line)
yield(0, msg)
def factory(register):
register(check_oslo_namespace_imports)
| # Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
"""python-keystoneclient's pep8 extensions.
In order to make the review process faster and easier for core devs we are
adding some python-keystoneclient specific pep8 checks. This will catch common
errors so that core devs don't have to.
"""
import re
def check_oslo_namespace_imports(logical_line, blank_before, filename):
oslo_namespace_imports = re.compile(
r"(((from)|(import))\s+oslo\.)|(from\s+oslo\s+import\s+)")
if re.match(oslo_namespace_imports, logical_line):
msg = ("K333: '%s' must be used instead of '%s'.") % (
logical_line.replace('oslo.', 'oslo_'),
logical_line)
yield(0, msg)
def factory(register):
register(check_oslo_namespace_imports)
| Change hacking check to verify all oslo imports | Change hacking check to verify all oslo imports
The hacking check was verifying that specific oslo imports
weren't using the oslo-namespaced package. Since all the oslo
libraries used by keystoneclient are now changed to use the
new package name the hacking check can be simplified.
bp drop-namespace-packages
Change-Id: I6466e857c6eda0add6918e9fb14dc9296ed98600
| Python | apache-2.0 | Mercador/python-keystoneclient,Mercador/python-keystoneclient,sdpp/python-keystoneclient,magic0704/python-keystoneclient,ging/python-keystoneclient,darren-wang/ksc,magic0704/python-keystoneclient,klmitch/python-keystoneclient,darren-wang/ksc,klmitch/python-keystoneclient,sdpp/python-keystoneclient,ging/python-keystoneclient | ---
+++
@@ -24,9 +24,7 @@
def check_oslo_namespace_imports(logical_line, blank_before, filename):
oslo_namespace_imports = re.compile(
- r"(((from)|(import))\s+oslo\."
- "((config)|(serialization)|(utils)|(i18n)))|"
- "(from\s+oslo\s+import\s+((config)|(serialization)|(utils)|(i18n)))")
+ r"(((from)|(import))\s+oslo\.)|(from\s+oslo\s+import\s+)")
if re.match(oslo_namespace_imports, logical_line):
msg = ("K333: '%s' must be used instead of '%s'.") % ( |
36e0821fcd871935e48ae10926be7594d42f13b8 | knowledge_repo/converters/pdf.py | knowledge_repo/converters/pdf.py | from ..converter import KnowledgePostConverter
from .html import HTMLConverter
class PDFConverter(KnowledgePostConverter):
'''
Use this as a template for new KnowledgePostConverters.
'''
_registry_keys = ['pdf']
@property
def dependencies(self):
# Dependencies required for this converter on top of core knowledge-repo dependencies
return ['weasyprint']
def from_file(self, filename, **opts):
raise NotImplementedError
def from_string(self, filename, **opts):
raise NotImplementedError
def to_file(self, filename, **opts):
with open(filename, 'wb') as f:
f.write(self.to_string())
def to_string(self, **opts):
from weasyprint import HTML
html = HTMLConverter(self.kp).to_string()
return HTML(string=html).write_pdf()
| from ..converter import KnowledgePostConverter
from .html import HTMLConverter
class PDFConverter(KnowledgePostConverter):
'''
Use this as a template for new KnowledgePostConverters.
'''
_registry_keys = ['pdf']
@property
def dependencies(self):
# Dependencies required for this converter on top of core knowledge-repo dependencies
return ['weasyprint']
def from_file(self, filename, **opts):
raise NotImplementedError
def from_string(self, filename, **opts):
raise NotImplementedError
def to_file(self, filename, **opts):
with open(filename, 'wb') as f:
f.write(self.to_string())
def to_string(self, **opts):
from weasyprint import HTML, CSS
html = HTMLConverter(self.kp).to_string()
return HTML(string=html).write_pdf(stylesheets=[CSS(string='body { font-family: Helvetica, sans-serif !important }')])
| Change PDF font to Helvetica | Change PDF font to Helvetica
Changing the PDF font from the default to Helvetica | Python | apache-2.0 | airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo | ---
+++
@@ -25,6 +25,6 @@
f.write(self.to_string())
def to_string(self, **opts):
- from weasyprint import HTML
+ from weasyprint import HTML, CSS
html = HTMLConverter(self.kp).to_string()
- return HTML(string=html).write_pdf()
+ return HTML(string=html).write_pdf(stylesheets=[CSS(string='body { font-family: Helvetica, sans-serif !important }')]) |
985198a9ea569cc6f418c5b337632b91cdda7e37 | lib/rapidsms/backends/backend.py | lib/rapidsms/backends/backend.py | #!/usr/bin/env python
# vim: ai ts=4 sts=4 et sw=4
class Backend(object):
def log(self, level, message):
self.router.log(level, message)
def start(self):
raise NotImplementedError
def stop(self):
raise NotImplementedError
def send(self):
raise NotImplementedError
def receive(self):
raise NotImplementedError
| #!/usr/bin/env python
# vim: ai ts=4 sts=4 et sw=4
class Backend(object):
def __init__ (self, router):
self.router = router
def log(self, level, message):
self.router.log(level, message)
def start(self):
raise NotImplementedError
def stop(self):
raise NotImplementedError
def send(self):
raise NotImplementedError
def receive(self):
raise NotImplementedError
| Add a constructor method for Backend | Add a constructor method for Backend
| Python | bsd-3-clause | dimagi/rapidsms-core-dev,peterayeni/rapidsms,rapidsms/rapidsms-core-dev,lsgunth/rapidsms,catalpainternational/rapidsms,ken-muturi/rapidsms,ehealthafrica-ci/rapidsms,eHealthAfrica/rapidsms,unicefuganda/edtrac,unicefuganda/edtrac,eHealthAfrica/rapidsms,catalpainternational/rapidsms,catalpainternational/rapidsms,eHealthAfrica/rapidsms,caktus/rapidsms,caktus/rapidsms,lsgunth/rapidsms,peterayeni/rapidsms,peterayeni/rapidsms,ken-muturi/rapidsms,rapidsms/rapidsms-core-dev,ehealthafrica-ci/rapidsms,dimagi/rapidsms,dimagi/rapidsms-core-dev,lsgunth/rapidsms,ken-muturi/rapidsms,catalpainternational/rapidsms,caktus/rapidsms,ehealthafrica-ci/rapidsms,peterayeni/rapidsms,unicefuganda/edtrac,lsgunth/rapidsms,dimagi/rapidsms | ---
+++
@@ -1,8 +1,9 @@
#!/usr/bin/env python
# vim: ai ts=4 sts=4 et sw=4
-
class Backend(object):
+ def __init__ (self, router):
+ self.router = router
def log(self, level, message):
self.router.log(level, message) |
7d574570ad831a86758352ef98bf6192f7d3e6fa | setup.py | setup.py | import os
from setuptools import setup
install_requires = [
'mysql-python>=1.2.3',
'psycopg2>=2.4.2',
'pyyaml>=3.10.0',
'argparse',
'pytz',
]
if os.name == 'posix':
install_requires.append('termcolor>=1.1.0')
setup(
name='py-mysql2pgsql',
version='0.1.6',
description='Tool for migrating/converting from mysql to postgresql.',
long_description=open('README.rst').read(),
license='MIT License',
author='Philip Southam',
author_email='philipsoutham@gmail.com',
url='https://github.com/philipsoutham/py-mysql2pgsql',
zip_safe=False,
packages=['mysql2pgsql', 'mysql2pgsql.lib'],
scripts=['bin/py-mysql2pgsql'],
platforms='any',
install_requires=install_requires,
classifiers=[
'License :: OSI Approved :: MIT License',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: System Administrators',
'Intended Audience :: Developers',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2.7',
'Topic :: Database',
'Topic :: Utilities'
],
keywords = 'mysql postgres postgresql pgsql psql migration',
)
| import os
from setuptools import setup
install_requires = [
'mysql-python>=1.2.3',
'psycopg2>=2.4.2',
'pyyaml>=3.10.0',
'pytz',
]
if os.name == 'posix':
install_requires.append('termcolor>=1.1.0')
if version < (2,7) or (3,0) <= version <= (3,1):
install_requires += ['argparse']
setup(
name='py-mysql2pgsql',
version='0.1.6',
description='Tool for migrating/converting from mysql to postgresql.',
long_description=open('README.rst').read(),
license='MIT License',
author='Philip Southam',
author_email='philipsoutham@gmail.com',
url='https://github.com/philipsoutham/py-mysql2pgsql',
zip_safe=False,
packages=['mysql2pgsql', 'mysql2pgsql.lib'],
scripts=['bin/py-mysql2pgsql'],
platforms='any',
install_requires=install_requires,
classifiers=[
'License :: OSI Approved :: MIT License',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: System Administrators',
'Intended Audience :: Developers',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2.7',
'Topic :: Database',
'Topic :: Utilities'
],
keywords = 'mysql postgres postgresql pgsql psql migration',
)
| Correct argparse dependency - argparse already is a part of base python as of 2.7 and 3.2. | Correct argparse dependency - argparse already is a part of base python as of 2.7 and 3.2. | Python | mit | Varsha-Arun/py-mysql2pgsql,philipsoutham/py-mysql2pgsql,EdNeg/DB-Project,leture/py-mysql2pgsql,mr-tron/py-mysql2pgsql,EdNeg/DB-Project,emailgregn/py-mysql2pgsql,kevgathuku/py-mysql2pgsql,andela-kndungu/py-mysql2pgsql,EdNeg/DB-Project,SpootDev/py-mysql2pgsql,edwardsamuel/py-mysql2pgsql | ---
+++
@@ -5,13 +5,15 @@
'mysql-python>=1.2.3',
'psycopg2>=2.4.2',
'pyyaml>=3.10.0',
- 'argparse',
'pytz',
]
if os.name == 'posix':
install_requires.append('termcolor>=1.1.0')
+if version < (2,7) or (3,0) <= version <= (3,1):
+ install_requires += ['argparse']
+
setup(
name='py-mysql2pgsql',
version='0.1.6', |
329fedb27fe54d593f192912beda56588faec214 | tests/__init__.py | tests/__init__.py | from django.conf import settings
from mock import Mock, patch
from unittest2 import TestCase
settings.configure()
# Need to import this after configure()
from django.db.models import ForeignKey
class TestPreference(object):
_meta = Mock(fields=[ForeignKey('user', name='user')])
objects = Mock()
def __init__(self, name, value, user=None):
self.name = name
self.value = value
self.user = user
def __eq__(self, other):
return self.__dict__ == other.__dict__
def __repr__(self):
return '<{name}:{value}:{user}>'.format(**self.__dict__)
def __cmp__(self, other):
return cmp(self.name, other.name)
class TestUser(object):
@property
def preferences(self):
return Mock(all=Mock(return_value=self._preferences))
@preferences.setter
def preferences(self, value):
self._preferences = [
TestPreference(k, v) for k, v in value.iteritems()]
class SerializerTestCase(TestCase):
def patch_from_native(self):
patcher = patch(
'madprops.serializers.ModelSerializer.from_native',
new=lambda self, data, files: TestPreference(
data['name'], data['value'], data.get('user'))
)
self.patched_from_native = patcher.start()
self.addCleanup(patcher.stop)
| from django.conf import settings
from mock import Mock, patch
from unittest2 import TestCase
settings.configure()
# Need to import this after configure()
from django.db.models import ForeignKey
class TestPreference(object):
_meta = Mock(fields=[ForeignKey('user', name='user')])
objects = Mock()
def __init__(self, name, value, user=None):
self.name = name
self.value = value
self.user = user
def __eq__(self, other):
return self.__dict__ == other.__dict__
def __repr__(self):
return '<{name}:{value}:{user}>'.format(**self.__dict__)
def __cmp__(self, other):
return cmp(self.name, other.name)
class TestUser(object):
@property
def preferences(self):
return Mock(all=Mock(return_value=self._preferences))
@preferences.setter
def preferences(self, value):
self._preferences = [
TestPreference(k, v) for k, v in value.iteritems()]
class SerializerTestCase(TestCase):
def patch_from_native(self):
def from_native(self, data, files):
self._errors = {}
if data:
self.perform_validation(data)
return TestPreference(data['name'], data['value'],
data.get('user'))
patcher = patch(
'madprops.serializers.ModelSerializer.from_native',
new=from_native)
self.patched_from_native = patcher.start()
self.addCleanup(patcher.stop)
# get_fields inspects the model's _meta, deeply
patcher = patch(
'madprops.serializers.ModelSerializer.get_fields',
new=lambda self: {})
self.patched_get_fields = patcher.start()
self.addCleanup(patcher.stop)
| Extend mocking to run validation | Extend mocking to run validation
| Python | mit | yola/drf-madprops | ---
+++
@@ -43,10 +43,22 @@
class SerializerTestCase(TestCase):
def patch_from_native(self):
+ def from_native(self, data, files):
+ self._errors = {}
+ if data:
+ self.perform_validation(data)
+ return TestPreference(data['name'], data['value'],
+ data.get('user'))
+
patcher = patch(
'madprops.serializers.ModelSerializer.from_native',
- new=lambda self, data, files: TestPreference(
- data['name'], data['value'], data.get('user'))
- )
+ new=from_native)
self.patched_from_native = patcher.start()
self.addCleanup(patcher.stop)
+
+ # get_fields inspects the model's _meta, deeply
+ patcher = patch(
+ 'madprops.serializers.ModelSerializer.get_fields',
+ new=lambda self: {})
+ self.patched_get_fields = patcher.start()
+ self.addCleanup(patcher.stop) |
6d90adebbcd9085de4f0af503b7e0b55dc9424b2 | tests/conftest.py | tests/conftest.py | import numpy
import pytest
@pytest.fixture(scope="module")
def rgen():
return numpy.random.RandomState(seed=52973992)
| import numpy
import pytest
@pytest.fixture(scope="module")
def rgen():
return numpy.random.RandomState(seed=3476583865)
| Fix test by questionable means :) | Fix test by questionable means :)
| Python | bsd-3-clause | dseuss/mpnum,dseuss/mpnum | ---
+++
@@ -4,4 +4,4 @@
@pytest.fixture(scope="module")
def rgen():
- return numpy.random.RandomState(seed=52973992)
+ return numpy.random.RandomState(seed=3476583865) |
817976878b584086bedc296e5fd6d264006c8dcd | tests/conftest.py | tests/conftest.py | from __future__ import absolute_import
from __future__ import unicode_literals
import os
import subprocess
import mock
import pytest
@pytest.yield_fixture
def in_tmpdir(tmpdir):
pwd = os.getcwd()
os.chdir(tmpdir.strpath)
try:
yield
finally:
os.chdir(pwd)
@pytest.yield_fixture
def check_call_mock():
with mock.patch.object(subprocess, 'check_call') as mocked:
yield mocked
| from __future__ import absolute_import
from __future__ import unicode_literals
import os
import subprocess
import mock
import pytest
@pytest.fixture
def in_tmpdir(tmpdir):
pwd = os.getcwd()
os.chdir(tmpdir.strpath)
try:
yield
finally:
os.chdir(pwd)
@pytest.fixture
def check_call_mock():
with mock.patch.object(subprocess, 'check_call') as mocked:
yield mocked
| Replace deprecated yield_fixture with fixture | Replace deprecated yield_fixture with fixture
Committed via https://github.com/asottile/all-repos
| Python | mit | asottile/css-explore | ---
+++
@@ -8,7 +8,7 @@
import pytest
-@pytest.yield_fixture
+@pytest.fixture
def in_tmpdir(tmpdir):
pwd = os.getcwd()
os.chdir(tmpdir.strpath)
@@ -18,7 +18,7 @@
os.chdir(pwd)
-@pytest.yield_fixture
+@pytest.fixture
def check_call_mock():
with mock.patch.object(subprocess, 'check_call') as mocked:
yield mocked |
b778c0192cabc652fc06daf99f6b890b3297f83f | Lib/test/test_sqlite.py | Lib/test/test_sqlite.py | from test.support import run_unittest, import_module, verbose
# Skip test if _sqlite3 module not installed
import_module('_sqlite3')
import sqlite3
from sqlite3.test import (dbapi, types, userfunctions,
factory, transactions, hooks, regression,
dump)
def test_main():
if verbose:
print("test_sqlite: testing with version",
"{!r}, sqlite_version {!r}".format(sqlite3.version,
sqlite3.sqlite_version))
run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(),
factory.suite(), transactions.suite(),
hooks.suite(), regression.suite(), dump.suite())
if __name__ == "__main__":
test_main()
| import test.support
# Skip test if _sqlite3 module not installed
test.support.import_module('_sqlite3')
import sqlite3
from sqlite3.test import (dbapi, types, userfunctions,
factory, transactions, hooks, regression,
dump)
def test_main():
if test.support.verbose:
print("test_sqlite: testing with version",
"{!r}, sqlite_version {!r}".format(sqlite3.version,
sqlite3.sqlite_version))
test.support.run_unittest(dbapi.suite(), types.suite(),
userfunctions.suite(),
factory.suite(), transactions.suite(),
hooks.suite(), regression.suite(),
dump.suite())
if __name__ == "__main__":
test_main()
| Make the printing of sqlite version in verbose mode work with regrtest -w. | Make the printing of sqlite version in verbose mode work with regrtest -w.
| Python | mit | sk-/python2.7-type-annotator,sk-/python2.7-type-annotator,sk-/python2.7-type-annotator | ---
+++
@@ -1,7 +1,7 @@
-from test.support import run_unittest, import_module, verbose
+import test.support
# Skip test if _sqlite3 module not installed
-import_module('_sqlite3')
+test.support.import_module('_sqlite3')
import sqlite3
from sqlite3.test import (dbapi, types, userfunctions,
@@ -9,13 +9,15 @@
dump)
def test_main():
- if verbose:
+ if test.support.verbose:
print("test_sqlite: testing with version",
"{!r}, sqlite_version {!r}".format(sqlite3.version,
sqlite3.sqlite_version))
- run_unittest(dbapi.suite(), types.suite(), userfunctions.suite(),
- factory.suite(), transactions.suite(),
- hooks.suite(), regression.suite(), dump.suite())
+ test.support.run_unittest(dbapi.suite(), types.suite(),
+ userfunctions.suite(),
+ factory.suite(), transactions.suite(),
+ hooks.suite(), regression.suite(),
+ dump.suite())
if __name__ == "__main__":
test_main() |
d8e9201c627840c72a540a77425ec0c13ac48a22 | tests/test_cmd.py | tests/test_cmd.py | import unittest
from click.testing import CliRunner
from scuevals_api.cmd import cli
class CmdsTestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
cls.runner = CliRunner()
def cli_run(self, *cmds):
return self.runner.invoke(cli, cmds)
cls.cli_run = cli_run
def test_initdb(self):
result = self.cli_run('initdb')
self.assertEqual(0, result.exit_code)
| import unittest
from click.testing import CliRunner
from scuevals_api.cmd import cli
class CmdsTestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
cls.runner = CliRunner()
def cli_run(self, *cmds):
return self.runner.invoke(cli, cmds)
cls.cli_run = cli_run
def test_initdb(self):
result = self.cli_run('initdb')
self.assertEqual(0, result.exit_code, msg=result.output)
| Add detailed error for CLI test failure | Add detailed error for CLI test failure
| Python | agpl-3.0 | SCUEvals/scuevals-api,SCUEvals/scuevals-api | ---
+++
@@ -17,4 +17,4 @@
def test_initdb(self):
result = self.cli_run('initdb')
- self.assertEqual(0, result.exit_code)
+ self.assertEqual(0, result.exit_code, msg=result.output) |
e7d42847284ae73befad8bdf2fa035a6f95a82bb | tests/test_dow.py | tests/test_dow.py | from datetime import datetime
import pycron
def test_parser():
now = datetime(2015, 6, 18, 16, 7)
assert pycron.is_now('* * * * *', now)
assert pycron.is_now('* * * * 4', now)
assert pycron.is_now('* * * * */4', now)
assert pycron.is_now('* * * * 0,3,4', now)
assert pycron.is_now('* * * * 3', now) is False
assert pycron.is_now('* * * * */3', now) is False
assert pycron.is_now('* * * * 0,3,6', now) is False
assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday'
assert pycron.DOW_CHOICES[0][1] == 'Sunday'
now = datetime(2015, 6, 21, 16, 7)
assert pycron.is_now('* * * * 0', now)
| from datetime import datetime, timedelta
import pycron
def test_parser():
now = datetime(2015, 6, 18, 16, 7)
assert pycron.is_now('* * * * *', now)
assert pycron.is_now('* * * * 4', now)
assert pycron.is_now('* * * * */4', now)
assert pycron.is_now('* * * * 0,3,4', now)
assert pycron.is_now('* * * * 3', now) is False
assert pycron.is_now('* * * * */3', now) is False
assert pycron.is_now('* * * * 0,3,6', now) is False
assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday'
assert pycron.DOW_CHOICES[0][1] == 'Sunday'
now = datetime(2015, 6, 20, 16, 7)
for i in range(0, 7):
# Test day matching from Sunday onwards...
now += timedelta(days=1)
assert pycron.is_now('* * * * %i' % (i), now)
# Test weekdays
assert pycron.is_now('* * * * 1,2,3,4,5', now) is (True if i not in [0, 6] else False)
# Test weekends
assert pycron.is_now('* * * * 0,6', now) is (True if i in [0, 6] else False)
| Add more thorough testing of day of week. | Add more thorough testing of day of week.
| Python | mit | kipe/pycron | ---
+++
@@ -1,4 +1,4 @@
-from datetime import datetime
+from datetime import datetime, timedelta
import pycron
@@ -14,5 +14,12 @@
assert pycron.DOW_CHOICES[now.isoweekday()][1] == 'Thursday'
assert pycron.DOW_CHOICES[0][1] == 'Sunday'
- now = datetime(2015, 6, 21, 16, 7)
- assert pycron.is_now('* * * * 0', now)
+ now = datetime(2015, 6, 20, 16, 7)
+ for i in range(0, 7):
+ # Test day matching from Sunday onwards...
+ now += timedelta(days=1)
+ assert pycron.is_now('* * * * %i' % (i), now)
+ # Test weekdays
+ assert pycron.is_now('* * * * 1,2,3,4,5', now) is (True if i not in [0, 6] else False)
+ # Test weekends
+ assert pycron.is_now('* * * * 0,6', now) is (True if i in [0, 6] else False) |
a89e7f9f625427d558300eb5e5cbc2881cdcc207 | get_a_job/__init__.py | get_a_job/__init__.py | from flask import Flask
from flask.ext.restful import Api
from .models import db
from .api import configure_api
def create_app(object_name):
app = Flask(object_name)
app.config.from_object(object_name)
db.init_app(app)
configure_api(app)
return app
| from flask import Flask
from flask.ext.restful import Api
from .models import db
from .api import configure_api
def create_app(object_name, **kwargs):
app = Flask(object_name)
app.config.from_object(object_name)
app.config.update(kwargs)
db.init_app(app)
configure_api(app)
return app
| Add optional configuration customization of app. | Add optional configuration customization of app.
| Python | mit | smoynes/get_a_job | ---
+++
@@ -4,9 +4,10 @@
from .models import db
from .api import configure_api
-def create_app(object_name):
+def create_app(object_name, **kwargs):
app = Flask(object_name)
app.config.from_object(object_name)
+ app.config.update(kwargs)
db.init_app(app)
configure_api(app) |
d8cde079d6e8dd0dcd5a13a36a0bca9685ba7b1c | api/BucketListAPI.py | api/BucketListAPI.py | from flask import Flask, jsonify
from modals.modals import User, Bucket, Item
from api.__init__ import create_app, db
app = create_app('DevelopmentEnv')
@app.errorhandler(404)
def page_not_found(e):
response = jsonify({'error': 'The request can not be completed'})
response.status_code = 404
return response
if __name__ == '__main__':
app.run()
| from flask import Flask, jsonify
from modals.modals import User, Bucket, Item
from api.__init__ import create_app, db
app = create_app('DevelopmentEnv')
@app.errorhandler(404)
def page_not_found(e):
response = jsonify({'error': 'The request can not be completed'})
response.status_code = 404
return response
@app.errorhandler(401)
def invalid_token(e):
response = jsonify({'error': 'Invalid Token'})
response.status_code = 401
return response
if __name__ == '__main__':
app.run()
| Add error handler for invalid token | Add error handler for invalid token
| Python | mit | patlub/BucketListAPI,patlub/BucketListAPI | ---
+++
@@ -11,6 +11,12 @@
response.status_code = 404
return response
+@app.errorhandler(401)
+def invalid_token(e):
+ response = jsonify({'error': 'Invalid Token'})
+ response.status_code = 401
+ return response
+
if __name__ == '__main__':
app.run() |
45254f3a7401b4b63d829f38c426c0635485f1e0 | PRESUBMIT.py | PRESUBMIT.py | # Copyright (c) 2011 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""Top-level presubmit script for GYP.
See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts
for more details about the presubmit API built into gcl.
"""
def CheckChangeOnUpload(input_api, output_api):
report = []
report.extend(input_api.canned_checks.PanProjectChecks(
input_api, output_api))
return report
def CheckChangeOnCommit(input_api, output_api):
report = []
report.extend(input_api.canned_checks.PanProjectChecks(
input_api, output_api))
report.extend(input_api.canned_checks.CheckTreeIsOpen(
input_api, output_api,
'http://gyp-status.appspot.com/status',
'http://gyp-status.appspot.com/current'))
return report
def GetPreferredTrySlaves():
return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac']
| # Copyright (c) 2011 Google Inc. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""Top-level presubmit script for GYP.
See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts
for more details about the presubmit API built into gcl.
"""
def CheckChangeOnUpload(input_api, output_api):
report = []
report.extend(input_api.canned_checks.PanProjectChecks(
input_api, output_api))
return report
def CheckChangeOnCommit(input_api, output_api):
report = []
license = (
r'.*? Copyright \(c\) %(year)s Google Inc\. All rights reserved\.\n'
r'.*? Use of this source code is governed by a BSD-style license that '
r'can be\n'
r'.*? found in the LICENSE file\.\n'
) % {
'year': input_api.time.strftime('%Y'),
}
report.extend(input_api.canned_checks.PanProjectChecks(
input_api, output_api, license_header=license))
report.extend(input_api.canned_checks.CheckTreeIsOpen(
input_api, output_api,
'http://gyp-status.appspot.com/status',
'http://gyp-status.appspot.com/current'))
return report
def GetPreferredTrySlaves():
return ['gyp-win32', 'gyp-win64', 'gyp-linux', 'gyp-mac']
| Fix the license header regex. | Fix the license header regex.
Most of the files are attributed to Google Inc so I used this instead of
Chromium Authors.
R=mark@chromium.org
BUG=
TEST=
Review URL: http://codereview.chromium.org/7108074
git-svn-id: e7e1075985beda50ea81ac4472467b4f6e91fc78@936 78cadc50-ecff-11dd-a971-7dbc132099af
| Python | bsd-3-clause | enkripsi/gyp,mistydemeo/gyp,cysp/gyp,IllusionRom-deprecated/android_platform_external_chromium_org_tools_gyp,LazyCodingCat/gyp,IllusionRom-deprecated/android_platform_external_chromium_org_tools_gyp,ttyangf/gyp,saghul/gyn,ttyangf/pdfium_gyp,lukeweber/gyp-override,carlTLR/gyp,amoikevin/gyp,erikge/watch_gyp,LazyCodingCat/gyp,omasanori/gyp,saghul/gyn,Omegaphora/external_chromium_org_tools_gyp,channing/gyp,yinquan529/platform-external-chromium_org-tools-gyp,Jack-Q/GYP-copy,duanhjlt/gyp,bnq4ever/gypgoogle,Danath/gyp,yangrongwei/gyp,Chilledheart/gyp,enkripsi/gyp,chromium/gyp,Chilledheart/gyp,bnoordhuis/gyp,svn2github/gyp,msc-/gyp,sanyaade-teachings/gyp,okumura/gyp,carlTLR/gyp,alexcrichton/gyp,trafi/gyp,brson/gyp,sanyaade-teachings/gyp,brson/gyp,sport-monkey/GYP,sanyaade-teachings/gyp,bnoordhuis/gyp,okumura/gyp,bulldy80/gyp_unofficial,saghul/gyn,mgamer/gyp,AWhetter/gyp,cchamberlain/gyp,bulldy80/gyp_unofficial,clar/gyp,lukeweber/gyp-override,adblockplus/gyp,bdarnell/gyp,cysp/gyp,Phuehvk/gyp,svn2github/gyp,mgamer/gyp,trafi/gyp,mistydemeo/gyp,amoikevin/gyp,sloanyang/gyp,azunite/gyp_20150930,bdarnell/gyp,turbulenz/gyp,mkrautz/gyp-libmumble,sdklite/gyp,mapbox/gyp,mgamer/gyp,pandaxcl/gyp,bpsinc-native/src_tools_gyp,sanyaade-teachings/gyp,msc-/gyp,ryfx/gyp,openpeer/webrtc-gyp,msc-/gyp,enkripsi/gyp,MIPS/external-chromium_org-tools-gyp,yinquan529/platform-external-chromium_org-tools-gyp,turbulenz/gyp,kevinchen3315/gyp-git,Jack-Q/GYP-copy,cysp/gyp,LazyCodingCat/gyp,chromium/gyp,svn2github/kgyp,ttyangf/gyp,cysp/gyp,azunite/gyp_20150930,lianliuwei/gyp,dougbeal/gyp,cchamberlain/gyp,adblockplus/gyp,mkrautz/gyp-libmumble,bnoordhuis/gyp,svn2github/gyp,Omegaphora/external_chromium_org_tools_gyp,mapbox/gyp,pyokagan/gyp,pyokagan/gyp,bulldy80/gyp_unofficial,bnoordhuis/gyp,sloanyang/gyp,mkrautz/gyp-libmumble,saghul/gyn,ttyangf/pdfium_gyp,alexcrichton/gyp,bdarnell/gyp,enkripsi/gyp,springmeyer/gyp,mapbox/gyp,trafi/gyp,svn2github/kgyp,luvit/gyp,duanhjlt/gyp,geekboxzone/lollipop_external_chromium_org_tools_gyp,amoikevin/gyp,lianliuwei/gyp,pandaxcl/gyp,clar/gyp,okwasi/gyp,carlTLR/gyp,turbulenz/gyp,AWhetter/gyp,ryfx/gyp,lukeweber/gyp-override,alexcrichton/gyp,lukeweber/gyp-override,sport-monkey/GYP,trafi/gyp,duanhjlt/gyp,adblockplus/gyp,chromium/gyp,brson/gyp,dougbeal/gyp,azunite/gyp_20150930,omasanori/gyp,sdklite/gyp,amoikevin/gyp,erikge/watch_gyp,svn2github/kgyp,pyokagan/gyp,amoikevin/gyp,AOSPU/external_chromium_org_tools_gyp,LazyCodingCat/gyp,android-ia/platform_external_chromium_org_tools_gyp,Danath/gyp,mumble-voip/libmumble-gyp,Jack-Q/GYP-copy,luvit/gyp,azunite/gyp_20150930,turbulenz/gyp,ryfx/gyp,svn2github/kgyp,bnq4ever/gypgoogle,android-ia/platform_external_chromium_org_tools_gyp,carlTLR/gyp,okwasi/gyp,omasanori/gyp,channing/gyp,enkripsi/gyp,Omegaphora/external_chromium_org_tools_gyp,springmeyer/gyp,yangrongwei/gyp,MIPS/external-chromium_org-tools-gyp,mumble-voip/libmumble-gyp,bulldy80/gyp_unofficial,okumura/gyp,tarc/gyp,luvit/gyp,ttyangf/pdfium_gyp,yinquan529/platform-external-chromium_org-tools-gyp,yjhjstz/gyp,ttyangf/pdfium_gyp,sport-monkey/GYP,geekboxzone/lollipop_external_chromium_org_tools_gyp,mapbox/gyp,sanyaade-teachings/gyp,ttyangf/pdfium_gyp,kevinchen3315/gyp-git,tarc/gyp,saghul/gyn,bnq4ever/gypgoogle,mgamer/gyp,bnq4ever/gypgoogle,channing/gyp,ryfx/gyp,bnoordhuis/gyp,azunite/gyp,trafi/gyp,AWhetter/gyp,erikge/watch_gyp,sloanyang/gyp,erikge/watch_gyp,xin3liang/platform_external_chromium_org_tools_gyp,springmeyer/gyp,omasanori/gyp,azunite/gyp,openpeer/webrtc-gyp,msc-/gyp,yjhjstz/gyp,tarc/gyp,mistydemeo/gyp,Danath/gyp,azunite/gyp,clar/gyp,mapbox/gyp,MIPS/external-chromium_org-tools-gyp,Danath/gyp,Phuehvk/gyp,ttyangf/gyp,mkrautz/gyp-libmumble,okumura/gyp,Omegaphora/external_chromium_org_tools_gyp,svn2github/gyp,sloanyang/gyp,AOSPU/external_chromium_org_tools_gyp,mistydemeo/gyp,dougbeal/gyp,bpsinc-native/src_tools_gyp,pandaxcl/gyp,Phuehvk/gyp,Jack-Q/GYP-copy,azunite/gyp,kevinchen3315/gyp-git,chromium/gyp,Chilledheart/gyp,mgamer/gyp,cchamberlain/gyp,okwasi/gyp,svn2github/kgyp,bnq4ever/gypgoogle,xin3liang/platform_external_chromium_org_tools_gyp,erikge/watch_gyp,azunite/gyp_20150930,luvit/gyp,adblockplus/gyp,alexcrichton/gyp,yinquan529/platform-external-chromium_org-tools-gyp,Phuehvk/gyp,Chilledheart/gyp,dougbeal/gyp,bdarnell/gyp,Chilledheart/gyp,cchamberlain/gyp,sdklite/gyp,pyokagan/gyp,okwasi/gyp,springmeyer/gyp,AWhetter/gyp,clar/gyp,sport-monkey/GYP,android-ia/platform_external_chromium_org_tools_gyp,Danath/gyp,carlTLR/gyp,cchamberlain/gyp,openpeer/webrtc-gyp,xin3liang/platform_external_chromium_org_tools_gyp,yjhjstz/gyp,AOSPU/external_chromium_org_tools_gyp,yangrongwei/gyp,Phuehvk/gyp,pandaxcl/gyp,yjhjstz/gyp,sport-monkey/GYP,bpsinc-native/src_tools_gyp,yangrongwei/gyp,AOSPU/external_chromium_org_tools_gyp,Jack-Q/GYP-copy,kevinchen3315/gyp-git,pyokagan/gyp,ryfx/gyp,brson/gyp,sdklite/gyp,bpsinc-native/src_tools_gyp,turbulenz/gyp,chromium/gyp,mumble-voip/libmumble-gyp,svn2github/gyp,channing/gyp,tarc/gyp,mumble-voip/libmumble-gyp,ttyangf/gyp,openpeer/webrtc-gyp,yjhjstz/gyp,adblockplus/gyp,xin3liang/platform_external_chromium_org_tools_gyp,MIPS/external-chromium_org-tools-gyp,android-ia/platform_external_chromium_org_tools_gyp,AWhetter/gyp,lianliuwei/gyp,bulldy80/gyp_unofficial,tarc/gyp,geekboxzone/lollipop_external_chromium_org_tools_gyp,IllusionRom-deprecated/android_platform_external_chromium_org_tools_gyp,msc-/gyp,cysp/gyp,duanhjlt/gyp,dougbeal/gyp,clar/gyp,geekboxzone/lollipop_external_chromium_org_tools_gyp,openpeer/webrtc-gyp,pandaxcl/gyp,springmeyer/gyp,sdklite/gyp,LazyCodingCat/gyp,ttyangf/gyp,azunite/gyp,duanhjlt/gyp,IllusionRom-deprecated/android_platform_external_chromium_org_tools_gyp,lianliuwei/gyp | ---
+++
@@ -1,4 +1,4 @@
-# Copyright (c) 2011 The Chromium Authors. All rights reserved.
+# Copyright (c) 2011 Google Inc. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
@@ -19,8 +19,17 @@
def CheckChangeOnCommit(input_api, output_api):
report = []
+ license = (
+ r'.*? Copyright \(c\) %(year)s Google Inc\. All rights reserved\.\n'
+ r'.*? Use of this source code is governed by a BSD-style license that '
+ r'can be\n'
+ r'.*? found in the LICENSE file\.\n'
+ ) % {
+ 'year': input_api.time.strftime('%Y'),
+ }
+
report.extend(input_api.canned_checks.PanProjectChecks(
- input_api, output_api))
+ input_api, output_api, license_header=license))
report.extend(input_api.canned_checks.CheckTreeIsOpen(
input_api, output_api,
'http://gyp-status.appspot.com/status', |
ac3c855583a023fc76b8720aa7e38419b28a26d4 | falcom/api/hathi.py | falcom/api/hathi.py | # Copyright (c) 2017 The Regents of the University of Michigan.
# All Rights Reserved. Licensed according to the terms of the Revised
# BSD License. See LICENSE.txt for details.
import json
def get_counts_from_item_list (items, htid):
a = len([x for x in items if x["htid"] == htid])
b = len(items) - a
return a, b
def get_oclc_counts_from_json (json_data, htid = ""):
try:
data = json.loads(json_data)
return get_counts_from_item_list(data["items"], htid)
except:
return 0, 0
def get_hathi_data_from_json ():
return ()
| # Copyright (c) 2017 The Regents of the University of Michigan.
# All Rights Reserved. Licensed according to the terms of the Revised
# BSD License. See LICENSE.txt for details.
import json
class HathiItems:
def __init__ (self):
pass
def __len__ (self):
return 0
def get_counts_from_item_list (items, htid):
a = len([x for x in items if x["htid"] == htid])
b = len(items) - a
return a, b
def get_oclc_counts_from_json (json_data, htid = ""):
try:
data = json.loads(json_data)
return get_counts_from_item_list(data["items"], htid)
except:
return 0, 0
def get_hathi_data_from_json ():
return HathiItems()
| Refactor empty tuple into empty object with len() | Refactor empty tuple into empty object with len()
| Python | bsd-3-clause | mlibrary/image-conversion-and-validation,mlibrary/image-conversion-and-validation | ---
+++
@@ -2,6 +2,14 @@
# All Rights Reserved. Licensed according to the terms of the Revised
# BSD License. See LICENSE.txt for details.
import json
+
+class HathiItems:
+
+ def __init__ (self):
+ pass
+
+ def __len__ (self):
+ return 0
def get_counts_from_item_list (items, htid):
a = len([x for x in items if x["htid"] == htid])
@@ -18,4 +26,4 @@
return 0, 0
def get_hathi_data_from_json ():
- return ()
+ return HathiItems() |
7872a2327f9dea7d4c1f5a3054b6be6bba25fdd4 | scripts/migration/migrate_deleted_wikis.py | scripts/migration/migrate_deleted_wikis.py | import logging
import sys
from modularodm import Q
from framework.transactions.context import TokuTransaction
from website.app import init_app
from website.models import NodeLog
from scripts import utils as script_utils
logger = logging.getLogger(__name__)
def get_targets():
return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED))
def migrate(targets, dry_run=True):
# iterate over targets
for log in targets:
node = log.node
versions = node.wiki_pages_versions
current = node.wiki_pages_current
updated_versions = {}
for wiki in versions:
if wiki in current:
updated_versions[wiki] = versions[wiki]
with TokuTransaction():
node.wiki_pages_versions = updated_versions
node.save()
if dry_run:
raise RuntimeError('Dry run, transaction rolled back.')
def main():
dry_run = False
if '--dry' in sys.argv:
dry_run = True
if not dry_run:
script_utils.add_file_logger(logger, __file__)
init_app(set_backends=True, routes=False)
with TokuTransaction():
migrate(targets=get_targets(), dry_run=dry_run)
if __name__ == "__main__":
main()
| import logging
import sys
from modularodm import Q
from framework.transactions.context import TokuTransaction
from website.app import init_app
from website.models import NodeLog
from scripts import utils as script_utils
logger = logging.getLogger(__name__)
def get_targets():
return NodeLog.find(Q('action', 'eq', NodeLog.WIKI_DELETED))
def migrate(targets, dry_run=True):
# iterate over targets
for log in targets:
node = log.node
versions = node.wiki_pages_versions
current = node.wiki_pages_current
updated_versions = {}
for wiki in versions:
if wiki in current:
updated_versions[wiki] = versions[wiki]
node.wiki_pages_versions = updated_versions
node.save()
def main():
dry_run = False
if '--dry' in sys.argv:
dry_run = True
if not dry_run:
script_utils.add_file_logger(logger, __file__)
init_app(set_backends=True, routes=False)
with TokuTransaction():
migrate(targets=get_targets(), dry_run=dry_run)
if dry_run:
raise RuntimeError('Dry run, transaction rolled back.')
if __name__ == "__main__":
main()
| Remove TokuTransaction in migrate function | Remove TokuTransaction in migrate function
| Python | apache-2.0 | hmoco/osf.io,samchrisinger/osf.io,hmoco/osf.io,icereval/osf.io,caneruguz/osf.io,cwisecarver/osf.io,chrisseto/osf.io,erinspace/osf.io,SSJohns/osf.io,monikagrabowska/osf.io,crcresearch/osf.io,crcresearch/osf.io,laurenrevere/osf.io,leb2dg/osf.io,crcresearch/osf.io,baylee-d/osf.io,leb2dg/osf.io,saradbowman/osf.io,sloria/osf.io,felliott/osf.io,mluke93/osf.io,adlius/osf.io,SSJohns/osf.io,mluke93/osf.io,binoculars/osf.io,mluo613/osf.io,felliott/osf.io,amyshi188/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,mfraezz/osf.io,DanielSBrown/osf.io,mluo613/osf.io,erinspace/osf.io,adlius/osf.io,mluke93/osf.io,acshi/osf.io,abought/osf.io,wearpants/osf.io,laurenrevere/osf.io,wearpants/osf.io,cslzchen/osf.io,mattclark/osf.io,rdhyee/osf.io,binoculars/osf.io,hmoco/osf.io,laurenrevere/osf.io,monikagrabowska/osf.io,baylee-d/osf.io,hmoco/osf.io,emetsger/osf.io,saradbowman/osf.io,mfraezz/osf.io,TomBaxter/osf.io,Nesiehr/osf.io,rdhyee/osf.io,abought/osf.io,abought/osf.io,rdhyee/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,aaxelb/osf.io,samchrisinger/osf.io,chrisseto/osf.io,caseyrollins/osf.io,felliott/osf.io,alexschiller/osf.io,DanielSBrown/osf.io,Nesiehr/osf.io,amyshi188/osf.io,chrisseto/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,mluo613/osf.io,rdhyee/osf.io,felliott/osf.io,abought/osf.io,samchrisinger/osf.io,alexschiller/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,brianjgeiger/osf.io,mluo613/osf.io,DanielSBrown/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,acshi/osf.io,TomBaxter/osf.io,emetsger/osf.io,adlius/osf.io,caseyrollins/osf.io,CenterForOpenScience/osf.io,wearpants/osf.io,emetsger/osf.io,monikagrabowska/osf.io,acshi/osf.io,aaxelb/osf.io,SSJohns/osf.io,brianjgeiger/osf.io,amyshi188/osf.io,chennan47/osf.io,sloria/osf.io,icereval/osf.io,mattclark/osf.io,alexschiller/osf.io,monikagrabowska/osf.io,binoculars/osf.io,wearpants/osf.io,samchrisinger/osf.io,HalcyonChimera/osf.io,amyshi188/osf.io,chrisseto/osf.io,cwisecarver/osf.io,aaxelb/osf.io,erinspace/osf.io,cslzchen/osf.io,aaxelb/osf.io,caneruguz/osf.io,pattisdr/osf.io,mluo613/osf.io,cwisecarver/osf.io,alexschiller/osf.io,chennan47/osf.io,pattisdr/osf.io,CenterForOpenScience/osf.io,CenterForOpenScience/osf.io,sloria/osf.io,mattclark/osf.io,mfraezz/osf.io,Nesiehr/osf.io,emetsger/osf.io,HalcyonChimera/osf.io,Johnetordoff/osf.io,leb2dg/osf.io,cslzchen/osf.io,pattisdr/osf.io,SSJohns/osf.io,icereval/osf.io,caneruguz/osf.io,alexschiller/osf.io,chennan47/osf.io,acshi/osf.io,caseyrollins/osf.io,mluke93/osf.io,CenterForOpenScience/osf.io,adlius/osf.io,HalcyonChimera/osf.io,Nesiehr/osf.io,acshi/osf.io,caneruguz/osf.io,mfraezz/osf.io,cwisecarver/osf.io | ---
+++
@@ -23,11 +23,8 @@
for wiki in versions:
if wiki in current:
updated_versions[wiki] = versions[wiki]
- with TokuTransaction():
- node.wiki_pages_versions = updated_versions
- node.save()
- if dry_run:
- raise RuntimeError('Dry run, transaction rolled back.')
+ node.wiki_pages_versions = updated_versions
+ node.save()
def main():
@@ -39,6 +36,8 @@
init_app(set_backends=True, routes=False)
with TokuTransaction():
migrate(targets=get_targets(), dry_run=dry_run)
+ if dry_run:
+ raise RuntimeError('Dry run, transaction rolled back.')
if __name__ == "__main__":
main() |
c69d99ff9c102926d94d8bd2d55c5de40f5e2072 | application/senic/nuimo_hub/subprocess_run.py | application/senic/nuimo_hub/subprocess_run.py | """Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`."""
try:
from subprocess import run
except ImportError:
from collections import namedtuple
from subprocess import check_output
def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None):
stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, encoding=encoding, errors=errors, timeout=timeout)
Output = namedtuple('Output', ['stdout'])
return Output(stdout=stdout_bytes)
| """Provides `subprocess.run()` from Python 3.5+ if available. Otherwise falls back to `subprocess.check_output()`."""
try:
from subprocess import run
except ImportError:
from collections import namedtuple
from subprocess import check_output
def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None):
stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, timeout=timeout)
Output = namedtuple('Output', ['stdout'])
return Output(stdout=stdout_bytes)
| Remove parameters that are not supported | Remove parameters that are not supported
Apparently Python 3.4 doesn't have encoding and errors parameters
| Python | mit | grunskis/nuimo-hub-backend,grunskis/senic-hub,grunskis/senic-hub,getsenic/senic-hub,grunskis/senic-hub,grunskis/senic-hub,grunskis/nuimo-hub-backend,grunskis/nuimo-hub-backend,grunskis/senic-hub,getsenic/senic-hub,grunskis/nuimo-hub-backend,grunskis/senic-hub,grunskis/nuimo-hub-backend | ---
+++
@@ -6,6 +6,6 @@
from collections import namedtuple
from subprocess import check_output
def run(args, *, stdin=None, input=None, stdout=None, stderr=None, shell=False, timeout=None, check=False, encoding=None, errors=None):
- stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, encoding=encoding, errors=errors, timeout=timeout)
+ stdout_bytes = check_output(args, stdin=stdin, stderr=stderr, shell=shell, timeout=timeout)
Output = namedtuple('Output', ['stdout'])
return Output(stdout=stdout_bytes) |
7894b066cde13eca75479921531e9d005970e9c3 | go/billing/views.py | go/billing/views.py | # Create your views here.
| import os
from django.conf import settings
from django.contrib.auth.decorators import login_required
from django.http import HttpResponse
from django.shortcuts import get_object_or_404
from django.template import RequestContext, loader
from xhtml2pdf import pisa
from go.billing.models import Statement
@login_required
def statement_view(request, statement_id=None):
"""Send a PDF version of the statement with the given
``statement_id`` to the user's browser.
"""
statement = get_object_or_404(
Statement, pk=statement_id, account__user=request.user)
response = HttpResponse(mimetype='application/pdf')
filename = "%s (%s).pdf" % (statement.title,
statement.from_date.strftime('%B %Y'))
response['Content-Disposition'] = 'attachment; filename=%s' % (filename,)
template = loader.get_template('billing/invoice.html')
line_item_list = statement.lineitem_set.all()
context = RequestContext(request, {'item_list': line_item_list})
html_result = template.render(context)
pisa.CreatePDF(html_result, dest=response, link_callback=link_callback)
return response
# Convert HTML URIs to absolute system paths so xhtml2pdf can access those
# resources
def link_callback(uri, rel):
# use short variable names
sUrl = settings.STATIC_URL # Typically /static/
sRoot = settings.STATIC_ROOT # Typically /home/userX/project_static/
mUrl = settings.MEDIA_URL # Typically /static/media/
mRoot = settings.MEDIA_ROOT # Typically /home/userX/project_static/media/
# convert URIs to absolute system paths
if uri.startswith(mUrl):
path = os.path.join(mRoot, uri.replace(mUrl, ""))
elif uri.startswith(sUrl):
path = os.path.join(sRoot, uri.replace(sUrl, ""))
# make sure that file exists
if not os.path.isfile(path):
raise Exception(
'media URI must start with %s or %s' % (sUrl, mUrl))
return path
| Add basic PDF view for billing | Add basic PDF view for billing
| Python | bsd-3-clause | praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go | ---
+++
@@ -1 +1,57 @@
-# Create your views here.
+import os
+
+from django.conf import settings
+from django.contrib.auth.decorators import login_required
+from django.http import HttpResponse
+from django.shortcuts import get_object_or_404
+from django.template import RequestContext, loader
+from xhtml2pdf import pisa
+
+
+from go.billing.models import Statement
+
+
+@login_required
+def statement_view(request, statement_id=None):
+ """Send a PDF version of the statement with the given
+ ``statement_id`` to the user's browser.
+ """
+ statement = get_object_or_404(
+ Statement, pk=statement_id, account__user=request.user)
+
+ response = HttpResponse(mimetype='application/pdf')
+ filename = "%s (%s).pdf" % (statement.title,
+ statement.from_date.strftime('%B %Y'))
+
+ response['Content-Disposition'] = 'attachment; filename=%s' % (filename,)
+
+ template = loader.get_template('billing/invoice.html')
+ line_item_list = statement.lineitem_set.all()
+ context = RequestContext(request, {'item_list': line_item_list})
+ html_result = template.render(context)
+
+ pisa.CreatePDF(html_result, dest=response, link_callback=link_callback)
+
+ return response
+
+
+# Convert HTML URIs to absolute system paths so xhtml2pdf can access those
+# resources
+def link_callback(uri, rel):
+ # use short variable names
+ sUrl = settings.STATIC_URL # Typically /static/
+ sRoot = settings.STATIC_ROOT # Typically /home/userX/project_static/
+ mUrl = settings.MEDIA_URL # Typically /static/media/
+ mRoot = settings.MEDIA_ROOT # Typically /home/userX/project_static/media/
+
+ # convert URIs to absolute system paths
+ if uri.startswith(mUrl):
+ path = os.path.join(mRoot, uri.replace(mUrl, ""))
+ elif uri.startswith(sUrl):
+ path = os.path.join(sRoot, uri.replace(sUrl, ""))
+
+ # make sure that file exists
+ if not os.path.isfile(path):
+ raise Exception(
+ 'media URI must start with %s or %s' % (sUrl, mUrl))
+ return path |
a3b6aa273900c9d9c98a6ad9c9735a323086c64f | user_clipboard/urls.py | user_clipboard/urls.py | from django.conf.urls import url
from .views import ClipboardFileAPIView, ClipboardImageAPIView
urlpatterns = [
url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"),
url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"),
url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"),
url(r'^', ClipboardFileAPIView.as_view(), name="clipboard"),
]
| from django.conf.urls import url
from .views import ClipboardFileAPIView, ClipboardImageAPIView
urlpatterns = [
url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"),
url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"),
url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"),
url(r'^$', ClipboardFileAPIView.as_view(), name="clipboard"),
]
| Fix clipboard URLs to not be greedy | Fix clipboard URLs to not be greedy
| Python | mit | MagicSolutions/django-user-clipboard,IndustriaTech/django-user-clipboard,MagicSolutions/django-user-clipboard,IndustriaTech/django-user-clipboard | ---
+++
@@ -6,5 +6,5 @@
url(r'^images/(?P<pk>\d+)/$', ClipboardImageAPIView.as_view(), name="clipboard_images"),
url(r'^images/$', ClipboardImageAPIView.as_view(), name="clipboard_images"),
url(r'^(?P<pk>\d+)$', ClipboardFileAPIView.as_view(), name="clipboard"),
- url(r'^', ClipboardFileAPIView.as_view(), name="clipboard"),
+ url(r'^$', ClipboardFileAPIView.as_view(), name="clipboard"),
] |
9a1a05c33258461c5d474b014654464892cd7b90 | bake/bakedefaults.py | bake/bakedefaults.py | #!/usr/bin/env python
LABEL_KEY = 'label'
KEY_START = '@'
KEY_END = '@'
CFGFILE = 'bake.cfg'
| #!/usr/bin/env python
LABEL_KEY = 'label'
KEY_START = '@'
KEY_END = '@'
| Remove mention of bake.cfg file | Remove mention of bake.cfg file
| Python | mit | AlexSzatmary/bake | ---
+++
@@ -3,4 +3,3 @@
LABEL_KEY = 'label'
KEY_START = '@'
KEY_END = '@'
-CFGFILE = 'bake.cfg' |
c859416d2d35aab83fc9e8f400e00f8f07c0b8a9 | test/parser_test.py | test/parser_test.py | import socket
client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
client_socket.connect(("localhost", 5002))
with open("resources/Matrix.java", "r") as java_file:
source = java_file.read() + "\nEOS_BITSHIFT"
client_socket.send("%d\n%s" % (len(source), source));
data = ''
while True:
data = client_socket.recv(10000)
if data != '':
client_socket.close()
break;
print data;
| import socket, sys
file_name = 'resources/<name>.c'
server_socket_number = 5001
if __name__ == '__main__':
if len(sys.argv) == 1:
print "Please input a parser to test."
elif len(sys.argv) > 2:
print "Too many arguments."
else:
if sys.argv[1] == 'c':
pass
elif sys.argv[1] == 'java':
file_name = "resources/Matrix.java"
server_socket_number = 5002
elif sys.argv[1] == 'ruby':
file_name = "resources/<name>.rb"
server_socket_number = 5003
client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
client_socket.connect(("localhost", server_socket_number))
with open(file_name, "r") as source_file:
source = source_file.read()
client_socket.send("%d\n%s" % (len(source), source));
data = ''
while True:
data = client_socket.recv(10000)
if data != '':
client_socket.close()
break;
print data;
| Change test file to support different parsers | Change test file to support different parsers
| Python | mit | earwig/bitshift,earwig/bitshift,earwig/bitshift,earwig/bitshift,earwig/bitshift,earwig/bitshift | ---
+++
@@ -1,18 +1,40 @@
-import socket
+import socket, sys
-client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
-client_socket.connect(("localhost", 5002))
+file_name = 'resources/<name>.c'
+server_socket_number = 5001
-with open("resources/Matrix.java", "r") as java_file:
- source = java_file.read() + "\nEOS_BITSHIFT"
- client_socket.send("%d\n%s" % (len(source), source));
+if __name__ == '__main__':
+ if len(sys.argv) == 1:
+ print "Please input a parser to test."
-data = ''
-while True:
- data = client_socket.recv(10000)
+ elif len(sys.argv) > 2:
+ print "Too many arguments."
- if data != '':
- client_socket.close()
- break;
+ else:
+ if sys.argv[1] == 'c':
+ pass
-print data;
+ elif sys.argv[1] == 'java':
+ file_name = "resources/Matrix.java"
+ server_socket_number = 5002
+
+ elif sys.argv[1] == 'ruby':
+ file_name = "resources/<name>.rb"
+ server_socket_number = 5003
+
+ client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
+ client_socket.connect(("localhost", server_socket_number))
+
+ with open(file_name, "r") as source_file:
+ source = source_file.read()
+ client_socket.send("%d\n%s" % (len(source), source));
+
+ data = ''
+ while True:
+ data = client_socket.recv(10000)
+
+ if data != '':
+ client_socket.close()
+ break;
+
+ print data; |
6603d277eb84712a907008bad8bc8156ed9bdf40 | setup.py | setup.py | """A setuptools based setup module.
See:
https://packaging.python.org/en/latest/distributing.html
https://github.com/pypa/sampleproject
"""
from setuptools import setup, find_packages
from codecs import open
from os import path
here = path.abspath(path.dirname(__file__))
with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f:
long_description = f.read()
setup(
name='Graphistry',
version='1.1.0.dev1',
description='This is established as a Data Loader for Graphistry',
long_description=long_description,
url='https://github.com/graphistry/pygraphistry',
author='Graphistry',
author_email='xin@graphistry.com',
license='MIT',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Graphistry User',
'Topic :: Data Visualization Development :: Load Tools',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
keywords='Python Data Loader',
packages=find_packages(exclude=['contrib', 'docs', 'tests*']),
)
| """A setuptools based setup module.
See:
https://packaging.python.org/en/latest/distributing.html
https://github.com/pypa/sampleproject
"""
from setuptools import setup, find_packages
from codecs import open
from os import path
here = path.abspath(path.dirname(__file__))
with open(path.join(here, 'DESCRIPTION.rst'), encoding='utf-8') as f:
long_description = f.read()
setup(
name='Graphistry',
version='1.1.0.dev1',
description='This is established as a Data Loader for Graphistry',
long_description=long_description,
url='https://github.com/graphistry/pygraphistry',
author='Graphistry',
author_email='xin@graphistry.com',
license='MIT',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Graphistry User',
'Topic :: Data Visualization Development :: Load Tools',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
keywords='Python Data Loader',
packages=find_packages(exclude=['contrib', 'docs', 'tests*']),
)
| Fix license MIT to BSD | Fix license MIT to BSD
| Python | bsd-3-clause | graphistry/pygraphistry,kcompher/pygraphistry,graphistry/pygraphistry | ---
+++
@@ -34,7 +34,7 @@
'Intended Audience :: Graphistry User',
'Topic :: Data Visualization Development :: Load Tools',
- 'License :: OSI Approved :: MIT License',
+ 'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6', |
0db9b203dddefa6576bbcf357453587f05403180 | setup.py | setup.py | #!/usr/bin/env python
# coding=utf-8
from setuptools import setup
setup(
name='alfred-workflow-packager',
version='0.8b',
description='A CLI utility for packaging and exporting Alfred workflows',
url='https://github.com/caleb531/alfred-workflow-packager',
author='Caleb Evans',
author_email='caleb@calebevans.me',
license='MIT',
keywords='alfred workflow package export',
packages=['awp'],
install_requires=[
'biplist >= 1, < 2',
'jsonschema >= 2, < 3'
],
entry_points={
'console_scripts': [
'alfred-workflow-packager=awp.packager:main',
'workflow-packager=awp.packager:main'
]
}
)
| #!/usr/bin/env python
# coding=utf-8
from setuptools import setup
setup(
name='alfred-workflow-packager',
version='0.9',
description='A CLI utility for packaging and exporting Alfred workflows',
url='https://github.com/caleb531/alfred-workflow-packager',
author='Caleb Evans',
author_email='caleb@calebevans.me',
license='MIT',
keywords='alfred workflow package export',
packages=['awp'],
install_requires=[
'biplist >= 1, < 2',
'jsonschema >= 2, < 3'
],
entry_points={
'console_scripts': [
'alfred-workflow-packager=awp.packager:main',
'workflow-packager=awp.packager:main'
]
}
)
| Change initial version to 0.9 | Change initial version to 0.9
| Python | mit | caleb531/alfred-workflow-packager | ---
+++
@@ -5,7 +5,7 @@
setup(
name='alfred-workflow-packager',
- version='0.8b',
+ version='0.9',
description='A CLI utility for packaging and exporting Alfred workflows',
url='https://github.com/caleb531/alfred-workflow-packager',
author='Caleb Evans', |
41de24fa992939d8cf3daf762bd7c3235faeb5d0 | setup.py | setup.py | import os
from setuptools import setup # type: ignore
VERSION = '4.5'
setup(
name='conllu',
packages=["conllu"],
python_requires=">=3.6",
package_data={
"": ["py.typed"]
},
version=VERSION,
license='MIT License',
description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary',
long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(),
long_description_content_type="text/markdown",
author=u'Emil Stenström',
author_email="emil@emilstenstrom.se",
url='https://github.com/EmilStenstrom/conllu/',
keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'],
classifiers=[
"Programming Language :: Python",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3 :: Only",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Programming Language :: Python :: 3.8",
"Programming Language :: Python :: 3.9",
"Programming Language :: Python :: 3.10",
"Operating System :: OS Independent",
],
)
| import os
from setuptools import setup # type: ignore
VERSION = '4.5'
with open(os.path.join(os.path.dirname(__file__), 'README.md')) as f:
description = f.read()
setup(
name='conllu',
packages=["conllu"],
python_requires=">=3.6",
package_data={
"": ["py.typed"]
},
version=VERSION,
license='MIT License',
description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary',
long_description=description,
long_description_content_type="text/markdown",
author=u'Emil Stenström',
author_email="emil@emilstenstrom.se",
url='https://github.com/EmilStenstrom/conllu/',
keywords=['conllu', 'conll', 'conll-u', 'parser', 'nlp'],
classifiers=[
"Programming Language :: Python",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3 :: Only",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Programming Language :: Python :: 3.8",
"Programming Language :: Python :: 3.9",
"Programming Language :: Python :: 3.10",
"Operating System :: OS Independent",
],
)
| Make sure file is closed properly. | Make sure file is closed properly.
| Python | mit | EmilStenstrom/conllu | ---
+++
@@ -3,6 +3,9 @@
from setuptools import setup # type: ignore
VERSION = '4.5'
+
+with open(os.path.join(os.path.dirname(__file__), 'README.md')) as f:
+ description = f.read()
setup(
name='conllu',
@@ -14,7 +17,7 @@
version=VERSION,
license='MIT License',
description='CoNLL-U Parser parses a CoNLL-U formatted string into a nested python dictionary',
- long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(),
+ long_description=description,
long_description_content_type="text/markdown",
author=u'Emil Stenström',
author_email="emil@emilstenstrom.se", |
857133125fe5a633f209887b1b739f6f82c80f43 | setup.py | setup.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup
with open('README.rst') as readme_file:
readme = readme_file.read()
with open('HISTORY.rst') as history_file:
history = history_file.read()
setup(
name='learnregex',
version='0.4.2',
description='A pyschool story for learning regular expressions.',
long_description=readme + '\n\n' + history,
author='Sophilabs',
author_email='hi@sophilabs.co',
url='https://github.com/sophilabs/learnregex',
packages=['learnregex'],
entry_points={
'console_scripts': [
'learnregex=learnregex.story:Story.begin'
]
},
include_package_data=True,
install_requires=['story'],
license='MIT license',
zip_safe=False,
keywords='learnregex',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
test_suite='tests',
tests_require=[]
)
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup
with open('README.rst') as readme_file:
readme = readme_file.read()
with open('HISTORY.rst') as history_file:
history = history_file.read()
setup(
name='learnregex',
version='0.4.2',
description='A pyschool story for learning regular expressions.',
long_description=readme + '\n\n' + history,
author='Sophilabs',
author_email='hi@sophilabs.co',
url='https://github.com/sophilabs/learnregex',
packages=['learnregex'],
entry_points={
'console_scripts': [
'learnregex=learnregex.story:Story.begin'
]
},
include_package_data=True,
install_requires=[
'story>=1.1.2'
],
license='MIT license',
zip_safe=False,
keywords='learnregex',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
test_suite='tests',
tests_require=[]
)
| Update mininum story version required. | Update mininum story version required.
| Python | mit | sophilabs/learnregex | ---
+++
@@ -24,7 +24,9 @@
]
},
include_package_data=True,
- install_requires=['story'],
+ install_requires=[
+ 'story>=1.1.2'
+ ],
license='MIT license',
zip_safe=False,
keywords='learnregex', |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.