commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
8e0d28d23c7ceb6a200773dde035b85965273ac6
inbox/events/actions/base.py
inbox/events/actions/base.py
from inbox.models.account import Account from inbox.models.event import Event from inbox.events.actions.backends import module_registry def create_event(account_id, event_id, db_session): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_create_event = module_registry[account.provider].remote_create_event remote_create_event(account, event, db_session) def update_event(account_id, event_id, db_session): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_update_event = module_registry[account.provider].remote_update_event remote_update_event(account, event, db_session) def delete_event(account_id, event_id, db_session, args): account = db_session.query(Account).get(account_id) remote_delete_event = module_registry[account.provider].remote_delete_event event_uid = args.get('event_uid') calendar_name = args.get('calendar_name') # The calendar_uid argument is required for some providers, like EAS. calendar_uid = args.get('calendar_uid') remote_delete_event(account, event_uid, calendar_name, calendar_uid, db_session)
from inbox.models.account import Account from inbox.models.event import Event from inbox.events.actions.backends import module_registry def create_event(account_id, event_id, db_session, *args): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_create_event = module_registry[account.provider].remote_create_event remote_create_event(account, event, db_session) def update_event(account_id, event_id, db_session, *args): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_update_event = module_registry[account.provider].remote_update_event remote_update_event(account, event, db_session) def delete_event(account_id, event_id, db_session, args): account = db_session.query(Account).get(account_id) remote_delete_event = module_registry[account.provider].remote_delete_event event_uid = args.get('event_uid') calendar_name = args.get('calendar_name') # The calendar_uid argument is required for some providers, like EAS. calendar_uid = args.get('calendar_uid') remote_delete_event(account, event_uid, calendar_name, calendar_uid, db_session)
Add args flexibility to event create to match EAS requirements
Add args flexibility to event create to match EAS requirements
Python
agpl-3.0
PriviPK/privipk-sync-engine,ErinCall/sync-engine,nylas/sync-engine,Eagles2F/sync-engine,closeio/nylas,nylas/sync-engine,jobscore/sync-engine,PriviPK/privipk-sync-engine,PriviPK/privipk-sync-engine,wakermahmud/sync-engine,gale320/sync-engine,wakermahmud/sync-engine,jobscore/sync-engine,jobscore/sync-engine,closeio/nylas,gale320/sync-engine,jobscore/sync-engine,PriviPK/privipk-sync-engine,wakermahmud/sync-engine,ErinCall/sync-engine,closeio/nylas,gale320/sync-engine,PriviPK/privipk-sync-engine,nylas/sync-engine,ErinCall/sync-engine,nylas/sync-engine,gale320/sync-engine,ErinCall/sync-engine,wakermahmud/sync-engine,ErinCall/sync-engine,Eagles2F/sync-engine,Eagles2F/sync-engine,closeio/nylas,wakermahmud/sync-engine,Eagles2F/sync-engine,gale320/sync-engine,Eagles2F/sync-engine
from inbox.models.account import Account from inbox.models.event import Event from inbox.events.actions.backends import module_registry def create_event(account_id, event_id, db_session): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_create_event = module_registry[account.provider].remote_create_event remote_create_event(account, event, db_session) def update_event(account_id, event_id, db_session): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_update_event = module_registry[account.provider].remote_update_event remote_update_event(account, event, db_session) def delete_event(account_id, event_id, db_session, args): account = db_session.query(Account).get(account_id) remote_delete_event = module_registry[account.provider].remote_delete_event event_uid = args.get('event_uid') calendar_name = args.get('calendar_name') # The calendar_uid argument is required for some providers, like EAS. calendar_uid = args.get('calendar_uid') remote_delete_event(account, event_uid, calendar_name, calendar_uid, db_session) Add args flexibility to event create to match EAS requirements
from inbox.models.account import Account from inbox.models.event import Event from inbox.events.actions.backends import module_registry def create_event(account_id, event_id, db_session, *args): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_create_event = module_registry[account.provider].remote_create_event remote_create_event(account, event, db_session) def update_event(account_id, event_id, db_session, *args): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_update_event = module_registry[account.provider].remote_update_event remote_update_event(account, event, db_session) def delete_event(account_id, event_id, db_session, args): account = db_session.query(Account).get(account_id) remote_delete_event = module_registry[account.provider].remote_delete_event event_uid = args.get('event_uid') calendar_name = args.get('calendar_name') # The calendar_uid argument is required for some providers, like EAS. calendar_uid = args.get('calendar_uid') remote_delete_event(account, event_uid, calendar_name, calendar_uid, db_session)
<commit_before>from inbox.models.account import Account from inbox.models.event import Event from inbox.events.actions.backends import module_registry def create_event(account_id, event_id, db_session): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_create_event = module_registry[account.provider].remote_create_event remote_create_event(account, event, db_session) def update_event(account_id, event_id, db_session): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_update_event = module_registry[account.provider].remote_update_event remote_update_event(account, event, db_session) def delete_event(account_id, event_id, db_session, args): account = db_session.query(Account).get(account_id) remote_delete_event = module_registry[account.provider].remote_delete_event event_uid = args.get('event_uid') calendar_name = args.get('calendar_name') # The calendar_uid argument is required for some providers, like EAS. calendar_uid = args.get('calendar_uid') remote_delete_event(account, event_uid, calendar_name, calendar_uid, db_session) <commit_msg>Add args flexibility to event create to match EAS requirements<commit_after>
from inbox.models.account import Account from inbox.models.event import Event from inbox.events.actions.backends import module_registry def create_event(account_id, event_id, db_session, *args): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_create_event = module_registry[account.provider].remote_create_event remote_create_event(account, event, db_session) def update_event(account_id, event_id, db_session, *args): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_update_event = module_registry[account.provider].remote_update_event remote_update_event(account, event, db_session) def delete_event(account_id, event_id, db_session, args): account = db_session.query(Account).get(account_id) remote_delete_event = module_registry[account.provider].remote_delete_event event_uid = args.get('event_uid') calendar_name = args.get('calendar_name') # The calendar_uid argument is required for some providers, like EAS. calendar_uid = args.get('calendar_uid') remote_delete_event(account, event_uid, calendar_name, calendar_uid, db_session)
from inbox.models.account import Account from inbox.models.event import Event from inbox.events.actions.backends import module_registry def create_event(account_id, event_id, db_session): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_create_event = module_registry[account.provider].remote_create_event remote_create_event(account, event, db_session) def update_event(account_id, event_id, db_session): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_update_event = module_registry[account.provider].remote_update_event remote_update_event(account, event, db_session) def delete_event(account_id, event_id, db_session, args): account = db_session.query(Account).get(account_id) remote_delete_event = module_registry[account.provider].remote_delete_event event_uid = args.get('event_uid') calendar_name = args.get('calendar_name') # The calendar_uid argument is required for some providers, like EAS. calendar_uid = args.get('calendar_uid') remote_delete_event(account, event_uid, calendar_name, calendar_uid, db_session) Add args flexibility to event create to match EAS requirementsfrom inbox.models.account import Account from inbox.models.event import Event from inbox.events.actions.backends import module_registry def create_event(account_id, event_id, db_session, *args): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_create_event = module_registry[account.provider].remote_create_event remote_create_event(account, event, db_session) def update_event(account_id, event_id, db_session, *args): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_update_event = module_registry[account.provider].remote_update_event remote_update_event(account, event, db_session) def delete_event(account_id, event_id, db_session, args): account = db_session.query(Account).get(account_id) remote_delete_event = module_registry[account.provider].remote_delete_event event_uid = args.get('event_uid') calendar_name = args.get('calendar_name') # The calendar_uid argument is required for some providers, like EAS. calendar_uid = args.get('calendar_uid') remote_delete_event(account, event_uid, calendar_name, calendar_uid, db_session)
<commit_before>from inbox.models.account import Account from inbox.models.event import Event from inbox.events.actions.backends import module_registry def create_event(account_id, event_id, db_session): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_create_event = module_registry[account.provider].remote_create_event remote_create_event(account, event, db_session) def update_event(account_id, event_id, db_session): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_update_event = module_registry[account.provider].remote_update_event remote_update_event(account, event, db_session) def delete_event(account_id, event_id, db_session, args): account = db_session.query(Account).get(account_id) remote_delete_event = module_registry[account.provider].remote_delete_event event_uid = args.get('event_uid') calendar_name = args.get('calendar_name') # The calendar_uid argument is required for some providers, like EAS. calendar_uid = args.get('calendar_uid') remote_delete_event(account, event_uid, calendar_name, calendar_uid, db_session) <commit_msg>Add args flexibility to event create to match EAS requirements<commit_after>from inbox.models.account import Account from inbox.models.event import Event from inbox.events.actions.backends import module_registry def create_event(account_id, event_id, db_session, *args): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_create_event = module_registry[account.provider].remote_create_event remote_create_event(account, event, db_session) def update_event(account_id, event_id, db_session, *args): account = db_session.query(Account).get(account_id) event = db_session.query(Event).get(event_id) remote_update_event = module_registry[account.provider].remote_update_event remote_update_event(account, event, db_session) def delete_event(account_id, event_id, db_session, args): account = db_session.query(Account).get(account_id) remote_delete_event = module_registry[account.provider].remote_delete_event event_uid = args.get('event_uid') calendar_name = args.get('calendar_name') # The calendar_uid argument is required for some providers, like EAS. calendar_uid = args.get('calendar_uid') remote_delete_event(account, event_uid, calendar_name, calendar_uid, db_session)
ea22f4bf62204805e698965300b6d8dfa637a662
pybossa_discourse/globals.py
pybossa_discourse/globals.py
# -*- coding: utf8 -*- """Jinja globals module for pybossa-discourse.""" from flask import Markup, request class DiscourseGlobals(object): """A class to implement Discourse Global variables.""" def __init__(self, app): self.url = app.config['DISCOURSE_URL'] app.jinja_env.globals.update(discourse=self) def comments(self): """Return an HTML snippet used to embed Discourse comments.""" return Markup(""" <div id="discourse-comments"></div> <script type="text/javascript"> DiscourseEmbed = {{ discourseUrl: '{0}/', discourseEmbedUrl: '{1}' }}; window.onload = function() {{ let d = document.createElement('script'), head = document.getElementsByTagName('head')[0], body = document.getElementsByTagName('body')[0]; d.type = 'text/javascript'; d.async = true; d.src = '{0}/javascripts/embed.js'; (head || body).appendChild(d); }} </script> """).format(self.url, request.base_url)
# -*- coding: utf8 -*- """Jinja globals module for pybossa-discourse.""" from flask import Markup, request from . import discourse_client class DiscourseGlobals(object): """A class to implement Discourse Global variables.""" def __init__(self, app): self.url = app.config['DISCOURSE_URL'] app.jinja_env.globals.update(discourse=self) def comments(self): """Return an HTML snippet used to embed Discourse comments.""" return Markup(""" <div id="discourse-comments"></div> <script type="text/javascript"> DiscourseEmbed = {{ discourseUrl: '{0}/', discourseEmbedUrl: '{1}' }}; window.onload = function() {{ let d = document.createElement('script'), head = document.getElementsByTagName('head')[0], body = document.getElementsByTagName('body')[0]; d.type = 'text/javascript'; d.async = true; d.src = '{0}/javascripts/embed.js'; (head || body).appendChild(d); }} </script> """).format(self.url, request.base_url) def notifications(self): """Return a count of unread notifications for the current user.""" notifications = discourse_client.user_notifications() if not notifications: return 0 return sum([1 for n in notifications['notifications'] if not n['read']])
Add notifications count to global envar
Add notifications count to global envar
Python
bsd-3-clause
alexandermendes/pybossa-discourse
# -*- coding: utf8 -*- """Jinja globals module for pybossa-discourse.""" from flask import Markup, request class DiscourseGlobals(object): """A class to implement Discourse Global variables.""" def __init__(self, app): self.url = app.config['DISCOURSE_URL'] app.jinja_env.globals.update(discourse=self) def comments(self): """Return an HTML snippet used to embed Discourse comments.""" return Markup(""" <div id="discourse-comments"></div> <script type="text/javascript"> DiscourseEmbed = {{ discourseUrl: '{0}/', discourseEmbedUrl: '{1}' }}; window.onload = function() {{ let d = document.createElement('script'), head = document.getElementsByTagName('head')[0], body = document.getElementsByTagName('body')[0]; d.type = 'text/javascript'; d.async = true; d.src = '{0}/javascripts/embed.js'; (head || body).appendChild(d); }} </script> """).format(self.url, request.base_url) Add notifications count to global envar
# -*- coding: utf8 -*- """Jinja globals module for pybossa-discourse.""" from flask import Markup, request from . import discourse_client class DiscourseGlobals(object): """A class to implement Discourse Global variables.""" def __init__(self, app): self.url = app.config['DISCOURSE_URL'] app.jinja_env.globals.update(discourse=self) def comments(self): """Return an HTML snippet used to embed Discourse comments.""" return Markup(""" <div id="discourse-comments"></div> <script type="text/javascript"> DiscourseEmbed = {{ discourseUrl: '{0}/', discourseEmbedUrl: '{1}' }}; window.onload = function() {{ let d = document.createElement('script'), head = document.getElementsByTagName('head')[0], body = document.getElementsByTagName('body')[0]; d.type = 'text/javascript'; d.async = true; d.src = '{0}/javascripts/embed.js'; (head || body).appendChild(d); }} </script> """).format(self.url, request.base_url) def notifications(self): """Return a count of unread notifications for the current user.""" notifications = discourse_client.user_notifications() if not notifications: return 0 return sum([1 for n in notifications['notifications'] if not n['read']])
<commit_before># -*- coding: utf8 -*- """Jinja globals module for pybossa-discourse.""" from flask import Markup, request class DiscourseGlobals(object): """A class to implement Discourse Global variables.""" def __init__(self, app): self.url = app.config['DISCOURSE_URL'] app.jinja_env.globals.update(discourse=self) def comments(self): """Return an HTML snippet used to embed Discourse comments.""" return Markup(""" <div id="discourse-comments"></div> <script type="text/javascript"> DiscourseEmbed = {{ discourseUrl: '{0}/', discourseEmbedUrl: '{1}' }}; window.onload = function() {{ let d = document.createElement('script'), head = document.getElementsByTagName('head')[0], body = document.getElementsByTagName('body')[0]; d.type = 'text/javascript'; d.async = true; d.src = '{0}/javascripts/embed.js'; (head || body).appendChild(d); }} </script> """).format(self.url, request.base_url) <commit_msg>Add notifications count to global envar<commit_after>
# -*- coding: utf8 -*- """Jinja globals module for pybossa-discourse.""" from flask import Markup, request from . import discourse_client class DiscourseGlobals(object): """A class to implement Discourse Global variables.""" def __init__(self, app): self.url = app.config['DISCOURSE_URL'] app.jinja_env.globals.update(discourse=self) def comments(self): """Return an HTML snippet used to embed Discourse comments.""" return Markup(""" <div id="discourse-comments"></div> <script type="text/javascript"> DiscourseEmbed = {{ discourseUrl: '{0}/', discourseEmbedUrl: '{1}' }}; window.onload = function() {{ let d = document.createElement('script'), head = document.getElementsByTagName('head')[0], body = document.getElementsByTagName('body')[0]; d.type = 'text/javascript'; d.async = true; d.src = '{0}/javascripts/embed.js'; (head || body).appendChild(d); }} </script> """).format(self.url, request.base_url) def notifications(self): """Return a count of unread notifications for the current user.""" notifications = discourse_client.user_notifications() if not notifications: return 0 return sum([1 for n in notifications['notifications'] if not n['read']])
# -*- coding: utf8 -*- """Jinja globals module for pybossa-discourse.""" from flask import Markup, request class DiscourseGlobals(object): """A class to implement Discourse Global variables.""" def __init__(self, app): self.url = app.config['DISCOURSE_URL'] app.jinja_env.globals.update(discourse=self) def comments(self): """Return an HTML snippet used to embed Discourse comments.""" return Markup(""" <div id="discourse-comments"></div> <script type="text/javascript"> DiscourseEmbed = {{ discourseUrl: '{0}/', discourseEmbedUrl: '{1}' }}; window.onload = function() {{ let d = document.createElement('script'), head = document.getElementsByTagName('head')[0], body = document.getElementsByTagName('body')[0]; d.type = 'text/javascript'; d.async = true; d.src = '{0}/javascripts/embed.js'; (head || body).appendChild(d); }} </script> """).format(self.url, request.base_url) Add notifications count to global envar# -*- coding: utf8 -*- """Jinja globals module for pybossa-discourse.""" from flask import Markup, request from . import discourse_client class DiscourseGlobals(object): """A class to implement Discourse Global variables.""" def __init__(self, app): self.url = app.config['DISCOURSE_URL'] app.jinja_env.globals.update(discourse=self) def comments(self): """Return an HTML snippet used to embed Discourse comments.""" return Markup(""" <div id="discourse-comments"></div> <script type="text/javascript"> DiscourseEmbed = {{ discourseUrl: '{0}/', discourseEmbedUrl: '{1}' }}; window.onload = function() {{ let d = document.createElement('script'), head = document.getElementsByTagName('head')[0], body = document.getElementsByTagName('body')[0]; d.type = 'text/javascript'; d.async = true; d.src = '{0}/javascripts/embed.js'; (head || body).appendChild(d); }} </script> """).format(self.url, request.base_url) def notifications(self): """Return a count of unread notifications for the current user.""" notifications = discourse_client.user_notifications() if not notifications: return 0 return sum([1 for n in notifications['notifications'] if not n['read']])
<commit_before># -*- coding: utf8 -*- """Jinja globals module for pybossa-discourse.""" from flask import Markup, request class DiscourseGlobals(object): """A class to implement Discourse Global variables.""" def __init__(self, app): self.url = app.config['DISCOURSE_URL'] app.jinja_env.globals.update(discourse=self) def comments(self): """Return an HTML snippet used to embed Discourse comments.""" return Markup(""" <div id="discourse-comments"></div> <script type="text/javascript"> DiscourseEmbed = {{ discourseUrl: '{0}/', discourseEmbedUrl: '{1}' }}; window.onload = function() {{ let d = document.createElement('script'), head = document.getElementsByTagName('head')[0], body = document.getElementsByTagName('body')[0]; d.type = 'text/javascript'; d.async = true; d.src = '{0}/javascripts/embed.js'; (head || body).appendChild(d); }} </script> """).format(self.url, request.base_url) <commit_msg>Add notifications count to global envar<commit_after># -*- coding: utf8 -*- """Jinja globals module for pybossa-discourse.""" from flask import Markup, request from . import discourse_client class DiscourseGlobals(object): """A class to implement Discourse Global variables.""" def __init__(self, app): self.url = app.config['DISCOURSE_URL'] app.jinja_env.globals.update(discourse=self) def comments(self): """Return an HTML snippet used to embed Discourse comments.""" return Markup(""" <div id="discourse-comments"></div> <script type="text/javascript"> DiscourseEmbed = {{ discourseUrl: '{0}/', discourseEmbedUrl: '{1}' }}; window.onload = function() {{ let d = document.createElement('script'), head = document.getElementsByTagName('head')[0], body = document.getElementsByTagName('body')[0]; d.type = 'text/javascript'; d.async = true; d.src = '{0}/javascripts/embed.js'; (head || body).appendChild(d); }} </script> """).format(self.url, request.base_url) def notifications(self): """Return a count of unread notifications for the current user.""" notifications = discourse_client.user_notifications() if not notifications: return 0 return sum([1 for n in notifications['notifications'] if not n['read']])
1d292feebd2999eb042da1f606c0fdc33103225f
api/models.py
api/models.py
class MessageModel: def __init__(self, message, duration, creation_date, message_category): # We will automatically generate the new id self.id = 0 self.message = message self.duration = duration self.creation_date = creation_date self.message_category = message_category self.printed_times = 0 self.printed_once = False
class MessageModel: def __init__(self, message, duration, creation_date, message_category): # We will automatically generate the new id self.id = 0 self.message = message self.duration = duration self.creation_date = creation_date self.message_category = message_category self.printed_times = 0 self.printed_once = False class AccountModel: def __init__(self, account_type, account_number, name, first_name, address, birthdate): # We will automatically generate the new id self.id = 0 self.type = account_type self.number = account_number self.name = name self.first_name = first_name self.address = address self.birthdate = birthdate #We will automatically generate next 2 parameters based on client address. self.longitude = 0; self.latitude = 0;
Update model script to support task database schema
Update model script to support task database schema
Python
mit
candidate48661/BEA
class MessageModel: def __init__(self, message, duration, creation_date, message_category): # We will automatically generate the new id self.id = 0 self.message = message self.duration = duration self.creation_date = creation_date self.message_category = message_category self.printed_times = 0 self.printed_once = FalseUpdate model script to support task database schema
class MessageModel: def __init__(self, message, duration, creation_date, message_category): # We will automatically generate the new id self.id = 0 self.message = message self.duration = duration self.creation_date = creation_date self.message_category = message_category self.printed_times = 0 self.printed_once = False class AccountModel: def __init__(self, account_type, account_number, name, first_name, address, birthdate): # We will automatically generate the new id self.id = 0 self.type = account_type self.number = account_number self.name = name self.first_name = first_name self.address = address self.birthdate = birthdate #We will automatically generate next 2 parameters based on client address. self.longitude = 0; self.latitude = 0;
<commit_before>class MessageModel: def __init__(self, message, duration, creation_date, message_category): # We will automatically generate the new id self.id = 0 self.message = message self.duration = duration self.creation_date = creation_date self.message_category = message_category self.printed_times = 0 self.printed_once = False<commit_msg>Update model script to support task database schema<commit_after>
class MessageModel: def __init__(self, message, duration, creation_date, message_category): # We will automatically generate the new id self.id = 0 self.message = message self.duration = duration self.creation_date = creation_date self.message_category = message_category self.printed_times = 0 self.printed_once = False class AccountModel: def __init__(self, account_type, account_number, name, first_name, address, birthdate): # We will automatically generate the new id self.id = 0 self.type = account_type self.number = account_number self.name = name self.first_name = first_name self.address = address self.birthdate = birthdate #We will automatically generate next 2 parameters based on client address. self.longitude = 0; self.latitude = 0;
class MessageModel: def __init__(self, message, duration, creation_date, message_category): # We will automatically generate the new id self.id = 0 self.message = message self.duration = duration self.creation_date = creation_date self.message_category = message_category self.printed_times = 0 self.printed_once = FalseUpdate model script to support task database schemaclass MessageModel: def __init__(self, message, duration, creation_date, message_category): # We will automatically generate the new id self.id = 0 self.message = message self.duration = duration self.creation_date = creation_date self.message_category = message_category self.printed_times = 0 self.printed_once = False class AccountModel: def __init__(self, account_type, account_number, name, first_name, address, birthdate): # We will automatically generate the new id self.id = 0 self.type = account_type self.number = account_number self.name = name self.first_name = first_name self.address = address self.birthdate = birthdate #We will automatically generate next 2 parameters based on client address. self.longitude = 0; self.latitude = 0;
<commit_before>class MessageModel: def __init__(self, message, duration, creation_date, message_category): # We will automatically generate the new id self.id = 0 self.message = message self.duration = duration self.creation_date = creation_date self.message_category = message_category self.printed_times = 0 self.printed_once = False<commit_msg>Update model script to support task database schema<commit_after>class MessageModel: def __init__(self, message, duration, creation_date, message_category): # We will automatically generate the new id self.id = 0 self.message = message self.duration = duration self.creation_date = creation_date self.message_category = message_category self.printed_times = 0 self.printed_once = False class AccountModel: def __init__(self, account_type, account_number, name, first_name, address, birthdate): # We will automatically generate the new id self.id = 0 self.type = account_type self.number = account_number self.name = name self.first_name = first_name self.address = address self.birthdate = birthdate #We will automatically generate next 2 parameters based on client address. self.longitude = 0; self.latitude = 0;
e0d8099e57fb890649490b9e9bb201b98b041212
libcloud/__init__.py
libcloud/__init__.py
# Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # libcloud.org licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ libcloud provides a unified interface to the cloud computing resources. """
# Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # libcloud.org licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ libcloud provides a unified interface to the cloud computing resources. """ __version__ = "0.1.1-dev"
Add version string to libcloud
Add version string to libcloud git-svn-id: 353d90d4d8d13dcb4e0402680a9155a727f61a5a@895867 13f79535-47bb-0310-9956-ffa450edef68
Python
apache-2.0
cloudkick/libcloud,cloudkick/libcloud
# Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # libcloud.org licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ libcloud provides a unified interface to the cloud computing resources. """ Add version string to libcloud git-svn-id: 353d90d4d8d13dcb4e0402680a9155a727f61a5a@895867 13f79535-47bb-0310-9956-ffa450edef68
# Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # libcloud.org licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ libcloud provides a unified interface to the cloud computing resources. """ __version__ = "0.1.1-dev"
<commit_before># Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # libcloud.org licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ libcloud provides a unified interface to the cloud computing resources. """ <commit_msg>Add version string to libcloud git-svn-id: 353d90d4d8d13dcb4e0402680a9155a727f61a5a@895867 13f79535-47bb-0310-9956-ffa450edef68<commit_after>
# Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # libcloud.org licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ libcloud provides a unified interface to the cloud computing resources. """ __version__ = "0.1.1-dev"
# Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # libcloud.org licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ libcloud provides a unified interface to the cloud computing resources. """ Add version string to libcloud git-svn-id: 353d90d4d8d13dcb4e0402680a9155a727f61a5a@895867 13f79535-47bb-0310-9956-ffa450edef68# Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # libcloud.org licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ libcloud provides a unified interface to the cloud computing resources. """ __version__ = "0.1.1-dev"
<commit_before># Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # libcloud.org licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ libcloud provides a unified interface to the cloud computing resources. """ <commit_msg>Add version string to libcloud git-svn-id: 353d90d4d8d13dcb4e0402680a9155a727f61a5a@895867 13f79535-47bb-0310-9956-ffa450edef68<commit_after># Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # libcloud.org licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ libcloud provides a unified interface to the cloud computing resources. """ __version__ = "0.1.1-dev"
46a0caa1bc162d11b26a996379170b2fc49f2940
mcbench/client.py
mcbench/client.py
import collections import redis BENCHMARK_FIELDS = [ 'author', 'author_url', 'date_submitted', 'date_updated', 'name', 'summary', 'tags', 'title', 'url' ] Benchmark = collections.namedtuple('Benchmark', ' '.join(BENCHMARK_FIELDS)) class McBenchClient(object): def __init__(self, redis): self.redis = redis def get_benchmark_by_id(self, benchmark_id): return Benchmark(**self.redis.hgetall('benchmark:%s' % benchmark_id)) def get_benchmark_by_name(self, name): benchmark_id = self.redis.get('benchmark:%s:id', name) return self.get_benchmark_by_id(benchmark_id) def insert_benchmark(self, benchmark): benchmark_id = self.redis.incr('global:next_benchmark_id') self.redis.set('benchmark:%s:id' % benchmark.name, benchmark_id) self.redis.hmset('benchmark:%s' % benchmark_id, benchmark._asdict()) def from_redis_url(redis_url): return McBenchClient(redis.from_url(redis_url))
import redis class Benchmark(object): def __init__(self, author, author_url, date_submitted, date_updated, name, summary, tags, title, url): self.author = author self.author_url = author_url self.date_submitted = date_submitted self.date_updated = date_updated self.name = name self.summary = summary self.tags = tags self.title = title self.url = url def __repr__(self): return '<Benchmark: %s>' % self.name class BenchmarkDoesNotExist(Exception): pass class BenchmarkAlreadyExists(Exception): pass class McBenchClient(object): def __init__(self, redis): self.redis = redis def get_benchmark_by_id(self, benchmark_id): data = self.redis.hgetall('benchmark:%s' % benchmark_id) if not data: raise BenchmarkDoesNotExist return Benchmark(**data) def get_benchmark_by_name(self, name): benchmark_id = self.redis.get('name:%s:id' % name) if benchmark_id is None: raise BenchmarkDoesNotExist return self.get_benchmark_by_id(benchmark_id) def get_all_benchmarks(self): return [self.get_benchmark_by_id(key[len('benchmark:'):]) for key in self.redis.keys('benchmark:*')] def insert_benchmark(self, benchmark): benchmark_id = self.redis.get('name:%s:id' % benchmark.name) if benchmark_id is not None: raise BenchmarkAlreadyExists benchmark_id = self.redis.incr('global:next_benchmark_id') self.redis.set('name:%s:id' % benchmark.name, benchmark_id) self.redis.hmset('benchmark:%s' % benchmark_id, vars(benchmark)) def from_redis_url(redis_url): return McBenchClient(redis.from_url(redis_url))
Make Benchmark a class, not a namedtuple.
Make Benchmark a class, not a namedtuple.
Python
mit
isbadawi/mcbench,isbadawi/mcbench
import collections import redis BENCHMARK_FIELDS = [ 'author', 'author_url', 'date_submitted', 'date_updated', 'name', 'summary', 'tags', 'title', 'url' ] Benchmark = collections.namedtuple('Benchmark', ' '.join(BENCHMARK_FIELDS)) class McBenchClient(object): def __init__(self, redis): self.redis = redis def get_benchmark_by_id(self, benchmark_id): return Benchmark(**self.redis.hgetall('benchmark:%s' % benchmark_id)) def get_benchmark_by_name(self, name): benchmark_id = self.redis.get('benchmark:%s:id', name) return self.get_benchmark_by_id(benchmark_id) def insert_benchmark(self, benchmark): benchmark_id = self.redis.incr('global:next_benchmark_id') self.redis.set('benchmark:%s:id' % benchmark.name, benchmark_id) self.redis.hmset('benchmark:%s' % benchmark_id, benchmark._asdict()) def from_redis_url(redis_url): return McBenchClient(redis.from_url(redis_url)) Make Benchmark a class, not a namedtuple.
import redis class Benchmark(object): def __init__(self, author, author_url, date_submitted, date_updated, name, summary, tags, title, url): self.author = author self.author_url = author_url self.date_submitted = date_submitted self.date_updated = date_updated self.name = name self.summary = summary self.tags = tags self.title = title self.url = url def __repr__(self): return '<Benchmark: %s>' % self.name class BenchmarkDoesNotExist(Exception): pass class BenchmarkAlreadyExists(Exception): pass class McBenchClient(object): def __init__(self, redis): self.redis = redis def get_benchmark_by_id(self, benchmark_id): data = self.redis.hgetall('benchmark:%s' % benchmark_id) if not data: raise BenchmarkDoesNotExist return Benchmark(**data) def get_benchmark_by_name(self, name): benchmark_id = self.redis.get('name:%s:id' % name) if benchmark_id is None: raise BenchmarkDoesNotExist return self.get_benchmark_by_id(benchmark_id) def get_all_benchmarks(self): return [self.get_benchmark_by_id(key[len('benchmark:'):]) for key in self.redis.keys('benchmark:*')] def insert_benchmark(self, benchmark): benchmark_id = self.redis.get('name:%s:id' % benchmark.name) if benchmark_id is not None: raise BenchmarkAlreadyExists benchmark_id = self.redis.incr('global:next_benchmark_id') self.redis.set('name:%s:id' % benchmark.name, benchmark_id) self.redis.hmset('benchmark:%s' % benchmark_id, vars(benchmark)) def from_redis_url(redis_url): return McBenchClient(redis.from_url(redis_url))
<commit_before>import collections import redis BENCHMARK_FIELDS = [ 'author', 'author_url', 'date_submitted', 'date_updated', 'name', 'summary', 'tags', 'title', 'url' ] Benchmark = collections.namedtuple('Benchmark', ' '.join(BENCHMARK_FIELDS)) class McBenchClient(object): def __init__(self, redis): self.redis = redis def get_benchmark_by_id(self, benchmark_id): return Benchmark(**self.redis.hgetall('benchmark:%s' % benchmark_id)) def get_benchmark_by_name(self, name): benchmark_id = self.redis.get('benchmark:%s:id', name) return self.get_benchmark_by_id(benchmark_id) def insert_benchmark(self, benchmark): benchmark_id = self.redis.incr('global:next_benchmark_id') self.redis.set('benchmark:%s:id' % benchmark.name, benchmark_id) self.redis.hmset('benchmark:%s' % benchmark_id, benchmark._asdict()) def from_redis_url(redis_url): return McBenchClient(redis.from_url(redis_url)) <commit_msg>Make Benchmark a class, not a namedtuple.<commit_after>
import redis class Benchmark(object): def __init__(self, author, author_url, date_submitted, date_updated, name, summary, tags, title, url): self.author = author self.author_url = author_url self.date_submitted = date_submitted self.date_updated = date_updated self.name = name self.summary = summary self.tags = tags self.title = title self.url = url def __repr__(self): return '<Benchmark: %s>' % self.name class BenchmarkDoesNotExist(Exception): pass class BenchmarkAlreadyExists(Exception): pass class McBenchClient(object): def __init__(self, redis): self.redis = redis def get_benchmark_by_id(self, benchmark_id): data = self.redis.hgetall('benchmark:%s' % benchmark_id) if not data: raise BenchmarkDoesNotExist return Benchmark(**data) def get_benchmark_by_name(self, name): benchmark_id = self.redis.get('name:%s:id' % name) if benchmark_id is None: raise BenchmarkDoesNotExist return self.get_benchmark_by_id(benchmark_id) def get_all_benchmarks(self): return [self.get_benchmark_by_id(key[len('benchmark:'):]) for key in self.redis.keys('benchmark:*')] def insert_benchmark(self, benchmark): benchmark_id = self.redis.get('name:%s:id' % benchmark.name) if benchmark_id is not None: raise BenchmarkAlreadyExists benchmark_id = self.redis.incr('global:next_benchmark_id') self.redis.set('name:%s:id' % benchmark.name, benchmark_id) self.redis.hmset('benchmark:%s' % benchmark_id, vars(benchmark)) def from_redis_url(redis_url): return McBenchClient(redis.from_url(redis_url))
import collections import redis BENCHMARK_FIELDS = [ 'author', 'author_url', 'date_submitted', 'date_updated', 'name', 'summary', 'tags', 'title', 'url' ] Benchmark = collections.namedtuple('Benchmark', ' '.join(BENCHMARK_FIELDS)) class McBenchClient(object): def __init__(self, redis): self.redis = redis def get_benchmark_by_id(self, benchmark_id): return Benchmark(**self.redis.hgetall('benchmark:%s' % benchmark_id)) def get_benchmark_by_name(self, name): benchmark_id = self.redis.get('benchmark:%s:id', name) return self.get_benchmark_by_id(benchmark_id) def insert_benchmark(self, benchmark): benchmark_id = self.redis.incr('global:next_benchmark_id') self.redis.set('benchmark:%s:id' % benchmark.name, benchmark_id) self.redis.hmset('benchmark:%s' % benchmark_id, benchmark._asdict()) def from_redis_url(redis_url): return McBenchClient(redis.from_url(redis_url)) Make Benchmark a class, not a namedtuple.import redis class Benchmark(object): def __init__(self, author, author_url, date_submitted, date_updated, name, summary, tags, title, url): self.author = author self.author_url = author_url self.date_submitted = date_submitted self.date_updated = date_updated self.name = name self.summary = summary self.tags = tags self.title = title self.url = url def __repr__(self): return '<Benchmark: %s>' % self.name class BenchmarkDoesNotExist(Exception): pass class BenchmarkAlreadyExists(Exception): pass class McBenchClient(object): def __init__(self, redis): self.redis = redis def get_benchmark_by_id(self, benchmark_id): data = self.redis.hgetall('benchmark:%s' % benchmark_id) if not data: raise BenchmarkDoesNotExist return Benchmark(**data) def get_benchmark_by_name(self, name): benchmark_id = self.redis.get('name:%s:id' % name) if benchmark_id is None: raise BenchmarkDoesNotExist return self.get_benchmark_by_id(benchmark_id) def get_all_benchmarks(self): return [self.get_benchmark_by_id(key[len('benchmark:'):]) for key in self.redis.keys('benchmark:*')] def insert_benchmark(self, benchmark): benchmark_id = self.redis.get('name:%s:id' % benchmark.name) if benchmark_id is not None: raise BenchmarkAlreadyExists benchmark_id = self.redis.incr('global:next_benchmark_id') self.redis.set('name:%s:id' % benchmark.name, benchmark_id) self.redis.hmset('benchmark:%s' % benchmark_id, vars(benchmark)) def from_redis_url(redis_url): return McBenchClient(redis.from_url(redis_url))
<commit_before>import collections import redis BENCHMARK_FIELDS = [ 'author', 'author_url', 'date_submitted', 'date_updated', 'name', 'summary', 'tags', 'title', 'url' ] Benchmark = collections.namedtuple('Benchmark', ' '.join(BENCHMARK_FIELDS)) class McBenchClient(object): def __init__(self, redis): self.redis = redis def get_benchmark_by_id(self, benchmark_id): return Benchmark(**self.redis.hgetall('benchmark:%s' % benchmark_id)) def get_benchmark_by_name(self, name): benchmark_id = self.redis.get('benchmark:%s:id', name) return self.get_benchmark_by_id(benchmark_id) def insert_benchmark(self, benchmark): benchmark_id = self.redis.incr('global:next_benchmark_id') self.redis.set('benchmark:%s:id' % benchmark.name, benchmark_id) self.redis.hmset('benchmark:%s' % benchmark_id, benchmark._asdict()) def from_redis_url(redis_url): return McBenchClient(redis.from_url(redis_url)) <commit_msg>Make Benchmark a class, not a namedtuple.<commit_after>import redis class Benchmark(object): def __init__(self, author, author_url, date_submitted, date_updated, name, summary, tags, title, url): self.author = author self.author_url = author_url self.date_submitted = date_submitted self.date_updated = date_updated self.name = name self.summary = summary self.tags = tags self.title = title self.url = url def __repr__(self): return '<Benchmark: %s>' % self.name class BenchmarkDoesNotExist(Exception): pass class BenchmarkAlreadyExists(Exception): pass class McBenchClient(object): def __init__(self, redis): self.redis = redis def get_benchmark_by_id(self, benchmark_id): data = self.redis.hgetall('benchmark:%s' % benchmark_id) if not data: raise BenchmarkDoesNotExist return Benchmark(**data) def get_benchmark_by_name(self, name): benchmark_id = self.redis.get('name:%s:id' % name) if benchmark_id is None: raise BenchmarkDoesNotExist return self.get_benchmark_by_id(benchmark_id) def get_all_benchmarks(self): return [self.get_benchmark_by_id(key[len('benchmark:'):]) for key in self.redis.keys('benchmark:*')] def insert_benchmark(self, benchmark): benchmark_id = self.redis.get('name:%s:id' % benchmark.name) if benchmark_id is not None: raise BenchmarkAlreadyExists benchmark_id = self.redis.incr('global:next_benchmark_id') self.redis.set('name:%s:id' % benchmark.name, benchmark_id) self.redis.hmset('benchmark:%s' % benchmark_id, vars(benchmark)) def from_redis_url(redis_url): return McBenchClient(redis.from_url(redis_url))
1b9c4935b2edf6601c2d75d8a2d318266de2d456
circuits/tools/__init__.py
circuits/tools/__init__.py
# Module: __init__ # Date: 8th November 2008 # Author: James Mills, prologic at shortcircuit dot net dot au """Circuits Tools circuits.tools contains a standard set of tools for circuits. These tools are installed as executables with a prefix of "circuits." """ try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, children)) children = list(x.components) i = 0 else: if stack: i, children = stack.pop() d -= 1 else: done = True return s.getvalue()
# Module: __init__ # Date: 8th November 2008 # Author: James Mills, prologic at shortcircuit dot net dot au """Circuits Tools circuits.tools contains a standard set of tools for circuits. These tools are installed as executables with a prefix of "circuits." """ try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, d, children)) children = list(x.components) i = 0 else: if stack: i, d, children = stack.pop() else: done = True return s.getvalue()
Store the depth (d) on the stack and restore when backtracking
tools: Store the depth (d) on the stack and restore when backtracking
Python
mit
treemo/circuits,treemo/circuits,eriol/circuits,treemo/circuits,eriol/circuits,nizox/circuits,eriol/circuits
# Module: __init__ # Date: 8th November 2008 # Author: James Mills, prologic at shortcircuit dot net dot au """Circuits Tools circuits.tools contains a standard set of tools for circuits. These tools are installed as executables with a prefix of "circuits." """ try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, children)) children = list(x.components) i = 0 else: if stack: i, children = stack.pop() d -= 1 else: done = True return s.getvalue() tools: Store the depth (d) on the stack and restore when backtracking
# Module: __init__ # Date: 8th November 2008 # Author: James Mills, prologic at shortcircuit dot net dot au """Circuits Tools circuits.tools contains a standard set of tools for circuits. These tools are installed as executables with a prefix of "circuits." """ try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, d, children)) children = list(x.components) i = 0 else: if stack: i, d, children = stack.pop() else: done = True return s.getvalue()
<commit_before># Module: __init__ # Date: 8th November 2008 # Author: James Mills, prologic at shortcircuit dot net dot au """Circuits Tools circuits.tools contains a standard set of tools for circuits. These tools are installed as executables with a prefix of "circuits." """ try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, children)) children = list(x.components) i = 0 else: if stack: i, children = stack.pop() d -= 1 else: done = True return s.getvalue() <commit_msg>tools: Store the depth (d) on the stack and restore when backtracking<commit_after>
# Module: __init__ # Date: 8th November 2008 # Author: James Mills, prologic at shortcircuit dot net dot au """Circuits Tools circuits.tools contains a standard set of tools for circuits. These tools are installed as executables with a prefix of "circuits." """ try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, d, children)) children = list(x.components) i = 0 else: if stack: i, d, children = stack.pop() else: done = True return s.getvalue()
# Module: __init__ # Date: 8th November 2008 # Author: James Mills, prologic at shortcircuit dot net dot au """Circuits Tools circuits.tools contains a standard set of tools for circuits. These tools are installed as executables with a prefix of "circuits." """ try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, children)) children = list(x.components) i = 0 else: if stack: i, children = stack.pop() d -= 1 else: done = True return s.getvalue() tools: Store the depth (d) on the stack and restore when backtracking# Module: __init__ # Date: 8th November 2008 # Author: James Mills, prologic at shortcircuit dot net dot au """Circuits Tools circuits.tools contains a standard set of tools for circuits. These tools are installed as executables with a prefix of "circuits." """ try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, d, children)) children = list(x.components) i = 0 else: if stack: i, d, children = stack.pop() else: done = True return s.getvalue()
<commit_before># Module: __init__ # Date: 8th November 2008 # Author: James Mills, prologic at shortcircuit dot net dot au """Circuits Tools circuits.tools contains a standard set of tools for circuits. These tools are installed as executables with a prefix of "circuits." """ try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, children)) children = list(x.components) i = 0 else: if stack: i, children = stack.pop() d -= 1 else: done = True return s.getvalue() <commit_msg>tools: Store the depth (d) on the stack and restore when backtracking<commit_after># Module: __init__ # Date: 8th November 2008 # Author: James Mills, prologic at shortcircuit dot net dot au """Circuits Tools circuits.tools contains a standard set of tools for circuits. These tools are installed as executables with a prefix of "circuits." """ try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, d, children)) children = list(x.components) i = 0 else: if stack: i, d, children = stack.pop() else: done = True return s.getvalue()
7ba2299e2d429bd873539507b3edbe3cdd3de9d6
linkatos/firebase.py
linkatos/firebase.py
import pyrebase def initialise(FB_API_KEY, project_name): config = { "apiKey": FB_API_KEY, "authDomain": "{}.firebaseapp.com".format(project_name), "databaseURL": "https://{}.firebaseio.com".format(project_name), "storageBucket": "{}.appspot.com".format(project_name), } return pyrebase.initialize_app(config) def store_url(is_yes, url, FB_USER, FB_PASS, firebase): # do nothing if it's unnecessary if not is_yes: return False # creates token every time maybe worth doing it once every 30m as they # expire every hour auth = firebase.auth() user = auth.sign_in_with_email_and_password(FB_USER, FB_PASS) db = firebase.database() data = { "url": url } db.child("users").push(data, user['idToken']) return False
import pyrebase def initialise(api_key, project_name): config = { "apiKey": api_key, "authDomain": "{}.firebaseapp.com".format(project_name), "databaseURL": "https://{}.firebaseio.com".format(project_name), "storageBucket": "{}.appspot.com".format(project_name), } return pyrebase.initialize_app(config) def store_url(is_yes, url, user, password, firebase): # do nothing if it's unnecessary if not is_yes: return False # creates token every time maybe worth doing it once every 30m as they # expire every hour auth = firebase.auth() user = auth.sign_in_with_email_and_password(user, password) db = firebase.database() data = { "url": url } db.child("users").push(data, user['idToken']) return False
Change variables to lower case
style: Change variables to lower case
Python
mit
iwi/linkatos,iwi/linkatos
import pyrebase def initialise(FB_API_KEY, project_name): config = { "apiKey": FB_API_KEY, "authDomain": "{}.firebaseapp.com".format(project_name), "databaseURL": "https://{}.firebaseio.com".format(project_name), "storageBucket": "{}.appspot.com".format(project_name), } return pyrebase.initialize_app(config) def store_url(is_yes, url, FB_USER, FB_PASS, firebase): # do nothing if it's unnecessary if not is_yes: return False # creates token every time maybe worth doing it once every 30m as they # expire every hour auth = firebase.auth() user = auth.sign_in_with_email_and_password(FB_USER, FB_PASS) db = firebase.database() data = { "url": url } db.child("users").push(data, user['idToken']) return False style: Change variables to lower case
import pyrebase def initialise(api_key, project_name): config = { "apiKey": api_key, "authDomain": "{}.firebaseapp.com".format(project_name), "databaseURL": "https://{}.firebaseio.com".format(project_name), "storageBucket": "{}.appspot.com".format(project_name), } return pyrebase.initialize_app(config) def store_url(is_yes, url, user, password, firebase): # do nothing if it's unnecessary if not is_yes: return False # creates token every time maybe worth doing it once every 30m as they # expire every hour auth = firebase.auth() user = auth.sign_in_with_email_and_password(user, password) db = firebase.database() data = { "url": url } db.child("users").push(data, user['idToken']) return False
<commit_before>import pyrebase def initialise(FB_API_KEY, project_name): config = { "apiKey": FB_API_KEY, "authDomain": "{}.firebaseapp.com".format(project_name), "databaseURL": "https://{}.firebaseio.com".format(project_name), "storageBucket": "{}.appspot.com".format(project_name), } return pyrebase.initialize_app(config) def store_url(is_yes, url, FB_USER, FB_PASS, firebase): # do nothing if it's unnecessary if not is_yes: return False # creates token every time maybe worth doing it once every 30m as they # expire every hour auth = firebase.auth() user = auth.sign_in_with_email_and_password(FB_USER, FB_PASS) db = firebase.database() data = { "url": url } db.child("users").push(data, user['idToken']) return False <commit_msg>style: Change variables to lower case<commit_after>
import pyrebase def initialise(api_key, project_name): config = { "apiKey": api_key, "authDomain": "{}.firebaseapp.com".format(project_name), "databaseURL": "https://{}.firebaseio.com".format(project_name), "storageBucket": "{}.appspot.com".format(project_name), } return pyrebase.initialize_app(config) def store_url(is_yes, url, user, password, firebase): # do nothing if it's unnecessary if not is_yes: return False # creates token every time maybe worth doing it once every 30m as they # expire every hour auth = firebase.auth() user = auth.sign_in_with_email_and_password(user, password) db = firebase.database() data = { "url": url } db.child("users").push(data, user['idToken']) return False
import pyrebase def initialise(FB_API_KEY, project_name): config = { "apiKey": FB_API_KEY, "authDomain": "{}.firebaseapp.com".format(project_name), "databaseURL": "https://{}.firebaseio.com".format(project_name), "storageBucket": "{}.appspot.com".format(project_name), } return pyrebase.initialize_app(config) def store_url(is_yes, url, FB_USER, FB_PASS, firebase): # do nothing if it's unnecessary if not is_yes: return False # creates token every time maybe worth doing it once every 30m as they # expire every hour auth = firebase.auth() user = auth.sign_in_with_email_and_password(FB_USER, FB_PASS) db = firebase.database() data = { "url": url } db.child("users").push(data, user['idToken']) return False style: Change variables to lower caseimport pyrebase def initialise(api_key, project_name): config = { "apiKey": api_key, "authDomain": "{}.firebaseapp.com".format(project_name), "databaseURL": "https://{}.firebaseio.com".format(project_name), "storageBucket": "{}.appspot.com".format(project_name), } return pyrebase.initialize_app(config) def store_url(is_yes, url, user, password, firebase): # do nothing if it's unnecessary if not is_yes: return False # creates token every time maybe worth doing it once every 30m as they # expire every hour auth = firebase.auth() user = auth.sign_in_with_email_and_password(user, password) db = firebase.database() data = { "url": url } db.child("users").push(data, user['idToken']) return False
<commit_before>import pyrebase def initialise(FB_API_KEY, project_name): config = { "apiKey": FB_API_KEY, "authDomain": "{}.firebaseapp.com".format(project_name), "databaseURL": "https://{}.firebaseio.com".format(project_name), "storageBucket": "{}.appspot.com".format(project_name), } return pyrebase.initialize_app(config) def store_url(is_yes, url, FB_USER, FB_PASS, firebase): # do nothing if it's unnecessary if not is_yes: return False # creates token every time maybe worth doing it once every 30m as they # expire every hour auth = firebase.auth() user = auth.sign_in_with_email_and_password(FB_USER, FB_PASS) db = firebase.database() data = { "url": url } db.child("users").push(data, user['idToken']) return False <commit_msg>style: Change variables to lower case<commit_after>import pyrebase def initialise(api_key, project_name): config = { "apiKey": api_key, "authDomain": "{}.firebaseapp.com".format(project_name), "databaseURL": "https://{}.firebaseio.com".format(project_name), "storageBucket": "{}.appspot.com".format(project_name), } return pyrebase.initialize_app(config) def store_url(is_yes, url, user, password, firebase): # do nothing if it's unnecessary if not is_yes: return False # creates token every time maybe worth doing it once every 30m as they # expire every hour auth = firebase.auth() user = auth.sign_in_with_email_and_password(user, password) db = firebase.database() data = { "url": url } db.child("users").push(data, user['idToken']) return False
d320de7a66472036f2504f8b935747ff4a1e4e49
barf/setup.py
barf/setup.py
#! /usr/bin/env python from setuptools import setup from setuptools import find_packages setup( author = 'Christian Heitman', author_email = 'cnheitman@fundacionsadosky.org.ar', description = 'A multiplatform open source Binary Analysis and Reverse engineering Framework', install_requires = [ 'capstone', 'networkx', 'pefile', 'pybfd', 'pydot', 'pygments', 'pyparsing', 'sphinx', ], license = 'BSD 2-Clause', name = 'barf', packages = find_packages(), url = 'http://github.com/programa-stic/barf-project', scripts = [ 'tools/gadgets/BARFgadgets' ], version = '0.2', zip_safe = False )
#! /usr/bin/env python from setuptools import setup from setuptools import find_packages # https://github.com/aquynh/capstone/issues/583 def fix_setuptools(): """Work around bugs in setuptools. Some versions of setuptools are broken and raise SandboxViolation for normal operations in a virtualenv. We therefore disable the sandbox to avoid these issues. """ try: from setuptools.sandbox import DirectorySandbox def violation(operation, *args, **_): print "SandboxViolation: %s" % (args,) DirectorySandbox._violation = violation except ImportError: pass # Fix bugs in setuptools. fix_setuptools() setup( author = 'Christian Heitman', author_email = 'cnheitman@fundacionsadosky.org.ar', description = 'A multiplatform open source Binary Analysis and Reverse engineering Framework', install_requires = [ 'capstone', 'networkx', 'pefile', 'pybfd', 'pydot', 'pygments', 'pyparsing', 'sphinx', ], license = 'BSD 2-Clause', name = 'barf', packages = find_packages(), url = 'http://github.com/programa-stic/barf-project', scripts = [ 'tools/gadgets/BARFgadgets' ], version = '0.2', zip_safe = False )
Fix Capstone installation error on virtualenvs
Fix Capstone installation error on virtualenvs
Python
bsd-2-clause
chubbymaggie/barf-project,programa-stic/barf-project,cnheitman/barf-project,cnheitman/barf-project,chubbymaggie/barf-project,chubbymaggie/barf-project,programa-stic/barf-project,cnheitman/barf-project
#! /usr/bin/env python from setuptools import setup from setuptools import find_packages setup( author = 'Christian Heitman', author_email = 'cnheitman@fundacionsadosky.org.ar', description = 'A multiplatform open source Binary Analysis and Reverse engineering Framework', install_requires = [ 'capstone', 'networkx', 'pefile', 'pybfd', 'pydot', 'pygments', 'pyparsing', 'sphinx', ], license = 'BSD 2-Clause', name = 'barf', packages = find_packages(), url = 'http://github.com/programa-stic/barf-project', scripts = [ 'tools/gadgets/BARFgadgets' ], version = '0.2', zip_safe = False ) Fix Capstone installation error on virtualenvs
#! /usr/bin/env python from setuptools import setup from setuptools import find_packages # https://github.com/aquynh/capstone/issues/583 def fix_setuptools(): """Work around bugs in setuptools. Some versions of setuptools are broken and raise SandboxViolation for normal operations in a virtualenv. We therefore disable the sandbox to avoid these issues. """ try: from setuptools.sandbox import DirectorySandbox def violation(operation, *args, **_): print "SandboxViolation: %s" % (args,) DirectorySandbox._violation = violation except ImportError: pass # Fix bugs in setuptools. fix_setuptools() setup( author = 'Christian Heitman', author_email = 'cnheitman@fundacionsadosky.org.ar', description = 'A multiplatform open source Binary Analysis and Reverse engineering Framework', install_requires = [ 'capstone', 'networkx', 'pefile', 'pybfd', 'pydot', 'pygments', 'pyparsing', 'sphinx', ], license = 'BSD 2-Clause', name = 'barf', packages = find_packages(), url = 'http://github.com/programa-stic/barf-project', scripts = [ 'tools/gadgets/BARFgadgets' ], version = '0.2', zip_safe = False )
<commit_before>#! /usr/bin/env python from setuptools import setup from setuptools import find_packages setup( author = 'Christian Heitman', author_email = 'cnheitman@fundacionsadosky.org.ar', description = 'A multiplatform open source Binary Analysis and Reverse engineering Framework', install_requires = [ 'capstone', 'networkx', 'pefile', 'pybfd', 'pydot', 'pygments', 'pyparsing', 'sphinx', ], license = 'BSD 2-Clause', name = 'barf', packages = find_packages(), url = 'http://github.com/programa-stic/barf-project', scripts = [ 'tools/gadgets/BARFgadgets' ], version = '0.2', zip_safe = False ) <commit_msg>Fix Capstone installation error on virtualenvs<commit_after>
#! /usr/bin/env python from setuptools import setup from setuptools import find_packages # https://github.com/aquynh/capstone/issues/583 def fix_setuptools(): """Work around bugs in setuptools. Some versions of setuptools are broken and raise SandboxViolation for normal operations in a virtualenv. We therefore disable the sandbox to avoid these issues. """ try: from setuptools.sandbox import DirectorySandbox def violation(operation, *args, **_): print "SandboxViolation: %s" % (args,) DirectorySandbox._violation = violation except ImportError: pass # Fix bugs in setuptools. fix_setuptools() setup( author = 'Christian Heitman', author_email = 'cnheitman@fundacionsadosky.org.ar', description = 'A multiplatform open source Binary Analysis and Reverse engineering Framework', install_requires = [ 'capstone', 'networkx', 'pefile', 'pybfd', 'pydot', 'pygments', 'pyparsing', 'sphinx', ], license = 'BSD 2-Clause', name = 'barf', packages = find_packages(), url = 'http://github.com/programa-stic/barf-project', scripts = [ 'tools/gadgets/BARFgadgets' ], version = '0.2', zip_safe = False )
#! /usr/bin/env python from setuptools import setup from setuptools import find_packages setup( author = 'Christian Heitman', author_email = 'cnheitman@fundacionsadosky.org.ar', description = 'A multiplatform open source Binary Analysis and Reverse engineering Framework', install_requires = [ 'capstone', 'networkx', 'pefile', 'pybfd', 'pydot', 'pygments', 'pyparsing', 'sphinx', ], license = 'BSD 2-Clause', name = 'barf', packages = find_packages(), url = 'http://github.com/programa-stic/barf-project', scripts = [ 'tools/gadgets/BARFgadgets' ], version = '0.2', zip_safe = False ) Fix Capstone installation error on virtualenvs#! /usr/bin/env python from setuptools import setup from setuptools import find_packages # https://github.com/aquynh/capstone/issues/583 def fix_setuptools(): """Work around bugs in setuptools. Some versions of setuptools are broken and raise SandboxViolation for normal operations in a virtualenv. We therefore disable the sandbox to avoid these issues. """ try: from setuptools.sandbox import DirectorySandbox def violation(operation, *args, **_): print "SandboxViolation: %s" % (args,) DirectorySandbox._violation = violation except ImportError: pass # Fix bugs in setuptools. fix_setuptools() setup( author = 'Christian Heitman', author_email = 'cnheitman@fundacionsadosky.org.ar', description = 'A multiplatform open source Binary Analysis and Reverse engineering Framework', install_requires = [ 'capstone', 'networkx', 'pefile', 'pybfd', 'pydot', 'pygments', 'pyparsing', 'sphinx', ], license = 'BSD 2-Clause', name = 'barf', packages = find_packages(), url = 'http://github.com/programa-stic/barf-project', scripts = [ 'tools/gadgets/BARFgadgets' ], version = '0.2', zip_safe = False )
<commit_before>#! /usr/bin/env python from setuptools import setup from setuptools import find_packages setup( author = 'Christian Heitman', author_email = 'cnheitman@fundacionsadosky.org.ar', description = 'A multiplatform open source Binary Analysis and Reverse engineering Framework', install_requires = [ 'capstone', 'networkx', 'pefile', 'pybfd', 'pydot', 'pygments', 'pyparsing', 'sphinx', ], license = 'BSD 2-Clause', name = 'barf', packages = find_packages(), url = 'http://github.com/programa-stic/barf-project', scripts = [ 'tools/gadgets/BARFgadgets' ], version = '0.2', zip_safe = False ) <commit_msg>Fix Capstone installation error on virtualenvs<commit_after>#! /usr/bin/env python from setuptools import setup from setuptools import find_packages # https://github.com/aquynh/capstone/issues/583 def fix_setuptools(): """Work around bugs in setuptools. Some versions of setuptools are broken and raise SandboxViolation for normal operations in a virtualenv. We therefore disable the sandbox to avoid these issues. """ try: from setuptools.sandbox import DirectorySandbox def violation(operation, *args, **_): print "SandboxViolation: %s" % (args,) DirectorySandbox._violation = violation except ImportError: pass # Fix bugs in setuptools. fix_setuptools() setup( author = 'Christian Heitman', author_email = 'cnheitman@fundacionsadosky.org.ar', description = 'A multiplatform open source Binary Analysis and Reverse engineering Framework', install_requires = [ 'capstone', 'networkx', 'pefile', 'pybfd', 'pydot', 'pygments', 'pyparsing', 'sphinx', ], license = 'BSD 2-Clause', name = 'barf', packages = find_packages(), url = 'http://github.com/programa-stic/barf-project', scripts = [ 'tools/gadgets/BARFgadgets' ], version = '0.2', zip_safe = False )
56edfe1bacff53eec22b05d43f32063f83f43ea5
studies/helpers.py
studies/helpers.py
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL, OSF_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL context['osf_url'] = OSF_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
Add OSF_URL to send_email helper.
Add OSF_URL to send_email helper.
Python
apache-2.0
CenterForOpenScience/lookit-api,pattisdr/lookit-api,pattisdr/lookit-api,pattisdr/lookit-api,CenterForOpenScience/lookit-api,CenterForOpenScience/lookit-api
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send() Add OSF_URL to send_email helper.
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL, OSF_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL context['osf_url'] = OSF_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
<commit_before>from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send() <commit_msg>Add OSF_URL to send_email helper.<commit_after>
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL, OSF_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL context['osf_url'] = OSF_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send() Add OSF_URL to send_email helper.from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL, OSF_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL context['osf_url'] = OSF_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
<commit_before>from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send() <commit_msg>Add OSF_URL to send_email helper.<commit_after>from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL, OSF_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL context['osf_url'] = OSF_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
6fcf03532dcc549a3a95390b7c999482a64fc6c6
tests/unit/utils/test_pycrypto.py
tests/unit/utils/test_pycrypto.py
# -*- coding: utf-8 -*- # Import python libs from __future__ import absolute_import, print_function, unicode_literals import logging import re # Import Salt Libs import salt.utils.pycrypto import salt.utils.platform # Import Salt Testing Libs from tests.support.unit import TestCase, skipIf log = logging.getLogger(__name__) class PycryptoTestCase(TestCase): ''' TestCase for salt.utils.pycrypto module ''' @skipIf(salt.utils.platform.is_windows(), 'No crypto module for Windows') def test_gen_hash(self): ''' Test gen_hash ''' passwd = 'test_password' id = '$' if salt.utils.platform.is_darwin(): id = '' ret = salt.utils.pycrypto.gen_hash(password=passwd) self.assertTrue(ret.startswith('$6{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='md5') self.assertTrue(ret.startswith('$1{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='sha256') self.assertTrue(ret.startswith('$5{0}'.format(id))) def test_secure_password(self): ''' test secure_password ''' ret = salt.utils.pycrypto.secure_password() check = re.compile(r'[!@#$%^&*()_=+]') assert check.search(ret) is None assert ret
# -*- coding: utf-8 -*- # Import python libs from __future__ import absolute_import, print_function, unicode_literals import logging import re # Import Salt Libs import salt.utils.pycrypto import salt.utils.platform # Import Salt Testing Libs from tests.support.unit import TestCase, skipIf log = logging.getLogger(__name__) class PycryptoTestCase(TestCase): ''' TestCase for salt.utils.pycrypto module ''' # The crypt module is only available on Unix systems # https://docs.python.org/dev/library/crypt.html @skipIf(not salt.utils.pycrypto.HAS_CRYPT, 'crypt module not available') def test_gen_hash(self): ''' Test gen_hash ''' passwd = 'test_password' id = '$' if salt.utils.platform.is_darwin(): id = '' ret = salt.utils.pycrypto.gen_hash(password=passwd) self.assertTrue(ret.startswith('$6{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='md5') self.assertTrue(ret.startswith('$1{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='sha256') self.assertTrue(ret.startswith('$5{0}'.format(id))) def test_secure_password(self): ''' test secure_password ''' ret = salt.utils.pycrypto.secure_password() check = re.compile(r'[!@#$%^&*()_=+]') assert check.search(ret) is None assert ret
Make the skip apply to any system missing crypt
Make the skip apply to any system missing crypt
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
# -*- coding: utf-8 -*- # Import python libs from __future__ import absolute_import, print_function, unicode_literals import logging import re # Import Salt Libs import salt.utils.pycrypto import salt.utils.platform # Import Salt Testing Libs from tests.support.unit import TestCase, skipIf log = logging.getLogger(__name__) class PycryptoTestCase(TestCase): ''' TestCase for salt.utils.pycrypto module ''' @skipIf(salt.utils.platform.is_windows(), 'No crypto module for Windows') def test_gen_hash(self): ''' Test gen_hash ''' passwd = 'test_password' id = '$' if salt.utils.platform.is_darwin(): id = '' ret = salt.utils.pycrypto.gen_hash(password=passwd) self.assertTrue(ret.startswith('$6{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='md5') self.assertTrue(ret.startswith('$1{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='sha256') self.assertTrue(ret.startswith('$5{0}'.format(id))) def test_secure_password(self): ''' test secure_password ''' ret = salt.utils.pycrypto.secure_password() check = re.compile(r'[!@#$%^&*()_=+]') assert check.search(ret) is None assert ret Make the skip apply to any system missing crypt
# -*- coding: utf-8 -*- # Import python libs from __future__ import absolute_import, print_function, unicode_literals import logging import re # Import Salt Libs import salt.utils.pycrypto import salt.utils.platform # Import Salt Testing Libs from tests.support.unit import TestCase, skipIf log = logging.getLogger(__name__) class PycryptoTestCase(TestCase): ''' TestCase for salt.utils.pycrypto module ''' # The crypt module is only available on Unix systems # https://docs.python.org/dev/library/crypt.html @skipIf(not salt.utils.pycrypto.HAS_CRYPT, 'crypt module not available') def test_gen_hash(self): ''' Test gen_hash ''' passwd = 'test_password' id = '$' if salt.utils.platform.is_darwin(): id = '' ret = salt.utils.pycrypto.gen_hash(password=passwd) self.assertTrue(ret.startswith('$6{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='md5') self.assertTrue(ret.startswith('$1{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='sha256') self.assertTrue(ret.startswith('$5{0}'.format(id))) def test_secure_password(self): ''' test secure_password ''' ret = salt.utils.pycrypto.secure_password() check = re.compile(r'[!@#$%^&*()_=+]') assert check.search(ret) is None assert ret
<commit_before># -*- coding: utf-8 -*- # Import python libs from __future__ import absolute_import, print_function, unicode_literals import logging import re # Import Salt Libs import salt.utils.pycrypto import salt.utils.platform # Import Salt Testing Libs from tests.support.unit import TestCase, skipIf log = logging.getLogger(__name__) class PycryptoTestCase(TestCase): ''' TestCase for salt.utils.pycrypto module ''' @skipIf(salt.utils.platform.is_windows(), 'No crypto module for Windows') def test_gen_hash(self): ''' Test gen_hash ''' passwd = 'test_password' id = '$' if salt.utils.platform.is_darwin(): id = '' ret = salt.utils.pycrypto.gen_hash(password=passwd) self.assertTrue(ret.startswith('$6{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='md5') self.assertTrue(ret.startswith('$1{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='sha256') self.assertTrue(ret.startswith('$5{0}'.format(id))) def test_secure_password(self): ''' test secure_password ''' ret = salt.utils.pycrypto.secure_password() check = re.compile(r'[!@#$%^&*()_=+]') assert check.search(ret) is None assert ret <commit_msg>Make the skip apply to any system missing crypt<commit_after>
# -*- coding: utf-8 -*- # Import python libs from __future__ import absolute_import, print_function, unicode_literals import logging import re # Import Salt Libs import salt.utils.pycrypto import salt.utils.platform # Import Salt Testing Libs from tests.support.unit import TestCase, skipIf log = logging.getLogger(__name__) class PycryptoTestCase(TestCase): ''' TestCase for salt.utils.pycrypto module ''' # The crypt module is only available on Unix systems # https://docs.python.org/dev/library/crypt.html @skipIf(not salt.utils.pycrypto.HAS_CRYPT, 'crypt module not available') def test_gen_hash(self): ''' Test gen_hash ''' passwd = 'test_password' id = '$' if salt.utils.platform.is_darwin(): id = '' ret = salt.utils.pycrypto.gen_hash(password=passwd) self.assertTrue(ret.startswith('$6{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='md5') self.assertTrue(ret.startswith('$1{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='sha256') self.assertTrue(ret.startswith('$5{0}'.format(id))) def test_secure_password(self): ''' test secure_password ''' ret = salt.utils.pycrypto.secure_password() check = re.compile(r'[!@#$%^&*()_=+]') assert check.search(ret) is None assert ret
# -*- coding: utf-8 -*- # Import python libs from __future__ import absolute_import, print_function, unicode_literals import logging import re # Import Salt Libs import salt.utils.pycrypto import salt.utils.platform # Import Salt Testing Libs from tests.support.unit import TestCase, skipIf log = logging.getLogger(__name__) class PycryptoTestCase(TestCase): ''' TestCase for salt.utils.pycrypto module ''' @skipIf(salt.utils.platform.is_windows(), 'No crypto module for Windows') def test_gen_hash(self): ''' Test gen_hash ''' passwd = 'test_password' id = '$' if salt.utils.platform.is_darwin(): id = '' ret = salt.utils.pycrypto.gen_hash(password=passwd) self.assertTrue(ret.startswith('$6{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='md5') self.assertTrue(ret.startswith('$1{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='sha256') self.assertTrue(ret.startswith('$5{0}'.format(id))) def test_secure_password(self): ''' test secure_password ''' ret = salt.utils.pycrypto.secure_password() check = re.compile(r'[!@#$%^&*()_=+]') assert check.search(ret) is None assert ret Make the skip apply to any system missing crypt# -*- coding: utf-8 -*- # Import python libs from __future__ import absolute_import, print_function, unicode_literals import logging import re # Import Salt Libs import salt.utils.pycrypto import salt.utils.platform # Import Salt Testing Libs from tests.support.unit import TestCase, skipIf log = logging.getLogger(__name__) class PycryptoTestCase(TestCase): ''' TestCase for salt.utils.pycrypto module ''' # The crypt module is only available on Unix systems # https://docs.python.org/dev/library/crypt.html @skipIf(not salt.utils.pycrypto.HAS_CRYPT, 'crypt module not available') def test_gen_hash(self): ''' Test gen_hash ''' passwd = 'test_password' id = '$' if salt.utils.platform.is_darwin(): id = '' ret = salt.utils.pycrypto.gen_hash(password=passwd) self.assertTrue(ret.startswith('$6{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='md5') self.assertTrue(ret.startswith('$1{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='sha256') self.assertTrue(ret.startswith('$5{0}'.format(id))) def test_secure_password(self): ''' test secure_password ''' ret = salt.utils.pycrypto.secure_password() check = re.compile(r'[!@#$%^&*()_=+]') assert check.search(ret) is None assert ret
<commit_before># -*- coding: utf-8 -*- # Import python libs from __future__ import absolute_import, print_function, unicode_literals import logging import re # Import Salt Libs import salt.utils.pycrypto import salt.utils.platform # Import Salt Testing Libs from tests.support.unit import TestCase, skipIf log = logging.getLogger(__name__) class PycryptoTestCase(TestCase): ''' TestCase for salt.utils.pycrypto module ''' @skipIf(salt.utils.platform.is_windows(), 'No crypto module for Windows') def test_gen_hash(self): ''' Test gen_hash ''' passwd = 'test_password' id = '$' if salt.utils.platform.is_darwin(): id = '' ret = salt.utils.pycrypto.gen_hash(password=passwd) self.assertTrue(ret.startswith('$6{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='md5') self.assertTrue(ret.startswith('$1{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='sha256') self.assertTrue(ret.startswith('$5{0}'.format(id))) def test_secure_password(self): ''' test secure_password ''' ret = salt.utils.pycrypto.secure_password() check = re.compile(r'[!@#$%^&*()_=+]') assert check.search(ret) is None assert ret <commit_msg>Make the skip apply to any system missing crypt<commit_after># -*- coding: utf-8 -*- # Import python libs from __future__ import absolute_import, print_function, unicode_literals import logging import re # Import Salt Libs import salt.utils.pycrypto import salt.utils.platform # Import Salt Testing Libs from tests.support.unit import TestCase, skipIf log = logging.getLogger(__name__) class PycryptoTestCase(TestCase): ''' TestCase for salt.utils.pycrypto module ''' # The crypt module is only available on Unix systems # https://docs.python.org/dev/library/crypt.html @skipIf(not salt.utils.pycrypto.HAS_CRYPT, 'crypt module not available') def test_gen_hash(self): ''' Test gen_hash ''' passwd = 'test_password' id = '$' if salt.utils.platform.is_darwin(): id = '' ret = salt.utils.pycrypto.gen_hash(password=passwd) self.assertTrue(ret.startswith('$6{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='md5') self.assertTrue(ret.startswith('$1{0}'.format(id))) ret = salt.utils.pycrypto.gen_hash(password=passwd, algorithm='sha256') self.assertTrue(ret.startswith('$5{0}'.format(id))) def test_secure_password(self): ''' test secure_password ''' ret = salt.utils.pycrypto.secure_password() check = re.compile(r'[!@#$%^&*()_=+]') assert check.search(ret) is None assert ret
edebe37458da391723e3206c63102cbb69606c5b
ideascube/conf/idb_irq_bardarash.py
ideascube/conf/idb_irq_bardarash.py
"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = []
"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] HOME_CARDS = STAFF_HOME_CARDS + [ { 'id': 'blog', }, { 'id': 'mediacenter', }, { 'id': 'library', } ]
Remove Kalite until Arabic language is available
Remove Kalite until Arabic language is available
Python
agpl-3.0
ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube
"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] Remove Kalite until Arabic language is available
"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] HOME_CARDS = STAFF_HOME_CARDS + [ { 'id': 'blog', }, { 'id': 'mediacenter', }, { 'id': 'library', } ]
<commit_before>"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] <commit_msg>Remove Kalite until Arabic language is available<commit_after>
"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] HOME_CARDS = STAFF_HOME_CARDS + [ { 'id': 'blog', }, { 'id': 'mediacenter', }, { 'id': 'library', } ]
"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] Remove Kalite until Arabic language is available"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] HOME_CARDS = STAFF_HOME_CARDS + [ { 'id': 'blog', }, { 'id': 'mediacenter', }, { 'id': 'library', } ]
<commit_before>"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] <commit_msg>Remove Kalite until Arabic language is available<commit_after>"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] HOME_CARDS = STAFF_HOME_CARDS + [ { 'id': 'blog', }, { 'id': 'mediacenter', }, { 'id': 'library', } ]
066833caebddb9a6e0735e635ff214448e078405
check_env.py
check_env.py
""" Run this file to check your python installation. """ def test_import_pandas(): import pandas def test_pandas_version(): import pandas version_found = pandas.__version__.split(".") version_found = tuple(int(num) for num in version_found) assert version_found > (0, 15) def test_import_numpy(): import numpy def test_import_matplotlib(): import matplotlib def test_scrape_web(): import pandas as pd pd.read_html("http://en.wikipedia.org/wiki/World_population") if __name__ == "__main__": import nose nose.run(defaultTest=__name__)
""" Run this file to check your python installation. """ from os.path import dirname, join HERE = dirname(__file__) def test_import_pandas(): import pandas def test_pandas_version(): import pandas version_found = pandas.__version__.split(".") version_found = tuple(int(num) for num in version_found) assert version_found > (0, 15) def test_import_numpy(): import numpy def test_import_matplotlib(): import matplotlib.pyplot as plt plt.figure plt.plot plt.legend plt.imshow def test_import_statsmodels(): import statsmodels as sm from statsmodels.formula.api import ols from statsmodels.tsa.ar_model import AR def test_read_html(): import pandas pandas.read_html(join(HERE, "demos", "climate_timeseries", "data", "sea_levels", "Obtaining Tide Gauge Data.html")) def test_scrape_web(): import pandas as pd pd.read_html("http://en.wikipedia.org/wiki/World_population") if __name__ == "__main__": import nose nose.run(defaultTest=__name__)
Add some more content in tests including with statsmodels.
Add some more content in tests including with statsmodels.
Python
mit
wateryhcho/pandas_tutorial,linan7788626/pandas_tutorial,jonathanrocher/pandas_tutorial,wateryhcho/pandas_tutorial,Sandor-PRA/pandas_tutorial,Sandor-PRA/pandas_tutorial,ajaykliyara/pandas_tutorial,ajaykliyara/pandas_tutorial,jonathanrocher/pandas_tutorial,jonathanrocher/pandas_tutorial,ajaykliyara/pandas_tutorial,linan7788626/pandas_tutorial,Sandor-PRA/pandas_tutorial,linan7788626/pandas_tutorial,wateryhcho/pandas_tutorial
""" Run this file to check your python installation. """ def test_import_pandas(): import pandas def test_pandas_version(): import pandas version_found = pandas.__version__.split(".") version_found = tuple(int(num) for num in version_found) assert version_found > (0, 15) def test_import_numpy(): import numpy def test_import_matplotlib(): import matplotlib def test_scrape_web(): import pandas as pd pd.read_html("http://en.wikipedia.org/wiki/World_population") if __name__ == "__main__": import nose nose.run(defaultTest=__name__) Add some more content in tests including with statsmodels.
""" Run this file to check your python installation. """ from os.path import dirname, join HERE = dirname(__file__) def test_import_pandas(): import pandas def test_pandas_version(): import pandas version_found = pandas.__version__.split(".") version_found = tuple(int(num) for num in version_found) assert version_found > (0, 15) def test_import_numpy(): import numpy def test_import_matplotlib(): import matplotlib.pyplot as plt plt.figure plt.plot plt.legend plt.imshow def test_import_statsmodels(): import statsmodels as sm from statsmodels.formula.api import ols from statsmodels.tsa.ar_model import AR def test_read_html(): import pandas pandas.read_html(join(HERE, "demos", "climate_timeseries", "data", "sea_levels", "Obtaining Tide Gauge Data.html")) def test_scrape_web(): import pandas as pd pd.read_html("http://en.wikipedia.org/wiki/World_population") if __name__ == "__main__": import nose nose.run(defaultTest=__name__)
<commit_before>""" Run this file to check your python installation. """ def test_import_pandas(): import pandas def test_pandas_version(): import pandas version_found = pandas.__version__.split(".") version_found = tuple(int(num) for num in version_found) assert version_found > (0, 15) def test_import_numpy(): import numpy def test_import_matplotlib(): import matplotlib def test_scrape_web(): import pandas as pd pd.read_html("http://en.wikipedia.org/wiki/World_population") if __name__ == "__main__": import nose nose.run(defaultTest=__name__) <commit_msg>Add some more content in tests including with statsmodels.<commit_after>
""" Run this file to check your python installation. """ from os.path import dirname, join HERE = dirname(__file__) def test_import_pandas(): import pandas def test_pandas_version(): import pandas version_found = pandas.__version__.split(".") version_found = tuple(int(num) for num in version_found) assert version_found > (0, 15) def test_import_numpy(): import numpy def test_import_matplotlib(): import matplotlib.pyplot as plt plt.figure plt.plot plt.legend plt.imshow def test_import_statsmodels(): import statsmodels as sm from statsmodels.formula.api import ols from statsmodels.tsa.ar_model import AR def test_read_html(): import pandas pandas.read_html(join(HERE, "demos", "climate_timeseries", "data", "sea_levels", "Obtaining Tide Gauge Data.html")) def test_scrape_web(): import pandas as pd pd.read_html("http://en.wikipedia.org/wiki/World_population") if __name__ == "__main__": import nose nose.run(defaultTest=__name__)
""" Run this file to check your python installation. """ def test_import_pandas(): import pandas def test_pandas_version(): import pandas version_found = pandas.__version__.split(".") version_found = tuple(int(num) for num in version_found) assert version_found > (0, 15) def test_import_numpy(): import numpy def test_import_matplotlib(): import matplotlib def test_scrape_web(): import pandas as pd pd.read_html("http://en.wikipedia.org/wiki/World_population") if __name__ == "__main__": import nose nose.run(defaultTest=__name__) Add some more content in tests including with statsmodels.""" Run this file to check your python installation. """ from os.path import dirname, join HERE = dirname(__file__) def test_import_pandas(): import pandas def test_pandas_version(): import pandas version_found = pandas.__version__.split(".") version_found = tuple(int(num) for num in version_found) assert version_found > (0, 15) def test_import_numpy(): import numpy def test_import_matplotlib(): import matplotlib.pyplot as plt plt.figure plt.plot plt.legend plt.imshow def test_import_statsmodels(): import statsmodels as sm from statsmodels.formula.api import ols from statsmodels.tsa.ar_model import AR def test_read_html(): import pandas pandas.read_html(join(HERE, "demos", "climate_timeseries", "data", "sea_levels", "Obtaining Tide Gauge Data.html")) def test_scrape_web(): import pandas as pd pd.read_html("http://en.wikipedia.org/wiki/World_population") if __name__ == "__main__": import nose nose.run(defaultTest=__name__)
<commit_before>""" Run this file to check your python installation. """ def test_import_pandas(): import pandas def test_pandas_version(): import pandas version_found = pandas.__version__.split(".") version_found = tuple(int(num) for num in version_found) assert version_found > (0, 15) def test_import_numpy(): import numpy def test_import_matplotlib(): import matplotlib def test_scrape_web(): import pandas as pd pd.read_html("http://en.wikipedia.org/wiki/World_population") if __name__ == "__main__": import nose nose.run(defaultTest=__name__) <commit_msg>Add some more content in tests including with statsmodels.<commit_after>""" Run this file to check your python installation. """ from os.path import dirname, join HERE = dirname(__file__) def test_import_pandas(): import pandas def test_pandas_version(): import pandas version_found = pandas.__version__.split(".") version_found = tuple(int(num) for num in version_found) assert version_found > (0, 15) def test_import_numpy(): import numpy def test_import_matplotlib(): import matplotlib.pyplot as plt plt.figure plt.plot plt.legend plt.imshow def test_import_statsmodels(): import statsmodels as sm from statsmodels.formula.api import ols from statsmodels.tsa.ar_model import AR def test_read_html(): import pandas pandas.read_html(join(HERE, "demos", "climate_timeseries", "data", "sea_levels", "Obtaining Tide Gauge Data.html")) def test_scrape_web(): import pandas as pd pd.read_html("http://en.wikipedia.org/wiki/World_population") if __name__ == "__main__": import nose nose.run(defaultTest=__name__)
5301f5a641426cd223cb528696495ee1df70258a
prefixlist/api.py
prefixlist/api.py
from flask import Flask from flask_restful import abort, Api, Resource app = Flask("pre-fixlist") api = Api(app) class PrefixListList(Resource): def get(self): # Return a list of all prefix lists we know about return None class PrefixList(Resource): def get(self, listid): # Return a list of versions return None def post(self): # Create a new prefixlist return None def put(self, listid): # Update a prefixlist return None def delete(self, listid): return None class PrefixListVersion(Resource): def get(self, listid, version): # Return a list of prefixes contained in the specific version of the prefixlist return None api.add_resource(PrefixListList, "/prefixlist") api.add_resource(PrefixList, "/prefixlist/<int:listid>") api.add_resource(PrefixListVersion, "/prefixlist/<int:listid>/<string:version>") if __name__ == '__main__': app.run(debug=True)
from flask import Flask from flask_restful import abort, Api, Resource from flask_cors import CORS app = Flask("pre-fixlist") api = Api(app) CORS(app) class PrefixListList(Resource): def get(self): # Return a list of all prefix lists we know about return None class PrefixList(Resource): def get(self, listid): # Return a list of versions return None def post(self): # Create a new prefixlist return None def put(self, listid): # Update a prefixlist return None def delete(self, listid): return None class PrefixListVersion(Resource): def get(self, listid, version): # Return a list of prefixes contained in the specific version of the prefixlist return None api.add_resource(PrefixListList, "/prefixlist") api.add_resource(PrefixList, "/prefixlist/<int:listid>") api.add_resource(PrefixListVersion, "/prefixlist/<int:listid>/<string:version>") if __name__ == '__main__': app.run(debug=True)
Add CORS header to API responses
Add CORS header to API responses
Python
bsd-2-clause
emjemj/pre-fixlist
from flask import Flask from flask_restful import abort, Api, Resource app = Flask("pre-fixlist") api = Api(app) class PrefixListList(Resource): def get(self): # Return a list of all prefix lists we know about return None class PrefixList(Resource): def get(self, listid): # Return a list of versions return None def post(self): # Create a new prefixlist return None def put(self, listid): # Update a prefixlist return None def delete(self, listid): return None class PrefixListVersion(Resource): def get(self, listid, version): # Return a list of prefixes contained in the specific version of the prefixlist return None api.add_resource(PrefixListList, "/prefixlist") api.add_resource(PrefixList, "/prefixlist/<int:listid>") api.add_resource(PrefixListVersion, "/prefixlist/<int:listid>/<string:version>") if __name__ == '__main__': app.run(debug=True) Add CORS header to API responses
from flask import Flask from flask_restful import abort, Api, Resource from flask_cors import CORS app = Flask("pre-fixlist") api = Api(app) CORS(app) class PrefixListList(Resource): def get(self): # Return a list of all prefix lists we know about return None class PrefixList(Resource): def get(self, listid): # Return a list of versions return None def post(self): # Create a new prefixlist return None def put(self, listid): # Update a prefixlist return None def delete(self, listid): return None class PrefixListVersion(Resource): def get(self, listid, version): # Return a list of prefixes contained in the specific version of the prefixlist return None api.add_resource(PrefixListList, "/prefixlist") api.add_resource(PrefixList, "/prefixlist/<int:listid>") api.add_resource(PrefixListVersion, "/prefixlist/<int:listid>/<string:version>") if __name__ == '__main__': app.run(debug=True)
<commit_before>from flask import Flask from flask_restful import abort, Api, Resource app = Flask("pre-fixlist") api = Api(app) class PrefixListList(Resource): def get(self): # Return a list of all prefix lists we know about return None class PrefixList(Resource): def get(self, listid): # Return a list of versions return None def post(self): # Create a new prefixlist return None def put(self, listid): # Update a prefixlist return None def delete(self, listid): return None class PrefixListVersion(Resource): def get(self, listid, version): # Return a list of prefixes contained in the specific version of the prefixlist return None api.add_resource(PrefixListList, "/prefixlist") api.add_resource(PrefixList, "/prefixlist/<int:listid>") api.add_resource(PrefixListVersion, "/prefixlist/<int:listid>/<string:version>") if __name__ == '__main__': app.run(debug=True) <commit_msg>Add CORS header to API responses<commit_after>
from flask import Flask from flask_restful import abort, Api, Resource from flask_cors import CORS app = Flask("pre-fixlist") api = Api(app) CORS(app) class PrefixListList(Resource): def get(self): # Return a list of all prefix lists we know about return None class PrefixList(Resource): def get(self, listid): # Return a list of versions return None def post(self): # Create a new prefixlist return None def put(self, listid): # Update a prefixlist return None def delete(self, listid): return None class PrefixListVersion(Resource): def get(self, listid, version): # Return a list of prefixes contained in the specific version of the prefixlist return None api.add_resource(PrefixListList, "/prefixlist") api.add_resource(PrefixList, "/prefixlist/<int:listid>") api.add_resource(PrefixListVersion, "/prefixlist/<int:listid>/<string:version>") if __name__ == '__main__': app.run(debug=True)
from flask import Flask from flask_restful import abort, Api, Resource app = Flask("pre-fixlist") api = Api(app) class PrefixListList(Resource): def get(self): # Return a list of all prefix lists we know about return None class PrefixList(Resource): def get(self, listid): # Return a list of versions return None def post(self): # Create a new prefixlist return None def put(self, listid): # Update a prefixlist return None def delete(self, listid): return None class PrefixListVersion(Resource): def get(self, listid, version): # Return a list of prefixes contained in the specific version of the prefixlist return None api.add_resource(PrefixListList, "/prefixlist") api.add_resource(PrefixList, "/prefixlist/<int:listid>") api.add_resource(PrefixListVersion, "/prefixlist/<int:listid>/<string:version>") if __name__ == '__main__': app.run(debug=True) Add CORS header to API responsesfrom flask import Flask from flask_restful import abort, Api, Resource from flask_cors import CORS app = Flask("pre-fixlist") api = Api(app) CORS(app) class PrefixListList(Resource): def get(self): # Return a list of all prefix lists we know about return None class PrefixList(Resource): def get(self, listid): # Return a list of versions return None def post(self): # Create a new prefixlist return None def put(self, listid): # Update a prefixlist return None def delete(self, listid): return None class PrefixListVersion(Resource): def get(self, listid, version): # Return a list of prefixes contained in the specific version of the prefixlist return None api.add_resource(PrefixListList, "/prefixlist") api.add_resource(PrefixList, "/prefixlist/<int:listid>") api.add_resource(PrefixListVersion, "/prefixlist/<int:listid>/<string:version>") if __name__ == '__main__': app.run(debug=True)
<commit_before>from flask import Flask from flask_restful import abort, Api, Resource app = Flask("pre-fixlist") api = Api(app) class PrefixListList(Resource): def get(self): # Return a list of all prefix lists we know about return None class PrefixList(Resource): def get(self, listid): # Return a list of versions return None def post(self): # Create a new prefixlist return None def put(self, listid): # Update a prefixlist return None def delete(self, listid): return None class PrefixListVersion(Resource): def get(self, listid, version): # Return a list of prefixes contained in the specific version of the prefixlist return None api.add_resource(PrefixListList, "/prefixlist") api.add_resource(PrefixList, "/prefixlist/<int:listid>") api.add_resource(PrefixListVersion, "/prefixlist/<int:listid>/<string:version>") if __name__ == '__main__': app.run(debug=True) <commit_msg>Add CORS header to API responses<commit_after>from flask import Flask from flask_restful import abort, Api, Resource from flask_cors import CORS app = Flask("pre-fixlist") api = Api(app) CORS(app) class PrefixListList(Resource): def get(self): # Return a list of all prefix lists we know about return None class PrefixList(Resource): def get(self, listid): # Return a list of versions return None def post(self): # Create a new prefixlist return None def put(self, listid): # Update a prefixlist return None def delete(self, listid): return None class PrefixListVersion(Resource): def get(self, listid, version): # Return a list of prefixes contained in the specific version of the prefixlist return None api.add_resource(PrefixListList, "/prefixlist") api.add_resource(PrefixList, "/prefixlist/<int:listid>") api.add_resource(PrefixListVersion, "/prefixlist/<int:listid>/<string:version>") if __name__ == '__main__': app.run(debug=True)
9b5b6514ace9d08d2dca563b71c8b6a1ca3f4f70
wordcloud.py
wordcloud.py
import falcon, json, pymongo MONGO_DB_USER_FILE = '/home/frank/word_cloud-backend/config/mongodb-user' mongo_db_user_config = open(MONGO_DB_USER_FILE, 'r').read() user_name = mongo_db_user_config.split(':')[0] password = mongo_db_user_config.split(':')[1] mongo_db_client = pymongo.MongoClient('wordcloud-mongo.home.franks-reich.net', 27017) mongo_db_client.admin.authenticate(user_name, password) wordcloud_database = mongo_db_client.wordcloud class WordCloudResource: def on_get(self, request, response, name): """Handles GET requests""" wordcloud = { 'name': name, 'wordcloud' : { 'blub' : 40, 'blah' : 34, 'derp' : 57 } } response.body = json.dumps(wordcloud) app = falcon.API() app.add_route('/wordcloud/{name}', WordCloudResource())
import falcon, json, pymongo MONGO_DB_USER_FILE = '/home/frank/word_cloud-backend/config/mongodb-user' mongo_db_user_config = open(MONGO_DB_USER_FILE, 'r').read() user_name = mongo_db_user_config.rstrip('\n').split(':')[0] password = mongo_db_user_config.rstrip('\n').split(':')[1] mongo_db_client = pymongo.MongoClient('wordcloud-mongo.home.franks-reich.net', 27017) mongo_db_client.admin.authenticate(user_name, password) wordcloud_database = mongo_db_client.wordcloud class WordCloudResource: def on_get(self, request, response, name): """Handles GET requests""" wordcloud = { 'name': name, 'wordcloud' : { 'blub' : 40, 'blah' : 34, 'derp' : 57 } } response.body = json.dumps(wordcloud) app = falcon.API() app.add_route('/wordcloud/{name}', WordCloudResource())
Read mongo user and passwd from config file
Read mongo user and passwd from config file
Python
apache-2.0
Frank-Krick/word_cloud-backend
import falcon, json, pymongo MONGO_DB_USER_FILE = '/home/frank/word_cloud-backend/config/mongodb-user' mongo_db_user_config = open(MONGO_DB_USER_FILE, 'r').read() user_name = mongo_db_user_config.split(':')[0] password = mongo_db_user_config.split(':')[1] mongo_db_client = pymongo.MongoClient('wordcloud-mongo.home.franks-reich.net', 27017) mongo_db_client.admin.authenticate(user_name, password) wordcloud_database = mongo_db_client.wordcloud class WordCloudResource: def on_get(self, request, response, name): """Handles GET requests""" wordcloud = { 'name': name, 'wordcloud' : { 'blub' : 40, 'blah' : 34, 'derp' : 57 } } response.body = json.dumps(wordcloud) app = falcon.API() app.add_route('/wordcloud/{name}', WordCloudResource()) Read mongo user and passwd from config file
import falcon, json, pymongo MONGO_DB_USER_FILE = '/home/frank/word_cloud-backend/config/mongodb-user' mongo_db_user_config = open(MONGO_DB_USER_FILE, 'r').read() user_name = mongo_db_user_config.rstrip('\n').split(':')[0] password = mongo_db_user_config.rstrip('\n').split(':')[1] mongo_db_client = pymongo.MongoClient('wordcloud-mongo.home.franks-reich.net', 27017) mongo_db_client.admin.authenticate(user_name, password) wordcloud_database = mongo_db_client.wordcloud class WordCloudResource: def on_get(self, request, response, name): """Handles GET requests""" wordcloud = { 'name': name, 'wordcloud' : { 'blub' : 40, 'blah' : 34, 'derp' : 57 } } response.body = json.dumps(wordcloud) app = falcon.API() app.add_route('/wordcloud/{name}', WordCloudResource())
<commit_before>import falcon, json, pymongo MONGO_DB_USER_FILE = '/home/frank/word_cloud-backend/config/mongodb-user' mongo_db_user_config = open(MONGO_DB_USER_FILE, 'r').read() user_name = mongo_db_user_config.split(':')[0] password = mongo_db_user_config.split(':')[1] mongo_db_client = pymongo.MongoClient('wordcloud-mongo.home.franks-reich.net', 27017) mongo_db_client.admin.authenticate(user_name, password) wordcloud_database = mongo_db_client.wordcloud class WordCloudResource: def on_get(self, request, response, name): """Handles GET requests""" wordcloud = { 'name': name, 'wordcloud' : { 'blub' : 40, 'blah' : 34, 'derp' : 57 } } response.body = json.dumps(wordcloud) app = falcon.API() app.add_route('/wordcloud/{name}', WordCloudResource()) <commit_msg>Read mongo user and passwd from config file<commit_after>
import falcon, json, pymongo MONGO_DB_USER_FILE = '/home/frank/word_cloud-backend/config/mongodb-user' mongo_db_user_config = open(MONGO_DB_USER_FILE, 'r').read() user_name = mongo_db_user_config.rstrip('\n').split(':')[0] password = mongo_db_user_config.rstrip('\n').split(':')[1] mongo_db_client = pymongo.MongoClient('wordcloud-mongo.home.franks-reich.net', 27017) mongo_db_client.admin.authenticate(user_name, password) wordcloud_database = mongo_db_client.wordcloud class WordCloudResource: def on_get(self, request, response, name): """Handles GET requests""" wordcloud = { 'name': name, 'wordcloud' : { 'blub' : 40, 'blah' : 34, 'derp' : 57 } } response.body = json.dumps(wordcloud) app = falcon.API() app.add_route('/wordcloud/{name}', WordCloudResource())
import falcon, json, pymongo MONGO_DB_USER_FILE = '/home/frank/word_cloud-backend/config/mongodb-user' mongo_db_user_config = open(MONGO_DB_USER_FILE, 'r').read() user_name = mongo_db_user_config.split(':')[0] password = mongo_db_user_config.split(':')[1] mongo_db_client = pymongo.MongoClient('wordcloud-mongo.home.franks-reich.net', 27017) mongo_db_client.admin.authenticate(user_name, password) wordcloud_database = mongo_db_client.wordcloud class WordCloudResource: def on_get(self, request, response, name): """Handles GET requests""" wordcloud = { 'name': name, 'wordcloud' : { 'blub' : 40, 'blah' : 34, 'derp' : 57 } } response.body = json.dumps(wordcloud) app = falcon.API() app.add_route('/wordcloud/{name}', WordCloudResource()) Read mongo user and passwd from config fileimport falcon, json, pymongo MONGO_DB_USER_FILE = '/home/frank/word_cloud-backend/config/mongodb-user' mongo_db_user_config = open(MONGO_DB_USER_FILE, 'r').read() user_name = mongo_db_user_config.rstrip('\n').split(':')[0] password = mongo_db_user_config.rstrip('\n').split(':')[1] mongo_db_client = pymongo.MongoClient('wordcloud-mongo.home.franks-reich.net', 27017) mongo_db_client.admin.authenticate(user_name, password) wordcloud_database = mongo_db_client.wordcloud class WordCloudResource: def on_get(self, request, response, name): """Handles GET requests""" wordcloud = { 'name': name, 'wordcloud' : { 'blub' : 40, 'blah' : 34, 'derp' : 57 } } response.body = json.dumps(wordcloud) app = falcon.API() app.add_route('/wordcloud/{name}', WordCloudResource())
<commit_before>import falcon, json, pymongo MONGO_DB_USER_FILE = '/home/frank/word_cloud-backend/config/mongodb-user' mongo_db_user_config = open(MONGO_DB_USER_FILE, 'r').read() user_name = mongo_db_user_config.split(':')[0] password = mongo_db_user_config.split(':')[1] mongo_db_client = pymongo.MongoClient('wordcloud-mongo.home.franks-reich.net', 27017) mongo_db_client.admin.authenticate(user_name, password) wordcloud_database = mongo_db_client.wordcloud class WordCloudResource: def on_get(self, request, response, name): """Handles GET requests""" wordcloud = { 'name': name, 'wordcloud' : { 'blub' : 40, 'blah' : 34, 'derp' : 57 } } response.body = json.dumps(wordcloud) app = falcon.API() app.add_route('/wordcloud/{name}', WordCloudResource()) <commit_msg>Read mongo user and passwd from config file<commit_after>import falcon, json, pymongo MONGO_DB_USER_FILE = '/home/frank/word_cloud-backend/config/mongodb-user' mongo_db_user_config = open(MONGO_DB_USER_FILE, 'r').read() user_name = mongo_db_user_config.rstrip('\n').split(':')[0] password = mongo_db_user_config.rstrip('\n').split(':')[1] mongo_db_client = pymongo.MongoClient('wordcloud-mongo.home.franks-reich.net', 27017) mongo_db_client.admin.authenticate(user_name, password) wordcloud_database = mongo_db_client.wordcloud class WordCloudResource: def on_get(self, request, response, name): """Handles GET requests""" wordcloud = { 'name': name, 'wordcloud' : { 'blub' : 40, 'blah' : 34, 'derp' : 57 } } response.body = json.dumps(wordcloud) app = falcon.API() app.add_route('/wordcloud/{name}', WordCloudResource())
2b0edbadec80300d20a280db0f06281040e00e25
radar/radar/validation/consultants.py
radar/radar/validation/consultants.py
from radar.validation.core import Field, Validation, ListField, ValidationError from radar.validation.meta import MetaValidationMixin from radar.validation.validators import not_empty, none_if_blank, optional, email_address, max_length, required, upper, lower from radar.validation.number_validators import gmc_number class GroupConsultantValidation(MetaValidationMixin, Validation): group = Field([required()]) def validate_group(self, group): if group.type != 'HOSPITAL': raise ValidationError('Must be a hospital.') return group class ConsultantValidation(MetaValidationMixin, Validation): first_name = Field([not_empty(), upper(), max_length(100)]) last_name = Field([not_empty(), upper(), max_length(100)]) email = Field([none_if_blank(), optional(), lower(), email_address()]) telephone_number = Field([none_if_blank(), optional(), max_length(100)]) gmc_number = Field([optional(), gmc_number()]) group_consultants = ListField(GroupConsultantValidation())
from radar.validation.core import Field, Validation, ListField, ValidationError from radar.validation.meta import MetaValidationMixin from radar.validation.validators import not_empty, none_if_blank, optional, email_address, max_length, required, upper, lower from radar.validation.number_validators import gmc_number class GroupConsultantValidation(MetaValidationMixin, Validation): group = Field([required()]) def validate_group(self, group): if group.type != 'HOSPITAL': raise ValidationError('Must be a hospital.') return group class GroupConsultantListField(ListField): def __init__(self, chain=None): super(GroupConsultantListField, self).__init__(GroupConsultantValidation(), chain=chain) def validate(self, group_consultants): print 'hello!' group_ids = set() for i, group_consultant in enumerate(group_consultants): group_id = group_consultant.group.id if group_id in group_ids: raise ValidationError({i: {'group': 'Consultant already in group.'}}) else: group_ids.add(group_id) return group_consultants class ConsultantValidation(MetaValidationMixin, Validation): first_name = Field([not_empty(), upper(), max_length(100)]) last_name = Field([not_empty(), upper(), max_length(100)]) email = Field([none_if_blank(), optional(), lower(), email_address()]) telephone_number = Field([none_if_blank(), optional(), max_length(100)]) gmc_number = Field([optional(), gmc_number()]) group_consultants = GroupConsultantListField()
Check consultant groups aren't duplicated
Check consultant groups aren't duplicated
Python
agpl-3.0
renalreg/radar,renalreg/radar,renalreg/radar,renalreg/radar
from radar.validation.core import Field, Validation, ListField, ValidationError from radar.validation.meta import MetaValidationMixin from radar.validation.validators import not_empty, none_if_blank, optional, email_address, max_length, required, upper, lower from radar.validation.number_validators import gmc_number class GroupConsultantValidation(MetaValidationMixin, Validation): group = Field([required()]) def validate_group(self, group): if group.type != 'HOSPITAL': raise ValidationError('Must be a hospital.') return group class ConsultantValidation(MetaValidationMixin, Validation): first_name = Field([not_empty(), upper(), max_length(100)]) last_name = Field([not_empty(), upper(), max_length(100)]) email = Field([none_if_blank(), optional(), lower(), email_address()]) telephone_number = Field([none_if_blank(), optional(), max_length(100)]) gmc_number = Field([optional(), gmc_number()]) group_consultants = ListField(GroupConsultantValidation()) Check consultant groups aren't duplicated
from radar.validation.core import Field, Validation, ListField, ValidationError from radar.validation.meta import MetaValidationMixin from radar.validation.validators import not_empty, none_if_blank, optional, email_address, max_length, required, upper, lower from radar.validation.number_validators import gmc_number class GroupConsultantValidation(MetaValidationMixin, Validation): group = Field([required()]) def validate_group(self, group): if group.type != 'HOSPITAL': raise ValidationError('Must be a hospital.') return group class GroupConsultantListField(ListField): def __init__(self, chain=None): super(GroupConsultantListField, self).__init__(GroupConsultantValidation(), chain=chain) def validate(self, group_consultants): print 'hello!' group_ids = set() for i, group_consultant in enumerate(group_consultants): group_id = group_consultant.group.id if group_id in group_ids: raise ValidationError({i: {'group': 'Consultant already in group.'}}) else: group_ids.add(group_id) return group_consultants class ConsultantValidation(MetaValidationMixin, Validation): first_name = Field([not_empty(), upper(), max_length(100)]) last_name = Field([not_empty(), upper(), max_length(100)]) email = Field([none_if_blank(), optional(), lower(), email_address()]) telephone_number = Field([none_if_blank(), optional(), max_length(100)]) gmc_number = Field([optional(), gmc_number()]) group_consultants = GroupConsultantListField()
<commit_before>from radar.validation.core import Field, Validation, ListField, ValidationError from radar.validation.meta import MetaValidationMixin from radar.validation.validators import not_empty, none_if_blank, optional, email_address, max_length, required, upper, lower from radar.validation.number_validators import gmc_number class GroupConsultantValidation(MetaValidationMixin, Validation): group = Field([required()]) def validate_group(self, group): if group.type != 'HOSPITAL': raise ValidationError('Must be a hospital.') return group class ConsultantValidation(MetaValidationMixin, Validation): first_name = Field([not_empty(), upper(), max_length(100)]) last_name = Field([not_empty(), upper(), max_length(100)]) email = Field([none_if_blank(), optional(), lower(), email_address()]) telephone_number = Field([none_if_blank(), optional(), max_length(100)]) gmc_number = Field([optional(), gmc_number()]) group_consultants = ListField(GroupConsultantValidation()) <commit_msg>Check consultant groups aren't duplicated<commit_after>
from radar.validation.core import Field, Validation, ListField, ValidationError from radar.validation.meta import MetaValidationMixin from radar.validation.validators import not_empty, none_if_blank, optional, email_address, max_length, required, upper, lower from radar.validation.number_validators import gmc_number class GroupConsultantValidation(MetaValidationMixin, Validation): group = Field([required()]) def validate_group(self, group): if group.type != 'HOSPITAL': raise ValidationError('Must be a hospital.') return group class GroupConsultantListField(ListField): def __init__(self, chain=None): super(GroupConsultantListField, self).__init__(GroupConsultantValidation(), chain=chain) def validate(self, group_consultants): print 'hello!' group_ids = set() for i, group_consultant in enumerate(group_consultants): group_id = group_consultant.group.id if group_id in group_ids: raise ValidationError({i: {'group': 'Consultant already in group.'}}) else: group_ids.add(group_id) return group_consultants class ConsultantValidation(MetaValidationMixin, Validation): first_name = Field([not_empty(), upper(), max_length(100)]) last_name = Field([not_empty(), upper(), max_length(100)]) email = Field([none_if_blank(), optional(), lower(), email_address()]) telephone_number = Field([none_if_blank(), optional(), max_length(100)]) gmc_number = Field([optional(), gmc_number()]) group_consultants = GroupConsultantListField()
from radar.validation.core import Field, Validation, ListField, ValidationError from radar.validation.meta import MetaValidationMixin from radar.validation.validators import not_empty, none_if_blank, optional, email_address, max_length, required, upper, lower from radar.validation.number_validators import gmc_number class GroupConsultantValidation(MetaValidationMixin, Validation): group = Field([required()]) def validate_group(self, group): if group.type != 'HOSPITAL': raise ValidationError('Must be a hospital.') return group class ConsultantValidation(MetaValidationMixin, Validation): first_name = Field([not_empty(), upper(), max_length(100)]) last_name = Field([not_empty(), upper(), max_length(100)]) email = Field([none_if_blank(), optional(), lower(), email_address()]) telephone_number = Field([none_if_blank(), optional(), max_length(100)]) gmc_number = Field([optional(), gmc_number()]) group_consultants = ListField(GroupConsultantValidation()) Check consultant groups aren't duplicatedfrom radar.validation.core import Field, Validation, ListField, ValidationError from radar.validation.meta import MetaValidationMixin from radar.validation.validators import not_empty, none_if_blank, optional, email_address, max_length, required, upper, lower from radar.validation.number_validators import gmc_number class GroupConsultantValidation(MetaValidationMixin, Validation): group = Field([required()]) def validate_group(self, group): if group.type != 'HOSPITAL': raise ValidationError('Must be a hospital.') return group class GroupConsultantListField(ListField): def __init__(self, chain=None): super(GroupConsultantListField, self).__init__(GroupConsultantValidation(), chain=chain) def validate(self, group_consultants): print 'hello!' group_ids = set() for i, group_consultant in enumerate(group_consultants): group_id = group_consultant.group.id if group_id in group_ids: raise ValidationError({i: {'group': 'Consultant already in group.'}}) else: group_ids.add(group_id) return group_consultants class ConsultantValidation(MetaValidationMixin, Validation): first_name = Field([not_empty(), upper(), max_length(100)]) last_name = Field([not_empty(), upper(), max_length(100)]) email = Field([none_if_blank(), optional(), lower(), email_address()]) telephone_number = Field([none_if_blank(), optional(), max_length(100)]) gmc_number = Field([optional(), gmc_number()]) group_consultants = GroupConsultantListField()
<commit_before>from radar.validation.core import Field, Validation, ListField, ValidationError from radar.validation.meta import MetaValidationMixin from radar.validation.validators import not_empty, none_if_blank, optional, email_address, max_length, required, upper, lower from radar.validation.number_validators import gmc_number class GroupConsultantValidation(MetaValidationMixin, Validation): group = Field([required()]) def validate_group(self, group): if group.type != 'HOSPITAL': raise ValidationError('Must be a hospital.') return group class ConsultantValidation(MetaValidationMixin, Validation): first_name = Field([not_empty(), upper(), max_length(100)]) last_name = Field([not_empty(), upper(), max_length(100)]) email = Field([none_if_blank(), optional(), lower(), email_address()]) telephone_number = Field([none_if_blank(), optional(), max_length(100)]) gmc_number = Field([optional(), gmc_number()]) group_consultants = ListField(GroupConsultantValidation()) <commit_msg>Check consultant groups aren't duplicated<commit_after>from radar.validation.core import Field, Validation, ListField, ValidationError from radar.validation.meta import MetaValidationMixin from radar.validation.validators import not_empty, none_if_blank, optional, email_address, max_length, required, upper, lower from radar.validation.number_validators import gmc_number class GroupConsultantValidation(MetaValidationMixin, Validation): group = Field([required()]) def validate_group(self, group): if group.type != 'HOSPITAL': raise ValidationError('Must be a hospital.') return group class GroupConsultantListField(ListField): def __init__(self, chain=None): super(GroupConsultantListField, self).__init__(GroupConsultantValidation(), chain=chain) def validate(self, group_consultants): print 'hello!' group_ids = set() for i, group_consultant in enumerate(group_consultants): group_id = group_consultant.group.id if group_id in group_ids: raise ValidationError({i: {'group': 'Consultant already in group.'}}) else: group_ids.add(group_id) return group_consultants class ConsultantValidation(MetaValidationMixin, Validation): first_name = Field([not_empty(), upper(), max_length(100)]) last_name = Field([not_empty(), upper(), max_length(100)]) email = Field([none_if_blank(), optional(), lower(), email_address()]) telephone_number = Field([none_if_blank(), optional(), max_length(100)]) gmc_number = Field([optional(), gmc_number()]) group_consultants = GroupConsultantListField()
b8359e6b04d13f550aec308308f2e91e194bc372
uberlogs/handlers/kill_process.py
uberlogs/handlers/kill_process.py
import sys import os from logging import Handler as LoggingHandler class KillProcessHandler(LoggingHandler): @profile def emit(self, record): if record.levelno != self.level: return # flush text before exiting for fd in [sys.stdout, sys.stderr]: fd.flush() # Twisted writes unhandled errors in different calls # If we exit on the first call, we'd lose the actual error for log_to_ignore in ["Unhandled error in Deferred"]: if log_to_ignore.lower() in record.getMessage().lower(): return os._exit(1)
import sys import os from logging import Handler as LoggingHandler class KillProcessHandler(LoggingHandler): @profile def emit(self, record): if record.levelno != self.level: return # flush text before exiting for fd in [sys.stdout, sys.stderr]: fd.flush() msg = record.getMessage() # Twisted writes unhandled errors in different calls # If we exit on the first call, we'd lose the actual error for log_to_ignore in ["Unhandled error in Deferred"]: if log_to_ignore.lower() in msg.lower(): return os._exit(1)
Remove repetitive getMessage calls in KillProcesshandler
Remove repetitive getMessage calls in KillProcesshandler
Python
mit
odedlaz/uberlogs,odedlaz/uberlogs
import sys import os from logging import Handler as LoggingHandler class KillProcessHandler(LoggingHandler): @profile def emit(self, record): if record.levelno != self.level: return # flush text before exiting for fd in [sys.stdout, sys.stderr]: fd.flush() # Twisted writes unhandled errors in different calls # If we exit on the first call, we'd lose the actual error for log_to_ignore in ["Unhandled error in Deferred"]: if log_to_ignore.lower() in record.getMessage().lower(): return os._exit(1) Remove repetitive getMessage calls in KillProcesshandler
import sys import os from logging import Handler as LoggingHandler class KillProcessHandler(LoggingHandler): @profile def emit(self, record): if record.levelno != self.level: return # flush text before exiting for fd in [sys.stdout, sys.stderr]: fd.flush() msg = record.getMessage() # Twisted writes unhandled errors in different calls # If we exit on the first call, we'd lose the actual error for log_to_ignore in ["Unhandled error in Deferred"]: if log_to_ignore.lower() in msg.lower(): return os._exit(1)
<commit_before>import sys import os from logging import Handler as LoggingHandler class KillProcessHandler(LoggingHandler): @profile def emit(self, record): if record.levelno != self.level: return # flush text before exiting for fd in [sys.stdout, sys.stderr]: fd.flush() # Twisted writes unhandled errors in different calls # If we exit on the first call, we'd lose the actual error for log_to_ignore in ["Unhandled error in Deferred"]: if log_to_ignore.lower() in record.getMessage().lower(): return os._exit(1) <commit_msg>Remove repetitive getMessage calls in KillProcesshandler<commit_after>
import sys import os from logging import Handler as LoggingHandler class KillProcessHandler(LoggingHandler): @profile def emit(self, record): if record.levelno != self.level: return # flush text before exiting for fd in [sys.stdout, sys.stderr]: fd.flush() msg = record.getMessage() # Twisted writes unhandled errors in different calls # If we exit on the first call, we'd lose the actual error for log_to_ignore in ["Unhandled error in Deferred"]: if log_to_ignore.lower() in msg.lower(): return os._exit(1)
import sys import os from logging import Handler as LoggingHandler class KillProcessHandler(LoggingHandler): @profile def emit(self, record): if record.levelno != self.level: return # flush text before exiting for fd in [sys.stdout, sys.stderr]: fd.flush() # Twisted writes unhandled errors in different calls # If we exit on the first call, we'd lose the actual error for log_to_ignore in ["Unhandled error in Deferred"]: if log_to_ignore.lower() in record.getMessage().lower(): return os._exit(1) Remove repetitive getMessage calls in KillProcesshandlerimport sys import os from logging import Handler as LoggingHandler class KillProcessHandler(LoggingHandler): @profile def emit(self, record): if record.levelno != self.level: return # flush text before exiting for fd in [sys.stdout, sys.stderr]: fd.flush() msg = record.getMessage() # Twisted writes unhandled errors in different calls # If we exit on the first call, we'd lose the actual error for log_to_ignore in ["Unhandled error in Deferred"]: if log_to_ignore.lower() in msg.lower(): return os._exit(1)
<commit_before>import sys import os from logging import Handler as LoggingHandler class KillProcessHandler(LoggingHandler): @profile def emit(self, record): if record.levelno != self.level: return # flush text before exiting for fd in [sys.stdout, sys.stderr]: fd.flush() # Twisted writes unhandled errors in different calls # If we exit on the first call, we'd lose the actual error for log_to_ignore in ["Unhandled error in Deferred"]: if log_to_ignore.lower() in record.getMessage().lower(): return os._exit(1) <commit_msg>Remove repetitive getMessage calls in KillProcesshandler<commit_after>import sys import os from logging import Handler as LoggingHandler class KillProcessHandler(LoggingHandler): @profile def emit(self, record): if record.levelno != self.level: return # flush text before exiting for fd in [sys.stdout, sys.stderr]: fd.flush() msg = record.getMessage() # Twisted writes unhandled errors in different calls # If we exit on the first call, we'd lose the actual error for log_to_ignore in ["Unhandled error in Deferred"]: if log_to_ignore.lower() in msg.lower(): return os._exit(1)
be5eecf2a043abf7585022f1dda3b79572eba192
tests/test__pycompat.py
tests/test__pycompat.py
#!/usr/bin/env python # -*- coding: utf-8 -*-
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import dask_ndmeasure._pycompat def test_irange(): r = dask_ndmeasure._pycompat.irange(5) assert not isinstance(r, list) assert list(r) == [0, 1, 2, 3, 4]
Add a basic test for irange
Add a basic test for irange Make sure `irange` is there, it doesn't return a list, and it acts like `range` on some test arguments.
Python
bsd-3-clause
dask-image/dask-ndmeasure
#!/usr/bin/env python # -*- coding: utf-8 -*- Add a basic test for irange Make sure `irange` is there, it doesn't return a list, and it acts like `range` on some test arguments.
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import dask_ndmeasure._pycompat def test_irange(): r = dask_ndmeasure._pycompat.irange(5) assert not isinstance(r, list) assert list(r) == [0, 1, 2, 3, 4]
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- <commit_msg>Add a basic test for irange Make sure `irange` is there, it doesn't return a list, and it acts like `range` on some test arguments.<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import dask_ndmeasure._pycompat def test_irange(): r = dask_ndmeasure._pycompat.irange(5) assert not isinstance(r, list) assert list(r) == [0, 1, 2, 3, 4]
#!/usr/bin/env python # -*- coding: utf-8 -*- Add a basic test for irange Make sure `irange` is there, it doesn't return a list, and it acts like `range` on some test arguments.#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import dask_ndmeasure._pycompat def test_irange(): r = dask_ndmeasure._pycompat.irange(5) assert not isinstance(r, list) assert list(r) == [0, 1, 2, 3, 4]
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- <commit_msg>Add a basic test for irange Make sure `irange` is there, it doesn't return a list, and it acts like `range` on some test arguments.<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import dask_ndmeasure._pycompat def test_irange(): r = dask_ndmeasure._pycompat.irange(5) assert not isinstance(r, list) assert list(r) == [0, 1, 2, 3, 4]
e90a60b1da00a6c6a5e1b4235c4009d7477986ca
conanfile.py
conanfile.py
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class CPPCheckTargetCMakeConan(ConanFile): name = "cppcheck-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "tooling-find-pkg-util/master@smspillaz/tooling-find-pkg-util", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit") url = "http://github.com/polysquare/cppcheck-target-cmake" license = "MIT" def source(self): zip_name = "cppcheck-target-cmake.zip" download("https://github.com/polysquare/" "cppcheck-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cppcheck-target-cmake", src="cppcheck-target-cmake-" + VERSION, keep_path=True)
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class CPPCheckTargetCMakeConan(ConanFile): name = "cppcheck-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "tooling-find-pkg-util/master@smspillaz/tooling-find-pkg-util", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit") url = "http://github.com/polysquare/cppcheck-target-cmake" license = "MIT" def source(self): zip_name = "cppcheck-target-cmake.zip" download("https://github.com/polysquare/" "cppcheck-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="Find*.cmake", dst="", src="cppcheck-target-cmake-" + VERSION, keep_path=True) self.copy(pattern="*.cmake", dst="cmake/cppcheck-target-cmake", src="cppcheck-target-cmake-" + VERSION, keep_path=True)
Copy find modules to root of module path
conan: Copy find modules to root of module path
Python
mit
polysquare/cppcheck-target-cmake
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class CPPCheckTargetCMakeConan(ConanFile): name = "cppcheck-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "tooling-find-pkg-util/master@smspillaz/tooling-find-pkg-util", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit") url = "http://github.com/polysquare/cppcheck-target-cmake" license = "MIT" def source(self): zip_name = "cppcheck-target-cmake.zip" download("https://github.com/polysquare/" "cppcheck-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cppcheck-target-cmake", src="cppcheck-target-cmake-" + VERSION, keep_path=True) conan: Copy find modules to root of module path
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class CPPCheckTargetCMakeConan(ConanFile): name = "cppcheck-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "tooling-find-pkg-util/master@smspillaz/tooling-find-pkg-util", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit") url = "http://github.com/polysquare/cppcheck-target-cmake" license = "MIT" def source(self): zip_name = "cppcheck-target-cmake.zip" download("https://github.com/polysquare/" "cppcheck-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="Find*.cmake", dst="", src="cppcheck-target-cmake-" + VERSION, keep_path=True) self.copy(pattern="*.cmake", dst="cmake/cppcheck-target-cmake", src="cppcheck-target-cmake-" + VERSION, keep_path=True)
<commit_before>from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class CPPCheckTargetCMakeConan(ConanFile): name = "cppcheck-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "tooling-find-pkg-util/master@smspillaz/tooling-find-pkg-util", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit") url = "http://github.com/polysquare/cppcheck-target-cmake" license = "MIT" def source(self): zip_name = "cppcheck-target-cmake.zip" download("https://github.com/polysquare/" "cppcheck-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cppcheck-target-cmake", src="cppcheck-target-cmake-" + VERSION, keep_path=True) <commit_msg>conan: Copy find modules to root of module path<commit_after>
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class CPPCheckTargetCMakeConan(ConanFile): name = "cppcheck-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "tooling-find-pkg-util/master@smspillaz/tooling-find-pkg-util", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit") url = "http://github.com/polysquare/cppcheck-target-cmake" license = "MIT" def source(self): zip_name = "cppcheck-target-cmake.zip" download("https://github.com/polysquare/" "cppcheck-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="Find*.cmake", dst="", src="cppcheck-target-cmake-" + VERSION, keep_path=True) self.copy(pattern="*.cmake", dst="cmake/cppcheck-target-cmake", src="cppcheck-target-cmake-" + VERSION, keep_path=True)
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class CPPCheckTargetCMakeConan(ConanFile): name = "cppcheck-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "tooling-find-pkg-util/master@smspillaz/tooling-find-pkg-util", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit") url = "http://github.com/polysquare/cppcheck-target-cmake" license = "MIT" def source(self): zip_name = "cppcheck-target-cmake.zip" download("https://github.com/polysquare/" "cppcheck-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cppcheck-target-cmake", src="cppcheck-target-cmake-" + VERSION, keep_path=True) conan: Copy find modules to root of module pathfrom conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class CPPCheckTargetCMakeConan(ConanFile): name = "cppcheck-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "tooling-find-pkg-util/master@smspillaz/tooling-find-pkg-util", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit") url = "http://github.com/polysquare/cppcheck-target-cmake" license = "MIT" def source(self): zip_name = "cppcheck-target-cmake.zip" download("https://github.com/polysquare/" "cppcheck-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="Find*.cmake", dst="", src="cppcheck-target-cmake-" + VERSION, keep_path=True) self.copy(pattern="*.cmake", dst="cmake/cppcheck-target-cmake", src="cppcheck-target-cmake-" + VERSION, keep_path=True)
<commit_before>from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class CPPCheckTargetCMakeConan(ConanFile): name = "cppcheck-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "tooling-find-pkg-util/master@smspillaz/tooling-find-pkg-util", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit") url = "http://github.com/polysquare/cppcheck-target-cmake" license = "MIT" def source(self): zip_name = "cppcheck-target-cmake.zip" download("https://github.com/polysquare/" "cppcheck-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/cppcheck-target-cmake", src="cppcheck-target-cmake-" + VERSION, keep_path=True) <commit_msg>conan: Copy find modules to root of module path<commit_after>from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class CPPCheckTargetCMakeConan(ConanFile): name = "cppcheck-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "tooling-find-pkg-util/master@smspillaz/tooling-find-pkg-util", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit") url = "http://github.com/polysquare/cppcheck-target-cmake" license = "MIT" def source(self): zip_name = "cppcheck-target-cmake.zip" download("https://github.com/polysquare/" "cppcheck-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="Find*.cmake", dst="", src="cppcheck-target-cmake-" + VERSION, keep_path=True) self.copy(pattern="*.cmake", dst="cmake/cppcheck-target-cmake", src="cppcheck-target-cmake-" + VERSION, keep_path=True)
59afb96f2211983ee2a2786c60791074b13c3e7f
ni/__main__.py
ni/__main__.py
"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a background queue, one could add the update as a work item # and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a work queue, one could make the updating of the # contribution a work item and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
Tweak comment about 202 response
Tweak comment about 202 response
Python
apache-2.0
python/the-knights-who-say-ni,python/the-knights-who-say-ni
"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a background queue, one could add the update as a work item # and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port()) Tweak comment about 202 response
"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a work queue, one could make the updating of the # contribution a work item and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
<commit_before>"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a background queue, one could add the update as a work item # and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port()) <commit_msg>Tweak comment about 202 response<commit_after>
"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a work queue, one could make the updating of the # contribution a work item and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a background queue, one could add the update as a work item # and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port()) Tweak comment about 202 response"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a work queue, one could make the updating of the # contribution a work item and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
<commit_before>"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a background queue, one could add the update as a work item # and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port()) <commit_msg>Tweak comment about 202 response<commit_after>"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a work queue, one could make the updating of the # contribution a work item and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
9681b2f31163e24e72121afc6195262376891220
marketpulse/main/__init__.py
marketpulse/main/__init__.py
import moneyed FFXOS_ACTIVITY_NAME = 'Submit FirefoxOS device price' def get_currency_choices(): return sorted(((currency, data.name) for currency, data in moneyed.CURRENCIES.items()))
import moneyed FFXOS_ACTIVITY_NAME = 'Submit FirefoxOS device price' def get_currency_choices(): return sorted(((currency, data.code) for currency, data in moneyed.CURRENCIES.items()))
Replace currency full name with currency code.
Replace currency full name with currency code.
Python
mpl-2.0
akatsoulas/marketpulse,johngian/marketpulse,akatsoulas/marketpulse,mozilla/marketpulse,akatsoulas/marketpulse,johngian/marketpulse,akatsoulas/marketpulse,johngian/marketpulse,mozilla/marketpulse,mozilla/marketpulse,johngian/marketpulse,mozilla/marketpulse
import moneyed FFXOS_ACTIVITY_NAME = 'Submit FirefoxOS device price' def get_currency_choices(): return sorted(((currency, data.name) for currency, data in moneyed.CURRENCIES.items())) Replace currency full name with currency code.
import moneyed FFXOS_ACTIVITY_NAME = 'Submit FirefoxOS device price' def get_currency_choices(): return sorted(((currency, data.code) for currency, data in moneyed.CURRENCIES.items()))
<commit_before>import moneyed FFXOS_ACTIVITY_NAME = 'Submit FirefoxOS device price' def get_currency_choices(): return sorted(((currency, data.name) for currency, data in moneyed.CURRENCIES.items())) <commit_msg>Replace currency full name with currency code.<commit_after>
import moneyed FFXOS_ACTIVITY_NAME = 'Submit FirefoxOS device price' def get_currency_choices(): return sorted(((currency, data.code) for currency, data in moneyed.CURRENCIES.items()))
import moneyed FFXOS_ACTIVITY_NAME = 'Submit FirefoxOS device price' def get_currency_choices(): return sorted(((currency, data.name) for currency, data in moneyed.CURRENCIES.items())) Replace currency full name with currency code.import moneyed FFXOS_ACTIVITY_NAME = 'Submit FirefoxOS device price' def get_currency_choices(): return sorted(((currency, data.code) for currency, data in moneyed.CURRENCIES.items()))
<commit_before>import moneyed FFXOS_ACTIVITY_NAME = 'Submit FirefoxOS device price' def get_currency_choices(): return sorted(((currency, data.name) for currency, data in moneyed.CURRENCIES.items())) <commit_msg>Replace currency full name with currency code.<commit_after>import moneyed FFXOS_ACTIVITY_NAME = 'Submit FirefoxOS device price' def get_currency_choices(): return sorted(((currency, data.code) for currency, data in moneyed.CURRENCIES.items()))
9cfe03ab06f126406a51c0945e990fc849d8dfb9
scripts/crontab/gen-crons.py
scripts/crontab/gen-crons.py
#!/usr/bin/env python import os from optparse import OptionParser from jinja2 import Template TEMPLATE = open(os.path.join(os.path.dirname(__file__), 'crontab.tpl')).read() def main(): parser = OptionParser() parser.add_option("-k", "--kitsune", help="Location of kitsune (required)") parser.add_option("-u", "--user", help=("Prefix cron with this user. " "Only define for cron.d style crontabs")) parser.add_option("-p", "--python", default="/usr/bin/python2.6", help="Python interpreter to use") (opts, args) = parser.parse_args() if not opts.kitsune: parser.error("-k must be defined") ctx = {'django': 'cd %s; %s manage.py' % (opts.kitsune, opts.python),} ctx['cron'] = '%s cron' % ctx['django'] if opts.user: for k, v in ctx.iteritems(): ctx[k] = '%s %s' % (opts.user, v) # Needs to stay below the opts.user injection. ctx['python'] = opts.python print Template(TEMPLATE).render(**ctx) if __name__ == "__main__": main()
#!/usr/bin/env python import os from optparse import OptionParser from jinja2 import Template TEMPLATE = open(os.path.join(os.path.dirname(__file__), 'crontab.tpl')).read() def main(): parser = OptionParser() parser.add_option("-k", "--kitsune", help="Location of kitsune (required)") parser.add_option("-u", "--user", help=("Prefix cron with this user. " "Only define for cron.d style crontabs")) parser.add_option("-p", "--python", default="/usr/bin/python2.6", help="Python interpreter to use") (opts, args) = parser.parse_args() if not opts.kitsune: parser.error("-k must be defined") # To pick up the right PyOpenSSL: python_path = 'PYTHONPATH=/usr/local/lib64/python2.6/site-packages' ctx = {'django': 'cd %s; %s %s manage.py' % ( opts.kitsune, python_path, opts.python),} ctx['cron'] = '%s cron' % ctx['django'] if opts.user: for k, v in ctx.iteritems(): ctx[k] = '%s %s' % (opts.user, v) # Needs to stay below the opts.user injection. ctx['python'] = opts.python print Template(TEMPLATE).render(**ctx) if __name__ == "__main__": main()
Add local site-packages to PYTHONPATH.
Add local site-packages to PYTHONPATH. To pick up the local version of PyOpenSSL.
Python
bsd-3-clause
philipp-sumo/kitsune,iDTLabssl/kitsune,orvi2014/kitsune,silentbob73/kitsune,NewPresident1/kitsune,MziRintu/kitsune,YOTOV-LIMITED/kitsune,iDTLabssl/kitsune,YOTOV-LIMITED/kitsune,safwanrahman/kitsune,silentbob73/kitsune,safwanrahman/linuxdesh,mythmon/kitsune,YOTOV-LIMITED/kitsune,iDTLabssl/kitsune,turtleloveshoes/kitsune,chirilo/kitsune,NewPresident1/kitsune,rlr/kitsune,turtleloveshoes/kitsune,MikkCZ/kitsune,brittanystoroz/kitsune,NewPresident1/kitsune,NewPresident1/kitsune,orvi2014/kitsune,silentbob73/kitsune,Osmose/kitsune,silentbob73/kitsune,MikkCZ/kitsune,chirilo/kitsune,mythmon/kitsune,dbbhattacharya/kitsune,philipp-sumo/kitsune,safwanrahman/linuxdesh,anushbmx/kitsune,dbbhattacharya/kitsune,asdofindia/kitsune,safwanrahman/kitsune,MziRintu/kitsune,asdofindia/kitsune,dbbhattacharya/kitsune,mythmon/kitsune,Osmose/kitsune,feer56/Kitsune2,feer56/Kitsune1,H1ghT0p/kitsune,mozilla/kitsune,feer56/Kitsune1,mozilla/kitsune,Osmose/kitsune,turtleloveshoes/kitsune,iDTLabssl/kitsune,safwanrahman/kitsune,brittanystoroz/kitsune,mozilla/kitsune,brittanystoroz/kitsune,MikkCZ/kitsune,mozilla/kitsune,Osmose/kitsune,dbbhattacharya/kitsune,rlr/kitsune,mythmon/kitsune,feer56/Kitsune2,turtleloveshoes/kitsune,YOTOV-LIMITED/kitsune,philipp-sumo/kitsune,chirilo/kitsune,feer56/Kitsune2,asdofindia/kitsune,orvi2014/kitsune,H1ghT0p/kitsune,anushbmx/kitsune,orvi2014/kitsune,MziRintu/kitsune,brittanystoroz/kitsune,H1ghT0p/kitsune,feer56/Kitsune1,safwanrahman/linuxdesh,safwanrahman/kitsune,chirilo/kitsune,MikkCZ/kitsune,anushbmx/kitsune,feer56/Kitsune2,MziRintu/kitsune,rlr/kitsune,rlr/kitsune,H1ghT0p/kitsune,anushbmx/kitsune,asdofindia/kitsune
#!/usr/bin/env python import os from optparse import OptionParser from jinja2 import Template TEMPLATE = open(os.path.join(os.path.dirname(__file__), 'crontab.tpl')).read() def main(): parser = OptionParser() parser.add_option("-k", "--kitsune", help="Location of kitsune (required)") parser.add_option("-u", "--user", help=("Prefix cron with this user. " "Only define for cron.d style crontabs")) parser.add_option("-p", "--python", default="/usr/bin/python2.6", help="Python interpreter to use") (opts, args) = parser.parse_args() if not opts.kitsune: parser.error("-k must be defined") ctx = {'django': 'cd %s; %s manage.py' % (opts.kitsune, opts.python),} ctx['cron'] = '%s cron' % ctx['django'] if opts.user: for k, v in ctx.iteritems(): ctx[k] = '%s %s' % (opts.user, v) # Needs to stay below the opts.user injection. ctx['python'] = opts.python print Template(TEMPLATE).render(**ctx) if __name__ == "__main__": main() Add local site-packages to PYTHONPATH. To pick up the local version of PyOpenSSL.
#!/usr/bin/env python import os from optparse import OptionParser from jinja2 import Template TEMPLATE = open(os.path.join(os.path.dirname(__file__), 'crontab.tpl')).read() def main(): parser = OptionParser() parser.add_option("-k", "--kitsune", help="Location of kitsune (required)") parser.add_option("-u", "--user", help=("Prefix cron with this user. " "Only define for cron.d style crontabs")) parser.add_option("-p", "--python", default="/usr/bin/python2.6", help="Python interpreter to use") (opts, args) = parser.parse_args() if not opts.kitsune: parser.error("-k must be defined") # To pick up the right PyOpenSSL: python_path = 'PYTHONPATH=/usr/local/lib64/python2.6/site-packages' ctx = {'django': 'cd %s; %s %s manage.py' % ( opts.kitsune, python_path, opts.python),} ctx['cron'] = '%s cron' % ctx['django'] if opts.user: for k, v in ctx.iteritems(): ctx[k] = '%s %s' % (opts.user, v) # Needs to stay below the opts.user injection. ctx['python'] = opts.python print Template(TEMPLATE).render(**ctx) if __name__ == "__main__": main()
<commit_before>#!/usr/bin/env python import os from optparse import OptionParser from jinja2 import Template TEMPLATE = open(os.path.join(os.path.dirname(__file__), 'crontab.tpl')).read() def main(): parser = OptionParser() parser.add_option("-k", "--kitsune", help="Location of kitsune (required)") parser.add_option("-u", "--user", help=("Prefix cron with this user. " "Only define for cron.d style crontabs")) parser.add_option("-p", "--python", default="/usr/bin/python2.6", help="Python interpreter to use") (opts, args) = parser.parse_args() if not opts.kitsune: parser.error("-k must be defined") ctx = {'django': 'cd %s; %s manage.py' % (opts.kitsune, opts.python),} ctx['cron'] = '%s cron' % ctx['django'] if opts.user: for k, v in ctx.iteritems(): ctx[k] = '%s %s' % (opts.user, v) # Needs to stay below the opts.user injection. ctx['python'] = opts.python print Template(TEMPLATE).render(**ctx) if __name__ == "__main__": main() <commit_msg>Add local site-packages to PYTHONPATH. To pick up the local version of PyOpenSSL.<commit_after>
#!/usr/bin/env python import os from optparse import OptionParser from jinja2 import Template TEMPLATE = open(os.path.join(os.path.dirname(__file__), 'crontab.tpl')).read() def main(): parser = OptionParser() parser.add_option("-k", "--kitsune", help="Location of kitsune (required)") parser.add_option("-u", "--user", help=("Prefix cron with this user. " "Only define for cron.d style crontabs")) parser.add_option("-p", "--python", default="/usr/bin/python2.6", help="Python interpreter to use") (opts, args) = parser.parse_args() if not opts.kitsune: parser.error("-k must be defined") # To pick up the right PyOpenSSL: python_path = 'PYTHONPATH=/usr/local/lib64/python2.6/site-packages' ctx = {'django': 'cd %s; %s %s manage.py' % ( opts.kitsune, python_path, opts.python),} ctx['cron'] = '%s cron' % ctx['django'] if opts.user: for k, v in ctx.iteritems(): ctx[k] = '%s %s' % (opts.user, v) # Needs to stay below the opts.user injection. ctx['python'] = opts.python print Template(TEMPLATE).render(**ctx) if __name__ == "__main__": main()
#!/usr/bin/env python import os from optparse import OptionParser from jinja2 import Template TEMPLATE = open(os.path.join(os.path.dirname(__file__), 'crontab.tpl')).read() def main(): parser = OptionParser() parser.add_option("-k", "--kitsune", help="Location of kitsune (required)") parser.add_option("-u", "--user", help=("Prefix cron with this user. " "Only define for cron.d style crontabs")) parser.add_option("-p", "--python", default="/usr/bin/python2.6", help="Python interpreter to use") (opts, args) = parser.parse_args() if not opts.kitsune: parser.error("-k must be defined") ctx = {'django': 'cd %s; %s manage.py' % (opts.kitsune, opts.python),} ctx['cron'] = '%s cron' % ctx['django'] if opts.user: for k, v in ctx.iteritems(): ctx[k] = '%s %s' % (opts.user, v) # Needs to stay below the opts.user injection. ctx['python'] = opts.python print Template(TEMPLATE).render(**ctx) if __name__ == "__main__": main() Add local site-packages to PYTHONPATH. To pick up the local version of PyOpenSSL.#!/usr/bin/env python import os from optparse import OptionParser from jinja2 import Template TEMPLATE = open(os.path.join(os.path.dirname(__file__), 'crontab.tpl')).read() def main(): parser = OptionParser() parser.add_option("-k", "--kitsune", help="Location of kitsune (required)") parser.add_option("-u", "--user", help=("Prefix cron with this user. " "Only define for cron.d style crontabs")) parser.add_option("-p", "--python", default="/usr/bin/python2.6", help="Python interpreter to use") (opts, args) = parser.parse_args() if not opts.kitsune: parser.error("-k must be defined") # To pick up the right PyOpenSSL: python_path = 'PYTHONPATH=/usr/local/lib64/python2.6/site-packages' ctx = {'django': 'cd %s; %s %s manage.py' % ( opts.kitsune, python_path, opts.python),} ctx['cron'] = '%s cron' % ctx['django'] if opts.user: for k, v in ctx.iteritems(): ctx[k] = '%s %s' % (opts.user, v) # Needs to stay below the opts.user injection. ctx['python'] = opts.python print Template(TEMPLATE).render(**ctx) if __name__ == "__main__": main()
<commit_before>#!/usr/bin/env python import os from optparse import OptionParser from jinja2 import Template TEMPLATE = open(os.path.join(os.path.dirname(__file__), 'crontab.tpl')).read() def main(): parser = OptionParser() parser.add_option("-k", "--kitsune", help="Location of kitsune (required)") parser.add_option("-u", "--user", help=("Prefix cron with this user. " "Only define for cron.d style crontabs")) parser.add_option("-p", "--python", default="/usr/bin/python2.6", help="Python interpreter to use") (opts, args) = parser.parse_args() if not opts.kitsune: parser.error("-k must be defined") ctx = {'django': 'cd %s; %s manage.py' % (opts.kitsune, opts.python),} ctx['cron'] = '%s cron' % ctx['django'] if opts.user: for k, v in ctx.iteritems(): ctx[k] = '%s %s' % (opts.user, v) # Needs to stay below the opts.user injection. ctx['python'] = opts.python print Template(TEMPLATE).render(**ctx) if __name__ == "__main__": main() <commit_msg>Add local site-packages to PYTHONPATH. To pick up the local version of PyOpenSSL.<commit_after>#!/usr/bin/env python import os from optparse import OptionParser from jinja2 import Template TEMPLATE = open(os.path.join(os.path.dirname(__file__), 'crontab.tpl')).read() def main(): parser = OptionParser() parser.add_option("-k", "--kitsune", help="Location of kitsune (required)") parser.add_option("-u", "--user", help=("Prefix cron with this user. " "Only define for cron.d style crontabs")) parser.add_option("-p", "--python", default="/usr/bin/python2.6", help="Python interpreter to use") (opts, args) = parser.parse_args() if not opts.kitsune: parser.error("-k must be defined") # To pick up the right PyOpenSSL: python_path = 'PYTHONPATH=/usr/local/lib64/python2.6/site-packages' ctx = {'django': 'cd %s; %s %s manage.py' % ( opts.kitsune, python_path, opts.python),} ctx['cron'] = '%s cron' % ctx['django'] if opts.user: for k, v in ctx.iteritems(): ctx[k] = '%s %s' % (opts.user, v) # Needs to stay below the opts.user injection. ctx['python'] = opts.python print Template(TEMPLATE).render(**ctx) if __name__ == "__main__": main()
49493f11e20e4cb033ae17d29bfa14fa8473d145
qutebrowser/.qutebrowser/config.py
qutebrowser/.qutebrowser/config.py
c.fonts.completion.category = "bold 16pt monospace" c.fonts.completion.entry = "16pt monospace" c.fonts.debug_console = "16pt monospace" c.fonts.downloads = "16pt monospace" c.fonts.hints = "16pt monospace" c.fonts.keyhint = "16pt monospace" c.fonts.messages.error = "16pt monospace" c.fonts.messages.info = "16pt monospace" c.fonts.messages.warning = "16pt monospace" c.fonts.monospace = "\"xos4 Terminus\", Terminus, monospace" c.fonts.prompts = "16pt sans-serif" c.fonts.statusbar = "16pt monospace" c.fonts.tabs = "16pt monospace"
# Adjust font c.fonts.completion.category = "bold 16pt monospace" c.fonts.completion.entry = "16pt monospace" c.fonts.debug_console = "16pt monospace" c.fonts.downloads = "16pt monospace" c.fonts.hints = "16pt monospace" c.fonts.keyhint = "16pt monospace" c.fonts.messages.error = "16pt monospace" c.fonts.messages.info = "16pt monospace" c.fonts.messages.warning = "16pt monospace" c.fonts.monospace = "\"xos4 Terminus\", Terminus, monospace" c.fonts.prompts = "16pt sans-serif" c.fonts.statusbar = "16pt monospace" c.fonts.tabs = "16pt monospace" # Play videos with mpv config.bind('e', 'spawn mpv {url}') config.bind('E', 'hint links spawn mpv {hint-url}')
Add mpv spawning to qutebrowser
Add mpv spawning to qutebrowser Signed-off-by: Tomas <5fad2aa041dd6f0fec7c9125f282be028fd2d7b0@Tomass-MacBook-Pro.local>
Python
mit
deathbeam/dotfiles,deathbeam/dotfiles,deathbeam/awesomedotrc,deathbeam/dotfiles,deathbeam/dotfiles
c.fonts.completion.category = "bold 16pt monospace" c.fonts.completion.entry = "16pt monospace" c.fonts.debug_console = "16pt monospace" c.fonts.downloads = "16pt monospace" c.fonts.hints = "16pt monospace" c.fonts.keyhint = "16pt monospace" c.fonts.messages.error = "16pt monospace" c.fonts.messages.info = "16pt monospace" c.fonts.messages.warning = "16pt monospace" c.fonts.monospace = "\"xos4 Terminus\", Terminus, monospace" c.fonts.prompts = "16pt sans-serif" c.fonts.statusbar = "16pt monospace" c.fonts.tabs = "16pt monospace" Add mpv spawning to qutebrowser Signed-off-by: Tomas <5fad2aa041dd6f0fec7c9125f282be028fd2d7b0@Tomass-MacBook-Pro.local>
# Adjust font c.fonts.completion.category = "bold 16pt monospace" c.fonts.completion.entry = "16pt monospace" c.fonts.debug_console = "16pt monospace" c.fonts.downloads = "16pt monospace" c.fonts.hints = "16pt monospace" c.fonts.keyhint = "16pt monospace" c.fonts.messages.error = "16pt monospace" c.fonts.messages.info = "16pt monospace" c.fonts.messages.warning = "16pt monospace" c.fonts.monospace = "\"xos4 Terminus\", Terminus, monospace" c.fonts.prompts = "16pt sans-serif" c.fonts.statusbar = "16pt monospace" c.fonts.tabs = "16pt monospace" # Play videos with mpv config.bind('e', 'spawn mpv {url}') config.bind('E', 'hint links spawn mpv {hint-url}')
<commit_before>c.fonts.completion.category = "bold 16pt monospace" c.fonts.completion.entry = "16pt monospace" c.fonts.debug_console = "16pt monospace" c.fonts.downloads = "16pt monospace" c.fonts.hints = "16pt monospace" c.fonts.keyhint = "16pt monospace" c.fonts.messages.error = "16pt monospace" c.fonts.messages.info = "16pt monospace" c.fonts.messages.warning = "16pt monospace" c.fonts.monospace = "\"xos4 Terminus\", Terminus, monospace" c.fonts.prompts = "16pt sans-serif" c.fonts.statusbar = "16pt monospace" c.fonts.tabs = "16pt monospace" <commit_msg>Add mpv spawning to qutebrowser Signed-off-by: Tomas <5fad2aa041dd6f0fec7c9125f282be028fd2d7b0@Tomass-MacBook-Pro.local><commit_after>
# Adjust font c.fonts.completion.category = "bold 16pt monospace" c.fonts.completion.entry = "16pt monospace" c.fonts.debug_console = "16pt monospace" c.fonts.downloads = "16pt monospace" c.fonts.hints = "16pt monospace" c.fonts.keyhint = "16pt monospace" c.fonts.messages.error = "16pt monospace" c.fonts.messages.info = "16pt monospace" c.fonts.messages.warning = "16pt monospace" c.fonts.monospace = "\"xos4 Terminus\", Terminus, monospace" c.fonts.prompts = "16pt sans-serif" c.fonts.statusbar = "16pt monospace" c.fonts.tabs = "16pt monospace" # Play videos with mpv config.bind('e', 'spawn mpv {url}') config.bind('E', 'hint links spawn mpv {hint-url}')
c.fonts.completion.category = "bold 16pt monospace" c.fonts.completion.entry = "16pt monospace" c.fonts.debug_console = "16pt monospace" c.fonts.downloads = "16pt monospace" c.fonts.hints = "16pt monospace" c.fonts.keyhint = "16pt monospace" c.fonts.messages.error = "16pt monospace" c.fonts.messages.info = "16pt monospace" c.fonts.messages.warning = "16pt monospace" c.fonts.monospace = "\"xos4 Terminus\", Terminus, monospace" c.fonts.prompts = "16pt sans-serif" c.fonts.statusbar = "16pt monospace" c.fonts.tabs = "16pt monospace" Add mpv spawning to qutebrowser Signed-off-by: Tomas <5fad2aa041dd6f0fec7c9125f282be028fd2d7b0@Tomass-MacBook-Pro.local># Adjust font c.fonts.completion.category = "bold 16pt monospace" c.fonts.completion.entry = "16pt monospace" c.fonts.debug_console = "16pt monospace" c.fonts.downloads = "16pt monospace" c.fonts.hints = "16pt monospace" c.fonts.keyhint = "16pt monospace" c.fonts.messages.error = "16pt monospace" c.fonts.messages.info = "16pt monospace" c.fonts.messages.warning = "16pt monospace" c.fonts.monospace = "\"xos4 Terminus\", Terminus, monospace" c.fonts.prompts = "16pt sans-serif" c.fonts.statusbar = "16pt monospace" c.fonts.tabs = "16pt monospace" # Play videos with mpv config.bind('e', 'spawn mpv {url}') config.bind('E', 'hint links spawn mpv {hint-url}')
<commit_before>c.fonts.completion.category = "bold 16pt monospace" c.fonts.completion.entry = "16pt monospace" c.fonts.debug_console = "16pt monospace" c.fonts.downloads = "16pt monospace" c.fonts.hints = "16pt monospace" c.fonts.keyhint = "16pt monospace" c.fonts.messages.error = "16pt monospace" c.fonts.messages.info = "16pt monospace" c.fonts.messages.warning = "16pt monospace" c.fonts.monospace = "\"xos4 Terminus\", Terminus, monospace" c.fonts.prompts = "16pt sans-serif" c.fonts.statusbar = "16pt monospace" c.fonts.tabs = "16pt monospace" <commit_msg>Add mpv spawning to qutebrowser Signed-off-by: Tomas <5fad2aa041dd6f0fec7c9125f282be028fd2d7b0@Tomass-MacBook-Pro.local><commit_after># Adjust font c.fonts.completion.category = "bold 16pt monospace" c.fonts.completion.entry = "16pt monospace" c.fonts.debug_console = "16pt monospace" c.fonts.downloads = "16pt monospace" c.fonts.hints = "16pt monospace" c.fonts.keyhint = "16pt monospace" c.fonts.messages.error = "16pt monospace" c.fonts.messages.info = "16pt monospace" c.fonts.messages.warning = "16pt monospace" c.fonts.monospace = "\"xos4 Terminus\", Terminus, monospace" c.fonts.prompts = "16pt sans-serif" c.fonts.statusbar = "16pt monospace" c.fonts.tabs = "16pt monospace" # Play videos with mpv config.bind('e', 'spawn mpv {url}') config.bind('E', 'hint links spawn mpv {hint-url}')
cac3099b9ab07d5ac2180e0b2796f55668ddda1e
generate_keyczart.py
generate_keyczart.py
import keyczar from keyczar import keyczart import os directory = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'keyset') if not os.listdir(directory): os.makedirs(directory) keyczart.main(['create','--location=keyset','--purpose=crypt','--name=crypt']) keyczart.main(['addkey','--location=keyset' ,'--status=primary']) else: print 'Keyset directory already has something in there. Skipping key generation.'
import keyczar from keyczar import keyczart import os directory = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'keyset') if not os.path.exists(directory): os.makedirs(directory) if not os.listdir(directory): keyczart.main(['create','--location=keyset','--purpose=crypt','--name=crypt']) keyczart.main(['addkey','--location=keyset' ,'--status=primary']) else: print 'Keyset directory already has something in there. Skipping key generation.'
Make the directory if it doesn't exist
Make the directory if it doesn't exist
Python
apache-2.0
arubdesu/Crypt-Server,arubdesu/Crypt-Server,arubdesu/Crypt-Server,grahamgilbert/Crypt-Server,squarit/Crypt-Server,grahamgilbert/Crypt-Server,squarit/Crypt-Server,squarit/Crypt-Server,grahamgilbert/Crypt-Server,squarit/Crypt-Server,grahamgilbert/Crypt-Server
import keyczar from keyczar import keyczart import os directory = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'keyset') if not os.listdir(directory): os.makedirs(directory) keyczart.main(['create','--location=keyset','--purpose=crypt','--name=crypt']) keyczart.main(['addkey','--location=keyset' ,'--status=primary']) else: print 'Keyset directory already has something in there. Skipping key generation.' Make the directory if it doesn't exist
import keyczar from keyczar import keyczart import os directory = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'keyset') if not os.path.exists(directory): os.makedirs(directory) if not os.listdir(directory): keyczart.main(['create','--location=keyset','--purpose=crypt','--name=crypt']) keyczart.main(['addkey','--location=keyset' ,'--status=primary']) else: print 'Keyset directory already has something in there. Skipping key generation.'
<commit_before>import keyczar from keyczar import keyczart import os directory = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'keyset') if not os.listdir(directory): os.makedirs(directory) keyczart.main(['create','--location=keyset','--purpose=crypt','--name=crypt']) keyczart.main(['addkey','--location=keyset' ,'--status=primary']) else: print 'Keyset directory already has something in there. Skipping key generation.' <commit_msg>Make the directory if it doesn't exist<commit_after>
import keyczar from keyczar import keyczart import os directory = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'keyset') if not os.path.exists(directory): os.makedirs(directory) if not os.listdir(directory): keyczart.main(['create','--location=keyset','--purpose=crypt','--name=crypt']) keyczart.main(['addkey','--location=keyset' ,'--status=primary']) else: print 'Keyset directory already has something in there. Skipping key generation.'
import keyczar from keyczar import keyczart import os directory = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'keyset') if not os.listdir(directory): os.makedirs(directory) keyczart.main(['create','--location=keyset','--purpose=crypt','--name=crypt']) keyczart.main(['addkey','--location=keyset' ,'--status=primary']) else: print 'Keyset directory already has something in there. Skipping key generation.' Make the directory if it doesn't existimport keyczar from keyczar import keyczart import os directory = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'keyset') if not os.path.exists(directory): os.makedirs(directory) if not os.listdir(directory): keyczart.main(['create','--location=keyset','--purpose=crypt','--name=crypt']) keyczart.main(['addkey','--location=keyset' ,'--status=primary']) else: print 'Keyset directory already has something in there. Skipping key generation.'
<commit_before>import keyczar from keyczar import keyczart import os directory = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'keyset') if not os.listdir(directory): os.makedirs(directory) keyczart.main(['create','--location=keyset','--purpose=crypt','--name=crypt']) keyczart.main(['addkey','--location=keyset' ,'--status=primary']) else: print 'Keyset directory already has something in there. Skipping key generation.' <commit_msg>Make the directory if it doesn't exist<commit_after>import keyczar from keyczar import keyczart import os directory = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'keyset') if not os.path.exists(directory): os.makedirs(directory) if not os.listdir(directory): keyczart.main(['create','--location=keyset','--purpose=crypt','--name=crypt']) keyczart.main(['addkey','--location=keyset' ,'--status=primary']) else: print 'Keyset directory already has something in there. Skipping key generation.'
62d5682fa3be9dfbae80b2acae9839cd1278dcb6
whatismyip.py
whatismyip.py
#! /usr/bin/python import requests from bs4 import BeautifulSoup def main(): r = requests.get('http://www.whatismyip.com') soup = BeautifulSoup(r.text) ip_address = '' for span in soup.find('div', 'the-ip'): ip_address += span.text print(ip_address) if __name__ == '__main__': main()
#! /usr/bin/python import requests from bs4 import BeautifulSoup def main(): r = requests.get('http://www.whatismyip.com') soup = BeautifulSoup(r.text, 'lxml') ip_address = '' for span in soup.find('div', 'the-ip'): ip_address += span.text print(ip_address) if __name__ == '__main__': main()
Use lxml as a parsing engine for bs4
Use lxml as a parsing engine for bs4
Python
apache-2.0
MichaelAquilina/whatismyip
#! /usr/bin/python import requests from bs4 import BeautifulSoup def main(): r = requests.get('http://www.whatismyip.com') soup = BeautifulSoup(r.text) ip_address = '' for span in soup.find('div', 'the-ip'): ip_address += span.text print(ip_address) if __name__ == '__main__': main() Use lxml as a parsing engine for bs4
#! /usr/bin/python import requests from bs4 import BeautifulSoup def main(): r = requests.get('http://www.whatismyip.com') soup = BeautifulSoup(r.text, 'lxml') ip_address = '' for span in soup.find('div', 'the-ip'): ip_address += span.text print(ip_address) if __name__ == '__main__': main()
<commit_before>#! /usr/bin/python import requests from bs4 import BeautifulSoup def main(): r = requests.get('http://www.whatismyip.com') soup = BeautifulSoup(r.text) ip_address = '' for span in soup.find('div', 'the-ip'): ip_address += span.text print(ip_address) if __name__ == '__main__': main() <commit_msg>Use lxml as a parsing engine for bs4<commit_after>
#! /usr/bin/python import requests from bs4 import BeautifulSoup def main(): r = requests.get('http://www.whatismyip.com') soup = BeautifulSoup(r.text, 'lxml') ip_address = '' for span in soup.find('div', 'the-ip'): ip_address += span.text print(ip_address) if __name__ == '__main__': main()
#! /usr/bin/python import requests from bs4 import BeautifulSoup def main(): r = requests.get('http://www.whatismyip.com') soup = BeautifulSoup(r.text) ip_address = '' for span in soup.find('div', 'the-ip'): ip_address += span.text print(ip_address) if __name__ == '__main__': main() Use lxml as a parsing engine for bs4#! /usr/bin/python import requests from bs4 import BeautifulSoup def main(): r = requests.get('http://www.whatismyip.com') soup = BeautifulSoup(r.text, 'lxml') ip_address = '' for span in soup.find('div', 'the-ip'): ip_address += span.text print(ip_address) if __name__ == '__main__': main()
<commit_before>#! /usr/bin/python import requests from bs4 import BeautifulSoup def main(): r = requests.get('http://www.whatismyip.com') soup = BeautifulSoup(r.text) ip_address = '' for span in soup.find('div', 'the-ip'): ip_address += span.text print(ip_address) if __name__ == '__main__': main() <commit_msg>Use lxml as a parsing engine for bs4<commit_after>#! /usr/bin/python import requests from bs4 import BeautifulSoup def main(): r = requests.get('http://www.whatismyip.com') soup = BeautifulSoup(r.text, 'lxml') ip_address = '' for span in soup.find('div', 'the-ip'): ip_address += span.text print(ip_address) if __name__ == '__main__': main()
c485876017c92ac01af733945f04d36a21da8a6d
newsroom/settings.py
newsroom/settings.py
from django.conf import settings ARTICLE_COPYRIGHT = getattr(settings, 'NEWSROOM_ARTICLE_COPYRIGHT', "") ARTICLES_PER_PAGE = getattr(settings, 'NEWSROOM_ARTICLES_PER_PAGE', 12) BEAUTIFUL_SOUP_PARSER = getattr(settings, 'NEWSROOM_BEAUTIFUL_SOUP_PARSER', "lxml") ARTICLE_SUMMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "big") ARTICLE_PRIMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "large") CACHE_PERIOD = getattr(settings, 'NEWSROOM_CACHE_PERIOD', 500) ADVERT_CODE = getattr(settings, 'NEWSROOM_ADVERT_CODE', '')
from django.conf import settings ARTICLE_COPYRIGHT = getattr(settings, 'NEWSROOM_ARTICLE_COPYRIGHT', "") ARTICLES_PER_PAGE = getattr(settings, 'NEWSROOM_ARTICLES_PER_PAGE', 16) BEAUTIFUL_SOUP_PARSER = getattr(settings, 'NEWSROOM_BEAUTIFUL_SOUP_PARSER', "lxml") ARTICLE_SUMMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "big") ARTICLE_PRIMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "large") CACHE_PERIOD = getattr(settings, 'NEWSROOM_CACHE_PERIOD', 500) ADVERT_CODE = getattr(settings, 'NEWSROOM_ADVERT_CODE', '')
Set default number of articles on page to 16, up from 12.
Set default number of articles on page to 16, up from 12.
Python
bsd-3-clause
groundupnews/gu,groundupnews/gu,groundupnews/gu,groundupnews/gu,groundupnews/gu
from django.conf import settings ARTICLE_COPYRIGHT = getattr(settings, 'NEWSROOM_ARTICLE_COPYRIGHT', "") ARTICLES_PER_PAGE = getattr(settings, 'NEWSROOM_ARTICLES_PER_PAGE', 12) BEAUTIFUL_SOUP_PARSER = getattr(settings, 'NEWSROOM_BEAUTIFUL_SOUP_PARSER', "lxml") ARTICLE_SUMMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "big") ARTICLE_PRIMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "large") CACHE_PERIOD = getattr(settings, 'NEWSROOM_CACHE_PERIOD', 500) ADVERT_CODE = getattr(settings, 'NEWSROOM_ADVERT_CODE', '') Set default number of articles on page to 16, up from 12.
from django.conf import settings ARTICLE_COPYRIGHT = getattr(settings, 'NEWSROOM_ARTICLE_COPYRIGHT', "") ARTICLES_PER_PAGE = getattr(settings, 'NEWSROOM_ARTICLES_PER_PAGE', 16) BEAUTIFUL_SOUP_PARSER = getattr(settings, 'NEWSROOM_BEAUTIFUL_SOUP_PARSER', "lxml") ARTICLE_SUMMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "big") ARTICLE_PRIMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "large") CACHE_PERIOD = getattr(settings, 'NEWSROOM_CACHE_PERIOD', 500) ADVERT_CODE = getattr(settings, 'NEWSROOM_ADVERT_CODE', '')
<commit_before>from django.conf import settings ARTICLE_COPYRIGHT = getattr(settings, 'NEWSROOM_ARTICLE_COPYRIGHT', "") ARTICLES_PER_PAGE = getattr(settings, 'NEWSROOM_ARTICLES_PER_PAGE', 12) BEAUTIFUL_SOUP_PARSER = getattr(settings, 'NEWSROOM_BEAUTIFUL_SOUP_PARSER', "lxml") ARTICLE_SUMMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "big") ARTICLE_PRIMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "large") CACHE_PERIOD = getattr(settings, 'NEWSROOM_CACHE_PERIOD', 500) ADVERT_CODE = getattr(settings, 'NEWSROOM_ADVERT_CODE', '') <commit_msg>Set default number of articles on page to 16, up from 12.<commit_after>
from django.conf import settings ARTICLE_COPYRIGHT = getattr(settings, 'NEWSROOM_ARTICLE_COPYRIGHT', "") ARTICLES_PER_PAGE = getattr(settings, 'NEWSROOM_ARTICLES_PER_PAGE', 16) BEAUTIFUL_SOUP_PARSER = getattr(settings, 'NEWSROOM_BEAUTIFUL_SOUP_PARSER', "lxml") ARTICLE_SUMMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "big") ARTICLE_PRIMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "large") CACHE_PERIOD = getattr(settings, 'NEWSROOM_CACHE_PERIOD', 500) ADVERT_CODE = getattr(settings, 'NEWSROOM_ADVERT_CODE', '')
from django.conf import settings ARTICLE_COPYRIGHT = getattr(settings, 'NEWSROOM_ARTICLE_COPYRIGHT', "") ARTICLES_PER_PAGE = getattr(settings, 'NEWSROOM_ARTICLES_PER_PAGE', 12) BEAUTIFUL_SOUP_PARSER = getattr(settings, 'NEWSROOM_BEAUTIFUL_SOUP_PARSER', "lxml") ARTICLE_SUMMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "big") ARTICLE_PRIMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "large") CACHE_PERIOD = getattr(settings, 'NEWSROOM_CACHE_PERIOD', 500) ADVERT_CODE = getattr(settings, 'NEWSROOM_ADVERT_CODE', '') Set default number of articles on page to 16, up from 12.from django.conf import settings ARTICLE_COPYRIGHT = getattr(settings, 'NEWSROOM_ARTICLE_COPYRIGHT', "") ARTICLES_PER_PAGE = getattr(settings, 'NEWSROOM_ARTICLES_PER_PAGE', 16) BEAUTIFUL_SOUP_PARSER = getattr(settings, 'NEWSROOM_BEAUTIFUL_SOUP_PARSER', "lxml") ARTICLE_SUMMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "big") ARTICLE_PRIMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "large") CACHE_PERIOD = getattr(settings, 'NEWSROOM_CACHE_PERIOD', 500) ADVERT_CODE = getattr(settings, 'NEWSROOM_ADVERT_CODE', '')
<commit_before>from django.conf import settings ARTICLE_COPYRIGHT = getattr(settings, 'NEWSROOM_ARTICLE_COPYRIGHT', "") ARTICLES_PER_PAGE = getattr(settings, 'NEWSROOM_ARTICLES_PER_PAGE', 12) BEAUTIFUL_SOUP_PARSER = getattr(settings, 'NEWSROOM_BEAUTIFUL_SOUP_PARSER', "lxml") ARTICLE_SUMMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "big") ARTICLE_PRIMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "large") CACHE_PERIOD = getattr(settings, 'NEWSROOM_CACHE_PERIOD', 500) ADVERT_CODE = getattr(settings, 'NEWSROOM_ADVERT_CODE', '') <commit_msg>Set default number of articles on page to 16, up from 12.<commit_after>from django.conf import settings ARTICLE_COPYRIGHT = getattr(settings, 'NEWSROOM_ARTICLE_COPYRIGHT', "") ARTICLES_PER_PAGE = getattr(settings, 'NEWSROOM_ARTICLES_PER_PAGE', 16) BEAUTIFUL_SOUP_PARSER = getattr(settings, 'NEWSROOM_BEAUTIFUL_SOUP_PARSER', "lxml") ARTICLE_SUMMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "big") ARTICLE_PRIMARY_IMAGE_SIZE = getattr(settings, 'NEWSROOM_ARTICLE_TEASER_IMAGE_SIZE', "large") CACHE_PERIOD = getattr(settings, 'NEWSROOM_CACHE_PERIOD', 500) ADVERT_CODE = getattr(settings, 'NEWSROOM_ADVERT_CODE', '')
9d0798904160f86d7f580dde3bfba8cc28b5a23f
troposphere/ram.py
troposphere/ram.py
# Copyright (c) 2012-2019, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from troposphere import Tags from . import AWSObject from .validators import boolean class ResourceShare(AWSObject): resource_type = "AWS::RAM::ResourceShare" props = { "AllowExternalPrincipals": (boolean, False), "Name": (str, True), "Principals": ([str], False), "ResourceArns": ([str], False), "Tags": (Tags, False), }
# Copyright (c) 2012-2021, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 39.3.0 from troposphere import Tags from . import AWSObject from .validators import boolean class ResourceShare(AWSObject): resource_type = "AWS::RAM::ResourceShare" props = { "AllowExternalPrincipals": (boolean, False), "Name": (str, True), "PermissionArns": ([str], False), "Principals": ([str], False), "ResourceArns": ([str], False), "Tags": (Tags, False), }
Update RAM per 2021-06-10 changes
Update RAM per 2021-06-10 changes
Python
bsd-2-clause
cloudtools/troposphere,cloudtools/troposphere
# Copyright (c) 2012-2019, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from troposphere import Tags from . import AWSObject from .validators import boolean class ResourceShare(AWSObject): resource_type = "AWS::RAM::ResourceShare" props = { "AllowExternalPrincipals": (boolean, False), "Name": (str, True), "Principals": ([str], False), "ResourceArns": ([str], False), "Tags": (Tags, False), } Update RAM per 2021-06-10 changes
# Copyright (c) 2012-2021, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 39.3.0 from troposphere import Tags from . import AWSObject from .validators import boolean class ResourceShare(AWSObject): resource_type = "AWS::RAM::ResourceShare" props = { "AllowExternalPrincipals": (boolean, False), "Name": (str, True), "PermissionArns": ([str], False), "Principals": ([str], False), "ResourceArns": ([str], False), "Tags": (Tags, False), }
<commit_before># Copyright (c) 2012-2019, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from troposphere import Tags from . import AWSObject from .validators import boolean class ResourceShare(AWSObject): resource_type = "AWS::RAM::ResourceShare" props = { "AllowExternalPrincipals": (boolean, False), "Name": (str, True), "Principals": ([str], False), "ResourceArns": ([str], False), "Tags": (Tags, False), } <commit_msg>Update RAM per 2021-06-10 changes<commit_after>
# Copyright (c) 2012-2021, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 39.3.0 from troposphere import Tags from . import AWSObject from .validators import boolean class ResourceShare(AWSObject): resource_type = "AWS::RAM::ResourceShare" props = { "AllowExternalPrincipals": (boolean, False), "Name": (str, True), "PermissionArns": ([str], False), "Principals": ([str], False), "ResourceArns": ([str], False), "Tags": (Tags, False), }
# Copyright (c) 2012-2019, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from troposphere import Tags from . import AWSObject from .validators import boolean class ResourceShare(AWSObject): resource_type = "AWS::RAM::ResourceShare" props = { "AllowExternalPrincipals": (boolean, False), "Name": (str, True), "Principals": ([str], False), "ResourceArns": ([str], False), "Tags": (Tags, False), } Update RAM per 2021-06-10 changes# Copyright (c) 2012-2021, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 39.3.0 from troposphere import Tags from . import AWSObject from .validators import boolean class ResourceShare(AWSObject): resource_type = "AWS::RAM::ResourceShare" props = { "AllowExternalPrincipals": (boolean, False), "Name": (str, True), "PermissionArns": ([str], False), "Principals": ([str], False), "ResourceArns": ([str], False), "Tags": (Tags, False), }
<commit_before># Copyright (c) 2012-2019, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. from troposphere import Tags from . import AWSObject from .validators import boolean class ResourceShare(AWSObject): resource_type = "AWS::RAM::ResourceShare" props = { "AllowExternalPrincipals": (boolean, False), "Name": (str, True), "Principals": ([str], False), "ResourceArns": ([str], False), "Tags": (Tags, False), } <commit_msg>Update RAM per 2021-06-10 changes<commit_after># Copyright (c) 2012-2021, Mark Peek <mark@peek.org> # All rights reserved. # # See LICENSE file for full license. # # *** Do not modify - this file is autogenerated *** # Resource specification version: 39.3.0 from troposphere import Tags from . import AWSObject from .validators import boolean class ResourceShare(AWSObject): resource_type = "AWS::RAM::ResourceShare" props = { "AllowExternalPrincipals": (boolean, False), "Name": (str, True), "PermissionArns": ([str], False), "Principals": ([str], False), "ResourceArns": ([str], False), "Tags": (Tags, False), }
eb4456b752313383a573bacfc102db9149ee1854
django_transfer/urls.py
django_transfer/urls.py
from __future__ import unicode_literals try: from django.conf.urls import patterns, url except ImportError: from django.conf.urls.defaults import patterns, url urlpatterns = patterns( '', url(r'^download/.*$', 'django_transfer.views.download', name='download'), url(r'^upload/$', 'django_transfer.views.upload', name='upload'), )
from __future__ import unicode_literals try: from django.conf.urls import url def patterns(*args): return args except ImportError: from django.conf.urls.defaults import patterns, url from django_transfer.views import download, upload urlpatterns = patterns( url(r'^download/.*$', download, name='download'), url(r'^upload/$', upload, name='upload'), )
Fix URL patterns for different Django versions.
Fix URL patterns for different Django versions.
Python
mit
smartfile/django-transfer
from __future__ import unicode_literals try: from django.conf.urls import patterns, url except ImportError: from django.conf.urls.defaults import patterns, url urlpatterns = patterns( '', url(r'^download/.*$', 'django_transfer.views.download', name='download'), url(r'^upload/$', 'django_transfer.views.upload', name='upload'), ) Fix URL patterns for different Django versions.
from __future__ import unicode_literals try: from django.conf.urls import url def patterns(*args): return args except ImportError: from django.conf.urls.defaults import patterns, url from django_transfer.views import download, upload urlpatterns = patterns( url(r'^download/.*$', download, name='download'), url(r'^upload/$', upload, name='upload'), )
<commit_before>from __future__ import unicode_literals try: from django.conf.urls import patterns, url except ImportError: from django.conf.urls.defaults import patterns, url urlpatterns = patterns( '', url(r'^download/.*$', 'django_transfer.views.download', name='download'), url(r'^upload/$', 'django_transfer.views.upload', name='upload'), ) <commit_msg>Fix URL patterns for different Django versions.<commit_after>
from __future__ import unicode_literals try: from django.conf.urls import url def patterns(*args): return args except ImportError: from django.conf.urls.defaults import patterns, url from django_transfer.views import download, upload urlpatterns = patterns( url(r'^download/.*$', download, name='download'), url(r'^upload/$', upload, name='upload'), )
from __future__ import unicode_literals try: from django.conf.urls import patterns, url except ImportError: from django.conf.urls.defaults import patterns, url urlpatterns = patterns( '', url(r'^download/.*$', 'django_transfer.views.download', name='download'), url(r'^upload/$', 'django_transfer.views.upload', name='upload'), ) Fix URL patterns for different Django versions.from __future__ import unicode_literals try: from django.conf.urls import url def patterns(*args): return args except ImportError: from django.conf.urls.defaults import patterns, url from django_transfer.views import download, upload urlpatterns = patterns( url(r'^download/.*$', download, name='download'), url(r'^upload/$', upload, name='upload'), )
<commit_before>from __future__ import unicode_literals try: from django.conf.urls import patterns, url except ImportError: from django.conf.urls.defaults import patterns, url urlpatterns = patterns( '', url(r'^download/.*$', 'django_transfer.views.download', name='download'), url(r'^upload/$', 'django_transfer.views.upload', name='upload'), ) <commit_msg>Fix URL patterns for different Django versions.<commit_after>from __future__ import unicode_literals try: from django.conf.urls import url def patterns(*args): return args except ImportError: from django.conf.urls.defaults import patterns, url from django_transfer.views import download, upload urlpatterns = patterns( url(r'^download/.*$', download, name='download'), url(r'^upload/$', upload, name='upload'), )
4cff7dd08fdb345b6e091570a2ca5500ef871318
flask_authorization_panda/basic_auth.py
flask_authorization_panda/basic_auth.py
""" Functions related to HTTP Basic Authorization """ from functools import wraps from flask import request, Response, current_app def basic_auth(original_function): """ Wrapper. Verify that request.authorization exists and that its contents match the application's config.basic_auth_credentials dict. Args: original_function (function): The function to wrap. Returns: flask.Response: When credentials are missing or don't match. original_function (function): The original function. """ @wraps(original_function) def decorated(*args, **kwargs): try: assert request.authorization.username == \ current_app.config['APP_USERNAME'] assert request.authorization.password == \ current_app.config['APP_PASSWORD'] except AttributeError: return Response( 'You must provide access credentials for this url.', 401, {'WWW-Authenticate': 'Basic'}) except AssertionError: return Response( 'Could not verify your access level for that URL.\n' 'You have to login with proper credentials', 401, {'WWW-Authenticate': 'Basic'}) return original_function(*args, **kwargs) return decorated
""" Functions related to HTTP Basic Authorization """ from functools import wraps from flask import request, jsonify, current_app def basic_auth(original_function): """ Wrapper. Verify that request.authorization exists and that its contents match the application's config.basic_auth_credentials dict. Args: original_function (function): The function to wrap. Returns: flask.Response: When credentials are missing or don't match. original_function (function): The original function. """ @wraps(original_function) def decorated(*args, **kwargs): try: if not (request.authorization.username, request.authorization.password) == ( current_app.config.basic_auth_credentials['username'], current_app.config.basic_auth_credentials['password']): unauthorized_response = jsonify( {'message': 'Could not verify your access level ' 'for that URL. \nYou have to login ' 'with proper credentials', 'statusCode': 401}) unauthorized_response.status_code = 401 return unauthorized_response except AttributeError: unauthorized_response = jsonify( {'message': 'Could not verify your access level ' 'for that URL. \nYou have to login ' 'with proper credentials', 'statusCode': 401}) unauthorized_response.status_code = 401 return original_function(*args, **kwargs) return decorated
Remove use of assert statements since this does not conform to general best practice.
Remove use of assert statements since this does not conform to general best practice. This is unfortunate, because the code is much more verbose than before and NOT as clear.
Python
mit
eikonomega/flask-authorization-panda
""" Functions related to HTTP Basic Authorization """ from functools import wraps from flask import request, Response, current_app def basic_auth(original_function): """ Wrapper. Verify that request.authorization exists and that its contents match the application's config.basic_auth_credentials dict. Args: original_function (function): The function to wrap. Returns: flask.Response: When credentials are missing or don't match. original_function (function): The original function. """ @wraps(original_function) def decorated(*args, **kwargs): try: assert request.authorization.username == \ current_app.config['APP_USERNAME'] assert request.authorization.password == \ current_app.config['APP_PASSWORD'] except AttributeError: return Response( 'You must provide access credentials for this url.', 401, {'WWW-Authenticate': 'Basic'}) except AssertionError: return Response( 'Could not verify your access level for that URL.\n' 'You have to login with proper credentials', 401, {'WWW-Authenticate': 'Basic'}) return original_function(*args, **kwargs) return decoratedRemove use of assert statements since this does not conform to general best practice. This is unfortunate, because the code is much more verbose than before and NOT as clear.
""" Functions related to HTTP Basic Authorization """ from functools import wraps from flask import request, jsonify, current_app def basic_auth(original_function): """ Wrapper. Verify that request.authorization exists and that its contents match the application's config.basic_auth_credentials dict. Args: original_function (function): The function to wrap. Returns: flask.Response: When credentials are missing or don't match. original_function (function): The original function. """ @wraps(original_function) def decorated(*args, **kwargs): try: if not (request.authorization.username, request.authorization.password) == ( current_app.config.basic_auth_credentials['username'], current_app.config.basic_auth_credentials['password']): unauthorized_response = jsonify( {'message': 'Could not verify your access level ' 'for that URL. \nYou have to login ' 'with proper credentials', 'statusCode': 401}) unauthorized_response.status_code = 401 return unauthorized_response except AttributeError: unauthorized_response = jsonify( {'message': 'Could not verify your access level ' 'for that URL. \nYou have to login ' 'with proper credentials', 'statusCode': 401}) unauthorized_response.status_code = 401 return original_function(*args, **kwargs) return decorated
<commit_before>""" Functions related to HTTP Basic Authorization """ from functools import wraps from flask import request, Response, current_app def basic_auth(original_function): """ Wrapper. Verify that request.authorization exists and that its contents match the application's config.basic_auth_credentials dict. Args: original_function (function): The function to wrap. Returns: flask.Response: When credentials are missing or don't match. original_function (function): The original function. """ @wraps(original_function) def decorated(*args, **kwargs): try: assert request.authorization.username == \ current_app.config['APP_USERNAME'] assert request.authorization.password == \ current_app.config['APP_PASSWORD'] except AttributeError: return Response( 'You must provide access credentials for this url.', 401, {'WWW-Authenticate': 'Basic'}) except AssertionError: return Response( 'Could not verify your access level for that URL.\n' 'You have to login with proper credentials', 401, {'WWW-Authenticate': 'Basic'}) return original_function(*args, **kwargs) return decorated<commit_msg>Remove use of assert statements since this does not conform to general best practice. This is unfortunate, because the code is much more verbose than before and NOT as clear.<commit_after>
""" Functions related to HTTP Basic Authorization """ from functools import wraps from flask import request, jsonify, current_app def basic_auth(original_function): """ Wrapper. Verify that request.authorization exists and that its contents match the application's config.basic_auth_credentials dict. Args: original_function (function): The function to wrap. Returns: flask.Response: When credentials are missing or don't match. original_function (function): The original function. """ @wraps(original_function) def decorated(*args, **kwargs): try: if not (request.authorization.username, request.authorization.password) == ( current_app.config.basic_auth_credentials['username'], current_app.config.basic_auth_credentials['password']): unauthorized_response = jsonify( {'message': 'Could not verify your access level ' 'for that URL. \nYou have to login ' 'with proper credentials', 'statusCode': 401}) unauthorized_response.status_code = 401 return unauthorized_response except AttributeError: unauthorized_response = jsonify( {'message': 'Could not verify your access level ' 'for that URL. \nYou have to login ' 'with proper credentials', 'statusCode': 401}) unauthorized_response.status_code = 401 return original_function(*args, **kwargs) return decorated
""" Functions related to HTTP Basic Authorization """ from functools import wraps from flask import request, Response, current_app def basic_auth(original_function): """ Wrapper. Verify that request.authorization exists and that its contents match the application's config.basic_auth_credentials dict. Args: original_function (function): The function to wrap. Returns: flask.Response: When credentials are missing or don't match. original_function (function): The original function. """ @wraps(original_function) def decorated(*args, **kwargs): try: assert request.authorization.username == \ current_app.config['APP_USERNAME'] assert request.authorization.password == \ current_app.config['APP_PASSWORD'] except AttributeError: return Response( 'You must provide access credentials for this url.', 401, {'WWW-Authenticate': 'Basic'}) except AssertionError: return Response( 'Could not verify your access level for that URL.\n' 'You have to login with proper credentials', 401, {'WWW-Authenticate': 'Basic'}) return original_function(*args, **kwargs) return decoratedRemove use of assert statements since this does not conform to general best practice. This is unfortunate, because the code is much more verbose than before and NOT as clear.""" Functions related to HTTP Basic Authorization """ from functools import wraps from flask import request, jsonify, current_app def basic_auth(original_function): """ Wrapper. Verify that request.authorization exists and that its contents match the application's config.basic_auth_credentials dict. Args: original_function (function): The function to wrap. Returns: flask.Response: When credentials are missing or don't match. original_function (function): The original function. """ @wraps(original_function) def decorated(*args, **kwargs): try: if not (request.authorization.username, request.authorization.password) == ( current_app.config.basic_auth_credentials['username'], current_app.config.basic_auth_credentials['password']): unauthorized_response = jsonify( {'message': 'Could not verify your access level ' 'for that URL. \nYou have to login ' 'with proper credentials', 'statusCode': 401}) unauthorized_response.status_code = 401 return unauthorized_response except AttributeError: unauthorized_response = jsonify( {'message': 'Could not verify your access level ' 'for that URL. \nYou have to login ' 'with proper credentials', 'statusCode': 401}) unauthorized_response.status_code = 401 return original_function(*args, **kwargs) return decorated
<commit_before>""" Functions related to HTTP Basic Authorization """ from functools import wraps from flask import request, Response, current_app def basic_auth(original_function): """ Wrapper. Verify that request.authorization exists and that its contents match the application's config.basic_auth_credentials dict. Args: original_function (function): The function to wrap. Returns: flask.Response: When credentials are missing or don't match. original_function (function): The original function. """ @wraps(original_function) def decorated(*args, **kwargs): try: assert request.authorization.username == \ current_app.config['APP_USERNAME'] assert request.authorization.password == \ current_app.config['APP_PASSWORD'] except AttributeError: return Response( 'You must provide access credentials for this url.', 401, {'WWW-Authenticate': 'Basic'}) except AssertionError: return Response( 'Could not verify your access level for that URL.\n' 'You have to login with proper credentials', 401, {'WWW-Authenticate': 'Basic'}) return original_function(*args, **kwargs) return decorated<commit_msg>Remove use of assert statements since this does not conform to general best practice. This is unfortunate, because the code is much more verbose than before and NOT as clear.<commit_after>""" Functions related to HTTP Basic Authorization """ from functools import wraps from flask import request, jsonify, current_app def basic_auth(original_function): """ Wrapper. Verify that request.authorization exists and that its contents match the application's config.basic_auth_credentials dict. Args: original_function (function): The function to wrap. Returns: flask.Response: When credentials are missing or don't match. original_function (function): The original function. """ @wraps(original_function) def decorated(*args, **kwargs): try: if not (request.authorization.username, request.authorization.password) == ( current_app.config.basic_auth_credentials['username'], current_app.config.basic_auth_credentials['password']): unauthorized_response = jsonify( {'message': 'Could not verify your access level ' 'for that URL. \nYou have to login ' 'with proper credentials', 'statusCode': 401}) unauthorized_response.status_code = 401 return unauthorized_response except AttributeError: unauthorized_response = jsonify( {'message': 'Could not verify your access level ' 'for that URL. \nYou have to login ' 'with proper credentials', 'statusCode': 401}) unauthorized_response.status_code = 401 return original_function(*args, **kwargs) return decorated
e33a2879fbafe36c8c29d48042dad8277b068e91
umap/tests/test_plot.py
umap/tests/test_plot.py
from nose.tools import assert_less from nose.tools import assert_greater_equal import os.path import numpy as np from nose import SkipTest from sklearn import datasets import umap import umap.plot np.random.seed(42) iris = datasets.load_iris() mapper = umap.UMAP(n_epochs=100).fit(iris.data) def test_plot_runs_at_all(): umap.plot.points(mapper, labels=iris.target) umap.plot.points(mapper, values=iris.data[:, 0]) umap.plot.diagnostic(mapper, diagnostic_type="all") umap.plot.connectivity(mapper)
from nose.tools import assert_less from nose.tools import assert_greater_equal import os.path import numpy as np from nose import SkipTest from sklearn import datasets import umap import umap.plot np.random.seed(42) iris = datasets.load_iris() mapper = umap.UMAP(n_epochs=100).fit(iris.data) def test_plot_runs_at_all(): umap.plot.points(mapper) umap.plot.points(mapper, labels=iris.target) umap.plot.points(mapper, values=iris.data[:, 0]) umap.plot.points(mapper, theme='fire') umap.plot.diagnostic(mapper, diagnostic_type='all') umap.plot.diagnostic(mapper, diagnostic_type='neighborhood') umap.plot.connectivity(mapper) umap.plot.interactive(mapper) umap.plot.interactive(mapper, labels=iris.target) umap.plot.interactive(mapper, values=iris.data[:, 0]) umap.plot.interactive(mapper, theme='fire') umap.plot._datashade_points(mapper.embedding_)
Add more basic plotting tests
Add more basic plotting tests
Python
bsd-3-clause
lmcinnes/umap,lmcinnes/umap
from nose.tools import assert_less from nose.tools import assert_greater_equal import os.path import numpy as np from nose import SkipTest from sklearn import datasets import umap import umap.plot np.random.seed(42) iris = datasets.load_iris() mapper = umap.UMAP(n_epochs=100).fit(iris.data) def test_plot_runs_at_all(): umap.plot.points(mapper, labels=iris.target) umap.plot.points(mapper, values=iris.data[:, 0]) umap.plot.diagnostic(mapper, diagnostic_type="all") umap.plot.connectivity(mapper) Add more basic plotting tests
from nose.tools import assert_less from nose.tools import assert_greater_equal import os.path import numpy as np from nose import SkipTest from sklearn import datasets import umap import umap.plot np.random.seed(42) iris = datasets.load_iris() mapper = umap.UMAP(n_epochs=100).fit(iris.data) def test_plot_runs_at_all(): umap.plot.points(mapper) umap.plot.points(mapper, labels=iris.target) umap.plot.points(mapper, values=iris.data[:, 0]) umap.plot.points(mapper, theme='fire') umap.plot.diagnostic(mapper, diagnostic_type='all') umap.plot.diagnostic(mapper, diagnostic_type='neighborhood') umap.plot.connectivity(mapper) umap.plot.interactive(mapper) umap.plot.interactive(mapper, labels=iris.target) umap.plot.interactive(mapper, values=iris.data[:, 0]) umap.plot.interactive(mapper, theme='fire') umap.plot._datashade_points(mapper.embedding_)
<commit_before>from nose.tools import assert_less from nose.tools import assert_greater_equal import os.path import numpy as np from nose import SkipTest from sklearn import datasets import umap import umap.plot np.random.seed(42) iris = datasets.load_iris() mapper = umap.UMAP(n_epochs=100).fit(iris.data) def test_plot_runs_at_all(): umap.plot.points(mapper, labels=iris.target) umap.plot.points(mapper, values=iris.data[:, 0]) umap.plot.diagnostic(mapper, diagnostic_type="all") umap.plot.connectivity(mapper) <commit_msg>Add more basic plotting tests<commit_after>
from nose.tools import assert_less from nose.tools import assert_greater_equal import os.path import numpy as np from nose import SkipTest from sklearn import datasets import umap import umap.plot np.random.seed(42) iris = datasets.load_iris() mapper = umap.UMAP(n_epochs=100).fit(iris.data) def test_plot_runs_at_all(): umap.plot.points(mapper) umap.plot.points(mapper, labels=iris.target) umap.plot.points(mapper, values=iris.data[:, 0]) umap.plot.points(mapper, theme='fire') umap.plot.diagnostic(mapper, diagnostic_type='all') umap.plot.diagnostic(mapper, diagnostic_type='neighborhood') umap.plot.connectivity(mapper) umap.plot.interactive(mapper) umap.plot.interactive(mapper, labels=iris.target) umap.plot.interactive(mapper, values=iris.data[:, 0]) umap.plot.interactive(mapper, theme='fire') umap.plot._datashade_points(mapper.embedding_)
from nose.tools import assert_less from nose.tools import assert_greater_equal import os.path import numpy as np from nose import SkipTest from sklearn import datasets import umap import umap.plot np.random.seed(42) iris = datasets.load_iris() mapper = umap.UMAP(n_epochs=100).fit(iris.data) def test_plot_runs_at_all(): umap.plot.points(mapper, labels=iris.target) umap.plot.points(mapper, values=iris.data[:, 0]) umap.plot.diagnostic(mapper, diagnostic_type="all") umap.plot.connectivity(mapper) Add more basic plotting testsfrom nose.tools import assert_less from nose.tools import assert_greater_equal import os.path import numpy as np from nose import SkipTest from sklearn import datasets import umap import umap.plot np.random.seed(42) iris = datasets.load_iris() mapper = umap.UMAP(n_epochs=100).fit(iris.data) def test_plot_runs_at_all(): umap.plot.points(mapper) umap.plot.points(mapper, labels=iris.target) umap.plot.points(mapper, values=iris.data[:, 0]) umap.plot.points(mapper, theme='fire') umap.plot.diagnostic(mapper, diagnostic_type='all') umap.plot.diagnostic(mapper, diagnostic_type='neighborhood') umap.plot.connectivity(mapper) umap.plot.interactive(mapper) umap.plot.interactive(mapper, labels=iris.target) umap.plot.interactive(mapper, values=iris.data[:, 0]) umap.plot.interactive(mapper, theme='fire') umap.plot._datashade_points(mapper.embedding_)
<commit_before>from nose.tools import assert_less from nose.tools import assert_greater_equal import os.path import numpy as np from nose import SkipTest from sklearn import datasets import umap import umap.plot np.random.seed(42) iris = datasets.load_iris() mapper = umap.UMAP(n_epochs=100).fit(iris.data) def test_plot_runs_at_all(): umap.plot.points(mapper, labels=iris.target) umap.plot.points(mapper, values=iris.data[:, 0]) umap.plot.diagnostic(mapper, diagnostic_type="all") umap.plot.connectivity(mapper) <commit_msg>Add more basic plotting tests<commit_after>from nose.tools import assert_less from nose.tools import assert_greater_equal import os.path import numpy as np from nose import SkipTest from sklearn import datasets import umap import umap.plot np.random.seed(42) iris = datasets.load_iris() mapper = umap.UMAP(n_epochs=100).fit(iris.data) def test_plot_runs_at_all(): umap.plot.points(mapper) umap.plot.points(mapper, labels=iris.target) umap.plot.points(mapper, values=iris.data[:, 0]) umap.plot.points(mapper, theme='fire') umap.plot.diagnostic(mapper, diagnostic_type='all') umap.plot.diagnostic(mapper, diagnostic_type='neighborhood') umap.plot.connectivity(mapper) umap.plot.interactive(mapper) umap.plot.interactive(mapper, labels=iris.target) umap.plot.interactive(mapper, values=iris.data[:, 0]) umap.plot.interactive(mapper, theme='fire') umap.plot._datashade_points(mapper.embedding_)
bf23c87c7606cbcf8afcd2c5120c25ede92675c0
irc/functools.py
irc/functools.py
from __future__ import absolute_import, print_function, unicode_literals import functools import collections def save_method_args(method): """ Wrap a method such that when it is called, we save the args and kwargs with which it was called. >>> class MyClass(object): ... @save_method_args ... def method(self, a, b): ... print(a, b) >>> my_ob = MyClass() >>> my_ob.method(1, 2) 1 2 >>> my_ob._saved_method.args (1, 2) >>> my_ob._saved_method.kwargs {} >>> my_ob.method(a=3, b='foo') 3 foo >>> my_ob._saved_method.args () >>> my_ob._saved_method.kwargs == dict(a=3, b='foo') True """ args_and_kwargs = collections.namedtuple('args_and_kwargs', 'args kwargs') @functools.wraps(method) def wrapper(self, *args, **kwargs): attr_name = '_saved_' + method.__name__ attr = args_and_kwargs(args, kwargs) setattr(self, attr_name, attr) return method(self, *args, **kwargs) return wrapper
from __future__ import absolute_import, print_function, unicode_literals import functools import collections def save_method_args(method): """ Wrap a method such that when it is called, the args and kwargs are saved on the method. >>> class MyClass(object): ... @save_method_args ... def method(self, a, b): ... print(a, b) >>> my_ob = MyClass() >>> my_ob.method(1, 2) 1 2 >>> my_ob._saved_method.args (1, 2) >>> my_ob._saved_method.kwargs {} >>> my_ob.method(a=3, b='foo') 3 foo >>> my_ob._saved_method.args () >>> my_ob._saved_method.kwargs == dict(a=3, b='foo') True The arguments are stored on the instance, allowing for different instance to save different args. >>> your_ob = MyClass() >>> your_ob.method({3}, b=[4]) {3} [4] >>> your_ob._saved_method.args ({3},) >>> my_ob._saved_method.args () """ args_and_kwargs = collections.namedtuple('args_and_kwargs', 'args kwargs') @functools.wraps(method) def wrapper(self, *args, **kwargs): attr_name = '_saved_' + method.__name__ attr = args_and_kwargs(args, kwargs) setattr(self, attr_name, attr) return method(self, *args, **kwargs) return wrapper
Update docstring and expand doctests.
Update docstring and expand doctests.
Python
mit
jaraco/irc
from __future__ import absolute_import, print_function, unicode_literals import functools import collections def save_method_args(method): """ Wrap a method such that when it is called, we save the args and kwargs with which it was called. >>> class MyClass(object): ... @save_method_args ... def method(self, a, b): ... print(a, b) >>> my_ob = MyClass() >>> my_ob.method(1, 2) 1 2 >>> my_ob._saved_method.args (1, 2) >>> my_ob._saved_method.kwargs {} >>> my_ob.method(a=3, b='foo') 3 foo >>> my_ob._saved_method.args () >>> my_ob._saved_method.kwargs == dict(a=3, b='foo') True """ args_and_kwargs = collections.namedtuple('args_and_kwargs', 'args kwargs') @functools.wraps(method) def wrapper(self, *args, **kwargs): attr_name = '_saved_' + method.__name__ attr = args_and_kwargs(args, kwargs) setattr(self, attr_name, attr) return method(self, *args, **kwargs) return wrapper Update docstring and expand doctests.
from __future__ import absolute_import, print_function, unicode_literals import functools import collections def save_method_args(method): """ Wrap a method such that when it is called, the args and kwargs are saved on the method. >>> class MyClass(object): ... @save_method_args ... def method(self, a, b): ... print(a, b) >>> my_ob = MyClass() >>> my_ob.method(1, 2) 1 2 >>> my_ob._saved_method.args (1, 2) >>> my_ob._saved_method.kwargs {} >>> my_ob.method(a=3, b='foo') 3 foo >>> my_ob._saved_method.args () >>> my_ob._saved_method.kwargs == dict(a=3, b='foo') True The arguments are stored on the instance, allowing for different instance to save different args. >>> your_ob = MyClass() >>> your_ob.method({3}, b=[4]) {3} [4] >>> your_ob._saved_method.args ({3},) >>> my_ob._saved_method.args () """ args_and_kwargs = collections.namedtuple('args_and_kwargs', 'args kwargs') @functools.wraps(method) def wrapper(self, *args, **kwargs): attr_name = '_saved_' + method.__name__ attr = args_and_kwargs(args, kwargs) setattr(self, attr_name, attr) return method(self, *args, **kwargs) return wrapper
<commit_before>from __future__ import absolute_import, print_function, unicode_literals import functools import collections def save_method_args(method): """ Wrap a method such that when it is called, we save the args and kwargs with which it was called. >>> class MyClass(object): ... @save_method_args ... def method(self, a, b): ... print(a, b) >>> my_ob = MyClass() >>> my_ob.method(1, 2) 1 2 >>> my_ob._saved_method.args (1, 2) >>> my_ob._saved_method.kwargs {} >>> my_ob.method(a=3, b='foo') 3 foo >>> my_ob._saved_method.args () >>> my_ob._saved_method.kwargs == dict(a=3, b='foo') True """ args_and_kwargs = collections.namedtuple('args_and_kwargs', 'args kwargs') @functools.wraps(method) def wrapper(self, *args, **kwargs): attr_name = '_saved_' + method.__name__ attr = args_and_kwargs(args, kwargs) setattr(self, attr_name, attr) return method(self, *args, **kwargs) return wrapper <commit_msg>Update docstring and expand doctests.<commit_after>
from __future__ import absolute_import, print_function, unicode_literals import functools import collections def save_method_args(method): """ Wrap a method such that when it is called, the args and kwargs are saved on the method. >>> class MyClass(object): ... @save_method_args ... def method(self, a, b): ... print(a, b) >>> my_ob = MyClass() >>> my_ob.method(1, 2) 1 2 >>> my_ob._saved_method.args (1, 2) >>> my_ob._saved_method.kwargs {} >>> my_ob.method(a=3, b='foo') 3 foo >>> my_ob._saved_method.args () >>> my_ob._saved_method.kwargs == dict(a=3, b='foo') True The arguments are stored on the instance, allowing for different instance to save different args. >>> your_ob = MyClass() >>> your_ob.method({3}, b=[4]) {3} [4] >>> your_ob._saved_method.args ({3},) >>> my_ob._saved_method.args () """ args_and_kwargs = collections.namedtuple('args_and_kwargs', 'args kwargs') @functools.wraps(method) def wrapper(self, *args, **kwargs): attr_name = '_saved_' + method.__name__ attr = args_and_kwargs(args, kwargs) setattr(self, attr_name, attr) return method(self, *args, **kwargs) return wrapper
from __future__ import absolute_import, print_function, unicode_literals import functools import collections def save_method_args(method): """ Wrap a method such that when it is called, we save the args and kwargs with which it was called. >>> class MyClass(object): ... @save_method_args ... def method(self, a, b): ... print(a, b) >>> my_ob = MyClass() >>> my_ob.method(1, 2) 1 2 >>> my_ob._saved_method.args (1, 2) >>> my_ob._saved_method.kwargs {} >>> my_ob.method(a=3, b='foo') 3 foo >>> my_ob._saved_method.args () >>> my_ob._saved_method.kwargs == dict(a=3, b='foo') True """ args_and_kwargs = collections.namedtuple('args_and_kwargs', 'args kwargs') @functools.wraps(method) def wrapper(self, *args, **kwargs): attr_name = '_saved_' + method.__name__ attr = args_and_kwargs(args, kwargs) setattr(self, attr_name, attr) return method(self, *args, **kwargs) return wrapper Update docstring and expand doctests.from __future__ import absolute_import, print_function, unicode_literals import functools import collections def save_method_args(method): """ Wrap a method such that when it is called, the args and kwargs are saved on the method. >>> class MyClass(object): ... @save_method_args ... def method(self, a, b): ... print(a, b) >>> my_ob = MyClass() >>> my_ob.method(1, 2) 1 2 >>> my_ob._saved_method.args (1, 2) >>> my_ob._saved_method.kwargs {} >>> my_ob.method(a=3, b='foo') 3 foo >>> my_ob._saved_method.args () >>> my_ob._saved_method.kwargs == dict(a=3, b='foo') True The arguments are stored on the instance, allowing for different instance to save different args. >>> your_ob = MyClass() >>> your_ob.method({3}, b=[4]) {3} [4] >>> your_ob._saved_method.args ({3},) >>> my_ob._saved_method.args () """ args_and_kwargs = collections.namedtuple('args_and_kwargs', 'args kwargs') @functools.wraps(method) def wrapper(self, *args, **kwargs): attr_name = '_saved_' + method.__name__ attr = args_and_kwargs(args, kwargs) setattr(self, attr_name, attr) return method(self, *args, **kwargs) return wrapper
<commit_before>from __future__ import absolute_import, print_function, unicode_literals import functools import collections def save_method_args(method): """ Wrap a method such that when it is called, we save the args and kwargs with which it was called. >>> class MyClass(object): ... @save_method_args ... def method(self, a, b): ... print(a, b) >>> my_ob = MyClass() >>> my_ob.method(1, 2) 1 2 >>> my_ob._saved_method.args (1, 2) >>> my_ob._saved_method.kwargs {} >>> my_ob.method(a=3, b='foo') 3 foo >>> my_ob._saved_method.args () >>> my_ob._saved_method.kwargs == dict(a=3, b='foo') True """ args_and_kwargs = collections.namedtuple('args_and_kwargs', 'args kwargs') @functools.wraps(method) def wrapper(self, *args, **kwargs): attr_name = '_saved_' + method.__name__ attr = args_and_kwargs(args, kwargs) setattr(self, attr_name, attr) return method(self, *args, **kwargs) return wrapper <commit_msg>Update docstring and expand doctests.<commit_after>from __future__ import absolute_import, print_function, unicode_literals import functools import collections def save_method_args(method): """ Wrap a method such that when it is called, the args and kwargs are saved on the method. >>> class MyClass(object): ... @save_method_args ... def method(self, a, b): ... print(a, b) >>> my_ob = MyClass() >>> my_ob.method(1, 2) 1 2 >>> my_ob._saved_method.args (1, 2) >>> my_ob._saved_method.kwargs {} >>> my_ob.method(a=3, b='foo') 3 foo >>> my_ob._saved_method.args () >>> my_ob._saved_method.kwargs == dict(a=3, b='foo') True The arguments are stored on the instance, allowing for different instance to save different args. >>> your_ob = MyClass() >>> your_ob.method({3}, b=[4]) {3} [4] >>> your_ob._saved_method.args ({3},) >>> my_ob._saved_method.args () """ args_and_kwargs = collections.namedtuple('args_and_kwargs', 'args kwargs') @functools.wraps(method) def wrapper(self, *args, **kwargs): attr_name = '_saved_' + method.__name__ attr = args_and_kwargs(args, kwargs) setattr(self, attr_name, attr) return method(self, *args, **kwargs) return wrapper
dfc7d629956f708cf5b69e464fe3a7298ffb6cfa
BotHandler.py
BotHandler.py
from twisted.internet import reactor from Hubbot import Hubbot, HubbotFactory import GlobalVars class BotHandler: botfactories = {} def __init__(self): for (server_with_port,channels) in GlobalVars.connections.items(): server = server_with_port.split(":")[0] port = int(server_with_port.split(":")[1]) self.startBotFactory(server, port, channels) GlobalVars.bothandler = self reactor.run() def startBotFactory(self, server, port, channels): if server in self.botfactories: print "Already on server '{}'.".format(server) return False print "Joining server '{}'.".format(server) botfactory = HubbotFactory(server, port, channels) self.botfactories[server] = botfactory return True def stopBotFactory(self, server, quitmessage="ohok"): quitmessage = quitmessage.encode("utf-8") if server not in self.botfactories: print "ERROR: Bot for '{}' does not exist yet was asked to stop.".format(server) else: self.botfactories[server].protocol.quit(quitmessage) self.unregisterFactory(server) def unregisterFactory(self, server): if server in self.botfactories: del self.botfactories[server] if len(self.botfactories)==0: print "No more running bots, shutting down." reactor.callLater(2.0, reactor.stop) if __name__=="__main__": bothandler = BotHandler()
from twisted.internet import reactor from Hubbot import Hubbot, HubbotFactory import GlobalVars class BotHandler: botfactories = {} def __init__(self): for (server_with_port,channels) in GlobalVars.connections.items(): server = server_with_port.split(":")[0] port = int(server_with_port.split(":")[1]) self.startBotFactory(server, port, channels) GlobalVars.bothandler = self reactor.run() def startBotFactory(self, server, port, channels): if server in self.botfactories: print "Already on server '{}'.".format(server) return False print "Joining server '{}'.".format(server) botfactory = HubbotFactory(server, port, channels) self.botfactories[server] = botfactory return True def stopBotFactory(self, server, quitmessage="ohok"): quitmessage = quitmessage.encode("utf-8") if server not in self.botfactories: print "ERROR: Bot for '{}' does not exist yet was asked to stop.".format(server) else: print "Shutting down bot for server '{}'".format(server) self.botfactories[server].protocol.quit(quitmessage) self.unregisterFactory(server) def unregisterFactory(self, server): if server in self.botfactories: del self.botfactories[server] if len(self.botfactories)==0: print "No more running bots, shutting down." reactor.callLater(2.0, reactor.stop) if __name__=="__main__": bothandler = BotHandler()
Print a quit message when shutting down a bot
Print a quit message when shutting down a bot
Python
mit
HubbeKing/Hubbot_Twisted
from twisted.internet import reactor from Hubbot import Hubbot, HubbotFactory import GlobalVars class BotHandler: botfactories = {} def __init__(self): for (server_with_port,channels) in GlobalVars.connections.items(): server = server_with_port.split(":")[0] port = int(server_with_port.split(":")[1]) self.startBotFactory(server, port, channels) GlobalVars.bothandler = self reactor.run() def startBotFactory(self, server, port, channels): if server in self.botfactories: print "Already on server '{}'.".format(server) return False print "Joining server '{}'.".format(server) botfactory = HubbotFactory(server, port, channels) self.botfactories[server] = botfactory return True def stopBotFactory(self, server, quitmessage="ohok"): quitmessage = quitmessage.encode("utf-8") if server not in self.botfactories: print "ERROR: Bot for '{}' does not exist yet was asked to stop.".format(server) else: self.botfactories[server].protocol.quit(quitmessage) self.unregisterFactory(server) def unregisterFactory(self, server): if server in self.botfactories: del self.botfactories[server] if len(self.botfactories)==0: print "No more running bots, shutting down." reactor.callLater(2.0, reactor.stop) if __name__=="__main__": bothandler = BotHandler() Print a quit message when shutting down a bot
from twisted.internet import reactor from Hubbot import Hubbot, HubbotFactory import GlobalVars class BotHandler: botfactories = {} def __init__(self): for (server_with_port,channels) in GlobalVars.connections.items(): server = server_with_port.split(":")[0] port = int(server_with_port.split(":")[1]) self.startBotFactory(server, port, channels) GlobalVars.bothandler = self reactor.run() def startBotFactory(self, server, port, channels): if server in self.botfactories: print "Already on server '{}'.".format(server) return False print "Joining server '{}'.".format(server) botfactory = HubbotFactory(server, port, channels) self.botfactories[server] = botfactory return True def stopBotFactory(self, server, quitmessage="ohok"): quitmessage = quitmessage.encode("utf-8") if server not in self.botfactories: print "ERROR: Bot for '{}' does not exist yet was asked to stop.".format(server) else: print "Shutting down bot for server '{}'".format(server) self.botfactories[server].protocol.quit(quitmessage) self.unregisterFactory(server) def unregisterFactory(self, server): if server in self.botfactories: del self.botfactories[server] if len(self.botfactories)==0: print "No more running bots, shutting down." reactor.callLater(2.0, reactor.stop) if __name__=="__main__": bothandler = BotHandler()
<commit_before>from twisted.internet import reactor from Hubbot import Hubbot, HubbotFactory import GlobalVars class BotHandler: botfactories = {} def __init__(self): for (server_with_port,channels) in GlobalVars.connections.items(): server = server_with_port.split(":")[0] port = int(server_with_port.split(":")[1]) self.startBotFactory(server, port, channels) GlobalVars.bothandler = self reactor.run() def startBotFactory(self, server, port, channels): if server in self.botfactories: print "Already on server '{}'.".format(server) return False print "Joining server '{}'.".format(server) botfactory = HubbotFactory(server, port, channels) self.botfactories[server] = botfactory return True def stopBotFactory(self, server, quitmessage="ohok"): quitmessage = quitmessage.encode("utf-8") if server not in self.botfactories: print "ERROR: Bot for '{}' does not exist yet was asked to stop.".format(server) else: self.botfactories[server].protocol.quit(quitmessage) self.unregisterFactory(server) def unregisterFactory(self, server): if server in self.botfactories: del self.botfactories[server] if len(self.botfactories)==0: print "No more running bots, shutting down." reactor.callLater(2.0, reactor.stop) if __name__=="__main__": bothandler = BotHandler() <commit_msg>Print a quit message when shutting down a bot<commit_after>
from twisted.internet import reactor from Hubbot import Hubbot, HubbotFactory import GlobalVars class BotHandler: botfactories = {} def __init__(self): for (server_with_port,channels) in GlobalVars.connections.items(): server = server_with_port.split(":")[0] port = int(server_with_port.split(":")[1]) self.startBotFactory(server, port, channels) GlobalVars.bothandler = self reactor.run() def startBotFactory(self, server, port, channels): if server in self.botfactories: print "Already on server '{}'.".format(server) return False print "Joining server '{}'.".format(server) botfactory = HubbotFactory(server, port, channels) self.botfactories[server] = botfactory return True def stopBotFactory(self, server, quitmessage="ohok"): quitmessage = quitmessage.encode("utf-8") if server not in self.botfactories: print "ERROR: Bot for '{}' does not exist yet was asked to stop.".format(server) else: print "Shutting down bot for server '{}'".format(server) self.botfactories[server].protocol.quit(quitmessage) self.unregisterFactory(server) def unregisterFactory(self, server): if server in self.botfactories: del self.botfactories[server] if len(self.botfactories)==0: print "No more running bots, shutting down." reactor.callLater(2.0, reactor.stop) if __name__=="__main__": bothandler = BotHandler()
from twisted.internet import reactor from Hubbot import Hubbot, HubbotFactory import GlobalVars class BotHandler: botfactories = {} def __init__(self): for (server_with_port,channels) in GlobalVars.connections.items(): server = server_with_port.split(":")[0] port = int(server_with_port.split(":")[1]) self.startBotFactory(server, port, channels) GlobalVars.bothandler = self reactor.run() def startBotFactory(self, server, port, channels): if server in self.botfactories: print "Already on server '{}'.".format(server) return False print "Joining server '{}'.".format(server) botfactory = HubbotFactory(server, port, channels) self.botfactories[server] = botfactory return True def stopBotFactory(self, server, quitmessage="ohok"): quitmessage = quitmessage.encode("utf-8") if server not in self.botfactories: print "ERROR: Bot for '{}' does not exist yet was asked to stop.".format(server) else: self.botfactories[server].protocol.quit(quitmessage) self.unregisterFactory(server) def unregisterFactory(self, server): if server in self.botfactories: del self.botfactories[server] if len(self.botfactories)==0: print "No more running bots, shutting down." reactor.callLater(2.0, reactor.stop) if __name__=="__main__": bothandler = BotHandler() Print a quit message when shutting down a botfrom twisted.internet import reactor from Hubbot import Hubbot, HubbotFactory import GlobalVars class BotHandler: botfactories = {} def __init__(self): for (server_with_port,channels) in GlobalVars.connections.items(): server = server_with_port.split(":")[0] port = int(server_with_port.split(":")[1]) self.startBotFactory(server, port, channels) GlobalVars.bothandler = self reactor.run() def startBotFactory(self, server, port, channels): if server in self.botfactories: print "Already on server '{}'.".format(server) return False print "Joining server '{}'.".format(server) botfactory = HubbotFactory(server, port, channels) self.botfactories[server] = botfactory return True def stopBotFactory(self, server, quitmessage="ohok"): quitmessage = quitmessage.encode("utf-8") if server not in self.botfactories: print "ERROR: Bot for '{}' does not exist yet was asked to stop.".format(server) else: print "Shutting down bot for server '{}'".format(server) self.botfactories[server].protocol.quit(quitmessage) self.unregisterFactory(server) def unregisterFactory(self, server): if server in self.botfactories: del self.botfactories[server] if len(self.botfactories)==0: print "No more running bots, shutting down." reactor.callLater(2.0, reactor.stop) if __name__=="__main__": bothandler = BotHandler()
<commit_before>from twisted.internet import reactor from Hubbot import Hubbot, HubbotFactory import GlobalVars class BotHandler: botfactories = {} def __init__(self): for (server_with_port,channels) in GlobalVars.connections.items(): server = server_with_port.split(":")[0] port = int(server_with_port.split(":")[1]) self.startBotFactory(server, port, channels) GlobalVars.bothandler = self reactor.run() def startBotFactory(self, server, port, channels): if server in self.botfactories: print "Already on server '{}'.".format(server) return False print "Joining server '{}'.".format(server) botfactory = HubbotFactory(server, port, channels) self.botfactories[server] = botfactory return True def stopBotFactory(self, server, quitmessage="ohok"): quitmessage = quitmessage.encode("utf-8") if server not in self.botfactories: print "ERROR: Bot for '{}' does not exist yet was asked to stop.".format(server) else: self.botfactories[server].protocol.quit(quitmessage) self.unregisterFactory(server) def unregisterFactory(self, server): if server in self.botfactories: del self.botfactories[server] if len(self.botfactories)==0: print "No more running bots, shutting down." reactor.callLater(2.0, reactor.stop) if __name__=="__main__": bothandler = BotHandler() <commit_msg>Print a quit message when shutting down a bot<commit_after>from twisted.internet import reactor from Hubbot import Hubbot, HubbotFactory import GlobalVars class BotHandler: botfactories = {} def __init__(self): for (server_with_port,channels) in GlobalVars.connections.items(): server = server_with_port.split(":")[0] port = int(server_with_port.split(":")[1]) self.startBotFactory(server, port, channels) GlobalVars.bothandler = self reactor.run() def startBotFactory(self, server, port, channels): if server in self.botfactories: print "Already on server '{}'.".format(server) return False print "Joining server '{}'.".format(server) botfactory = HubbotFactory(server, port, channels) self.botfactories[server] = botfactory return True def stopBotFactory(self, server, quitmessage="ohok"): quitmessage = quitmessage.encode("utf-8") if server not in self.botfactories: print "ERROR: Bot for '{}' does not exist yet was asked to stop.".format(server) else: print "Shutting down bot for server '{}'".format(server) self.botfactories[server].protocol.quit(quitmessage) self.unregisterFactory(server) def unregisterFactory(self, server): if server in self.botfactories: del self.botfactories[server] if len(self.botfactories)==0: print "No more running bots, shutting down." reactor.callLater(2.0, reactor.stop) if __name__=="__main__": bothandler = BotHandler()
982f4af638e83ee49c87a0dffad2b47daf872749
workers/data_refinery_workers/downloaders/test_utils.py
workers/data_refinery_workers/downloaders/test_utils.py
import os from django.test import TestCase, tag from typing import List from unittest.mock import patch, call from urllib.error import URLError from data_refinery_workers.downloaders import utils class UtilsTestCase(TestCase): def test_no_jobs_to_create(self): """Make sure this function doesn't raise an exception with no files.""" create_processor_job_for_original_files([]) self.assertTrue(True)
import os from django.test import TestCase, tag from typing import List from unittest.mock import patch, call from urllib.error import URLError from data_refinery_workers.downloaders import utils class UtilsTestCase(TestCase): @tag('downloaders') def test_no_jobs_to_create(self): """Make sure this function doesn't raise an exception with no files.""" create_processor_job_for_original_files([]) self.assertTrue(True)
Add tag to downloaders test so it is actually run.
Add tag to downloaders test so it is actually run.
Python
bsd-3-clause
data-refinery/data_refinery,data-refinery/data_refinery,data-refinery/data_refinery
import os from django.test import TestCase, tag from typing import List from unittest.mock import patch, call from urllib.error import URLError from data_refinery_workers.downloaders import utils class UtilsTestCase(TestCase): def test_no_jobs_to_create(self): """Make sure this function doesn't raise an exception with no files.""" create_processor_job_for_original_files([]) self.assertTrue(True) Add tag to downloaders test so it is actually run.
import os from django.test import TestCase, tag from typing import List from unittest.mock import patch, call from urllib.error import URLError from data_refinery_workers.downloaders import utils class UtilsTestCase(TestCase): @tag('downloaders') def test_no_jobs_to_create(self): """Make sure this function doesn't raise an exception with no files.""" create_processor_job_for_original_files([]) self.assertTrue(True)
<commit_before>import os from django.test import TestCase, tag from typing import List from unittest.mock import patch, call from urllib.error import URLError from data_refinery_workers.downloaders import utils class UtilsTestCase(TestCase): def test_no_jobs_to_create(self): """Make sure this function doesn't raise an exception with no files.""" create_processor_job_for_original_files([]) self.assertTrue(True) <commit_msg>Add tag to downloaders test so it is actually run.<commit_after>
import os from django.test import TestCase, tag from typing import List from unittest.mock import patch, call from urllib.error import URLError from data_refinery_workers.downloaders import utils class UtilsTestCase(TestCase): @tag('downloaders') def test_no_jobs_to_create(self): """Make sure this function doesn't raise an exception with no files.""" create_processor_job_for_original_files([]) self.assertTrue(True)
import os from django.test import TestCase, tag from typing import List from unittest.mock import patch, call from urllib.error import URLError from data_refinery_workers.downloaders import utils class UtilsTestCase(TestCase): def test_no_jobs_to_create(self): """Make sure this function doesn't raise an exception with no files.""" create_processor_job_for_original_files([]) self.assertTrue(True) Add tag to downloaders test so it is actually run.import os from django.test import TestCase, tag from typing import List from unittest.mock import patch, call from urllib.error import URLError from data_refinery_workers.downloaders import utils class UtilsTestCase(TestCase): @tag('downloaders') def test_no_jobs_to_create(self): """Make sure this function doesn't raise an exception with no files.""" create_processor_job_for_original_files([]) self.assertTrue(True)
<commit_before>import os from django.test import TestCase, tag from typing import List from unittest.mock import patch, call from urllib.error import URLError from data_refinery_workers.downloaders import utils class UtilsTestCase(TestCase): def test_no_jobs_to_create(self): """Make sure this function doesn't raise an exception with no files.""" create_processor_job_for_original_files([]) self.assertTrue(True) <commit_msg>Add tag to downloaders test so it is actually run.<commit_after>import os from django.test import TestCase, tag from typing import List from unittest.mock import patch, call from urllib.error import URLError from data_refinery_workers.downloaders import utils class UtilsTestCase(TestCase): @tag('downloaders') def test_no_jobs_to_create(self): """Make sure this function doesn't raise an exception with no files.""" create_processor_job_for_original_files([]) self.assertTrue(True)
813970150109450e53be44715913e5fb3c680c77
uscampgrounds/models.py
uscampgrounds/models.py
from django.conf import settings from django.contrib.gis.db import models class Campground(models.Model): campground_code = models.CharField(max_length=64) name = models.CharField(max_length=128) campground_type = models.CharField(max_length=128) phone = models.CharField(max_length=128) comments = models.TextField() sites = models.CharField(max_length=128) elevation = models.CharField(max_length=128) hookups = models.CharField(max_length=128) amenities = models.TextField() point = models.PointField(srid=4326) def locator_point(self): return self.point def __unicode__(self): return self.name # integrate with the django-locator app for easy geo lookups if it's installed if 'locator.objects' in settings.INSTALLED_APPS: from locator.objects.models import create_locator_object models.signals.post_save.connect(create_locator_object, sender=Campground)
from django.conf import settings from django.contrib.gis.db import models class Campground(models.Model): campground_code = models.CharField(max_length=64) name = models.CharField(max_length=128) campground_type = models.CharField(max_length=128) phone = models.CharField(max_length=128) comments = models.TextField() sites = models.CharField(max_length=128) elevation = models.CharField(max_length=128) hookups = models.CharField(max_length=128) amenities = models.TextField() point = models.PointField(srid=4326) objects = models.GeoManager() def locator_point(self): return self.point def __unicode__(self): return self.name # integrate with the django-locator app for easy geo lookups if it's installed if 'locator.objects' in settings.INSTALLED_APPS: from locator.objects.models import create_locator_object models.signals.post_save.connect(create_locator_object, sender=Campground)
Include an override to the default manager to allow geospatial querying.
Include an override to the default manager to allow geospatial querying.
Python
bsd-3-clause
adamfast/geodjango-uscampgrounds,adamfast/geodjango-uscampgrounds
from django.conf import settings from django.contrib.gis.db import models class Campground(models.Model): campground_code = models.CharField(max_length=64) name = models.CharField(max_length=128) campground_type = models.CharField(max_length=128) phone = models.CharField(max_length=128) comments = models.TextField() sites = models.CharField(max_length=128) elevation = models.CharField(max_length=128) hookups = models.CharField(max_length=128) amenities = models.TextField() point = models.PointField(srid=4326) def locator_point(self): return self.point def __unicode__(self): return self.name # integrate with the django-locator app for easy geo lookups if it's installed if 'locator.objects' in settings.INSTALLED_APPS: from locator.objects.models import create_locator_object models.signals.post_save.connect(create_locator_object, sender=Campground) Include an override to the default manager to allow geospatial querying.
from django.conf import settings from django.contrib.gis.db import models class Campground(models.Model): campground_code = models.CharField(max_length=64) name = models.CharField(max_length=128) campground_type = models.CharField(max_length=128) phone = models.CharField(max_length=128) comments = models.TextField() sites = models.CharField(max_length=128) elevation = models.CharField(max_length=128) hookups = models.CharField(max_length=128) amenities = models.TextField() point = models.PointField(srid=4326) objects = models.GeoManager() def locator_point(self): return self.point def __unicode__(self): return self.name # integrate with the django-locator app for easy geo lookups if it's installed if 'locator.objects' in settings.INSTALLED_APPS: from locator.objects.models import create_locator_object models.signals.post_save.connect(create_locator_object, sender=Campground)
<commit_before>from django.conf import settings from django.contrib.gis.db import models class Campground(models.Model): campground_code = models.CharField(max_length=64) name = models.CharField(max_length=128) campground_type = models.CharField(max_length=128) phone = models.CharField(max_length=128) comments = models.TextField() sites = models.CharField(max_length=128) elevation = models.CharField(max_length=128) hookups = models.CharField(max_length=128) amenities = models.TextField() point = models.PointField(srid=4326) def locator_point(self): return self.point def __unicode__(self): return self.name # integrate with the django-locator app for easy geo lookups if it's installed if 'locator.objects' in settings.INSTALLED_APPS: from locator.objects.models import create_locator_object models.signals.post_save.connect(create_locator_object, sender=Campground) <commit_msg>Include an override to the default manager to allow geospatial querying.<commit_after>
from django.conf import settings from django.contrib.gis.db import models class Campground(models.Model): campground_code = models.CharField(max_length=64) name = models.CharField(max_length=128) campground_type = models.CharField(max_length=128) phone = models.CharField(max_length=128) comments = models.TextField() sites = models.CharField(max_length=128) elevation = models.CharField(max_length=128) hookups = models.CharField(max_length=128) amenities = models.TextField() point = models.PointField(srid=4326) objects = models.GeoManager() def locator_point(self): return self.point def __unicode__(self): return self.name # integrate with the django-locator app for easy geo lookups if it's installed if 'locator.objects' in settings.INSTALLED_APPS: from locator.objects.models import create_locator_object models.signals.post_save.connect(create_locator_object, sender=Campground)
from django.conf import settings from django.contrib.gis.db import models class Campground(models.Model): campground_code = models.CharField(max_length=64) name = models.CharField(max_length=128) campground_type = models.CharField(max_length=128) phone = models.CharField(max_length=128) comments = models.TextField() sites = models.CharField(max_length=128) elevation = models.CharField(max_length=128) hookups = models.CharField(max_length=128) amenities = models.TextField() point = models.PointField(srid=4326) def locator_point(self): return self.point def __unicode__(self): return self.name # integrate with the django-locator app for easy geo lookups if it's installed if 'locator.objects' in settings.INSTALLED_APPS: from locator.objects.models import create_locator_object models.signals.post_save.connect(create_locator_object, sender=Campground) Include an override to the default manager to allow geospatial querying.from django.conf import settings from django.contrib.gis.db import models class Campground(models.Model): campground_code = models.CharField(max_length=64) name = models.CharField(max_length=128) campground_type = models.CharField(max_length=128) phone = models.CharField(max_length=128) comments = models.TextField() sites = models.CharField(max_length=128) elevation = models.CharField(max_length=128) hookups = models.CharField(max_length=128) amenities = models.TextField() point = models.PointField(srid=4326) objects = models.GeoManager() def locator_point(self): return self.point def __unicode__(self): return self.name # integrate with the django-locator app for easy geo lookups if it's installed if 'locator.objects' in settings.INSTALLED_APPS: from locator.objects.models import create_locator_object models.signals.post_save.connect(create_locator_object, sender=Campground)
<commit_before>from django.conf import settings from django.contrib.gis.db import models class Campground(models.Model): campground_code = models.CharField(max_length=64) name = models.CharField(max_length=128) campground_type = models.CharField(max_length=128) phone = models.CharField(max_length=128) comments = models.TextField() sites = models.CharField(max_length=128) elevation = models.CharField(max_length=128) hookups = models.CharField(max_length=128) amenities = models.TextField() point = models.PointField(srid=4326) def locator_point(self): return self.point def __unicode__(self): return self.name # integrate with the django-locator app for easy geo lookups if it's installed if 'locator.objects' in settings.INSTALLED_APPS: from locator.objects.models import create_locator_object models.signals.post_save.connect(create_locator_object, sender=Campground) <commit_msg>Include an override to the default manager to allow geospatial querying.<commit_after>from django.conf import settings from django.contrib.gis.db import models class Campground(models.Model): campground_code = models.CharField(max_length=64) name = models.CharField(max_length=128) campground_type = models.CharField(max_length=128) phone = models.CharField(max_length=128) comments = models.TextField() sites = models.CharField(max_length=128) elevation = models.CharField(max_length=128) hookups = models.CharField(max_length=128) amenities = models.TextField() point = models.PointField(srid=4326) objects = models.GeoManager() def locator_point(self): return self.point def __unicode__(self): return self.name # integrate with the django-locator app for easy geo lookups if it's installed if 'locator.objects' in settings.INSTALLED_APPS: from locator.objects.models import create_locator_object models.signals.post_save.connect(create_locator_object, sender=Campground)
f0dd6411458a19985408404d511584f7c7e26e38
useful/tests/tasks/call_management_command.py
useful/tests/tasks/call_management_command.py
from django.test import TestCase from useful.tasks import call_management_command class ManagementCommandTestCase(TestCase): """Test calling a management command as a Celery task""" def test_management_command(self): t = call_management_command.delay('validate') self.assertEquals(t.status, 'SUCCESS')
from django.test import TestCase from useful.tasks import call_management_command class ManagementCommandTestCase(TestCase): """Test calling a management command as a Celery task""" def test_success(self): t = call_management_command.delay('validate') self.assertEquals(t.status, 'SUCCESS') def test_failure(self): t = call_management_command.delay('somethingrandomthatdoesntexist') self.assertEquals(t.status, 'FAILURE')
Add failed management command test
Add failed management command test
Python
isc
yprez/django-useful,yprez/django-useful
from django.test import TestCase from useful.tasks import call_management_command class ManagementCommandTestCase(TestCase): """Test calling a management command as a Celery task""" def test_management_command(self): t = call_management_command.delay('validate') self.assertEquals(t.status, 'SUCCESS') Add failed management command test
from django.test import TestCase from useful.tasks import call_management_command class ManagementCommandTestCase(TestCase): """Test calling a management command as a Celery task""" def test_success(self): t = call_management_command.delay('validate') self.assertEquals(t.status, 'SUCCESS') def test_failure(self): t = call_management_command.delay('somethingrandomthatdoesntexist') self.assertEquals(t.status, 'FAILURE')
<commit_before>from django.test import TestCase from useful.tasks import call_management_command class ManagementCommandTestCase(TestCase): """Test calling a management command as a Celery task""" def test_management_command(self): t = call_management_command.delay('validate') self.assertEquals(t.status, 'SUCCESS') <commit_msg>Add failed management command test<commit_after>
from django.test import TestCase from useful.tasks import call_management_command class ManagementCommandTestCase(TestCase): """Test calling a management command as a Celery task""" def test_success(self): t = call_management_command.delay('validate') self.assertEquals(t.status, 'SUCCESS') def test_failure(self): t = call_management_command.delay('somethingrandomthatdoesntexist') self.assertEquals(t.status, 'FAILURE')
from django.test import TestCase from useful.tasks import call_management_command class ManagementCommandTestCase(TestCase): """Test calling a management command as a Celery task""" def test_management_command(self): t = call_management_command.delay('validate') self.assertEquals(t.status, 'SUCCESS') Add failed management command testfrom django.test import TestCase from useful.tasks import call_management_command class ManagementCommandTestCase(TestCase): """Test calling a management command as a Celery task""" def test_success(self): t = call_management_command.delay('validate') self.assertEquals(t.status, 'SUCCESS') def test_failure(self): t = call_management_command.delay('somethingrandomthatdoesntexist') self.assertEquals(t.status, 'FAILURE')
<commit_before>from django.test import TestCase from useful.tasks import call_management_command class ManagementCommandTestCase(TestCase): """Test calling a management command as a Celery task""" def test_management_command(self): t = call_management_command.delay('validate') self.assertEquals(t.status, 'SUCCESS') <commit_msg>Add failed management command test<commit_after>from django.test import TestCase from useful.tasks import call_management_command class ManagementCommandTestCase(TestCase): """Test calling a management command as a Celery task""" def test_success(self): t = call_management_command.delay('validate') self.assertEquals(t.status, 'SUCCESS') def test_failure(self): t = call_management_command.delay('somethingrandomthatdoesntexist') self.assertEquals(t.status, 'FAILURE')
bfa98f350a3da827da8359a4e6e7373fc11626cd
changeling/models.py
changeling/models.py
import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None): self.id = id or str(uuid.uuid4()) self.name = name @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None, description=None): self.id = id or str(uuid.uuid4()) self.name = name self.description = description @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
Add description attr to Change
Add description attr to Change
Python
apache-2.0
bcwaldon/changeling,bcwaldon/changeling
import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None): self.id = id or str(uuid.uuid4()) self.name = name @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True Add description attr to Change
import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None, description=None): self.id = id or str(uuid.uuid4()) self.name = name self.description = description @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
<commit_before>import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None): self.id = id or str(uuid.uuid4()) self.name = name @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True <commit_msg>Add description attr to Change<commit_after>
import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None, description=None): self.id = id or str(uuid.uuid4()) self.name = name self.description = description @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None): self.id = id or str(uuid.uuid4()) self.name = name @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True Add description attr to Changeimport uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None, description=None): self.id = id or str(uuid.uuid4()) self.name = name self.description = description @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
<commit_before>import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None): self.id = id or str(uuid.uuid4()) self.name = name @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True <commit_msg>Add description attr to Change<commit_after>import uuid import jsonschema import changeling.exception class Change(object): schema = { 'name': 'change', 'properties': { 'id': {'type': 'string'}, 'name': {'type': 'string'}, 'description': {'type': 'string'}, }, 'additionalProperties': False, } def __init__(self, id=None, name=None, description=None): self.id = id or str(uuid.uuid4()) self.name = name self.description = description @classmethod def from_dict(self, data): self.validate(data) return Change(**data) def to_dict(self): def _generate_set_attributes(): for k in Change.schema['properties'].keys(): val = getattr(self, k) if val is not None: yield (k, val) return dict(_generate_set_attributes()) def __str__(self): return "<Change id=%s name=%s>" % (self.id, self.name) @classmethod def validate(cls, data): try: jsonschema.validate(data, cls.schema) except jsonschema.ValidationError as exc: raise changeling.exception.ValidationError(exc) def is_valid(self): try: self.validate(self.to_dict()) except changeling.exception.ValidationError: return False else: return True
0165e350a8b49e11b119c6393de93b28f9523dca
fedoracommunity/mokshaapps/packages/controllers/bugs.py
fedoracommunity/mokshaapps/packages/controllers/bugs.py
from moksha.lib.base import Controller from moksha.lib.helpers import Category, MokshaApp, Not, not_anonymous, MokshaWidget from moksha.api.widgets.containers import DashboardContainer from moksha.api.widgets import ContextAwareWidget from tg import expose, tmpl_context, require, request class BugsDashboard(DashboardContainer, ContextAwareWidget): template = 'mako:fedoracommunity.mokshaapps.packages.templates.single_col_dashboard' layout = [Category('content-col-apps',[])] bugs_dashboard = BugsDashboard class BugsController(Controller): @expose('mako:moksha.templates.widget') def index(self, package): tmpl_context.widget = bugs_dashboard return {'package': package}
from tw.api import Widget as TWWidget from pylons import cache from moksha.lib.base import Controller from moksha.lib.helpers import Category, MokshaApp, Not, not_anonymous, MokshaWidget from moksha.lib.helpers import Widget from moksha.api.widgets.containers import DashboardContainer from moksha.api.widgets import ContextAwareWidget, Grid from moksha.api.connectors import get_connector from tg import expose, tmpl_context, require, request class BugStatsWidget(TWWidget): template='mako:fedoracommunity.mokshaapps.packages.templates.bugs_stats_widget' params = ['id', 'product', 'component', 'version', 'num_closed', 'num_open', 'num_new'] product = 'Fedora' version = 'rawhide' component = None num_closed = num_open = num_new = '-' bug_stats_widget = BugStatsWidget('bug_stats') class BugsGrid(Grid): template='mako:fedoracommunity.mokshaapps.packages.templates.bugs_table_widget' def update_params(self, d): d['resource'] = 'bugzilla' d['resource_path'] = 'query_bugs' super(BugsGrid, self).update_params(d) bugs_grid = BugsGrid('bugs_grid') class BugsDashboard(DashboardContainer, ContextAwareWidget): template = 'mako:fedoracommunity.mokshaapps.packages.templates.single_col_dashboard' layout = [Category('content-col-apps',[ Widget('Dashboard', bug_stats_widget, params={'filters':{'package': ''}}), Widget('Recently Filed Bugs', bugs_grid, params={'filters':{'package': ''}}), ])] def update_params(self, d): package = d.get('package') conn = get_connector('pkgdb') info = conn.get_basic_package_info(package) d['pkg_summary'] = info['summary'] super(BugsDashboard, self).update_params(d) bugs_dashboard = BugsDashboard('bugs_dashboard') class BugsController(Controller): @expose('mako:moksha.templates.widget') def index(self, package): tmpl_context.widget = bugs_dashboard return {'options': {'package': package}}
Add BugStats and BugsGrid widgets
Add BugStats and BugsGrid widgets
Python
agpl-3.0
Fale/fedora-packages,fedora-infra/fedora-packages,fedora-infra/fedora-packages,fedora-infra/fedora-packages,Fale/fedora-packages,fedora-infra/fedora-packages,Fale/fedora-packages
from moksha.lib.base import Controller from moksha.lib.helpers import Category, MokshaApp, Not, not_anonymous, MokshaWidget from moksha.api.widgets.containers import DashboardContainer from moksha.api.widgets import ContextAwareWidget from tg import expose, tmpl_context, require, request class BugsDashboard(DashboardContainer, ContextAwareWidget): template = 'mako:fedoracommunity.mokshaapps.packages.templates.single_col_dashboard' layout = [Category('content-col-apps',[])] bugs_dashboard = BugsDashboard class BugsController(Controller): @expose('mako:moksha.templates.widget') def index(self, package): tmpl_context.widget = bugs_dashboard return {'package': package} Add BugStats and BugsGrid widgets
from tw.api import Widget as TWWidget from pylons import cache from moksha.lib.base import Controller from moksha.lib.helpers import Category, MokshaApp, Not, not_anonymous, MokshaWidget from moksha.lib.helpers import Widget from moksha.api.widgets.containers import DashboardContainer from moksha.api.widgets import ContextAwareWidget, Grid from moksha.api.connectors import get_connector from tg import expose, tmpl_context, require, request class BugStatsWidget(TWWidget): template='mako:fedoracommunity.mokshaapps.packages.templates.bugs_stats_widget' params = ['id', 'product', 'component', 'version', 'num_closed', 'num_open', 'num_new'] product = 'Fedora' version = 'rawhide' component = None num_closed = num_open = num_new = '-' bug_stats_widget = BugStatsWidget('bug_stats') class BugsGrid(Grid): template='mako:fedoracommunity.mokshaapps.packages.templates.bugs_table_widget' def update_params(self, d): d['resource'] = 'bugzilla' d['resource_path'] = 'query_bugs' super(BugsGrid, self).update_params(d) bugs_grid = BugsGrid('bugs_grid') class BugsDashboard(DashboardContainer, ContextAwareWidget): template = 'mako:fedoracommunity.mokshaapps.packages.templates.single_col_dashboard' layout = [Category('content-col-apps',[ Widget('Dashboard', bug_stats_widget, params={'filters':{'package': ''}}), Widget('Recently Filed Bugs', bugs_grid, params={'filters':{'package': ''}}), ])] def update_params(self, d): package = d.get('package') conn = get_connector('pkgdb') info = conn.get_basic_package_info(package) d['pkg_summary'] = info['summary'] super(BugsDashboard, self).update_params(d) bugs_dashboard = BugsDashboard('bugs_dashboard') class BugsController(Controller): @expose('mako:moksha.templates.widget') def index(self, package): tmpl_context.widget = bugs_dashboard return {'options': {'package': package}}
<commit_before>from moksha.lib.base import Controller from moksha.lib.helpers import Category, MokshaApp, Not, not_anonymous, MokshaWidget from moksha.api.widgets.containers import DashboardContainer from moksha.api.widgets import ContextAwareWidget from tg import expose, tmpl_context, require, request class BugsDashboard(DashboardContainer, ContextAwareWidget): template = 'mako:fedoracommunity.mokshaapps.packages.templates.single_col_dashboard' layout = [Category('content-col-apps',[])] bugs_dashboard = BugsDashboard class BugsController(Controller): @expose('mako:moksha.templates.widget') def index(self, package): tmpl_context.widget = bugs_dashboard return {'package': package} <commit_msg>Add BugStats and BugsGrid widgets<commit_after>
from tw.api import Widget as TWWidget from pylons import cache from moksha.lib.base import Controller from moksha.lib.helpers import Category, MokshaApp, Not, not_anonymous, MokshaWidget from moksha.lib.helpers import Widget from moksha.api.widgets.containers import DashboardContainer from moksha.api.widgets import ContextAwareWidget, Grid from moksha.api.connectors import get_connector from tg import expose, tmpl_context, require, request class BugStatsWidget(TWWidget): template='mako:fedoracommunity.mokshaapps.packages.templates.bugs_stats_widget' params = ['id', 'product', 'component', 'version', 'num_closed', 'num_open', 'num_new'] product = 'Fedora' version = 'rawhide' component = None num_closed = num_open = num_new = '-' bug_stats_widget = BugStatsWidget('bug_stats') class BugsGrid(Grid): template='mako:fedoracommunity.mokshaapps.packages.templates.bugs_table_widget' def update_params(self, d): d['resource'] = 'bugzilla' d['resource_path'] = 'query_bugs' super(BugsGrid, self).update_params(d) bugs_grid = BugsGrid('bugs_grid') class BugsDashboard(DashboardContainer, ContextAwareWidget): template = 'mako:fedoracommunity.mokshaapps.packages.templates.single_col_dashboard' layout = [Category('content-col-apps',[ Widget('Dashboard', bug_stats_widget, params={'filters':{'package': ''}}), Widget('Recently Filed Bugs', bugs_grid, params={'filters':{'package': ''}}), ])] def update_params(self, d): package = d.get('package') conn = get_connector('pkgdb') info = conn.get_basic_package_info(package) d['pkg_summary'] = info['summary'] super(BugsDashboard, self).update_params(d) bugs_dashboard = BugsDashboard('bugs_dashboard') class BugsController(Controller): @expose('mako:moksha.templates.widget') def index(self, package): tmpl_context.widget = bugs_dashboard return {'options': {'package': package}}
from moksha.lib.base import Controller from moksha.lib.helpers import Category, MokshaApp, Not, not_anonymous, MokshaWidget from moksha.api.widgets.containers import DashboardContainer from moksha.api.widgets import ContextAwareWidget from tg import expose, tmpl_context, require, request class BugsDashboard(DashboardContainer, ContextAwareWidget): template = 'mako:fedoracommunity.mokshaapps.packages.templates.single_col_dashboard' layout = [Category('content-col-apps',[])] bugs_dashboard = BugsDashboard class BugsController(Controller): @expose('mako:moksha.templates.widget') def index(self, package): tmpl_context.widget = bugs_dashboard return {'package': package} Add BugStats and BugsGrid widgetsfrom tw.api import Widget as TWWidget from pylons import cache from moksha.lib.base import Controller from moksha.lib.helpers import Category, MokshaApp, Not, not_anonymous, MokshaWidget from moksha.lib.helpers import Widget from moksha.api.widgets.containers import DashboardContainer from moksha.api.widgets import ContextAwareWidget, Grid from moksha.api.connectors import get_connector from tg import expose, tmpl_context, require, request class BugStatsWidget(TWWidget): template='mako:fedoracommunity.mokshaapps.packages.templates.bugs_stats_widget' params = ['id', 'product', 'component', 'version', 'num_closed', 'num_open', 'num_new'] product = 'Fedora' version = 'rawhide' component = None num_closed = num_open = num_new = '-' bug_stats_widget = BugStatsWidget('bug_stats') class BugsGrid(Grid): template='mako:fedoracommunity.mokshaapps.packages.templates.bugs_table_widget' def update_params(self, d): d['resource'] = 'bugzilla' d['resource_path'] = 'query_bugs' super(BugsGrid, self).update_params(d) bugs_grid = BugsGrid('bugs_grid') class BugsDashboard(DashboardContainer, ContextAwareWidget): template = 'mako:fedoracommunity.mokshaapps.packages.templates.single_col_dashboard' layout = [Category('content-col-apps',[ Widget('Dashboard', bug_stats_widget, params={'filters':{'package': ''}}), Widget('Recently Filed Bugs', bugs_grid, params={'filters':{'package': ''}}), ])] def update_params(self, d): package = d.get('package') conn = get_connector('pkgdb') info = conn.get_basic_package_info(package) d['pkg_summary'] = info['summary'] super(BugsDashboard, self).update_params(d) bugs_dashboard = BugsDashboard('bugs_dashboard') class BugsController(Controller): @expose('mako:moksha.templates.widget') def index(self, package): tmpl_context.widget = bugs_dashboard return {'options': {'package': package}}
<commit_before>from moksha.lib.base import Controller from moksha.lib.helpers import Category, MokshaApp, Not, not_anonymous, MokshaWidget from moksha.api.widgets.containers import DashboardContainer from moksha.api.widgets import ContextAwareWidget from tg import expose, tmpl_context, require, request class BugsDashboard(DashboardContainer, ContextAwareWidget): template = 'mako:fedoracommunity.mokshaapps.packages.templates.single_col_dashboard' layout = [Category('content-col-apps',[])] bugs_dashboard = BugsDashboard class BugsController(Controller): @expose('mako:moksha.templates.widget') def index(self, package): tmpl_context.widget = bugs_dashboard return {'package': package} <commit_msg>Add BugStats and BugsGrid widgets<commit_after>from tw.api import Widget as TWWidget from pylons import cache from moksha.lib.base import Controller from moksha.lib.helpers import Category, MokshaApp, Not, not_anonymous, MokshaWidget from moksha.lib.helpers import Widget from moksha.api.widgets.containers import DashboardContainer from moksha.api.widgets import ContextAwareWidget, Grid from moksha.api.connectors import get_connector from tg import expose, tmpl_context, require, request class BugStatsWidget(TWWidget): template='mako:fedoracommunity.mokshaapps.packages.templates.bugs_stats_widget' params = ['id', 'product', 'component', 'version', 'num_closed', 'num_open', 'num_new'] product = 'Fedora' version = 'rawhide' component = None num_closed = num_open = num_new = '-' bug_stats_widget = BugStatsWidget('bug_stats') class BugsGrid(Grid): template='mako:fedoracommunity.mokshaapps.packages.templates.bugs_table_widget' def update_params(self, d): d['resource'] = 'bugzilla' d['resource_path'] = 'query_bugs' super(BugsGrid, self).update_params(d) bugs_grid = BugsGrid('bugs_grid') class BugsDashboard(DashboardContainer, ContextAwareWidget): template = 'mako:fedoracommunity.mokshaapps.packages.templates.single_col_dashboard' layout = [Category('content-col-apps',[ Widget('Dashboard', bug_stats_widget, params={'filters':{'package': ''}}), Widget('Recently Filed Bugs', bugs_grid, params={'filters':{'package': ''}}), ])] def update_params(self, d): package = d.get('package') conn = get_connector('pkgdb') info = conn.get_basic_package_info(package) d['pkg_summary'] = info['summary'] super(BugsDashboard, self).update_params(d) bugs_dashboard = BugsDashboard('bugs_dashboard') class BugsController(Controller): @expose('mako:moksha.templates.widget') def index(self, package): tmpl_context.widget = bugs_dashboard return {'options': {'package': package}}
0ff8c816d739deb19352b6e49cab86ddbde948fb
openelex/tasks/__init__.py
openelex/tasks/__init__.py
from invoke import Collection from mongoengine import ConnectionError from openelex.settings import init_db from fetch import fetch import archive, cache, datasource, load, load_metadata, transform, validate, bake # Build tasks namespace ns = Collection() ns.add_task(fetch) ns.add_collection(archive) ns.add_collection(cache) ns.add_collection(datasource) ns.add_collection(load) ns.add_collection(load_metadata) ns.add_collection(transform) ns.add_collection(validate) ns.add_collection(bake) # Initialize prod Mongo connection try: init_db() except ConnectionError: pass
from invoke import Collection from mongoengine import ConnectionError from openelex.settings import init_db from fetch import fetch import archive, cache, datasource, load, load_metadata, transform, validate # TODO: Add bake task back in # import bake # Build tasks namespace ns = Collection() ns.add_task(fetch) ns.add_collection(archive) ns.add_collection(cache) ns.add_collection(datasource) ns.add_collection(load) ns.add_collection(load_metadata) ns.add_collection(transform) ns.add_collection(validate) #ns.add_collection(bake) # Initialize prod Mongo connection try: init_db() except ConnectionError: pass
Disable bake for now because it doesn't match rawresult models
Disable bake for now because it doesn't match rawresult models
Python
mit
cathydeng/openelections-core,datamade/openelections-core,openelections/openelections-core,cathydeng/openelections-core,openelections/openelections-core,datamade/openelections-core
from invoke import Collection from mongoengine import ConnectionError from openelex.settings import init_db from fetch import fetch import archive, cache, datasource, load, load_metadata, transform, validate, bake # Build tasks namespace ns = Collection() ns.add_task(fetch) ns.add_collection(archive) ns.add_collection(cache) ns.add_collection(datasource) ns.add_collection(load) ns.add_collection(load_metadata) ns.add_collection(transform) ns.add_collection(validate) ns.add_collection(bake) # Initialize prod Mongo connection try: init_db() except ConnectionError: pass Disable bake for now because it doesn't match rawresult models
from invoke import Collection from mongoengine import ConnectionError from openelex.settings import init_db from fetch import fetch import archive, cache, datasource, load, load_metadata, transform, validate # TODO: Add bake task back in # import bake # Build tasks namespace ns = Collection() ns.add_task(fetch) ns.add_collection(archive) ns.add_collection(cache) ns.add_collection(datasource) ns.add_collection(load) ns.add_collection(load_metadata) ns.add_collection(transform) ns.add_collection(validate) #ns.add_collection(bake) # Initialize prod Mongo connection try: init_db() except ConnectionError: pass
<commit_before>from invoke import Collection from mongoengine import ConnectionError from openelex.settings import init_db from fetch import fetch import archive, cache, datasource, load, load_metadata, transform, validate, bake # Build tasks namespace ns = Collection() ns.add_task(fetch) ns.add_collection(archive) ns.add_collection(cache) ns.add_collection(datasource) ns.add_collection(load) ns.add_collection(load_metadata) ns.add_collection(transform) ns.add_collection(validate) ns.add_collection(bake) # Initialize prod Mongo connection try: init_db() except ConnectionError: pass <commit_msg>Disable bake for now because it doesn't match rawresult models<commit_after>
from invoke import Collection from mongoengine import ConnectionError from openelex.settings import init_db from fetch import fetch import archive, cache, datasource, load, load_metadata, transform, validate # TODO: Add bake task back in # import bake # Build tasks namespace ns = Collection() ns.add_task(fetch) ns.add_collection(archive) ns.add_collection(cache) ns.add_collection(datasource) ns.add_collection(load) ns.add_collection(load_metadata) ns.add_collection(transform) ns.add_collection(validate) #ns.add_collection(bake) # Initialize prod Mongo connection try: init_db() except ConnectionError: pass
from invoke import Collection from mongoengine import ConnectionError from openelex.settings import init_db from fetch import fetch import archive, cache, datasource, load, load_metadata, transform, validate, bake # Build tasks namespace ns = Collection() ns.add_task(fetch) ns.add_collection(archive) ns.add_collection(cache) ns.add_collection(datasource) ns.add_collection(load) ns.add_collection(load_metadata) ns.add_collection(transform) ns.add_collection(validate) ns.add_collection(bake) # Initialize prod Mongo connection try: init_db() except ConnectionError: pass Disable bake for now because it doesn't match rawresult modelsfrom invoke import Collection from mongoengine import ConnectionError from openelex.settings import init_db from fetch import fetch import archive, cache, datasource, load, load_metadata, transform, validate # TODO: Add bake task back in # import bake # Build tasks namespace ns = Collection() ns.add_task(fetch) ns.add_collection(archive) ns.add_collection(cache) ns.add_collection(datasource) ns.add_collection(load) ns.add_collection(load_metadata) ns.add_collection(transform) ns.add_collection(validate) #ns.add_collection(bake) # Initialize prod Mongo connection try: init_db() except ConnectionError: pass
<commit_before>from invoke import Collection from mongoengine import ConnectionError from openelex.settings import init_db from fetch import fetch import archive, cache, datasource, load, load_metadata, transform, validate, bake # Build tasks namespace ns = Collection() ns.add_task(fetch) ns.add_collection(archive) ns.add_collection(cache) ns.add_collection(datasource) ns.add_collection(load) ns.add_collection(load_metadata) ns.add_collection(transform) ns.add_collection(validate) ns.add_collection(bake) # Initialize prod Mongo connection try: init_db() except ConnectionError: pass <commit_msg>Disable bake for now because it doesn't match rawresult models<commit_after>from invoke import Collection from mongoengine import ConnectionError from openelex.settings import init_db from fetch import fetch import archive, cache, datasource, load, load_metadata, transform, validate # TODO: Add bake task back in # import bake # Build tasks namespace ns = Collection() ns.add_task(fetch) ns.add_collection(archive) ns.add_collection(cache) ns.add_collection(datasource) ns.add_collection(load) ns.add_collection(load_metadata) ns.add_collection(transform) ns.add_collection(validate) #ns.add_collection(bake) # Initialize prod Mongo connection try: init_db() except ConnectionError: pass
46ab04db26f6330e732abdf4284242bd83179684
lametro/management/commands/refresh_guid.py
lametro/management/commands/refresh_guid.py
from django.core.management.base import BaseCommand from django.conf import settings from legistar.bills import LegistarAPIBillScraper from lametro.models import SubjectGuid, Subject class Command(BaseCommand): def handle(self, *args, **options): scraper = LegistarAPIBillScraper() scraper.BASE_URL = 'https://webapi.legistar.com/v1/metrotest' scraper.retry_attempts = 0 scraper.requests_per_minute = 0 all_topics = scraper.topics() # Delete topics not currently in use current_topics = Subject.objects.values_list('subject', flat=True) deleted, _ = SubjectGuid.objects.exclude(name__in=current_topics).delete() self.stdout.write('Removed {0} stale topics'.format(deleted)) total_created = 0 for topic in all_topics: subject, created = SubjectGuid.objects.get_or_create( name=topic['IndexName'], guid=topic['api_metadata'] ) if created: total_created += 1 self.stdout.write('{0} topics created'.format(total_created))
from django.core.management.base import BaseCommand from django.conf import settings from django.db.utils import IntegrityError from legistar.bills import LegistarAPIBillScraper from lametro.models import SubjectGuid, Subject class Command(BaseCommand): def handle(self, *args, **options): scraper = LegistarAPIBillScraper() scraper.BASE_URL = 'https://webapi.legistar.com/v1/metro' scraper.retry_attempts = 0 scraper.requests_per_minute = 0 all_topics = scraper.topics() # Delete topics not currently in use current_topics = Subject.objects.values_list('subject', flat=True) deleted, _ = SubjectGuid.objects.exclude(name__in=current_topics).delete() self.stdout.write('Removed {0} stale topics'.format(deleted)) total_created = 0 total_updated = 0 total_noop = 0 for topic in all_topics: try: subject, created = SubjectGuid.objects.get_or_create( name=topic['IndexName'], guid=topic['api_metadata'] ) except IntegrityError as e: # This exception will be raised if get_or_create tries to create # a SubjectGuid with a name that already exists. The Legistar # API should not contain duplicates, i.e., the GUID has changed. # Update the GUID on the existing topic. subject = SubjectGuid.objects.get(name=topic['IndexName']) subject.guid = topic['api_metadata'] subject.save() total_updated += 1 else: if created: total_created += 1 else: total_noop += 1 self.stdout.write('Created {0} new topics'.format(total_created)) self.stdout.write('Updated {0} existing topics'.format(total_updated)) self.stdout.write("No-op'ed {0} topics".format(total_updated))
Update API path, handle updated SubjectGuid
Update API path, handle updated SubjectGuid
Python
mit
datamade/la-metro-councilmatic,datamade/la-metro-councilmatic,datamade/la-metro-councilmatic,datamade/la-metro-councilmatic
from django.core.management.base import BaseCommand from django.conf import settings from legistar.bills import LegistarAPIBillScraper from lametro.models import SubjectGuid, Subject class Command(BaseCommand): def handle(self, *args, **options): scraper = LegistarAPIBillScraper() scraper.BASE_URL = 'https://webapi.legistar.com/v1/metrotest' scraper.retry_attempts = 0 scraper.requests_per_minute = 0 all_topics = scraper.topics() # Delete topics not currently in use current_topics = Subject.objects.values_list('subject', flat=True) deleted, _ = SubjectGuid.objects.exclude(name__in=current_topics).delete() self.stdout.write('Removed {0} stale topics'.format(deleted)) total_created = 0 for topic in all_topics: subject, created = SubjectGuid.objects.get_or_create( name=topic['IndexName'], guid=topic['api_metadata'] ) if created: total_created += 1 self.stdout.write('{0} topics created'.format(total_created)) Update API path, handle updated SubjectGuid
from django.core.management.base import BaseCommand from django.conf import settings from django.db.utils import IntegrityError from legistar.bills import LegistarAPIBillScraper from lametro.models import SubjectGuid, Subject class Command(BaseCommand): def handle(self, *args, **options): scraper = LegistarAPIBillScraper() scraper.BASE_URL = 'https://webapi.legistar.com/v1/metro' scraper.retry_attempts = 0 scraper.requests_per_minute = 0 all_topics = scraper.topics() # Delete topics not currently in use current_topics = Subject.objects.values_list('subject', flat=True) deleted, _ = SubjectGuid.objects.exclude(name__in=current_topics).delete() self.stdout.write('Removed {0} stale topics'.format(deleted)) total_created = 0 total_updated = 0 total_noop = 0 for topic in all_topics: try: subject, created = SubjectGuid.objects.get_or_create( name=topic['IndexName'], guid=topic['api_metadata'] ) except IntegrityError as e: # This exception will be raised if get_or_create tries to create # a SubjectGuid with a name that already exists. The Legistar # API should not contain duplicates, i.e., the GUID has changed. # Update the GUID on the existing topic. subject = SubjectGuid.objects.get(name=topic['IndexName']) subject.guid = topic['api_metadata'] subject.save() total_updated += 1 else: if created: total_created += 1 else: total_noop += 1 self.stdout.write('Created {0} new topics'.format(total_created)) self.stdout.write('Updated {0} existing topics'.format(total_updated)) self.stdout.write("No-op'ed {0} topics".format(total_updated))
<commit_before>from django.core.management.base import BaseCommand from django.conf import settings from legistar.bills import LegistarAPIBillScraper from lametro.models import SubjectGuid, Subject class Command(BaseCommand): def handle(self, *args, **options): scraper = LegistarAPIBillScraper() scraper.BASE_URL = 'https://webapi.legistar.com/v1/metrotest' scraper.retry_attempts = 0 scraper.requests_per_minute = 0 all_topics = scraper.topics() # Delete topics not currently in use current_topics = Subject.objects.values_list('subject', flat=True) deleted, _ = SubjectGuid.objects.exclude(name__in=current_topics).delete() self.stdout.write('Removed {0} stale topics'.format(deleted)) total_created = 0 for topic in all_topics: subject, created = SubjectGuid.objects.get_or_create( name=topic['IndexName'], guid=topic['api_metadata'] ) if created: total_created += 1 self.stdout.write('{0} topics created'.format(total_created)) <commit_msg>Update API path, handle updated SubjectGuid<commit_after>
from django.core.management.base import BaseCommand from django.conf import settings from django.db.utils import IntegrityError from legistar.bills import LegistarAPIBillScraper from lametro.models import SubjectGuid, Subject class Command(BaseCommand): def handle(self, *args, **options): scraper = LegistarAPIBillScraper() scraper.BASE_URL = 'https://webapi.legistar.com/v1/metro' scraper.retry_attempts = 0 scraper.requests_per_minute = 0 all_topics = scraper.topics() # Delete topics not currently in use current_topics = Subject.objects.values_list('subject', flat=True) deleted, _ = SubjectGuid.objects.exclude(name__in=current_topics).delete() self.stdout.write('Removed {0} stale topics'.format(deleted)) total_created = 0 total_updated = 0 total_noop = 0 for topic in all_topics: try: subject, created = SubjectGuid.objects.get_or_create( name=topic['IndexName'], guid=topic['api_metadata'] ) except IntegrityError as e: # This exception will be raised if get_or_create tries to create # a SubjectGuid with a name that already exists. The Legistar # API should not contain duplicates, i.e., the GUID has changed. # Update the GUID on the existing topic. subject = SubjectGuid.objects.get(name=topic['IndexName']) subject.guid = topic['api_metadata'] subject.save() total_updated += 1 else: if created: total_created += 1 else: total_noop += 1 self.stdout.write('Created {0} new topics'.format(total_created)) self.stdout.write('Updated {0} existing topics'.format(total_updated)) self.stdout.write("No-op'ed {0} topics".format(total_updated))
from django.core.management.base import BaseCommand from django.conf import settings from legistar.bills import LegistarAPIBillScraper from lametro.models import SubjectGuid, Subject class Command(BaseCommand): def handle(self, *args, **options): scraper = LegistarAPIBillScraper() scraper.BASE_URL = 'https://webapi.legistar.com/v1/metrotest' scraper.retry_attempts = 0 scraper.requests_per_minute = 0 all_topics = scraper.topics() # Delete topics not currently in use current_topics = Subject.objects.values_list('subject', flat=True) deleted, _ = SubjectGuid.objects.exclude(name__in=current_topics).delete() self.stdout.write('Removed {0} stale topics'.format(deleted)) total_created = 0 for topic in all_topics: subject, created = SubjectGuid.objects.get_or_create( name=topic['IndexName'], guid=topic['api_metadata'] ) if created: total_created += 1 self.stdout.write('{0} topics created'.format(total_created)) Update API path, handle updated SubjectGuidfrom django.core.management.base import BaseCommand from django.conf import settings from django.db.utils import IntegrityError from legistar.bills import LegistarAPIBillScraper from lametro.models import SubjectGuid, Subject class Command(BaseCommand): def handle(self, *args, **options): scraper = LegistarAPIBillScraper() scraper.BASE_URL = 'https://webapi.legistar.com/v1/metro' scraper.retry_attempts = 0 scraper.requests_per_minute = 0 all_topics = scraper.topics() # Delete topics not currently in use current_topics = Subject.objects.values_list('subject', flat=True) deleted, _ = SubjectGuid.objects.exclude(name__in=current_topics).delete() self.stdout.write('Removed {0} stale topics'.format(deleted)) total_created = 0 total_updated = 0 total_noop = 0 for topic in all_topics: try: subject, created = SubjectGuid.objects.get_or_create( name=topic['IndexName'], guid=topic['api_metadata'] ) except IntegrityError as e: # This exception will be raised if get_or_create tries to create # a SubjectGuid with a name that already exists. The Legistar # API should not contain duplicates, i.e., the GUID has changed. # Update the GUID on the existing topic. subject = SubjectGuid.objects.get(name=topic['IndexName']) subject.guid = topic['api_metadata'] subject.save() total_updated += 1 else: if created: total_created += 1 else: total_noop += 1 self.stdout.write('Created {0} new topics'.format(total_created)) self.stdout.write('Updated {0} existing topics'.format(total_updated)) self.stdout.write("No-op'ed {0} topics".format(total_updated))
<commit_before>from django.core.management.base import BaseCommand from django.conf import settings from legistar.bills import LegistarAPIBillScraper from lametro.models import SubjectGuid, Subject class Command(BaseCommand): def handle(self, *args, **options): scraper = LegistarAPIBillScraper() scraper.BASE_URL = 'https://webapi.legistar.com/v1/metrotest' scraper.retry_attempts = 0 scraper.requests_per_minute = 0 all_topics = scraper.topics() # Delete topics not currently in use current_topics = Subject.objects.values_list('subject', flat=True) deleted, _ = SubjectGuid.objects.exclude(name__in=current_topics).delete() self.stdout.write('Removed {0} stale topics'.format(deleted)) total_created = 0 for topic in all_topics: subject, created = SubjectGuid.objects.get_or_create( name=topic['IndexName'], guid=topic['api_metadata'] ) if created: total_created += 1 self.stdout.write('{0} topics created'.format(total_created)) <commit_msg>Update API path, handle updated SubjectGuid<commit_after>from django.core.management.base import BaseCommand from django.conf import settings from django.db.utils import IntegrityError from legistar.bills import LegistarAPIBillScraper from lametro.models import SubjectGuid, Subject class Command(BaseCommand): def handle(self, *args, **options): scraper = LegistarAPIBillScraper() scraper.BASE_URL = 'https://webapi.legistar.com/v1/metro' scraper.retry_attempts = 0 scraper.requests_per_minute = 0 all_topics = scraper.topics() # Delete topics not currently in use current_topics = Subject.objects.values_list('subject', flat=True) deleted, _ = SubjectGuid.objects.exclude(name__in=current_topics).delete() self.stdout.write('Removed {0} stale topics'.format(deleted)) total_created = 0 total_updated = 0 total_noop = 0 for topic in all_topics: try: subject, created = SubjectGuid.objects.get_or_create( name=topic['IndexName'], guid=topic['api_metadata'] ) except IntegrityError as e: # This exception will be raised if get_or_create tries to create # a SubjectGuid with a name that already exists. The Legistar # API should not contain duplicates, i.e., the GUID has changed. # Update the GUID on the existing topic. subject = SubjectGuid.objects.get(name=topic['IndexName']) subject.guid = topic['api_metadata'] subject.save() total_updated += 1 else: if created: total_created += 1 else: total_noop += 1 self.stdout.write('Created {0} new topics'.format(total_created)) self.stdout.write('Updated {0} existing topics'.format(total_updated)) self.stdout.write("No-op'ed {0} topics".format(total_updated))
7798d5d8c625a4cbdb689cb7ffa38c2f90c0dc02
dduplicated/fileManager.py
dduplicated/fileManager.py
import os from threading import Thread def _delete(path: str, src: str, link: bool): os.remove(path) if link: os.symlink(src, path) def manager_files(paths, link): # The first file is preserved to not delete all files in directories. first = True src = "" deleted_files = [] linked_files = [] errors = [] for path in paths: if os.path.isfile(path): if first: first = False src = path else: Thread(target=_delete, args=(path, src, link)).start() deleted_files.append(path) if link: linked_files.append(path) else: errors.append("Not identified by file: \"{}\"".format(path)) return {"preserved": src, "linked_files": linked_files, "deleted_files": deleted_files, "errors": errors} # Try The Voight-Kampff if you not recognize if is a replicant or not, all is suspect def manager(duplicates, create_link=False): if len(duplicates) == 0: return None processed_files = [] for files_by_hash in duplicates.values(): processed_files.append(manager_files(files_by_hash, create_link)) return processed_files def delete(duplicates): return manager(duplicates) def link(duplicates): return manager(duplicates, True)
import os from threading import Thread def _delete(path: str, src: str, link: bool): os.remove(path) if link: os.symlink(src, path) def manager_files(paths, link): # The first file is preserved to not delete all files in directories. first = True src = "" deleted_files = [] linked_files = [] errors = [] for path in paths: if os.path.isfile(path): if first: first = False src = path else: Thread(target=_delete, args=(path, src, link)).start() deleted_files.append(path) if link: linked_files.append(path) else: errors.append("Not identified by file: \"{}\"".format(path)) return {"preserved": src, "linked_files": linked_files, "deleted_files": deleted_files, "errors": errors} # Try The Voight-Kampff if you not recognize if is a replicant or not, all is suspect def manager(duplicates, create_link=False): if len(duplicates) == 0: # Return empty list object return [] processed_files = [] for files_by_hash in duplicates.values(): processed_files.append(manager_files(files_by_hash, create_link)) return processed_files def delete(duplicates): return manager(duplicates) def link(duplicates): return manager(duplicates, True)
Fix in error when analyze directories without duplicates
Fix in error when analyze directories without duplicates
Python
mit
messiasthi/dduplicated-cli
import os from threading import Thread def _delete(path: str, src: str, link: bool): os.remove(path) if link: os.symlink(src, path) def manager_files(paths, link): # The first file is preserved to not delete all files in directories. first = True src = "" deleted_files = [] linked_files = [] errors = [] for path in paths: if os.path.isfile(path): if first: first = False src = path else: Thread(target=_delete, args=(path, src, link)).start() deleted_files.append(path) if link: linked_files.append(path) else: errors.append("Not identified by file: \"{}\"".format(path)) return {"preserved": src, "linked_files": linked_files, "deleted_files": deleted_files, "errors": errors} # Try The Voight-Kampff if you not recognize if is a replicant or not, all is suspect def manager(duplicates, create_link=False): if len(duplicates) == 0: return None processed_files = [] for files_by_hash in duplicates.values(): processed_files.append(manager_files(files_by_hash, create_link)) return processed_files def delete(duplicates): return manager(duplicates) def link(duplicates): return manager(duplicates, True) Fix in error when analyze directories without duplicates
import os from threading import Thread def _delete(path: str, src: str, link: bool): os.remove(path) if link: os.symlink(src, path) def manager_files(paths, link): # The first file is preserved to not delete all files in directories. first = True src = "" deleted_files = [] linked_files = [] errors = [] for path in paths: if os.path.isfile(path): if first: first = False src = path else: Thread(target=_delete, args=(path, src, link)).start() deleted_files.append(path) if link: linked_files.append(path) else: errors.append("Not identified by file: \"{}\"".format(path)) return {"preserved": src, "linked_files": linked_files, "deleted_files": deleted_files, "errors": errors} # Try The Voight-Kampff if you not recognize if is a replicant or not, all is suspect def manager(duplicates, create_link=False): if len(duplicates) == 0: # Return empty list object return [] processed_files = [] for files_by_hash in duplicates.values(): processed_files.append(manager_files(files_by_hash, create_link)) return processed_files def delete(duplicates): return manager(duplicates) def link(duplicates): return manager(duplicates, True)
<commit_before>import os from threading import Thread def _delete(path: str, src: str, link: bool): os.remove(path) if link: os.symlink(src, path) def manager_files(paths, link): # The first file is preserved to not delete all files in directories. first = True src = "" deleted_files = [] linked_files = [] errors = [] for path in paths: if os.path.isfile(path): if first: first = False src = path else: Thread(target=_delete, args=(path, src, link)).start() deleted_files.append(path) if link: linked_files.append(path) else: errors.append("Not identified by file: \"{}\"".format(path)) return {"preserved": src, "linked_files": linked_files, "deleted_files": deleted_files, "errors": errors} # Try The Voight-Kampff if you not recognize if is a replicant or not, all is suspect def manager(duplicates, create_link=False): if len(duplicates) == 0: return None processed_files = [] for files_by_hash in duplicates.values(): processed_files.append(manager_files(files_by_hash, create_link)) return processed_files def delete(duplicates): return manager(duplicates) def link(duplicates): return manager(duplicates, True) <commit_msg>Fix in error when analyze directories without duplicates<commit_after>
import os from threading import Thread def _delete(path: str, src: str, link: bool): os.remove(path) if link: os.symlink(src, path) def manager_files(paths, link): # The first file is preserved to not delete all files in directories. first = True src = "" deleted_files = [] linked_files = [] errors = [] for path in paths: if os.path.isfile(path): if first: first = False src = path else: Thread(target=_delete, args=(path, src, link)).start() deleted_files.append(path) if link: linked_files.append(path) else: errors.append("Not identified by file: \"{}\"".format(path)) return {"preserved": src, "linked_files": linked_files, "deleted_files": deleted_files, "errors": errors} # Try The Voight-Kampff if you not recognize if is a replicant or not, all is suspect def manager(duplicates, create_link=False): if len(duplicates) == 0: # Return empty list object return [] processed_files = [] for files_by_hash in duplicates.values(): processed_files.append(manager_files(files_by_hash, create_link)) return processed_files def delete(duplicates): return manager(duplicates) def link(duplicates): return manager(duplicates, True)
import os from threading import Thread def _delete(path: str, src: str, link: bool): os.remove(path) if link: os.symlink(src, path) def manager_files(paths, link): # The first file is preserved to not delete all files in directories. first = True src = "" deleted_files = [] linked_files = [] errors = [] for path in paths: if os.path.isfile(path): if first: first = False src = path else: Thread(target=_delete, args=(path, src, link)).start() deleted_files.append(path) if link: linked_files.append(path) else: errors.append("Not identified by file: \"{}\"".format(path)) return {"preserved": src, "linked_files": linked_files, "deleted_files": deleted_files, "errors": errors} # Try The Voight-Kampff if you not recognize if is a replicant or not, all is suspect def manager(duplicates, create_link=False): if len(duplicates) == 0: return None processed_files = [] for files_by_hash in duplicates.values(): processed_files.append(manager_files(files_by_hash, create_link)) return processed_files def delete(duplicates): return manager(duplicates) def link(duplicates): return manager(duplicates, True) Fix in error when analyze directories without duplicatesimport os from threading import Thread def _delete(path: str, src: str, link: bool): os.remove(path) if link: os.symlink(src, path) def manager_files(paths, link): # The first file is preserved to not delete all files in directories. first = True src = "" deleted_files = [] linked_files = [] errors = [] for path in paths: if os.path.isfile(path): if first: first = False src = path else: Thread(target=_delete, args=(path, src, link)).start() deleted_files.append(path) if link: linked_files.append(path) else: errors.append("Not identified by file: \"{}\"".format(path)) return {"preserved": src, "linked_files": linked_files, "deleted_files": deleted_files, "errors": errors} # Try The Voight-Kampff if you not recognize if is a replicant or not, all is suspect def manager(duplicates, create_link=False): if len(duplicates) == 0: # Return empty list object return [] processed_files = [] for files_by_hash in duplicates.values(): processed_files.append(manager_files(files_by_hash, create_link)) return processed_files def delete(duplicates): return manager(duplicates) def link(duplicates): return manager(duplicates, True)
<commit_before>import os from threading import Thread def _delete(path: str, src: str, link: bool): os.remove(path) if link: os.symlink(src, path) def manager_files(paths, link): # The first file is preserved to not delete all files in directories. first = True src = "" deleted_files = [] linked_files = [] errors = [] for path in paths: if os.path.isfile(path): if first: first = False src = path else: Thread(target=_delete, args=(path, src, link)).start() deleted_files.append(path) if link: linked_files.append(path) else: errors.append("Not identified by file: \"{}\"".format(path)) return {"preserved": src, "linked_files": linked_files, "deleted_files": deleted_files, "errors": errors} # Try The Voight-Kampff if you not recognize if is a replicant or not, all is suspect def manager(duplicates, create_link=False): if len(duplicates) == 0: return None processed_files = [] for files_by_hash in duplicates.values(): processed_files.append(manager_files(files_by_hash, create_link)) return processed_files def delete(duplicates): return manager(duplicates) def link(duplicates): return manager(duplicates, True) <commit_msg>Fix in error when analyze directories without duplicates<commit_after>import os from threading import Thread def _delete(path: str, src: str, link: bool): os.remove(path) if link: os.symlink(src, path) def manager_files(paths, link): # The first file is preserved to not delete all files in directories. first = True src = "" deleted_files = [] linked_files = [] errors = [] for path in paths: if os.path.isfile(path): if first: first = False src = path else: Thread(target=_delete, args=(path, src, link)).start() deleted_files.append(path) if link: linked_files.append(path) else: errors.append("Not identified by file: \"{}\"".format(path)) return {"preserved": src, "linked_files": linked_files, "deleted_files": deleted_files, "errors": errors} # Try The Voight-Kampff if you not recognize if is a replicant or not, all is suspect def manager(duplicates, create_link=False): if len(duplicates) == 0: # Return empty list object return [] processed_files = [] for files_by_hash in duplicates.values(): processed_files.append(manager_files(files_by_hash, create_link)) return processed_files def delete(duplicates): return manager(duplicates) def link(duplicates): return manager(duplicates, True)
bfea504373593bbfbe08ad423a8e98ecbd77565e
mule/utils/multithreading.py
mule/utils/multithreading.py
from collections import defaultdict from Queue import Queue from threading import Thread import thread _results = defaultdict(list) class Worker(Thread): """Thread executing tasks from a given tasks queue""" def __init__(self, tasks): Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: func, args, kwargs, ident = self.tasks.get() try: _results[ident].append({ 'func': func, 'args': args, 'kwargs': kwargs, 'result': func(*args, **kwargs), }) finally: self.tasks.task_done() class ThreadPool: """Pool of threads consuming tasks from a queue""" def __init__(self, num_threads): self.tasks = Queue(num_threads) for _ in xrange(num_threads): Worker(self.tasks) def add(self, func, *args, **kwargs): """Add a task to the queue""" self.tasks.put((func, args, kwargs, id(self)), False) def join(self): """Wait for completion of all the tasks in the queue""" self.tasks.join() return _results[id(self)]
from collections import defaultdict from Queue import Queue from threading import Thread import thread _results = defaultdict(list) class Worker(Thread): """Thread executing tasks from a given tasks queue""" def __init__(self, tasks): Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: func, args, kwargs, ident = self.tasks.get() try: _results[ident].append({ 'func': func, 'args': args, 'kwargs': kwargs, 'result': func(*args, **kwargs), }) finally: self.tasks.task_done() class ThreadPool: """Pool of threads consuming tasks from a queue""" def __init__(self, num_threads): self.tasks = Queue(num_threads) for _ in xrange(num_threads): Worker(self.tasks) def add(self, func, *args, **kwargs): """Add a task to the queue""" self.tasks.put((func, args, kwargs, id(self))) def join(self): """Wait for completion of all the tasks in the queue""" self.tasks.join() return _results[id(self)]
Add needs to block so the queue doesnt end up full
Add needs to block so the queue doesnt end up full
Python
apache-2.0
disqus/mule
from collections import defaultdict from Queue import Queue from threading import Thread import thread _results = defaultdict(list) class Worker(Thread): """Thread executing tasks from a given tasks queue""" def __init__(self, tasks): Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: func, args, kwargs, ident = self.tasks.get() try: _results[ident].append({ 'func': func, 'args': args, 'kwargs': kwargs, 'result': func(*args, **kwargs), }) finally: self.tasks.task_done() class ThreadPool: """Pool of threads consuming tasks from a queue""" def __init__(self, num_threads): self.tasks = Queue(num_threads) for _ in xrange(num_threads): Worker(self.tasks) def add(self, func, *args, **kwargs): """Add a task to the queue""" self.tasks.put((func, args, kwargs, id(self)), False) def join(self): """Wait for completion of all the tasks in the queue""" self.tasks.join() return _results[id(self)]Add needs to block so the queue doesnt end up full
from collections import defaultdict from Queue import Queue from threading import Thread import thread _results = defaultdict(list) class Worker(Thread): """Thread executing tasks from a given tasks queue""" def __init__(self, tasks): Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: func, args, kwargs, ident = self.tasks.get() try: _results[ident].append({ 'func': func, 'args': args, 'kwargs': kwargs, 'result': func(*args, **kwargs), }) finally: self.tasks.task_done() class ThreadPool: """Pool of threads consuming tasks from a queue""" def __init__(self, num_threads): self.tasks = Queue(num_threads) for _ in xrange(num_threads): Worker(self.tasks) def add(self, func, *args, **kwargs): """Add a task to the queue""" self.tasks.put((func, args, kwargs, id(self))) def join(self): """Wait for completion of all the tasks in the queue""" self.tasks.join() return _results[id(self)]
<commit_before>from collections import defaultdict from Queue import Queue from threading import Thread import thread _results = defaultdict(list) class Worker(Thread): """Thread executing tasks from a given tasks queue""" def __init__(self, tasks): Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: func, args, kwargs, ident = self.tasks.get() try: _results[ident].append({ 'func': func, 'args': args, 'kwargs': kwargs, 'result': func(*args, **kwargs), }) finally: self.tasks.task_done() class ThreadPool: """Pool of threads consuming tasks from a queue""" def __init__(self, num_threads): self.tasks = Queue(num_threads) for _ in xrange(num_threads): Worker(self.tasks) def add(self, func, *args, **kwargs): """Add a task to the queue""" self.tasks.put((func, args, kwargs, id(self)), False) def join(self): """Wait for completion of all the tasks in the queue""" self.tasks.join() return _results[id(self)]<commit_msg>Add needs to block so the queue doesnt end up full<commit_after>
from collections import defaultdict from Queue import Queue from threading import Thread import thread _results = defaultdict(list) class Worker(Thread): """Thread executing tasks from a given tasks queue""" def __init__(self, tasks): Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: func, args, kwargs, ident = self.tasks.get() try: _results[ident].append({ 'func': func, 'args': args, 'kwargs': kwargs, 'result': func(*args, **kwargs), }) finally: self.tasks.task_done() class ThreadPool: """Pool of threads consuming tasks from a queue""" def __init__(self, num_threads): self.tasks = Queue(num_threads) for _ in xrange(num_threads): Worker(self.tasks) def add(self, func, *args, **kwargs): """Add a task to the queue""" self.tasks.put((func, args, kwargs, id(self))) def join(self): """Wait for completion of all the tasks in the queue""" self.tasks.join() return _results[id(self)]
from collections import defaultdict from Queue import Queue from threading import Thread import thread _results = defaultdict(list) class Worker(Thread): """Thread executing tasks from a given tasks queue""" def __init__(self, tasks): Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: func, args, kwargs, ident = self.tasks.get() try: _results[ident].append({ 'func': func, 'args': args, 'kwargs': kwargs, 'result': func(*args, **kwargs), }) finally: self.tasks.task_done() class ThreadPool: """Pool of threads consuming tasks from a queue""" def __init__(self, num_threads): self.tasks = Queue(num_threads) for _ in xrange(num_threads): Worker(self.tasks) def add(self, func, *args, **kwargs): """Add a task to the queue""" self.tasks.put((func, args, kwargs, id(self)), False) def join(self): """Wait for completion of all the tasks in the queue""" self.tasks.join() return _results[id(self)]Add needs to block so the queue doesnt end up fullfrom collections import defaultdict from Queue import Queue from threading import Thread import thread _results = defaultdict(list) class Worker(Thread): """Thread executing tasks from a given tasks queue""" def __init__(self, tasks): Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: func, args, kwargs, ident = self.tasks.get() try: _results[ident].append({ 'func': func, 'args': args, 'kwargs': kwargs, 'result': func(*args, **kwargs), }) finally: self.tasks.task_done() class ThreadPool: """Pool of threads consuming tasks from a queue""" def __init__(self, num_threads): self.tasks = Queue(num_threads) for _ in xrange(num_threads): Worker(self.tasks) def add(self, func, *args, **kwargs): """Add a task to the queue""" self.tasks.put((func, args, kwargs, id(self))) def join(self): """Wait for completion of all the tasks in the queue""" self.tasks.join() return _results[id(self)]
<commit_before>from collections import defaultdict from Queue import Queue from threading import Thread import thread _results = defaultdict(list) class Worker(Thread): """Thread executing tasks from a given tasks queue""" def __init__(self, tasks): Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: func, args, kwargs, ident = self.tasks.get() try: _results[ident].append({ 'func': func, 'args': args, 'kwargs': kwargs, 'result': func(*args, **kwargs), }) finally: self.tasks.task_done() class ThreadPool: """Pool of threads consuming tasks from a queue""" def __init__(self, num_threads): self.tasks = Queue(num_threads) for _ in xrange(num_threads): Worker(self.tasks) def add(self, func, *args, **kwargs): """Add a task to the queue""" self.tasks.put((func, args, kwargs, id(self)), False) def join(self): """Wait for completion of all the tasks in the queue""" self.tasks.join() return _results[id(self)]<commit_msg>Add needs to block so the queue doesnt end up full<commit_after>from collections import defaultdict from Queue import Queue from threading import Thread import thread _results = defaultdict(list) class Worker(Thread): """Thread executing tasks from a given tasks queue""" def __init__(self, tasks): Thread.__init__(self) self.tasks = tasks self.daemon = True self.start() def run(self): while True: func, args, kwargs, ident = self.tasks.get() try: _results[ident].append({ 'func': func, 'args': args, 'kwargs': kwargs, 'result': func(*args, **kwargs), }) finally: self.tasks.task_done() class ThreadPool: """Pool of threads consuming tasks from a queue""" def __init__(self, num_threads): self.tasks = Queue(num_threads) for _ in xrange(num_threads): Worker(self.tasks) def add(self, func, *args, **kwargs): """Add a task to the queue""" self.tasks.put((func, args, kwargs, id(self))) def join(self): """Wait for completion of all the tasks in the queue""" self.tasks.join() return _results[id(self)]
63935d3c62dad19d2668d0e0633ebd4ce6e6ed26
actions/kvstore.py
actions/kvstore.py
from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = KeyValuePair() instance.id = client.keys.get_by_name(key).name instance.name = key instance.value = value try: kvstore = getattr(client.keys, action) kvp = kvstore(instance) except Exception as e: raise if action == 'delete': return kvp else: return kvp.serialize()
from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = client.keys.get_by_name(key) or KeyValuePair() instance.id = key instance.name = key instance.value = value kvp = client.keys.update(instance) if action in ['create', 'update'] else None if action == 'delete': return kvp else: return kvp.serialize()
Fix create action for key value pair
Fix create action for key value pair
Python
apache-2.0
StackStorm/st2cd,StackStorm/st2cd
from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = KeyValuePair() instance.id = client.keys.get_by_name(key).name instance.name = key instance.value = value try: kvstore = getattr(client.keys, action) kvp = kvstore(instance) except Exception as e: raise if action == 'delete': return kvp else: return kvp.serialize() Fix create action for key value pair
from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = client.keys.get_by_name(key) or KeyValuePair() instance.id = key instance.name = key instance.value = value kvp = client.keys.update(instance) if action in ['create', 'update'] else None if action == 'delete': return kvp else: return kvp.serialize()
<commit_before>from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = KeyValuePair() instance.id = client.keys.get_by_name(key).name instance.name = key instance.value = value try: kvstore = getattr(client.keys, action) kvp = kvstore(instance) except Exception as e: raise if action == 'delete': return kvp else: return kvp.serialize() <commit_msg>Fix create action for key value pair<commit_after>
from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = client.keys.get_by_name(key) or KeyValuePair() instance.id = key instance.name = key instance.value = value kvp = client.keys.update(instance) if action in ['create', 'update'] else None if action == 'delete': return kvp else: return kvp.serialize()
from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = KeyValuePair() instance.id = client.keys.get_by_name(key).name instance.name = key instance.value = value try: kvstore = getattr(client.keys, action) kvp = kvstore(instance) except Exception as e: raise if action == 'delete': return kvp else: return kvp.serialize() Fix create action for key value pairfrom st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = client.keys.get_by_name(key) or KeyValuePair() instance.id = key instance.name = key instance.value = value kvp = client.keys.update(instance) if action in ['create', 'update'] else None if action == 'delete': return kvp else: return kvp.serialize()
<commit_before>from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = KeyValuePair() instance.id = client.keys.get_by_name(key).name instance.name = key instance.value = value try: kvstore = getattr(client.keys, action) kvp = kvstore(instance) except Exception as e: raise if action == 'delete': return kvp else: return kvp.serialize() <commit_msg>Fix create action for key value pair<commit_after>from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = client.keys.get_by_name(key) or KeyValuePair() instance.id = key instance.name = key instance.value = value kvp = client.keys.update(instance) if action in ['create', 'update'] else None if action == 'delete': return kvp else: return kvp.serialize()
c18f21995ff76681fdfa7e511019f5f27bfea749
playserver/trackchecker.py
playserver/trackchecker.py
from threading import Timer from . import track _listeners = [] class TrackChecker(): def __init__(self, interval = 5): self.listeners = [] self.CHECK_INTERVAL = interval self.currentSong = "" self.currentArtist = "" self.currentAlbum = "" self.timer = None def checkSong(self): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() if (song != self.currentSong or artist != self.currentArtist or album != self.currentAlbum): self.currentSong = song self.currentArtist = artist self.currentAlbum = album self._callListeners() if self.timer != None: self.startTimer() def registerListener(self, function): _listeners.append(function) def _callListeners(self): for listener in _listeners: listener() def startTimer(self): self.timer = Timer(self.CHECK_INTERVAL, self.checkSong) self.timer.daemon = True self.timer.start() def cancelTimer(self): self.timer.cancel() self.timer = None
from threading import Timer from . import track _listeners = [] class TrackChecker(): def __init__(self, interval = 5): self.listeners = [] self.CHECK_INTERVAL = interval self.currentSong = "" self.currentArtist = "" self.currentAlbum = "" self.timer = None def checkSong(self): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() if (song != self.currentSong or artist != self.currentArtist or album != self.currentAlbum): self.currentSong = song self.currentArtist = artist self.currentAlbum = album self._callListeners() if self.timer != None: self.startTimer() def registerListener(self, function): _listeners.append(function) def _callListeners(self): data = { "song": track.getCurrentSong(), "artist": track.getCurrentArtist(), "album": track.getCurrentAlbum() } for listener in _listeners: listener(data) def startTimer(self): self.timer = Timer(self.CHECK_INTERVAL, self.checkSong) self.timer.daemon = True self.timer.start() def cancelTimer(self): self.timer.cancel() self.timer = None
Send data in trackChecker callbacks
Send data in trackChecker callbacks
Python
mit
ollien/playserver,ollien/playserver,ollien/playserver
from threading import Timer from . import track _listeners = [] class TrackChecker(): def __init__(self, interval = 5): self.listeners = [] self.CHECK_INTERVAL = interval self.currentSong = "" self.currentArtist = "" self.currentAlbum = "" self.timer = None def checkSong(self): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() if (song != self.currentSong or artist != self.currentArtist or album != self.currentAlbum): self.currentSong = song self.currentArtist = artist self.currentAlbum = album self._callListeners() if self.timer != None: self.startTimer() def registerListener(self, function): _listeners.append(function) def _callListeners(self): for listener in _listeners: listener() def startTimer(self): self.timer = Timer(self.CHECK_INTERVAL, self.checkSong) self.timer.daemon = True self.timer.start() def cancelTimer(self): self.timer.cancel() self.timer = None Send data in trackChecker callbacks
from threading import Timer from . import track _listeners = [] class TrackChecker(): def __init__(self, interval = 5): self.listeners = [] self.CHECK_INTERVAL = interval self.currentSong = "" self.currentArtist = "" self.currentAlbum = "" self.timer = None def checkSong(self): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() if (song != self.currentSong or artist != self.currentArtist or album != self.currentAlbum): self.currentSong = song self.currentArtist = artist self.currentAlbum = album self._callListeners() if self.timer != None: self.startTimer() def registerListener(self, function): _listeners.append(function) def _callListeners(self): data = { "song": track.getCurrentSong(), "artist": track.getCurrentArtist(), "album": track.getCurrentAlbum() } for listener in _listeners: listener(data) def startTimer(self): self.timer = Timer(self.CHECK_INTERVAL, self.checkSong) self.timer.daemon = True self.timer.start() def cancelTimer(self): self.timer.cancel() self.timer = None
<commit_before>from threading import Timer from . import track _listeners = [] class TrackChecker(): def __init__(self, interval = 5): self.listeners = [] self.CHECK_INTERVAL = interval self.currentSong = "" self.currentArtist = "" self.currentAlbum = "" self.timer = None def checkSong(self): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() if (song != self.currentSong or artist != self.currentArtist or album != self.currentAlbum): self.currentSong = song self.currentArtist = artist self.currentAlbum = album self._callListeners() if self.timer != None: self.startTimer() def registerListener(self, function): _listeners.append(function) def _callListeners(self): for listener in _listeners: listener() def startTimer(self): self.timer = Timer(self.CHECK_INTERVAL, self.checkSong) self.timer.daemon = True self.timer.start() def cancelTimer(self): self.timer.cancel() self.timer = None <commit_msg>Send data in trackChecker callbacks<commit_after>
from threading import Timer from . import track _listeners = [] class TrackChecker(): def __init__(self, interval = 5): self.listeners = [] self.CHECK_INTERVAL = interval self.currentSong = "" self.currentArtist = "" self.currentAlbum = "" self.timer = None def checkSong(self): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() if (song != self.currentSong or artist != self.currentArtist or album != self.currentAlbum): self.currentSong = song self.currentArtist = artist self.currentAlbum = album self._callListeners() if self.timer != None: self.startTimer() def registerListener(self, function): _listeners.append(function) def _callListeners(self): data = { "song": track.getCurrentSong(), "artist": track.getCurrentArtist(), "album": track.getCurrentAlbum() } for listener in _listeners: listener(data) def startTimer(self): self.timer = Timer(self.CHECK_INTERVAL, self.checkSong) self.timer.daemon = True self.timer.start() def cancelTimer(self): self.timer.cancel() self.timer = None
from threading import Timer from . import track _listeners = [] class TrackChecker(): def __init__(self, interval = 5): self.listeners = [] self.CHECK_INTERVAL = interval self.currentSong = "" self.currentArtist = "" self.currentAlbum = "" self.timer = None def checkSong(self): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() if (song != self.currentSong or artist != self.currentArtist or album != self.currentAlbum): self.currentSong = song self.currentArtist = artist self.currentAlbum = album self._callListeners() if self.timer != None: self.startTimer() def registerListener(self, function): _listeners.append(function) def _callListeners(self): for listener in _listeners: listener() def startTimer(self): self.timer = Timer(self.CHECK_INTERVAL, self.checkSong) self.timer.daemon = True self.timer.start() def cancelTimer(self): self.timer.cancel() self.timer = None Send data in trackChecker callbacksfrom threading import Timer from . import track _listeners = [] class TrackChecker(): def __init__(self, interval = 5): self.listeners = [] self.CHECK_INTERVAL = interval self.currentSong = "" self.currentArtist = "" self.currentAlbum = "" self.timer = None def checkSong(self): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() if (song != self.currentSong or artist != self.currentArtist or album != self.currentAlbum): self.currentSong = song self.currentArtist = artist self.currentAlbum = album self._callListeners() if self.timer != None: self.startTimer() def registerListener(self, function): _listeners.append(function) def _callListeners(self): data = { "song": track.getCurrentSong(), "artist": track.getCurrentArtist(), "album": track.getCurrentAlbum() } for listener in _listeners: listener(data) def startTimer(self): self.timer = Timer(self.CHECK_INTERVAL, self.checkSong) self.timer.daemon = True self.timer.start() def cancelTimer(self): self.timer.cancel() self.timer = None
<commit_before>from threading import Timer from . import track _listeners = [] class TrackChecker(): def __init__(self, interval = 5): self.listeners = [] self.CHECK_INTERVAL = interval self.currentSong = "" self.currentArtist = "" self.currentAlbum = "" self.timer = None def checkSong(self): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() if (song != self.currentSong or artist != self.currentArtist or album != self.currentAlbum): self.currentSong = song self.currentArtist = artist self.currentAlbum = album self._callListeners() if self.timer != None: self.startTimer() def registerListener(self, function): _listeners.append(function) def _callListeners(self): for listener in _listeners: listener() def startTimer(self): self.timer = Timer(self.CHECK_INTERVAL, self.checkSong) self.timer.daemon = True self.timer.start() def cancelTimer(self): self.timer.cancel() self.timer = None <commit_msg>Send data in trackChecker callbacks<commit_after>from threading import Timer from . import track _listeners = [] class TrackChecker(): def __init__(self, interval = 5): self.listeners = [] self.CHECK_INTERVAL = interval self.currentSong = "" self.currentArtist = "" self.currentAlbum = "" self.timer = None def checkSong(self): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() if (song != self.currentSong or artist != self.currentArtist or album != self.currentAlbum): self.currentSong = song self.currentArtist = artist self.currentAlbum = album self._callListeners() if self.timer != None: self.startTimer() def registerListener(self, function): _listeners.append(function) def _callListeners(self): data = { "song": track.getCurrentSong(), "artist": track.getCurrentArtist(), "album": track.getCurrentAlbum() } for listener in _listeners: listener(data) def startTimer(self): self.timer = Timer(self.CHECK_INTERVAL, self.checkSong) self.timer.daemon = True self.timer.start() def cancelTimer(self): self.timer.cancel() self.timer = None
51080ad6e8ad38ea8c22593b07d70b27965545bd
api/views/users.py
api/views/users.py
from rest_framework import viewsets, status from rest_framework.authtoken.models import Token from rest_framework.decorators import detail_route from rest_framework.response import Response from api.models import User from api.serializers import UserSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer @detail_route(methods=['put']) def follow(self, request, pk=None): # follows a given user pass @detail_route(methods=['put']) def unfollow(self, request, pk=None): # unfollows a given user pass # Override create to return token def create(self, request, *args, **kwargs): serializer = self.get_serializer(data=request.data) serializer.is_valid(raise_exception=True) user = serializer.save() headers = self.get_success_headers(serializer.data) response = serializer.data response['access_token'] = Token.objects.get(user=user).key return Response(response, status=status.HTTP_201_CREATED, headers=headers)
from rest_framework import viewsets, status from rest_framework.authtoken.models import Token from rest_framework.decorators import detail_route from rest_framework.response import Response from api.models import User from api.serializers import UserSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer permission_classes_by_action = {'create': []} @detail_route(methods=['put']) def follow(self, request, pk=None): # follows a given user pass @detail_route(methods=['put']) def unfollow(self, request, pk=None): # unfollows a given user pass # Override create to return token def create(self, request, *args, **kwargs): serializer = self.get_serializer(data=request.data) serializer.is_valid(raise_exception=True) user = serializer.save() headers = self.get_success_headers(serializer.data) response = serializer.data response['access_token'] = Token.objects.get(user=user).key return Response(response, status=status.HTTP_201_CREATED, headers=headers) # Token isn't required when creating user (signup) def get_permissions(self): try: return [permission() for permission in self.permission_classes_by_action[self.action]] except KeyError: return [permission() for permission in self.permission_classes]
Make signup not require token in API
Make signup not require token in API
Python
mit
frostblooded/kanq,frostblooded/kanq,frostblooded/kanq,frostblooded/kanq,frostblooded/kanq
from rest_framework import viewsets, status from rest_framework.authtoken.models import Token from rest_framework.decorators import detail_route from rest_framework.response import Response from api.models import User from api.serializers import UserSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer @detail_route(methods=['put']) def follow(self, request, pk=None): # follows a given user pass @detail_route(methods=['put']) def unfollow(self, request, pk=None): # unfollows a given user pass # Override create to return token def create(self, request, *args, **kwargs): serializer = self.get_serializer(data=request.data) serializer.is_valid(raise_exception=True) user = serializer.save() headers = self.get_success_headers(serializer.data) response = serializer.data response['access_token'] = Token.objects.get(user=user).key return Response(response, status=status.HTTP_201_CREATED, headers=headers) Make signup not require token in API
from rest_framework import viewsets, status from rest_framework.authtoken.models import Token from rest_framework.decorators import detail_route from rest_framework.response import Response from api.models import User from api.serializers import UserSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer permission_classes_by_action = {'create': []} @detail_route(methods=['put']) def follow(self, request, pk=None): # follows a given user pass @detail_route(methods=['put']) def unfollow(self, request, pk=None): # unfollows a given user pass # Override create to return token def create(self, request, *args, **kwargs): serializer = self.get_serializer(data=request.data) serializer.is_valid(raise_exception=True) user = serializer.save() headers = self.get_success_headers(serializer.data) response = serializer.data response['access_token'] = Token.objects.get(user=user).key return Response(response, status=status.HTTP_201_CREATED, headers=headers) # Token isn't required when creating user (signup) def get_permissions(self): try: return [permission() for permission in self.permission_classes_by_action[self.action]] except KeyError: return [permission() for permission in self.permission_classes]
<commit_before>from rest_framework import viewsets, status from rest_framework.authtoken.models import Token from rest_framework.decorators import detail_route from rest_framework.response import Response from api.models import User from api.serializers import UserSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer @detail_route(methods=['put']) def follow(self, request, pk=None): # follows a given user pass @detail_route(methods=['put']) def unfollow(self, request, pk=None): # unfollows a given user pass # Override create to return token def create(self, request, *args, **kwargs): serializer = self.get_serializer(data=request.data) serializer.is_valid(raise_exception=True) user = serializer.save() headers = self.get_success_headers(serializer.data) response = serializer.data response['access_token'] = Token.objects.get(user=user).key return Response(response, status=status.HTTP_201_CREATED, headers=headers) <commit_msg>Make signup not require token in API<commit_after>
from rest_framework import viewsets, status from rest_framework.authtoken.models import Token from rest_framework.decorators import detail_route from rest_framework.response import Response from api.models import User from api.serializers import UserSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer permission_classes_by_action = {'create': []} @detail_route(methods=['put']) def follow(self, request, pk=None): # follows a given user pass @detail_route(methods=['put']) def unfollow(self, request, pk=None): # unfollows a given user pass # Override create to return token def create(self, request, *args, **kwargs): serializer = self.get_serializer(data=request.data) serializer.is_valid(raise_exception=True) user = serializer.save() headers = self.get_success_headers(serializer.data) response = serializer.data response['access_token'] = Token.objects.get(user=user).key return Response(response, status=status.HTTP_201_CREATED, headers=headers) # Token isn't required when creating user (signup) def get_permissions(self): try: return [permission() for permission in self.permission_classes_by_action[self.action]] except KeyError: return [permission() for permission in self.permission_classes]
from rest_framework import viewsets, status from rest_framework.authtoken.models import Token from rest_framework.decorators import detail_route from rest_framework.response import Response from api.models import User from api.serializers import UserSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer @detail_route(methods=['put']) def follow(self, request, pk=None): # follows a given user pass @detail_route(methods=['put']) def unfollow(self, request, pk=None): # unfollows a given user pass # Override create to return token def create(self, request, *args, **kwargs): serializer = self.get_serializer(data=request.data) serializer.is_valid(raise_exception=True) user = serializer.save() headers = self.get_success_headers(serializer.data) response = serializer.data response['access_token'] = Token.objects.get(user=user).key return Response(response, status=status.HTTP_201_CREATED, headers=headers) Make signup not require token in APIfrom rest_framework import viewsets, status from rest_framework.authtoken.models import Token from rest_framework.decorators import detail_route from rest_framework.response import Response from api.models import User from api.serializers import UserSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer permission_classes_by_action = {'create': []} @detail_route(methods=['put']) def follow(self, request, pk=None): # follows a given user pass @detail_route(methods=['put']) def unfollow(self, request, pk=None): # unfollows a given user pass # Override create to return token def create(self, request, *args, **kwargs): serializer = self.get_serializer(data=request.data) serializer.is_valid(raise_exception=True) user = serializer.save() headers = self.get_success_headers(serializer.data) response = serializer.data response['access_token'] = Token.objects.get(user=user).key return Response(response, status=status.HTTP_201_CREATED, headers=headers) # Token isn't required when creating user (signup) def get_permissions(self): try: return [permission() for permission in self.permission_classes_by_action[self.action]] except KeyError: return [permission() for permission in self.permission_classes]
<commit_before>from rest_framework import viewsets, status from rest_framework.authtoken.models import Token from rest_framework.decorators import detail_route from rest_framework.response import Response from api.models import User from api.serializers import UserSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer @detail_route(methods=['put']) def follow(self, request, pk=None): # follows a given user pass @detail_route(methods=['put']) def unfollow(self, request, pk=None): # unfollows a given user pass # Override create to return token def create(self, request, *args, **kwargs): serializer = self.get_serializer(data=request.data) serializer.is_valid(raise_exception=True) user = serializer.save() headers = self.get_success_headers(serializer.data) response = serializer.data response['access_token'] = Token.objects.get(user=user).key return Response(response, status=status.HTTP_201_CREATED, headers=headers) <commit_msg>Make signup not require token in API<commit_after>from rest_framework import viewsets, status from rest_framework.authtoken.models import Token from rest_framework.decorators import detail_route from rest_framework.response import Response from api.models import User from api.serializers import UserSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all() serializer_class = UserSerializer permission_classes_by_action = {'create': []} @detail_route(methods=['put']) def follow(self, request, pk=None): # follows a given user pass @detail_route(methods=['put']) def unfollow(self, request, pk=None): # unfollows a given user pass # Override create to return token def create(self, request, *args, **kwargs): serializer = self.get_serializer(data=request.data) serializer.is_valid(raise_exception=True) user = serializer.save() headers = self.get_success_headers(serializer.data) response = serializer.data response['access_token'] = Token.objects.get(user=user).key return Response(response, status=status.HTTP_201_CREATED, headers=headers) # Token isn't required when creating user (signup) def get_permissions(self): try: return [permission() for permission in self.permission_classes_by_action[self.action]] except KeyError: return [permission() for permission in self.permission_classes]
6d7c1172ff156f376c61476bccf9912598059d19
rabbitpy/__init__.py
rabbitpy/__init__.py
__version__ = '0.10.0' version = __version__ from rabbitpy.connection import Connection from rabbitpy.exchange import Exchange from rabbitpy.message import Message from rabbitpy.amqp_queue import Queue from rabbitpy.tx import Tx from rabbitpy.simple import consume from rabbitpy.simple import get from rabbitpy.simple import publish from rabbitpy.simple import create_queue from rabbitpy.simple import delete_queue from rabbitpy.simple import create_direct_exchange from rabbitpy.simple import create_fanout_exchange from rabbitpy.simple import create_topic_exchange from rabbitpy.simple import delete_exchange import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('rabbitpy').addHandler(NullHandler()) DEBUG = False
__version__ = '0.10.0' version = __version__ DEBUG = False import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('rabbitpy').addHandler(NullHandler()) from rabbitpy.connection import Connection from rabbitpy.exchange import Exchange from rabbitpy.message import Message from rabbitpy.amqp_queue import Queue from rabbitpy.tx import Tx from rabbitpy.simple import consume from rabbitpy.simple import get from rabbitpy.simple import publish from rabbitpy.simple import create_queue from rabbitpy.simple import delete_queue from rabbitpy.simple import create_direct_exchange from rabbitpy.simple import create_fanout_exchange from rabbitpy.simple import create_topic_exchange from rabbitpy.simple import delete_exchange
Fix order of operations issues
Fix order of operations issues
Python
bsd-3-clause
jonahbull/rabbitpy,gmr/rabbitpy,gmr/rabbitpy
__version__ = '0.10.0' version = __version__ from rabbitpy.connection import Connection from rabbitpy.exchange import Exchange from rabbitpy.message import Message from rabbitpy.amqp_queue import Queue from rabbitpy.tx import Tx from rabbitpy.simple import consume from rabbitpy.simple import get from rabbitpy.simple import publish from rabbitpy.simple import create_queue from rabbitpy.simple import delete_queue from rabbitpy.simple import create_direct_exchange from rabbitpy.simple import create_fanout_exchange from rabbitpy.simple import create_topic_exchange from rabbitpy.simple import delete_exchange import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('rabbitpy').addHandler(NullHandler()) DEBUG = False Fix order of operations issues
__version__ = '0.10.0' version = __version__ DEBUG = False import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('rabbitpy').addHandler(NullHandler()) from rabbitpy.connection import Connection from rabbitpy.exchange import Exchange from rabbitpy.message import Message from rabbitpy.amqp_queue import Queue from rabbitpy.tx import Tx from rabbitpy.simple import consume from rabbitpy.simple import get from rabbitpy.simple import publish from rabbitpy.simple import create_queue from rabbitpy.simple import delete_queue from rabbitpy.simple import create_direct_exchange from rabbitpy.simple import create_fanout_exchange from rabbitpy.simple import create_topic_exchange from rabbitpy.simple import delete_exchange
<commit_before>__version__ = '0.10.0' version = __version__ from rabbitpy.connection import Connection from rabbitpy.exchange import Exchange from rabbitpy.message import Message from rabbitpy.amqp_queue import Queue from rabbitpy.tx import Tx from rabbitpy.simple import consume from rabbitpy.simple import get from rabbitpy.simple import publish from rabbitpy.simple import create_queue from rabbitpy.simple import delete_queue from rabbitpy.simple import create_direct_exchange from rabbitpy.simple import create_fanout_exchange from rabbitpy.simple import create_topic_exchange from rabbitpy.simple import delete_exchange import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('rabbitpy').addHandler(NullHandler()) DEBUG = False <commit_msg>Fix order of operations issues<commit_after>
__version__ = '0.10.0' version = __version__ DEBUG = False import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('rabbitpy').addHandler(NullHandler()) from rabbitpy.connection import Connection from rabbitpy.exchange import Exchange from rabbitpy.message import Message from rabbitpy.amqp_queue import Queue from rabbitpy.tx import Tx from rabbitpy.simple import consume from rabbitpy.simple import get from rabbitpy.simple import publish from rabbitpy.simple import create_queue from rabbitpy.simple import delete_queue from rabbitpy.simple import create_direct_exchange from rabbitpy.simple import create_fanout_exchange from rabbitpy.simple import create_topic_exchange from rabbitpy.simple import delete_exchange
__version__ = '0.10.0' version = __version__ from rabbitpy.connection import Connection from rabbitpy.exchange import Exchange from rabbitpy.message import Message from rabbitpy.amqp_queue import Queue from rabbitpy.tx import Tx from rabbitpy.simple import consume from rabbitpy.simple import get from rabbitpy.simple import publish from rabbitpy.simple import create_queue from rabbitpy.simple import delete_queue from rabbitpy.simple import create_direct_exchange from rabbitpy.simple import create_fanout_exchange from rabbitpy.simple import create_topic_exchange from rabbitpy.simple import delete_exchange import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('rabbitpy').addHandler(NullHandler()) DEBUG = False Fix order of operations issues__version__ = '0.10.0' version = __version__ DEBUG = False import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('rabbitpy').addHandler(NullHandler()) from rabbitpy.connection import Connection from rabbitpy.exchange import Exchange from rabbitpy.message import Message from rabbitpy.amqp_queue import Queue from rabbitpy.tx import Tx from rabbitpy.simple import consume from rabbitpy.simple import get from rabbitpy.simple import publish from rabbitpy.simple import create_queue from rabbitpy.simple import delete_queue from rabbitpy.simple import create_direct_exchange from rabbitpy.simple import create_fanout_exchange from rabbitpy.simple import create_topic_exchange from rabbitpy.simple import delete_exchange
<commit_before>__version__ = '0.10.0' version = __version__ from rabbitpy.connection import Connection from rabbitpy.exchange import Exchange from rabbitpy.message import Message from rabbitpy.amqp_queue import Queue from rabbitpy.tx import Tx from rabbitpy.simple import consume from rabbitpy.simple import get from rabbitpy.simple import publish from rabbitpy.simple import create_queue from rabbitpy.simple import delete_queue from rabbitpy.simple import create_direct_exchange from rabbitpy.simple import create_fanout_exchange from rabbitpy.simple import create_topic_exchange from rabbitpy.simple import delete_exchange import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('rabbitpy').addHandler(NullHandler()) DEBUG = False <commit_msg>Fix order of operations issues<commit_after>__version__ = '0.10.0' version = __version__ DEBUG = False import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger('rabbitpy').addHandler(NullHandler()) from rabbitpy.connection import Connection from rabbitpy.exchange import Exchange from rabbitpy.message import Message from rabbitpy.amqp_queue import Queue from rabbitpy.tx import Tx from rabbitpy.simple import consume from rabbitpy.simple import get from rabbitpy.simple import publish from rabbitpy.simple import create_queue from rabbitpy.simple import delete_queue from rabbitpy.simple import create_direct_exchange from rabbitpy.simple import create_fanout_exchange from rabbitpy.simple import create_topic_exchange from rabbitpy.simple import delete_exchange
7c60024684024b604eb19a02d119adab547ed0d1
ovp_organizations/migrations/0023_auto_20170627_0236.py
ovp_organizations/migrations/0023_auto_20170627_0236.py
# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2017-06-27 02:36 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('ovp_organizations', '0022_auto_20170613_1424'), ] operations = [ migrations.AlterField( model_name='organization', name='address', field=models.OneToOneField(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.CASCADE, to='ovp_core.SimpleAddress', verbose_name='address'), ), ]
# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2017-06-27 02:36 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('ovp_organizations', '0022_auto_20170613_1424'), ('ovp_core', '0011_simpleaddress'), ] operations = [ migrations.AlterField( model_name='organization', name='address', field=models.OneToOneField(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.CASCADE, to='ovp_core.SimpleAddress', verbose_name='address'), ), ]
Add ovp_core_0011 migration as dependency for ovp_organizations_0023
Add ovp_core_0011 migration as dependency for ovp_organizations_0023
Python
agpl-3.0
OpenVolunteeringPlatform/django-ovp-organizations,OpenVolunteeringPlatform/django-ovp-organizations
# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2017-06-27 02:36 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('ovp_organizations', '0022_auto_20170613_1424'), ] operations = [ migrations.AlterField( model_name='organization', name='address', field=models.OneToOneField(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.CASCADE, to='ovp_core.SimpleAddress', verbose_name='address'), ), ] Add ovp_core_0011 migration as dependency for ovp_organizations_0023
# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2017-06-27 02:36 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('ovp_organizations', '0022_auto_20170613_1424'), ('ovp_core', '0011_simpleaddress'), ] operations = [ migrations.AlterField( model_name='organization', name='address', field=models.OneToOneField(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.CASCADE, to='ovp_core.SimpleAddress', verbose_name='address'), ), ]
<commit_before># -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2017-06-27 02:36 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('ovp_organizations', '0022_auto_20170613_1424'), ] operations = [ migrations.AlterField( model_name='organization', name='address', field=models.OneToOneField(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.CASCADE, to='ovp_core.SimpleAddress', verbose_name='address'), ), ] <commit_msg>Add ovp_core_0011 migration as dependency for ovp_organizations_0023<commit_after>
# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2017-06-27 02:36 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('ovp_organizations', '0022_auto_20170613_1424'), ('ovp_core', '0011_simpleaddress'), ] operations = [ migrations.AlterField( model_name='organization', name='address', field=models.OneToOneField(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.CASCADE, to='ovp_core.SimpleAddress', verbose_name='address'), ), ]
# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2017-06-27 02:36 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('ovp_organizations', '0022_auto_20170613_1424'), ] operations = [ migrations.AlterField( model_name='organization', name='address', field=models.OneToOneField(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.CASCADE, to='ovp_core.SimpleAddress', verbose_name='address'), ), ] Add ovp_core_0011 migration as dependency for ovp_organizations_0023# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2017-06-27 02:36 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('ovp_organizations', '0022_auto_20170613_1424'), ('ovp_core', '0011_simpleaddress'), ] operations = [ migrations.AlterField( model_name='organization', name='address', field=models.OneToOneField(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.CASCADE, to='ovp_core.SimpleAddress', verbose_name='address'), ), ]
<commit_before># -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2017-06-27 02:36 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('ovp_organizations', '0022_auto_20170613_1424'), ] operations = [ migrations.AlterField( model_name='organization', name='address', field=models.OneToOneField(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.CASCADE, to='ovp_core.SimpleAddress', verbose_name='address'), ), ] <commit_msg>Add ovp_core_0011 migration as dependency for ovp_organizations_0023<commit_after># -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2017-06-27 02:36 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('ovp_organizations', '0022_auto_20170613_1424'), ('ovp_core', '0011_simpleaddress'), ] operations = [ migrations.AlterField( model_name='organization', name='address', field=models.OneToOneField(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.CASCADE, to='ovp_core.SimpleAddress', verbose_name='address'), ), ]
b2f40d9b1ed9d78a4fdc1f73e64575a26d117d0c
nuitka/tools/release/msi_create/__main__.py
nuitka/tools/release/msi_create/__main__.py
# Copyright 2017, Kay Hayen, mailto:kay.hayen@gmail.com # # Part of "Nuitka", an optimizing Python compiler that is compatible and # integrates with CPython, but also works on its own. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Release: Create Windows MSI files for Nuitka """ from nuitka.tools.release.MSI import createMSIPackage def main(): createMSIPackage() if __name__ == "__main__": main()
# Copyright 2017, Kay Hayen, mailto:kay.hayen@gmail.com # # Part of "Nuitka", an optimizing Python compiler that is compatible and # integrates with CPython, but also works on its own. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Release: Create Windows MSI files for Nuitka """ import os import shutil from nuitka.tools.Basics import goHome from nuitka.tools.release.MSI import createMSIPackage def main(): goHome() msi_filename = createMSIPackage() if not os.path.exists("msi"): os.makedirs("msi") shutil.move(msi_filename, "msi") if __name__ == "__main__": main()
Copy created MSI to dedicated folder.
Release: Copy created MSI to dedicated folder. * The "dist" folder is erased each time and we determine the result name from being the only MSI file.
Python
apache-2.0
kayhayen/Nuitka,kayhayen/Nuitka,kayhayen/Nuitka,kayhayen/Nuitka
# Copyright 2017, Kay Hayen, mailto:kay.hayen@gmail.com # # Part of "Nuitka", an optimizing Python compiler that is compatible and # integrates with CPython, but also works on its own. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Release: Create Windows MSI files for Nuitka """ from nuitka.tools.release.MSI import createMSIPackage def main(): createMSIPackage() if __name__ == "__main__": main() Release: Copy created MSI to dedicated folder. * The "dist" folder is erased each time and we determine the result name from being the only MSI file.
# Copyright 2017, Kay Hayen, mailto:kay.hayen@gmail.com # # Part of "Nuitka", an optimizing Python compiler that is compatible and # integrates with CPython, but also works on its own. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Release: Create Windows MSI files for Nuitka """ import os import shutil from nuitka.tools.Basics import goHome from nuitka.tools.release.MSI import createMSIPackage def main(): goHome() msi_filename = createMSIPackage() if not os.path.exists("msi"): os.makedirs("msi") shutil.move(msi_filename, "msi") if __name__ == "__main__": main()
<commit_before># Copyright 2017, Kay Hayen, mailto:kay.hayen@gmail.com # # Part of "Nuitka", an optimizing Python compiler that is compatible and # integrates with CPython, but also works on its own. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Release: Create Windows MSI files for Nuitka """ from nuitka.tools.release.MSI import createMSIPackage def main(): createMSIPackage() if __name__ == "__main__": main() <commit_msg>Release: Copy created MSI to dedicated folder. * The "dist" folder is erased each time and we determine the result name from being the only MSI file.<commit_after>
# Copyright 2017, Kay Hayen, mailto:kay.hayen@gmail.com # # Part of "Nuitka", an optimizing Python compiler that is compatible and # integrates with CPython, but also works on its own. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Release: Create Windows MSI files for Nuitka """ import os import shutil from nuitka.tools.Basics import goHome from nuitka.tools.release.MSI import createMSIPackage def main(): goHome() msi_filename = createMSIPackage() if not os.path.exists("msi"): os.makedirs("msi") shutil.move(msi_filename, "msi") if __name__ == "__main__": main()
# Copyright 2017, Kay Hayen, mailto:kay.hayen@gmail.com # # Part of "Nuitka", an optimizing Python compiler that is compatible and # integrates with CPython, but also works on its own. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Release: Create Windows MSI files for Nuitka """ from nuitka.tools.release.MSI import createMSIPackage def main(): createMSIPackage() if __name__ == "__main__": main() Release: Copy created MSI to dedicated folder. * The "dist" folder is erased each time and we determine the result name from being the only MSI file.# Copyright 2017, Kay Hayen, mailto:kay.hayen@gmail.com # # Part of "Nuitka", an optimizing Python compiler that is compatible and # integrates with CPython, but also works on its own. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Release: Create Windows MSI files for Nuitka """ import os import shutil from nuitka.tools.Basics import goHome from nuitka.tools.release.MSI import createMSIPackage def main(): goHome() msi_filename = createMSIPackage() if not os.path.exists("msi"): os.makedirs("msi") shutil.move(msi_filename, "msi") if __name__ == "__main__": main()
<commit_before># Copyright 2017, Kay Hayen, mailto:kay.hayen@gmail.com # # Part of "Nuitka", an optimizing Python compiler that is compatible and # integrates with CPython, but also works on its own. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Release: Create Windows MSI files for Nuitka """ from nuitka.tools.release.MSI import createMSIPackage def main(): createMSIPackage() if __name__ == "__main__": main() <commit_msg>Release: Copy created MSI to dedicated folder. * The "dist" folder is erased each time and we determine the result name from being the only MSI file.<commit_after># Copyright 2017, Kay Hayen, mailto:kay.hayen@gmail.com # # Part of "Nuitka", an optimizing Python compiler that is compatible and # integrates with CPython, but also works on its own. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Release: Create Windows MSI files for Nuitka """ import os import shutil from nuitka.tools.Basics import goHome from nuitka.tools.release.MSI import createMSIPackage def main(): goHome() msi_filename = createMSIPackage() if not os.path.exists("msi"): os.makedirs("msi") shutil.move(msi_filename, "msi") if __name__ == "__main__": main()
41a8d7e1e762133fb24608524e229b882fabda22
packages/pegasus-python/src/Pegasus/service/defaults.py
packages/pegasus-python/src/Pegasus/service/defaults.py
import os import tempfile # SERVER CONFIGURATION SERVER_HOST = "127.0.0.1" SERVER_PORT = 5000 # SSL config: path to certificate and private key files CERTIFICATE = None PRIVATE_KEY = None # Max number of processes to fork when handling requests MAX_PROCESSES = 10 # Enable debugging DEBUG = False # The secret key used by Flask to encrypt session keys SECRET_KEY = os.urandom(24) # Authentication method to use (NoAuthentication or PAMAuthentication) AUTHENTICATION = "PAMAuthentication" PROCESS_SWITCHING = True # Flask cache configuration CACHE_TYPE = "filesystem" CACHE_DIR = os.path.join(tempfile.gettempdir(), "pegasus-service") # # Authorization - # None, '', False -> User can only access their own data. # * -> All users are admin users and can access data of any other user. # {'u1', .., 'un'} OR ['u1', .., 'un'] -> Only users in the set/list are admin users. # ADMIN_USERS = None # CLIENT CONFIGURATION # User credentials USERNAME = "" PASSWORD = "" # ENSEMBLE MANAGER CONFIGURATION # Workflow processing interval in seconds EM_INTERVAL = 60 # Directory to store data STORAGE_DIRECTORY = "/var/pegasus" # Path to Pegasus home directory # PEGASUS_HOME = "/usr" # Path to Condor home directory # CONDOR_HOME = "/usr"
import os import tempfile # SERVER CONFIGURATION SERVER_HOST = "127.0.0.1" SERVER_PORT = 5000 # SSL config: path to certificate and private key files CERTIFICATE = None PRIVATE_KEY = None # Max number of processes to fork when handling requests MAX_PROCESSES = 10 # Enable debugging DEBUG = False # The secret key used by Flask to encrypt session keys SECRET_KEY = os.urandom(24) # Authentication method to use (NoAuthentication or PAMAuthentication) AUTHENTICATION = "PAMAuthentication" PROCESS_SWITCHING = True # Flask cache configuration CACHE_TYPE = "flask_caching.backends.SimpleCache" # # Authorization - # None, '', False -> User can only access their own data. # * -> All users are admin users and can access data of any other user. # {'u1', .., 'un'} OR ['u1', .., 'un'] -> Only users in the set/list are admin users. # ADMIN_USERS = None # CLIENT CONFIGURATION # User credentials USERNAME = "" PASSWORD = "" # ENSEMBLE MANAGER CONFIGURATION # Workflow processing interval in seconds EM_INTERVAL = 60 # Directory to store data STORAGE_DIRECTORY = "/var/pegasus" # Path to Pegasus home directory # PEGASUS_HOME = "/usr" # Path to Condor home directory # CONDOR_HOME = "/usr"
Use SimpleCache instead of Filesystemcache as it can cause errors due to use of different pickle versions.
Use SimpleCache instead of Filesystemcache as it can cause errors due to use of different pickle versions.
Python
apache-2.0
pegasus-isi/pegasus,pegasus-isi/pegasus,pegasus-isi/pegasus,pegasus-isi/pegasus,pegasus-isi/pegasus,pegasus-isi/pegasus,pegasus-isi/pegasus,pegasus-isi/pegasus,pegasus-isi/pegasus,pegasus-isi/pegasus
import os import tempfile # SERVER CONFIGURATION SERVER_HOST = "127.0.0.1" SERVER_PORT = 5000 # SSL config: path to certificate and private key files CERTIFICATE = None PRIVATE_KEY = None # Max number of processes to fork when handling requests MAX_PROCESSES = 10 # Enable debugging DEBUG = False # The secret key used by Flask to encrypt session keys SECRET_KEY = os.urandom(24) # Authentication method to use (NoAuthentication or PAMAuthentication) AUTHENTICATION = "PAMAuthentication" PROCESS_SWITCHING = True # Flask cache configuration CACHE_TYPE = "filesystem" CACHE_DIR = os.path.join(tempfile.gettempdir(), "pegasus-service") # # Authorization - # None, '', False -> User can only access their own data. # * -> All users are admin users and can access data of any other user. # {'u1', .., 'un'} OR ['u1', .., 'un'] -> Only users in the set/list are admin users. # ADMIN_USERS = None # CLIENT CONFIGURATION # User credentials USERNAME = "" PASSWORD = "" # ENSEMBLE MANAGER CONFIGURATION # Workflow processing interval in seconds EM_INTERVAL = 60 # Directory to store data STORAGE_DIRECTORY = "/var/pegasus" # Path to Pegasus home directory # PEGASUS_HOME = "/usr" # Path to Condor home directory # CONDOR_HOME = "/usr" Use SimpleCache instead of Filesystemcache as it can cause errors due to use of different pickle versions.
import os import tempfile # SERVER CONFIGURATION SERVER_HOST = "127.0.0.1" SERVER_PORT = 5000 # SSL config: path to certificate and private key files CERTIFICATE = None PRIVATE_KEY = None # Max number of processes to fork when handling requests MAX_PROCESSES = 10 # Enable debugging DEBUG = False # The secret key used by Flask to encrypt session keys SECRET_KEY = os.urandom(24) # Authentication method to use (NoAuthentication or PAMAuthentication) AUTHENTICATION = "PAMAuthentication" PROCESS_SWITCHING = True # Flask cache configuration CACHE_TYPE = "flask_caching.backends.SimpleCache" # # Authorization - # None, '', False -> User can only access their own data. # * -> All users are admin users and can access data of any other user. # {'u1', .., 'un'} OR ['u1', .., 'un'] -> Only users in the set/list are admin users. # ADMIN_USERS = None # CLIENT CONFIGURATION # User credentials USERNAME = "" PASSWORD = "" # ENSEMBLE MANAGER CONFIGURATION # Workflow processing interval in seconds EM_INTERVAL = 60 # Directory to store data STORAGE_DIRECTORY = "/var/pegasus" # Path to Pegasus home directory # PEGASUS_HOME = "/usr" # Path to Condor home directory # CONDOR_HOME = "/usr"
<commit_before>import os import tempfile # SERVER CONFIGURATION SERVER_HOST = "127.0.0.1" SERVER_PORT = 5000 # SSL config: path to certificate and private key files CERTIFICATE = None PRIVATE_KEY = None # Max number of processes to fork when handling requests MAX_PROCESSES = 10 # Enable debugging DEBUG = False # The secret key used by Flask to encrypt session keys SECRET_KEY = os.urandom(24) # Authentication method to use (NoAuthentication or PAMAuthentication) AUTHENTICATION = "PAMAuthentication" PROCESS_SWITCHING = True # Flask cache configuration CACHE_TYPE = "filesystem" CACHE_DIR = os.path.join(tempfile.gettempdir(), "pegasus-service") # # Authorization - # None, '', False -> User can only access their own data. # * -> All users are admin users and can access data of any other user. # {'u1', .., 'un'} OR ['u1', .., 'un'] -> Only users in the set/list are admin users. # ADMIN_USERS = None # CLIENT CONFIGURATION # User credentials USERNAME = "" PASSWORD = "" # ENSEMBLE MANAGER CONFIGURATION # Workflow processing interval in seconds EM_INTERVAL = 60 # Directory to store data STORAGE_DIRECTORY = "/var/pegasus" # Path to Pegasus home directory # PEGASUS_HOME = "/usr" # Path to Condor home directory # CONDOR_HOME = "/usr" <commit_msg>Use SimpleCache instead of Filesystemcache as it can cause errors due to use of different pickle versions.<commit_after>
import os import tempfile # SERVER CONFIGURATION SERVER_HOST = "127.0.0.1" SERVER_PORT = 5000 # SSL config: path to certificate and private key files CERTIFICATE = None PRIVATE_KEY = None # Max number of processes to fork when handling requests MAX_PROCESSES = 10 # Enable debugging DEBUG = False # The secret key used by Flask to encrypt session keys SECRET_KEY = os.urandom(24) # Authentication method to use (NoAuthentication or PAMAuthentication) AUTHENTICATION = "PAMAuthentication" PROCESS_SWITCHING = True # Flask cache configuration CACHE_TYPE = "flask_caching.backends.SimpleCache" # # Authorization - # None, '', False -> User can only access their own data. # * -> All users are admin users and can access data of any other user. # {'u1', .., 'un'} OR ['u1', .., 'un'] -> Only users in the set/list are admin users. # ADMIN_USERS = None # CLIENT CONFIGURATION # User credentials USERNAME = "" PASSWORD = "" # ENSEMBLE MANAGER CONFIGURATION # Workflow processing interval in seconds EM_INTERVAL = 60 # Directory to store data STORAGE_DIRECTORY = "/var/pegasus" # Path to Pegasus home directory # PEGASUS_HOME = "/usr" # Path to Condor home directory # CONDOR_HOME = "/usr"
import os import tempfile # SERVER CONFIGURATION SERVER_HOST = "127.0.0.1" SERVER_PORT = 5000 # SSL config: path to certificate and private key files CERTIFICATE = None PRIVATE_KEY = None # Max number of processes to fork when handling requests MAX_PROCESSES = 10 # Enable debugging DEBUG = False # The secret key used by Flask to encrypt session keys SECRET_KEY = os.urandom(24) # Authentication method to use (NoAuthentication or PAMAuthentication) AUTHENTICATION = "PAMAuthentication" PROCESS_SWITCHING = True # Flask cache configuration CACHE_TYPE = "filesystem" CACHE_DIR = os.path.join(tempfile.gettempdir(), "pegasus-service") # # Authorization - # None, '', False -> User can only access their own data. # * -> All users are admin users and can access data of any other user. # {'u1', .., 'un'} OR ['u1', .., 'un'] -> Only users in the set/list are admin users. # ADMIN_USERS = None # CLIENT CONFIGURATION # User credentials USERNAME = "" PASSWORD = "" # ENSEMBLE MANAGER CONFIGURATION # Workflow processing interval in seconds EM_INTERVAL = 60 # Directory to store data STORAGE_DIRECTORY = "/var/pegasus" # Path to Pegasus home directory # PEGASUS_HOME = "/usr" # Path to Condor home directory # CONDOR_HOME = "/usr" Use SimpleCache instead of Filesystemcache as it can cause errors due to use of different pickle versions.import os import tempfile # SERVER CONFIGURATION SERVER_HOST = "127.0.0.1" SERVER_PORT = 5000 # SSL config: path to certificate and private key files CERTIFICATE = None PRIVATE_KEY = None # Max number of processes to fork when handling requests MAX_PROCESSES = 10 # Enable debugging DEBUG = False # The secret key used by Flask to encrypt session keys SECRET_KEY = os.urandom(24) # Authentication method to use (NoAuthentication or PAMAuthentication) AUTHENTICATION = "PAMAuthentication" PROCESS_SWITCHING = True # Flask cache configuration CACHE_TYPE = "flask_caching.backends.SimpleCache" # # Authorization - # None, '', False -> User can only access their own data. # * -> All users are admin users and can access data of any other user. # {'u1', .., 'un'} OR ['u1', .., 'un'] -> Only users in the set/list are admin users. # ADMIN_USERS = None # CLIENT CONFIGURATION # User credentials USERNAME = "" PASSWORD = "" # ENSEMBLE MANAGER CONFIGURATION # Workflow processing interval in seconds EM_INTERVAL = 60 # Directory to store data STORAGE_DIRECTORY = "/var/pegasus" # Path to Pegasus home directory # PEGASUS_HOME = "/usr" # Path to Condor home directory # CONDOR_HOME = "/usr"
<commit_before>import os import tempfile # SERVER CONFIGURATION SERVER_HOST = "127.0.0.1" SERVER_PORT = 5000 # SSL config: path to certificate and private key files CERTIFICATE = None PRIVATE_KEY = None # Max number of processes to fork when handling requests MAX_PROCESSES = 10 # Enable debugging DEBUG = False # The secret key used by Flask to encrypt session keys SECRET_KEY = os.urandom(24) # Authentication method to use (NoAuthentication or PAMAuthentication) AUTHENTICATION = "PAMAuthentication" PROCESS_SWITCHING = True # Flask cache configuration CACHE_TYPE = "filesystem" CACHE_DIR = os.path.join(tempfile.gettempdir(), "pegasus-service") # # Authorization - # None, '', False -> User can only access their own data. # * -> All users are admin users and can access data of any other user. # {'u1', .., 'un'} OR ['u1', .., 'un'] -> Only users in the set/list are admin users. # ADMIN_USERS = None # CLIENT CONFIGURATION # User credentials USERNAME = "" PASSWORD = "" # ENSEMBLE MANAGER CONFIGURATION # Workflow processing interval in seconds EM_INTERVAL = 60 # Directory to store data STORAGE_DIRECTORY = "/var/pegasus" # Path to Pegasus home directory # PEGASUS_HOME = "/usr" # Path to Condor home directory # CONDOR_HOME = "/usr" <commit_msg>Use SimpleCache instead of Filesystemcache as it can cause errors due to use of different pickle versions.<commit_after>import os import tempfile # SERVER CONFIGURATION SERVER_HOST = "127.0.0.1" SERVER_PORT = 5000 # SSL config: path to certificate and private key files CERTIFICATE = None PRIVATE_KEY = None # Max number of processes to fork when handling requests MAX_PROCESSES = 10 # Enable debugging DEBUG = False # The secret key used by Flask to encrypt session keys SECRET_KEY = os.urandom(24) # Authentication method to use (NoAuthentication or PAMAuthentication) AUTHENTICATION = "PAMAuthentication" PROCESS_SWITCHING = True # Flask cache configuration CACHE_TYPE = "flask_caching.backends.SimpleCache" # # Authorization - # None, '', False -> User can only access their own data. # * -> All users are admin users and can access data of any other user. # {'u1', .., 'un'} OR ['u1', .., 'un'] -> Only users in the set/list are admin users. # ADMIN_USERS = None # CLIENT CONFIGURATION # User credentials USERNAME = "" PASSWORD = "" # ENSEMBLE MANAGER CONFIGURATION # Workflow processing interval in seconds EM_INTERVAL = 60 # Directory to store data STORAGE_DIRECTORY = "/var/pegasus" # Path to Pegasus home directory # PEGASUS_HOME = "/usr" # Path to Condor home directory # CONDOR_HOME = "/usr"
dda88345334985796dac2095f6e78bb106bc19b3
pullpush/pullpush.py
pullpush/pullpush.py
#!/usr/bin/env python3 import git class PullPush: def __init__(self, repo_dir): self.repo_dir = repo_dir self.repo = None def pull(self, source_repo): """ Pulls the remote source_repo and stores it in the repo_dir directory. """ self.repo = git.Repo.init(self.repo_dir) origin = self.repo.create_remote('origin', source_repo) origin.fetch() origin.pull(origin.refs[0].remote_head) def set_target_repo(self, new_url): """ Changes the target url of the previously pulled repo. """ origin = self.repo.remotes.origin cw = origin.config_writer cw.set("url", new_url) cw.release() def push(self, target_repo): """ Pushes the previously pulled repo to the target_repo. """ origin = self.repo.remotes.origin self.set_target_repo(target_repo) self.repo.create_head('master', origin.refs.master).set_tracking_branch(origin.refs.master) origin.push()
#!/usr/bin/env python3 import git class PullPush: def __init__(self, repo_dir): self.repo_dir = repo_dir self.repo = None def pull(self, source_repo): """ Pulls the remote source_repo and stores it in the repo_dir directory. """ self.repo = git.Repo.init(self.repo_dir) origin = self.repo.create_remote('origin', source_repo) origin.fetch() origin.pull(origin.refs[0].remote_head) def set_target_repo(self, new_url): """ Changes the target url of the previously pulled repo. """ origin = self.repo.remotes.origin cw = origin.config_writer cw.set("url", new_url) cw.release() def push(self, target_repo): """ Pushes the previously pulled repo to the target_repo. """ if self.repo is None: # TODO Better handling return origin = self.repo.remotes.origin self.set_target_repo(target_repo) self.repo.create_head('master', origin.refs.master).set_tracking_branch(origin.refs.master) origin.push()
Check of repo was pulled
Check of repo was pulled
Python
mit
martialblog/git-pullpush
#!/usr/bin/env python3 import git class PullPush: def __init__(self, repo_dir): self.repo_dir = repo_dir self.repo = None def pull(self, source_repo): """ Pulls the remote source_repo and stores it in the repo_dir directory. """ self.repo = git.Repo.init(self.repo_dir) origin = self.repo.create_remote('origin', source_repo) origin.fetch() origin.pull(origin.refs[0].remote_head) def set_target_repo(self, new_url): """ Changes the target url of the previously pulled repo. """ origin = self.repo.remotes.origin cw = origin.config_writer cw.set("url", new_url) cw.release() def push(self, target_repo): """ Pushes the previously pulled repo to the target_repo. """ origin = self.repo.remotes.origin self.set_target_repo(target_repo) self.repo.create_head('master', origin.refs.master).set_tracking_branch(origin.refs.master) origin.push() Check of repo was pulled
#!/usr/bin/env python3 import git class PullPush: def __init__(self, repo_dir): self.repo_dir = repo_dir self.repo = None def pull(self, source_repo): """ Pulls the remote source_repo and stores it in the repo_dir directory. """ self.repo = git.Repo.init(self.repo_dir) origin = self.repo.create_remote('origin', source_repo) origin.fetch() origin.pull(origin.refs[0].remote_head) def set_target_repo(self, new_url): """ Changes the target url of the previously pulled repo. """ origin = self.repo.remotes.origin cw = origin.config_writer cw.set("url", new_url) cw.release() def push(self, target_repo): """ Pushes the previously pulled repo to the target_repo. """ if self.repo is None: # TODO Better handling return origin = self.repo.remotes.origin self.set_target_repo(target_repo) self.repo.create_head('master', origin.refs.master).set_tracking_branch(origin.refs.master) origin.push()
<commit_before>#!/usr/bin/env python3 import git class PullPush: def __init__(self, repo_dir): self.repo_dir = repo_dir self.repo = None def pull(self, source_repo): """ Pulls the remote source_repo and stores it in the repo_dir directory. """ self.repo = git.Repo.init(self.repo_dir) origin = self.repo.create_remote('origin', source_repo) origin.fetch() origin.pull(origin.refs[0].remote_head) def set_target_repo(self, new_url): """ Changes the target url of the previously pulled repo. """ origin = self.repo.remotes.origin cw = origin.config_writer cw.set("url", new_url) cw.release() def push(self, target_repo): """ Pushes the previously pulled repo to the target_repo. """ origin = self.repo.remotes.origin self.set_target_repo(target_repo) self.repo.create_head('master', origin.refs.master).set_tracking_branch(origin.refs.master) origin.push() <commit_msg>Check of repo was pulled<commit_after>
#!/usr/bin/env python3 import git class PullPush: def __init__(self, repo_dir): self.repo_dir = repo_dir self.repo = None def pull(self, source_repo): """ Pulls the remote source_repo and stores it in the repo_dir directory. """ self.repo = git.Repo.init(self.repo_dir) origin = self.repo.create_remote('origin', source_repo) origin.fetch() origin.pull(origin.refs[0].remote_head) def set_target_repo(self, new_url): """ Changes the target url of the previously pulled repo. """ origin = self.repo.remotes.origin cw = origin.config_writer cw.set("url", new_url) cw.release() def push(self, target_repo): """ Pushes the previously pulled repo to the target_repo. """ if self.repo is None: # TODO Better handling return origin = self.repo.remotes.origin self.set_target_repo(target_repo) self.repo.create_head('master', origin.refs.master).set_tracking_branch(origin.refs.master) origin.push()
#!/usr/bin/env python3 import git class PullPush: def __init__(self, repo_dir): self.repo_dir = repo_dir self.repo = None def pull(self, source_repo): """ Pulls the remote source_repo and stores it in the repo_dir directory. """ self.repo = git.Repo.init(self.repo_dir) origin = self.repo.create_remote('origin', source_repo) origin.fetch() origin.pull(origin.refs[0].remote_head) def set_target_repo(self, new_url): """ Changes the target url of the previously pulled repo. """ origin = self.repo.remotes.origin cw = origin.config_writer cw.set("url", new_url) cw.release() def push(self, target_repo): """ Pushes the previously pulled repo to the target_repo. """ origin = self.repo.remotes.origin self.set_target_repo(target_repo) self.repo.create_head('master', origin.refs.master).set_tracking_branch(origin.refs.master) origin.push() Check of repo was pulled#!/usr/bin/env python3 import git class PullPush: def __init__(self, repo_dir): self.repo_dir = repo_dir self.repo = None def pull(self, source_repo): """ Pulls the remote source_repo and stores it in the repo_dir directory. """ self.repo = git.Repo.init(self.repo_dir) origin = self.repo.create_remote('origin', source_repo) origin.fetch() origin.pull(origin.refs[0].remote_head) def set_target_repo(self, new_url): """ Changes the target url of the previously pulled repo. """ origin = self.repo.remotes.origin cw = origin.config_writer cw.set("url", new_url) cw.release() def push(self, target_repo): """ Pushes the previously pulled repo to the target_repo. """ if self.repo is None: # TODO Better handling return origin = self.repo.remotes.origin self.set_target_repo(target_repo) self.repo.create_head('master', origin.refs.master).set_tracking_branch(origin.refs.master) origin.push()
<commit_before>#!/usr/bin/env python3 import git class PullPush: def __init__(self, repo_dir): self.repo_dir = repo_dir self.repo = None def pull(self, source_repo): """ Pulls the remote source_repo and stores it in the repo_dir directory. """ self.repo = git.Repo.init(self.repo_dir) origin = self.repo.create_remote('origin', source_repo) origin.fetch() origin.pull(origin.refs[0].remote_head) def set_target_repo(self, new_url): """ Changes the target url of the previously pulled repo. """ origin = self.repo.remotes.origin cw = origin.config_writer cw.set("url", new_url) cw.release() def push(self, target_repo): """ Pushes the previously pulled repo to the target_repo. """ origin = self.repo.remotes.origin self.set_target_repo(target_repo) self.repo.create_head('master', origin.refs.master).set_tracking_branch(origin.refs.master) origin.push() <commit_msg>Check of repo was pulled<commit_after>#!/usr/bin/env python3 import git class PullPush: def __init__(self, repo_dir): self.repo_dir = repo_dir self.repo = None def pull(self, source_repo): """ Pulls the remote source_repo and stores it in the repo_dir directory. """ self.repo = git.Repo.init(self.repo_dir) origin = self.repo.create_remote('origin', source_repo) origin.fetch() origin.pull(origin.refs[0].remote_head) def set_target_repo(self, new_url): """ Changes the target url of the previously pulled repo. """ origin = self.repo.remotes.origin cw = origin.config_writer cw.set("url", new_url) cw.release() def push(self, target_repo): """ Pushes the previously pulled repo to the target_repo. """ if self.repo is None: # TODO Better handling return origin = self.repo.remotes.origin self.set_target_repo(target_repo) self.repo.create_head('master', origin.refs.master).set_tracking_branch(origin.refs.master) origin.push()
beb8f12e4a8290d4107cdb91a321a6618a038ef9
rose_trellis/util.py
rose_trellis/util.py
from urllib.parse import urljoin import time import asyncio from typing import Any from typing import Callable TRELLO_URL_BASE = 'https://api.trello.com/1/' def join_url(part: str) -> str: """ Adds `part` to API base url. Always returns url without trailing slash. :param part: :return: url """ part = part.strip('/') newpath = urljoin(TRELLO_URL_BASE, part) while newpath.endswith('/'): newpath = newpath[:-1] return newpath def easy_run(func: Callable[Any], *args, **kwargs) -> Any: el = asyncio.get_event_loop() return el.run_until_complete(func(*args, **kwargs))
from urllib.parse import urljoin import time import asyncio from typing import Any from typing import Callable TRELLO_URL_BASE = 'https://api.trello.com/1/' def join_url(part: str) -> str: """ Adds `part` to API base url. Always returns url without trailing slash. :param part: :return: url """ part = part.strip('/') newpath = urljoin(TRELLO_URL_BASE, part) while newpath.endswith('/'): newpath = newpath[:-1] return newpath def easy_run(gen) -> Any: el = asyncio.get_event_loop() return el.run_until_complete(gen)
Make easy_run better by taking the generator from coroutines
Make easy_run better by taking the generator from coroutines
Python
mit
dmwyatt/rose_trellis
from urllib.parse import urljoin import time import asyncio from typing import Any from typing import Callable TRELLO_URL_BASE = 'https://api.trello.com/1/' def join_url(part: str) -> str: """ Adds `part` to API base url. Always returns url without trailing slash. :param part: :return: url """ part = part.strip('/') newpath = urljoin(TRELLO_URL_BASE, part) while newpath.endswith('/'): newpath = newpath[:-1] return newpath def easy_run(func: Callable[Any], *args, **kwargs) -> Any: el = asyncio.get_event_loop() return el.run_until_complete(func(*args, **kwargs)) Make easy_run better by taking the generator from coroutines
from urllib.parse import urljoin import time import asyncio from typing import Any from typing import Callable TRELLO_URL_BASE = 'https://api.trello.com/1/' def join_url(part: str) -> str: """ Adds `part` to API base url. Always returns url without trailing slash. :param part: :return: url """ part = part.strip('/') newpath = urljoin(TRELLO_URL_BASE, part) while newpath.endswith('/'): newpath = newpath[:-1] return newpath def easy_run(gen) -> Any: el = asyncio.get_event_loop() return el.run_until_complete(gen)
<commit_before>from urllib.parse import urljoin import time import asyncio from typing import Any from typing import Callable TRELLO_URL_BASE = 'https://api.trello.com/1/' def join_url(part: str) -> str: """ Adds `part` to API base url. Always returns url without trailing slash. :param part: :return: url """ part = part.strip('/') newpath = urljoin(TRELLO_URL_BASE, part) while newpath.endswith('/'): newpath = newpath[:-1] return newpath def easy_run(func: Callable[Any], *args, **kwargs) -> Any: el = asyncio.get_event_loop() return el.run_until_complete(func(*args, **kwargs)) <commit_msg>Make easy_run better by taking the generator from coroutines<commit_after>
from urllib.parse import urljoin import time import asyncio from typing import Any from typing import Callable TRELLO_URL_BASE = 'https://api.trello.com/1/' def join_url(part: str) -> str: """ Adds `part` to API base url. Always returns url without trailing slash. :param part: :return: url """ part = part.strip('/') newpath = urljoin(TRELLO_URL_BASE, part) while newpath.endswith('/'): newpath = newpath[:-1] return newpath def easy_run(gen) -> Any: el = asyncio.get_event_loop() return el.run_until_complete(gen)
from urllib.parse import urljoin import time import asyncio from typing import Any from typing import Callable TRELLO_URL_BASE = 'https://api.trello.com/1/' def join_url(part: str) -> str: """ Adds `part` to API base url. Always returns url without trailing slash. :param part: :return: url """ part = part.strip('/') newpath = urljoin(TRELLO_URL_BASE, part) while newpath.endswith('/'): newpath = newpath[:-1] return newpath def easy_run(func: Callable[Any], *args, **kwargs) -> Any: el = asyncio.get_event_loop() return el.run_until_complete(func(*args, **kwargs)) Make easy_run better by taking the generator from coroutinesfrom urllib.parse import urljoin import time import asyncio from typing import Any from typing import Callable TRELLO_URL_BASE = 'https://api.trello.com/1/' def join_url(part: str) -> str: """ Adds `part` to API base url. Always returns url without trailing slash. :param part: :return: url """ part = part.strip('/') newpath = urljoin(TRELLO_URL_BASE, part) while newpath.endswith('/'): newpath = newpath[:-1] return newpath def easy_run(gen) -> Any: el = asyncio.get_event_loop() return el.run_until_complete(gen)
<commit_before>from urllib.parse import urljoin import time import asyncio from typing import Any from typing import Callable TRELLO_URL_BASE = 'https://api.trello.com/1/' def join_url(part: str) -> str: """ Adds `part` to API base url. Always returns url without trailing slash. :param part: :return: url """ part = part.strip('/') newpath = urljoin(TRELLO_URL_BASE, part) while newpath.endswith('/'): newpath = newpath[:-1] return newpath def easy_run(func: Callable[Any], *args, **kwargs) -> Any: el = asyncio.get_event_loop() return el.run_until_complete(func(*args, **kwargs)) <commit_msg>Make easy_run better by taking the generator from coroutines<commit_after>from urllib.parse import urljoin import time import asyncio from typing import Any from typing import Callable TRELLO_URL_BASE = 'https://api.trello.com/1/' def join_url(part: str) -> str: """ Adds `part` to API base url. Always returns url without trailing slash. :param part: :return: url """ part = part.strip('/') newpath = urljoin(TRELLO_URL_BASE, part) while newpath.endswith('/'): newpath = newpath[:-1] return newpath def easy_run(gen) -> Any: el = asyncio.get_event_loop() return el.run_until_complete(gen)
2c15d96a7d77269fbe41e5b2940873fc849d411a
random_projection.py
random_projection.py
""" Random projection, Assignment 1c """
""" Random projection, Assignment 1c """ import numpy as np import matplotlib.pylab as plt import random, mnist_dataloader from numpy import dtype """ Generate random projection matrix R @param: k, the reduced number of dimensions @param: d, the original number of dimensions @return: R, the generated random projection matrix, k * d size """ def generate_random_projection_matrix(k, d): R = np.zeros((k, d), dtype = np.float64) for r in np.nditer(R, op_flags=['readwrite']): r[...] = random.randint(0, 1) if r[...] == 0: r[...] = -1 R *= 1.0 / np.sqrt(k) return R """ random projection matrix P into R @param R: random projection matrix @param P: matrix to be reduced in dimension @return: Q: projected matrix of P on R """ def random_projection(R, P): if R.shape[1] != P.shape[0]: return False print R.shape, P.shape return np.dot(R, P) if __name__ == "__main__": # load data set training_data, validation_data, test_data = mnist_dataloader.load_data() # row vector (matrix) training_data_instances = training_data[0] training_data_labels = training_data[1] # row vector (matrix) test_data_instances = test_data[0] test_data_labels = test_data[1] # dimension of a training data instance d = training_data_instances.shape[1] for k in [50, 100, 500]: random_projection_matrix = generate_random_projection_matrix(k, d) # transpose to column vector (matrix) before projection and recover after projection random_projected_matrix = np.transpose(random_projection(random_projection_matrix, np.transpose(training_data_instances[0:20]))) print random_projected_matrix[0], random_projected_matrix.shape
Add random projection matrix generation and random projection.
Add random projection matrix generation and random projection.
Python
mit
lidalei/DataMining
""" Random projection, Assignment 1c """ Add random projection matrix generation and random projection.
""" Random projection, Assignment 1c """ import numpy as np import matplotlib.pylab as plt import random, mnist_dataloader from numpy import dtype """ Generate random projection matrix R @param: k, the reduced number of dimensions @param: d, the original number of dimensions @return: R, the generated random projection matrix, k * d size """ def generate_random_projection_matrix(k, d): R = np.zeros((k, d), dtype = np.float64) for r in np.nditer(R, op_flags=['readwrite']): r[...] = random.randint(0, 1) if r[...] == 0: r[...] = -1 R *= 1.0 / np.sqrt(k) return R """ random projection matrix P into R @param R: random projection matrix @param P: matrix to be reduced in dimension @return: Q: projected matrix of P on R """ def random_projection(R, P): if R.shape[1] != P.shape[0]: return False print R.shape, P.shape return np.dot(R, P) if __name__ == "__main__": # load data set training_data, validation_data, test_data = mnist_dataloader.load_data() # row vector (matrix) training_data_instances = training_data[0] training_data_labels = training_data[1] # row vector (matrix) test_data_instances = test_data[0] test_data_labels = test_data[1] # dimension of a training data instance d = training_data_instances.shape[1] for k in [50, 100, 500]: random_projection_matrix = generate_random_projection_matrix(k, d) # transpose to column vector (matrix) before projection and recover after projection random_projected_matrix = np.transpose(random_projection(random_projection_matrix, np.transpose(training_data_instances[0:20]))) print random_projected_matrix[0], random_projected_matrix.shape
<commit_before>""" Random projection, Assignment 1c """ <commit_msg>Add random projection matrix generation and random projection.<commit_after>
""" Random projection, Assignment 1c """ import numpy as np import matplotlib.pylab as plt import random, mnist_dataloader from numpy import dtype """ Generate random projection matrix R @param: k, the reduced number of dimensions @param: d, the original number of dimensions @return: R, the generated random projection matrix, k * d size """ def generate_random_projection_matrix(k, d): R = np.zeros((k, d), dtype = np.float64) for r in np.nditer(R, op_flags=['readwrite']): r[...] = random.randint(0, 1) if r[...] == 0: r[...] = -1 R *= 1.0 / np.sqrt(k) return R """ random projection matrix P into R @param R: random projection matrix @param P: matrix to be reduced in dimension @return: Q: projected matrix of P on R """ def random_projection(R, P): if R.shape[1] != P.shape[0]: return False print R.shape, P.shape return np.dot(R, P) if __name__ == "__main__": # load data set training_data, validation_data, test_data = mnist_dataloader.load_data() # row vector (matrix) training_data_instances = training_data[0] training_data_labels = training_data[1] # row vector (matrix) test_data_instances = test_data[0] test_data_labels = test_data[1] # dimension of a training data instance d = training_data_instances.shape[1] for k in [50, 100, 500]: random_projection_matrix = generate_random_projection_matrix(k, d) # transpose to column vector (matrix) before projection and recover after projection random_projected_matrix = np.transpose(random_projection(random_projection_matrix, np.transpose(training_data_instances[0:20]))) print random_projected_matrix[0], random_projected_matrix.shape
""" Random projection, Assignment 1c """ Add random projection matrix generation and random projection.""" Random projection, Assignment 1c """ import numpy as np import matplotlib.pylab as plt import random, mnist_dataloader from numpy import dtype """ Generate random projection matrix R @param: k, the reduced number of dimensions @param: d, the original number of dimensions @return: R, the generated random projection matrix, k * d size """ def generate_random_projection_matrix(k, d): R = np.zeros((k, d), dtype = np.float64) for r in np.nditer(R, op_flags=['readwrite']): r[...] = random.randint(0, 1) if r[...] == 0: r[...] = -1 R *= 1.0 / np.sqrt(k) return R """ random projection matrix P into R @param R: random projection matrix @param P: matrix to be reduced in dimension @return: Q: projected matrix of P on R """ def random_projection(R, P): if R.shape[1] != P.shape[0]: return False print R.shape, P.shape return np.dot(R, P) if __name__ == "__main__": # load data set training_data, validation_data, test_data = mnist_dataloader.load_data() # row vector (matrix) training_data_instances = training_data[0] training_data_labels = training_data[1] # row vector (matrix) test_data_instances = test_data[0] test_data_labels = test_data[1] # dimension of a training data instance d = training_data_instances.shape[1] for k in [50, 100, 500]: random_projection_matrix = generate_random_projection_matrix(k, d) # transpose to column vector (matrix) before projection and recover after projection random_projected_matrix = np.transpose(random_projection(random_projection_matrix, np.transpose(training_data_instances[0:20]))) print random_projected_matrix[0], random_projected_matrix.shape
<commit_before>""" Random projection, Assignment 1c """ <commit_msg>Add random projection matrix generation and random projection.<commit_after>""" Random projection, Assignment 1c """ import numpy as np import matplotlib.pylab as plt import random, mnist_dataloader from numpy import dtype """ Generate random projection matrix R @param: k, the reduced number of dimensions @param: d, the original number of dimensions @return: R, the generated random projection matrix, k * d size """ def generate_random_projection_matrix(k, d): R = np.zeros((k, d), dtype = np.float64) for r in np.nditer(R, op_flags=['readwrite']): r[...] = random.randint(0, 1) if r[...] == 0: r[...] = -1 R *= 1.0 / np.sqrt(k) return R """ random projection matrix P into R @param R: random projection matrix @param P: matrix to be reduced in dimension @return: Q: projected matrix of P on R """ def random_projection(R, P): if R.shape[1] != P.shape[0]: return False print R.shape, P.shape return np.dot(R, P) if __name__ == "__main__": # load data set training_data, validation_data, test_data = mnist_dataloader.load_data() # row vector (matrix) training_data_instances = training_data[0] training_data_labels = training_data[1] # row vector (matrix) test_data_instances = test_data[0] test_data_labels = test_data[1] # dimension of a training data instance d = training_data_instances.shape[1] for k in [50, 100, 500]: random_projection_matrix = generate_random_projection_matrix(k, d) # transpose to column vector (matrix) before projection and recover after projection random_projected_matrix = np.transpose(random_projection(random_projection_matrix, np.transpose(training_data_instances[0:20]))) print random_projected_matrix[0], random_projected_matrix.shape
5596f599287d36126b3a6e30e7579eb00ed07d73
downstream_farmer/utils.py
downstream_farmer/utils.py
#!/usr/bin/env python # -*- coding: utf-8 -*- try: from urllib import quote_plus except ImportError: from urllib.parse import quote_plus def urlify(string): """ You might be wondering: why is this here at all, since it's basically doing exactly what the quote_plus function in urllib does. Well, to keep the 2 & 3 stuff all in one place, meaning rather than try to import the urllib stuff twice in each file where url-safe strings are needed, we keep it all in one file: here. Supporting multiple Pythons is hard. :param string: String to URLify :return: URLified string """ return quote_plus(string)
#!/usr/bin/env python # -*- coding: utf-8 -*- try: from urllib import quote except ImportError: from urllib.parse import quote def urlify(string): """ You might be wondering: why is this here at all, since it's basically doing exactly what the quote_plus function in urllib does. Well, to keep the 2 & 3 stuff all in one place, meaning rather than try to import the urllib stuff twice in each file where url-safe strings are needed, we keep it all in one file: here. Supporting multiple Pythons is hard. :param string: String to URLify :return: URLified string """ return quote(string)
Fix method to use quote, not quote_plus
Fix method to use quote, not quote_plus
Python
mit
Storj/downstream-farmer
#!/usr/bin/env python # -*- coding: utf-8 -*- try: from urllib import quote_plus except ImportError: from urllib.parse import quote_plus def urlify(string): """ You might be wondering: why is this here at all, since it's basically doing exactly what the quote_plus function in urllib does. Well, to keep the 2 & 3 stuff all in one place, meaning rather than try to import the urllib stuff twice in each file where url-safe strings are needed, we keep it all in one file: here. Supporting multiple Pythons is hard. :param string: String to URLify :return: URLified string """ return quote_plus(string) Fix method to use quote, not quote_plus
#!/usr/bin/env python # -*- coding: utf-8 -*- try: from urllib import quote except ImportError: from urllib.parse import quote def urlify(string): """ You might be wondering: why is this here at all, since it's basically doing exactly what the quote_plus function in urllib does. Well, to keep the 2 & 3 stuff all in one place, meaning rather than try to import the urllib stuff twice in each file where url-safe strings are needed, we keep it all in one file: here. Supporting multiple Pythons is hard. :param string: String to URLify :return: URLified string """ return quote(string)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- try: from urllib import quote_plus except ImportError: from urllib.parse import quote_plus def urlify(string): """ You might be wondering: why is this here at all, since it's basically doing exactly what the quote_plus function in urllib does. Well, to keep the 2 & 3 stuff all in one place, meaning rather than try to import the urllib stuff twice in each file where url-safe strings are needed, we keep it all in one file: here. Supporting multiple Pythons is hard. :param string: String to URLify :return: URLified string """ return quote_plus(string) <commit_msg>Fix method to use quote, not quote_plus<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- try: from urllib import quote except ImportError: from urllib.parse import quote def urlify(string): """ You might be wondering: why is this here at all, since it's basically doing exactly what the quote_plus function in urllib does. Well, to keep the 2 & 3 stuff all in one place, meaning rather than try to import the urllib stuff twice in each file where url-safe strings are needed, we keep it all in one file: here. Supporting multiple Pythons is hard. :param string: String to URLify :return: URLified string """ return quote(string)
#!/usr/bin/env python # -*- coding: utf-8 -*- try: from urllib import quote_plus except ImportError: from urllib.parse import quote_plus def urlify(string): """ You might be wondering: why is this here at all, since it's basically doing exactly what the quote_plus function in urllib does. Well, to keep the 2 & 3 stuff all in one place, meaning rather than try to import the urllib stuff twice in each file where url-safe strings are needed, we keep it all in one file: here. Supporting multiple Pythons is hard. :param string: String to URLify :return: URLified string """ return quote_plus(string) Fix method to use quote, not quote_plus#!/usr/bin/env python # -*- coding: utf-8 -*- try: from urllib import quote except ImportError: from urllib.parse import quote def urlify(string): """ You might be wondering: why is this here at all, since it's basically doing exactly what the quote_plus function in urllib does. Well, to keep the 2 & 3 stuff all in one place, meaning rather than try to import the urllib stuff twice in each file where url-safe strings are needed, we keep it all in one file: here. Supporting multiple Pythons is hard. :param string: String to URLify :return: URLified string """ return quote(string)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- try: from urllib import quote_plus except ImportError: from urllib.parse import quote_plus def urlify(string): """ You might be wondering: why is this here at all, since it's basically doing exactly what the quote_plus function in urllib does. Well, to keep the 2 & 3 stuff all in one place, meaning rather than try to import the urllib stuff twice in each file where url-safe strings are needed, we keep it all in one file: here. Supporting multiple Pythons is hard. :param string: String to URLify :return: URLified string """ return quote_plus(string) <commit_msg>Fix method to use quote, not quote_plus<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- try: from urllib import quote except ImportError: from urllib.parse import quote def urlify(string): """ You might be wondering: why is this here at all, since it's basically doing exactly what the quote_plus function in urllib does. Well, to keep the 2 & 3 stuff all in one place, meaning rather than try to import the urllib stuff twice in each file where url-safe strings are needed, we keep it all in one file: here. Supporting multiple Pythons is hard. :param string: String to URLify :return: URLified string """ return quote(string)
f842cf7605018e85b13f27f3a0886122e4cbb80c
expert_tourist/views.py
expert_tourist/views.py
from collections import namedtuple from flask import Blueprint, request, jsonify from flask_jwt_extended import jwt_required, get_jwt_identity from .models import User from .errors import APIException from .schemas import UserSchema api = Blueprint('api', __name__) @api.route('/whoami') @jwt_required def protected(): return jsonify(token=get_jwt_identity()) @api.route('/sign_up', methods=['POST']) def sign_up(): schema = UserSchema().loads(request.data) if schema.errors: raise APIException('Fields <username, email, password> are required') user = schema.data if not user.exists(): user.save() return jsonify(UserSchema().dump(user).data) else: raise APIException('A user with the email {} already exists. Forgot your password?'.format(user.email)) @api.route('/sign_in', methods=['POST']) def sign_in(): conditions = [request.json.get('username', None), request.json.get('password', None)] if not all(conditions): raise APIException('Fields <username, password> are required.') user = User.validate_login(conditions[0], conditions[1]) if user: print(user.token) return jsonify(token=user.token) else: raise APIException('Invalid login credentials', status_code=401)
from flask import Blueprint, request, jsonify from flask_jwt_extended import jwt_required, get_jwt_identity from .models import User from .errors import APIException from .schemas import UserSchema api = Blueprint('api', __name__) @api.route('/whoami') @jwt_required def protected(): return jsonify(token=get_jwt_identity()) @api.route('/sign_up', methods=['POST']) def sign_up(): schema = UserSchema().loads(request.data) if schema.errors: raise APIException('Fields <username, email, password> are required') user = schema.data if not user.exists(): user.save() return jsonify(UserSchema().dump(user).data) else: raise APIException('A user with the email {} already exists. Forgot your password?'.format(user.email)) @api.route('/sign_in', methods=['POST']) def sign_in(): conditions = [request.json.get('username', None), request.json.get('password', None)] if not all(conditions): raise APIException('Fields <username, password> are required.') user = User.validate_login(conditions[0], conditions[1]) if user: return jsonify(UserSchema().dump(user).data) else: raise APIException('Invalid login credentials', status_code=401)
Return the whole user in successful login
Return the whole user in successful login
Python
mit
richin13/expert-tourist
from collections import namedtuple from flask import Blueprint, request, jsonify from flask_jwt_extended import jwt_required, get_jwt_identity from .models import User from .errors import APIException from .schemas import UserSchema api = Blueprint('api', __name__) @api.route('/whoami') @jwt_required def protected(): return jsonify(token=get_jwt_identity()) @api.route('/sign_up', methods=['POST']) def sign_up(): schema = UserSchema().loads(request.data) if schema.errors: raise APIException('Fields <username, email, password> are required') user = schema.data if not user.exists(): user.save() return jsonify(UserSchema().dump(user).data) else: raise APIException('A user with the email {} already exists. Forgot your password?'.format(user.email)) @api.route('/sign_in', methods=['POST']) def sign_in(): conditions = [request.json.get('username', None), request.json.get('password', None)] if not all(conditions): raise APIException('Fields <username, password> are required.') user = User.validate_login(conditions[0], conditions[1]) if user: print(user.token) return jsonify(token=user.token) else: raise APIException('Invalid login credentials', status_code=401) Return the whole user in successful login
from flask import Blueprint, request, jsonify from flask_jwt_extended import jwt_required, get_jwt_identity from .models import User from .errors import APIException from .schemas import UserSchema api = Blueprint('api', __name__) @api.route('/whoami') @jwt_required def protected(): return jsonify(token=get_jwt_identity()) @api.route('/sign_up', methods=['POST']) def sign_up(): schema = UserSchema().loads(request.data) if schema.errors: raise APIException('Fields <username, email, password> are required') user = schema.data if not user.exists(): user.save() return jsonify(UserSchema().dump(user).data) else: raise APIException('A user with the email {} already exists. Forgot your password?'.format(user.email)) @api.route('/sign_in', methods=['POST']) def sign_in(): conditions = [request.json.get('username', None), request.json.get('password', None)] if not all(conditions): raise APIException('Fields <username, password> are required.') user = User.validate_login(conditions[0], conditions[1]) if user: return jsonify(UserSchema().dump(user).data) else: raise APIException('Invalid login credentials', status_code=401)
<commit_before>from collections import namedtuple from flask import Blueprint, request, jsonify from flask_jwt_extended import jwt_required, get_jwt_identity from .models import User from .errors import APIException from .schemas import UserSchema api = Blueprint('api', __name__) @api.route('/whoami') @jwt_required def protected(): return jsonify(token=get_jwt_identity()) @api.route('/sign_up', methods=['POST']) def sign_up(): schema = UserSchema().loads(request.data) if schema.errors: raise APIException('Fields <username, email, password> are required') user = schema.data if not user.exists(): user.save() return jsonify(UserSchema().dump(user).data) else: raise APIException('A user with the email {} already exists. Forgot your password?'.format(user.email)) @api.route('/sign_in', methods=['POST']) def sign_in(): conditions = [request.json.get('username', None), request.json.get('password', None)] if not all(conditions): raise APIException('Fields <username, password> are required.') user = User.validate_login(conditions[0], conditions[1]) if user: print(user.token) return jsonify(token=user.token) else: raise APIException('Invalid login credentials', status_code=401) <commit_msg>Return the whole user in successful login<commit_after>
from flask import Blueprint, request, jsonify from flask_jwt_extended import jwt_required, get_jwt_identity from .models import User from .errors import APIException from .schemas import UserSchema api = Blueprint('api', __name__) @api.route('/whoami') @jwt_required def protected(): return jsonify(token=get_jwt_identity()) @api.route('/sign_up', methods=['POST']) def sign_up(): schema = UserSchema().loads(request.data) if schema.errors: raise APIException('Fields <username, email, password> are required') user = schema.data if not user.exists(): user.save() return jsonify(UserSchema().dump(user).data) else: raise APIException('A user with the email {} already exists. Forgot your password?'.format(user.email)) @api.route('/sign_in', methods=['POST']) def sign_in(): conditions = [request.json.get('username', None), request.json.get('password', None)] if not all(conditions): raise APIException('Fields <username, password> are required.') user = User.validate_login(conditions[0], conditions[1]) if user: return jsonify(UserSchema().dump(user).data) else: raise APIException('Invalid login credentials', status_code=401)
from collections import namedtuple from flask import Blueprint, request, jsonify from flask_jwt_extended import jwt_required, get_jwt_identity from .models import User from .errors import APIException from .schemas import UserSchema api = Blueprint('api', __name__) @api.route('/whoami') @jwt_required def protected(): return jsonify(token=get_jwt_identity()) @api.route('/sign_up', methods=['POST']) def sign_up(): schema = UserSchema().loads(request.data) if schema.errors: raise APIException('Fields <username, email, password> are required') user = schema.data if not user.exists(): user.save() return jsonify(UserSchema().dump(user).data) else: raise APIException('A user with the email {} already exists. Forgot your password?'.format(user.email)) @api.route('/sign_in', methods=['POST']) def sign_in(): conditions = [request.json.get('username', None), request.json.get('password', None)] if not all(conditions): raise APIException('Fields <username, password> are required.') user = User.validate_login(conditions[0], conditions[1]) if user: print(user.token) return jsonify(token=user.token) else: raise APIException('Invalid login credentials', status_code=401) Return the whole user in successful loginfrom flask import Blueprint, request, jsonify from flask_jwt_extended import jwt_required, get_jwt_identity from .models import User from .errors import APIException from .schemas import UserSchema api = Blueprint('api', __name__) @api.route('/whoami') @jwt_required def protected(): return jsonify(token=get_jwt_identity()) @api.route('/sign_up', methods=['POST']) def sign_up(): schema = UserSchema().loads(request.data) if schema.errors: raise APIException('Fields <username, email, password> are required') user = schema.data if not user.exists(): user.save() return jsonify(UserSchema().dump(user).data) else: raise APIException('A user with the email {} already exists. Forgot your password?'.format(user.email)) @api.route('/sign_in', methods=['POST']) def sign_in(): conditions = [request.json.get('username', None), request.json.get('password', None)] if not all(conditions): raise APIException('Fields <username, password> are required.') user = User.validate_login(conditions[0], conditions[1]) if user: return jsonify(UserSchema().dump(user).data) else: raise APIException('Invalid login credentials', status_code=401)
<commit_before>from collections import namedtuple from flask import Blueprint, request, jsonify from flask_jwt_extended import jwt_required, get_jwt_identity from .models import User from .errors import APIException from .schemas import UserSchema api = Blueprint('api', __name__) @api.route('/whoami') @jwt_required def protected(): return jsonify(token=get_jwt_identity()) @api.route('/sign_up', methods=['POST']) def sign_up(): schema = UserSchema().loads(request.data) if schema.errors: raise APIException('Fields <username, email, password> are required') user = schema.data if not user.exists(): user.save() return jsonify(UserSchema().dump(user).data) else: raise APIException('A user with the email {} already exists. Forgot your password?'.format(user.email)) @api.route('/sign_in', methods=['POST']) def sign_in(): conditions = [request.json.get('username', None), request.json.get('password', None)] if not all(conditions): raise APIException('Fields <username, password> are required.') user = User.validate_login(conditions[0], conditions[1]) if user: print(user.token) return jsonify(token=user.token) else: raise APIException('Invalid login credentials', status_code=401) <commit_msg>Return the whole user in successful login<commit_after>from flask import Blueprint, request, jsonify from flask_jwt_extended import jwt_required, get_jwt_identity from .models import User from .errors import APIException from .schemas import UserSchema api = Blueprint('api', __name__) @api.route('/whoami') @jwt_required def protected(): return jsonify(token=get_jwt_identity()) @api.route('/sign_up', methods=['POST']) def sign_up(): schema = UserSchema().loads(request.data) if schema.errors: raise APIException('Fields <username, email, password> are required') user = schema.data if not user.exists(): user.save() return jsonify(UserSchema().dump(user).data) else: raise APIException('A user with the email {} already exists. Forgot your password?'.format(user.email)) @api.route('/sign_in', methods=['POST']) def sign_in(): conditions = [request.json.get('username', None), request.json.get('password', None)] if not all(conditions): raise APIException('Fields <username, password> are required.') user = User.validate_login(conditions[0], conditions[1]) if user: return jsonify(UserSchema().dump(user).data) else: raise APIException('Invalid login credentials', status_code=401)
01a9b6457d78dd583637bf8174edda40e2bd3276
django_website/blog/feeds.py
django_website/blog/feeds.py
from __future__ import absolute_import from django.contrib.syndication.views import Feed from .models import Entry class WeblogEntryFeed(Feed): title = "The Django weblog" link = "http://www.djangoproject.com/weblog/" description = "Latest news about Django, the Python Web framework." def items(self): return Entry.objects.published()[:10] def item_pubdate(self, item): return item.pub_date
from __future__ import absolute_import from django.contrib.syndication.views import Feed from .models import Entry class WeblogEntryFeed(Feed): title = "The Django weblog" link = "http://www.djangoproject.com/weblog/" description = "Latest news about Django, the Python Web framework." def items(self): return Entry.objects.published()[:10] def item_pubdate(self, item): return item.pub_date def item_author_name(self, item): return item.author def item_description(self, item): return item.body_html
Add author name and body to the weblog RSS feed.
Add author name and body to the weblog RSS feed.
Python
bsd-3-clause
alawnchen/djangoproject.com,django/djangoproject.com,nanuxbe/django,django/djangoproject.com,khkaminska/djangoproject.com,khkaminska/djangoproject.com,django/djangoproject.com,rmoorman/djangoproject.com,rmoorman/djangoproject.com,django/djangoproject.com,nanuxbe/django,django/djangoproject.com,xavierdutreilh/djangoproject.com,gnarf/djangoproject.com,relekang/djangoproject.com,rmoorman/djangoproject.com,relekang/djangoproject.com,rmoorman/djangoproject.com,hassanabidpk/djangoproject.com,vxvinh1511/djangoproject.com,vxvinh1511/djangoproject.com,khkaminska/djangoproject.com,nanuxbe/django,alawnchen/djangoproject.com,khkaminska/djangoproject.com,vxvinh1511/djangoproject.com,xavierdutreilh/djangoproject.com,xavierdutreilh/djangoproject.com,hassanabidpk/djangoproject.com,xavierdutreilh/djangoproject.com,gnarf/djangoproject.com,relekang/djangoproject.com,hassanabidpk/djangoproject.com,vxvinh1511/djangoproject.com,django/djangoproject.com,alawnchen/djangoproject.com,relekang/djangoproject.com,gnarf/djangoproject.com,gnarf/djangoproject.com,hassanabidpk/djangoproject.com,nanuxbe/django,alawnchen/djangoproject.com
from __future__ import absolute_import from django.contrib.syndication.views import Feed from .models import Entry class WeblogEntryFeed(Feed): title = "The Django weblog" link = "http://www.djangoproject.com/weblog/" description = "Latest news about Django, the Python Web framework." def items(self): return Entry.objects.published()[:10] def item_pubdate(self, item): return item.pub_date Add author name and body to the weblog RSS feed.
from __future__ import absolute_import from django.contrib.syndication.views import Feed from .models import Entry class WeblogEntryFeed(Feed): title = "The Django weblog" link = "http://www.djangoproject.com/weblog/" description = "Latest news about Django, the Python Web framework." def items(self): return Entry.objects.published()[:10] def item_pubdate(self, item): return item.pub_date def item_author_name(self, item): return item.author def item_description(self, item): return item.body_html
<commit_before>from __future__ import absolute_import from django.contrib.syndication.views import Feed from .models import Entry class WeblogEntryFeed(Feed): title = "The Django weblog" link = "http://www.djangoproject.com/weblog/" description = "Latest news about Django, the Python Web framework." def items(self): return Entry.objects.published()[:10] def item_pubdate(self, item): return item.pub_date <commit_msg>Add author name and body to the weblog RSS feed.<commit_after>
from __future__ import absolute_import from django.contrib.syndication.views import Feed from .models import Entry class WeblogEntryFeed(Feed): title = "The Django weblog" link = "http://www.djangoproject.com/weblog/" description = "Latest news about Django, the Python Web framework." def items(self): return Entry.objects.published()[:10] def item_pubdate(self, item): return item.pub_date def item_author_name(self, item): return item.author def item_description(self, item): return item.body_html
from __future__ import absolute_import from django.contrib.syndication.views import Feed from .models import Entry class WeblogEntryFeed(Feed): title = "The Django weblog" link = "http://www.djangoproject.com/weblog/" description = "Latest news about Django, the Python Web framework." def items(self): return Entry.objects.published()[:10] def item_pubdate(self, item): return item.pub_date Add author name and body to the weblog RSS feed.from __future__ import absolute_import from django.contrib.syndication.views import Feed from .models import Entry class WeblogEntryFeed(Feed): title = "The Django weblog" link = "http://www.djangoproject.com/weblog/" description = "Latest news about Django, the Python Web framework." def items(self): return Entry.objects.published()[:10] def item_pubdate(self, item): return item.pub_date def item_author_name(self, item): return item.author def item_description(self, item): return item.body_html
<commit_before>from __future__ import absolute_import from django.contrib.syndication.views import Feed from .models import Entry class WeblogEntryFeed(Feed): title = "The Django weblog" link = "http://www.djangoproject.com/weblog/" description = "Latest news about Django, the Python Web framework." def items(self): return Entry.objects.published()[:10] def item_pubdate(self, item): return item.pub_date <commit_msg>Add author name and body to the weblog RSS feed.<commit_after>from __future__ import absolute_import from django.contrib.syndication.views import Feed from .models import Entry class WeblogEntryFeed(Feed): title = "The Django weblog" link = "http://www.djangoproject.com/weblog/" description = "Latest news about Django, the Python Web framework." def items(self): return Entry.objects.published()[:10] def item_pubdate(self, item): return item.pub_date def item_author_name(self, item): return item.author def item_description(self, item): return item.body_html
5adb6de6b926a54dc9a6dd334a34d42dd2044481
src/pushover_complete/pushover_api.py
src/pushover_complete/pushover_api.py
import requests from .error import PushoverCompleteError class PushoverAPI(object): def __init__(self, token): self.token = token def send_message(self, user, message, device=None, title=None, url=None, url_title=None, priority=None, retry=None, expire=None, timestamp=None, sound=None, html=False): if priority == 2: if retry is None or expire is None: raise PushoverCompleteError('Must specify `retry` and `expire` with priority 2.') payload = { 'token': self.token, 'user': user, 'message': message, 'device': device, 'title': title, 'url': url, 'url_title': url_title, 'priority': priority, 'retry': retry, 'expire': expire, 'timestamp': timestamp, 'sound': sound, 'html': html } headers = {'Content-type': 'application/x-www-form-urlencoded'} r = requests.post( 'https://api.pushover.net/1/messages.json', data=payload, headers=headers ) return r
import requests from .error import PushoverCompleteError class PushoverAPI(object): def __init__(self, token): self.token = token def send_message(self, user, message, device=None, title=None, url=None, url_title=None, priority=None, retry=None, expire=None, timestamp=None, sound=None, html=False): if priority == 2: if retry is None or expire is None: raise PushoverCompleteError('Must specify `retry` and `expire` with priority 2.') payload = { 'token': self.token, 'user': user, 'message': message, 'device': device, 'title': title, 'url': url, 'url_title': url_title, 'priority': priority, 'retry': retry, 'expire': expire, 'timestamp': timestamp, 'sound': sound, 'html': html } headers = {'Content-type': 'application/x-www-form-urlencoded'} r = requests.post( 'https://api.pushover.net/1/messages.json', data=payload, headers=headers def get_sounds(self): resp = requests.get( urljoin(PUSHOVER_API_URL, 'sounds.json'), data={'token': self.token} ) sounds = resp.json().get('sounds', None) if sounds: return sounds else: raise PushoverCompleteError('Could not retrieve sounds')
Split send_message out in to _send_message to reuse functionality later
Split send_message out in to _send_message to reuse functionality later
Python
mit
scolby33/pushover_complete
import requests from .error import PushoverCompleteError class PushoverAPI(object): def __init__(self, token): self.token = token def send_message(self, user, message, device=None, title=None, url=None, url_title=None, priority=None, retry=None, expire=None, timestamp=None, sound=None, html=False): if priority == 2: if retry is None or expire is None: raise PushoverCompleteError('Must specify `retry` and `expire` with priority 2.') payload = { 'token': self.token, 'user': user, 'message': message, 'device': device, 'title': title, 'url': url, 'url_title': url_title, 'priority': priority, 'retry': retry, 'expire': expire, 'timestamp': timestamp, 'sound': sound, 'html': html } headers = {'Content-type': 'application/x-www-form-urlencoded'} r = requests.post( 'https://api.pushover.net/1/messages.json', data=payload, headers=headers ) return r Split send_message out in to _send_message to reuse functionality later
import requests from .error import PushoverCompleteError class PushoverAPI(object): def __init__(self, token): self.token = token def send_message(self, user, message, device=None, title=None, url=None, url_title=None, priority=None, retry=None, expire=None, timestamp=None, sound=None, html=False): if priority == 2: if retry is None or expire is None: raise PushoverCompleteError('Must specify `retry` and `expire` with priority 2.') payload = { 'token': self.token, 'user': user, 'message': message, 'device': device, 'title': title, 'url': url, 'url_title': url_title, 'priority': priority, 'retry': retry, 'expire': expire, 'timestamp': timestamp, 'sound': sound, 'html': html } headers = {'Content-type': 'application/x-www-form-urlencoded'} r = requests.post( 'https://api.pushover.net/1/messages.json', data=payload, headers=headers def get_sounds(self): resp = requests.get( urljoin(PUSHOVER_API_URL, 'sounds.json'), data={'token': self.token} ) sounds = resp.json().get('sounds', None) if sounds: return sounds else: raise PushoverCompleteError('Could not retrieve sounds')
<commit_before>import requests from .error import PushoverCompleteError class PushoverAPI(object): def __init__(self, token): self.token = token def send_message(self, user, message, device=None, title=None, url=None, url_title=None, priority=None, retry=None, expire=None, timestamp=None, sound=None, html=False): if priority == 2: if retry is None or expire is None: raise PushoverCompleteError('Must specify `retry` and `expire` with priority 2.') payload = { 'token': self.token, 'user': user, 'message': message, 'device': device, 'title': title, 'url': url, 'url_title': url_title, 'priority': priority, 'retry': retry, 'expire': expire, 'timestamp': timestamp, 'sound': sound, 'html': html } headers = {'Content-type': 'application/x-www-form-urlencoded'} r = requests.post( 'https://api.pushover.net/1/messages.json', data=payload, headers=headers ) return r <commit_msg>Split send_message out in to _send_message to reuse functionality later<commit_after>
import requests from .error import PushoverCompleteError class PushoverAPI(object): def __init__(self, token): self.token = token def send_message(self, user, message, device=None, title=None, url=None, url_title=None, priority=None, retry=None, expire=None, timestamp=None, sound=None, html=False): if priority == 2: if retry is None or expire is None: raise PushoverCompleteError('Must specify `retry` and `expire` with priority 2.') payload = { 'token': self.token, 'user': user, 'message': message, 'device': device, 'title': title, 'url': url, 'url_title': url_title, 'priority': priority, 'retry': retry, 'expire': expire, 'timestamp': timestamp, 'sound': sound, 'html': html } headers = {'Content-type': 'application/x-www-form-urlencoded'} r = requests.post( 'https://api.pushover.net/1/messages.json', data=payload, headers=headers def get_sounds(self): resp = requests.get( urljoin(PUSHOVER_API_URL, 'sounds.json'), data={'token': self.token} ) sounds = resp.json().get('sounds', None) if sounds: return sounds else: raise PushoverCompleteError('Could not retrieve sounds')
import requests from .error import PushoverCompleteError class PushoverAPI(object): def __init__(self, token): self.token = token def send_message(self, user, message, device=None, title=None, url=None, url_title=None, priority=None, retry=None, expire=None, timestamp=None, sound=None, html=False): if priority == 2: if retry is None or expire is None: raise PushoverCompleteError('Must specify `retry` and `expire` with priority 2.') payload = { 'token': self.token, 'user': user, 'message': message, 'device': device, 'title': title, 'url': url, 'url_title': url_title, 'priority': priority, 'retry': retry, 'expire': expire, 'timestamp': timestamp, 'sound': sound, 'html': html } headers = {'Content-type': 'application/x-www-form-urlencoded'} r = requests.post( 'https://api.pushover.net/1/messages.json', data=payload, headers=headers ) return r Split send_message out in to _send_message to reuse functionality laterimport requests from .error import PushoverCompleteError class PushoverAPI(object): def __init__(self, token): self.token = token def send_message(self, user, message, device=None, title=None, url=None, url_title=None, priority=None, retry=None, expire=None, timestamp=None, sound=None, html=False): if priority == 2: if retry is None or expire is None: raise PushoverCompleteError('Must specify `retry` and `expire` with priority 2.') payload = { 'token': self.token, 'user': user, 'message': message, 'device': device, 'title': title, 'url': url, 'url_title': url_title, 'priority': priority, 'retry': retry, 'expire': expire, 'timestamp': timestamp, 'sound': sound, 'html': html } headers = {'Content-type': 'application/x-www-form-urlencoded'} r = requests.post( 'https://api.pushover.net/1/messages.json', data=payload, headers=headers def get_sounds(self): resp = requests.get( urljoin(PUSHOVER_API_URL, 'sounds.json'), data={'token': self.token} ) sounds = resp.json().get('sounds', None) if sounds: return sounds else: raise PushoverCompleteError('Could not retrieve sounds')
<commit_before>import requests from .error import PushoverCompleteError class PushoverAPI(object): def __init__(self, token): self.token = token def send_message(self, user, message, device=None, title=None, url=None, url_title=None, priority=None, retry=None, expire=None, timestamp=None, sound=None, html=False): if priority == 2: if retry is None or expire is None: raise PushoverCompleteError('Must specify `retry` and `expire` with priority 2.') payload = { 'token': self.token, 'user': user, 'message': message, 'device': device, 'title': title, 'url': url, 'url_title': url_title, 'priority': priority, 'retry': retry, 'expire': expire, 'timestamp': timestamp, 'sound': sound, 'html': html } headers = {'Content-type': 'application/x-www-form-urlencoded'} r = requests.post( 'https://api.pushover.net/1/messages.json', data=payload, headers=headers ) return r <commit_msg>Split send_message out in to _send_message to reuse functionality later<commit_after>import requests from .error import PushoverCompleteError class PushoverAPI(object): def __init__(self, token): self.token = token def send_message(self, user, message, device=None, title=None, url=None, url_title=None, priority=None, retry=None, expire=None, timestamp=None, sound=None, html=False): if priority == 2: if retry is None or expire is None: raise PushoverCompleteError('Must specify `retry` and `expire` with priority 2.') payload = { 'token': self.token, 'user': user, 'message': message, 'device': device, 'title': title, 'url': url, 'url_title': url_title, 'priority': priority, 'retry': retry, 'expire': expire, 'timestamp': timestamp, 'sound': sound, 'html': html } headers = {'Content-type': 'application/x-www-form-urlencoded'} r = requests.post( 'https://api.pushover.net/1/messages.json', data=payload, headers=headers def get_sounds(self): resp = requests.get( urljoin(PUSHOVER_API_URL, 'sounds.json'), data={'token': self.token} ) sounds = resp.json().get('sounds', None) if sounds: return sounds else: raise PushoverCompleteError('Could not retrieve sounds')
f87b8c5b94e3e163f19ea0414d1fb2c42f09c166
test/test_genmidi.py
test/test_genmidi.py
import unittest import tempfile from pyknon.MidiFile import MIDIFile from pyknon.genmidi import Midi, MidiError from pyknon.music import NoteSeq, Note class TestMidi(unittest.TestCase): def test_init(self): midi = Midi(1, tempo=120) self.assertEqual(midi.number_tracks, 1) self.assertIsInstance(midi.midi_data, MIDIFile) def test_seq_notes_with_more_tracks_than_exists(self): midi = Midi(1) with self.assertRaises(MidiError): midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_notes(self): midi = Midi(2) midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) class TestWriteMidi(unittest.TestCase): def test_write_midifile(self): notes1 = NoteSeq("D4 F#8 R A") midi = Midi(1, tempo=133) midi.seq_notes(notes1, track=0) midi.write(tempfile.TemporaryFile())
import unittest import tempfile from pyknon.MidiFile import MIDIFile from pyknon.genmidi import Midi, MidiError from pyknon.music import NoteSeq, Note class TestMidi(unittest.TestCase): def test_init(self): midi = Midi(1, tempo=120) self.assertEqual(midi.number_tracks, 1) self.assertIsInstance(midi.midi_data, MIDIFile) def test_seq_notes_with_more_tracks_than_exists(self): midi = Midi(1) with self.assertRaises(MidiError): midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_notes(self): midi = Midi(2) midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_chords(self): chords = [NoteSeq("C E G"), NoteSeq("G B D")] midi = Midi() midi.seq_chords(chords) class TestWriteMidi(unittest.TestCase): def test_write_midifile(self): notes1 = NoteSeq("D4 F#8 R A") midi = Midi(1, tempo=133) midi.seq_notes(notes1, track=0) midi.write(tempfile.TemporaryFile())
Test for sequence of chords
Test for sequence of chords
Python
mit
palmerev/pyknon,kroger/pyknon
import unittest import tempfile from pyknon.MidiFile import MIDIFile from pyknon.genmidi import Midi, MidiError from pyknon.music import NoteSeq, Note class TestMidi(unittest.TestCase): def test_init(self): midi = Midi(1, tempo=120) self.assertEqual(midi.number_tracks, 1) self.assertIsInstance(midi.midi_data, MIDIFile) def test_seq_notes_with_more_tracks_than_exists(self): midi = Midi(1) with self.assertRaises(MidiError): midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_notes(self): midi = Midi(2) midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) class TestWriteMidi(unittest.TestCase): def test_write_midifile(self): notes1 = NoteSeq("D4 F#8 R A") midi = Midi(1, tempo=133) midi.seq_notes(notes1, track=0) midi.write(tempfile.TemporaryFile()) Test for sequence of chords
import unittest import tempfile from pyknon.MidiFile import MIDIFile from pyknon.genmidi import Midi, MidiError from pyknon.music import NoteSeq, Note class TestMidi(unittest.TestCase): def test_init(self): midi = Midi(1, tempo=120) self.assertEqual(midi.number_tracks, 1) self.assertIsInstance(midi.midi_data, MIDIFile) def test_seq_notes_with_more_tracks_than_exists(self): midi = Midi(1) with self.assertRaises(MidiError): midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_notes(self): midi = Midi(2) midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_chords(self): chords = [NoteSeq("C E G"), NoteSeq("G B D")] midi = Midi() midi.seq_chords(chords) class TestWriteMidi(unittest.TestCase): def test_write_midifile(self): notes1 = NoteSeq("D4 F#8 R A") midi = Midi(1, tempo=133) midi.seq_notes(notes1, track=0) midi.write(tempfile.TemporaryFile())
<commit_before>import unittest import tempfile from pyknon.MidiFile import MIDIFile from pyknon.genmidi import Midi, MidiError from pyknon.music import NoteSeq, Note class TestMidi(unittest.TestCase): def test_init(self): midi = Midi(1, tempo=120) self.assertEqual(midi.number_tracks, 1) self.assertIsInstance(midi.midi_data, MIDIFile) def test_seq_notes_with_more_tracks_than_exists(self): midi = Midi(1) with self.assertRaises(MidiError): midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_notes(self): midi = Midi(2) midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) class TestWriteMidi(unittest.TestCase): def test_write_midifile(self): notes1 = NoteSeq("D4 F#8 R A") midi = Midi(1, tempo=133) midi.seq_notes(notes1, track=0) midi.write(tempfile.TemporaryFile()) <commit_msg>Test for sequence of chords<commit_after>
import unittest import tempfile from pyknon.MidiFile import MIDIFile from pyknon.genmidi import Midi, MidiError from pyknon.music import NoteSeq, Note class TestMidi(unittest.TestCase): def test_init(self): midi = Midi(1, tempo=120) self.assertEqual(midi.number_tracks, 1) self.assertIsInstance(midi.midi_data, MIDIFile) def test_seq_notes_with_more_tracks_than_exists(self): midi = Midi(1) with self.assertRaises(MidiError): midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_notes(self): midi = Midi(2) midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_chords(self): chords = [NoteSeq("C E G"), NoteSeq("G B D")] midi = Midi() midi.seq_chords(chords) class TestWriteMidi(unittest.TestCase): def test_write_midifile(self): notes1 = NoteSeq("D4 F#8 R A") midi = Midi(1, tempo=133) midi.seq_notes(notes1, track=0) midi.write(tempfile.TemporaryFile())
import unittest import tempfile from pyknon.MidiFile import MIDIFile from pyknon.genmidi import Midi, MidiError from pyknon.music import NoteSeq, Note class TestMidi(unittest.TestCase): def test_init(self): midi = Midi(1, tempo=120) self.assertEqual(midi.number_tracks, 1) self.assertIsInstance(midi.midi_data, MIDIFile) def test_seq_notes_with_more_tracks_than_exists(self): midi = Midi(1) with self.assertRaises(MidiError): midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_notes(self): midi = Midi(2) midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) class TestWriteMidi(unittest.TestCase): def test_write_midifile(self): notes1 = NoteSeq("D4 F#8 R A") midi = Midi(1, tempo=133) midi.seq_notes(notes1, track=0) midi.write(tempfile.TemporaryFile()) Test for sequence of chordsimport unittest import tempfile from pyknon.MidiFile import MIDIFile from pyknon.genmidi import Midi, MidiError from pyknon.music import NoteSeq, Note class TestMidi(unittest.TestCase): def test_init(self): midi = Midi(1, tempo=120) self.assertEqual(midi.number_tracks, 1) self.assertIsInstance(midi.midi_data, MIDIFile) def test_seq_notes_with_more_tracks_than_exists(self): midi = Midi(1) with self.assertRaises(MidiError): midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_notes(self): midi = Midi(2) midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_chords(self): chords = [NoteSeq("C E G"), NoteSeq("G B D")] midi = Midi() midi.seq_chords(chords) class TestWriteMidi(unittest.TestCase): def test_write_midifile(self): notes1 = NoteSeq("D4 F#8 R A") midi = Midi(1, tempo=133) midi.seq_notes(notes1, track=0) midi.write(tempfile.TemporaryFile())
<commit_before>import unittest import tempfile from pyknon.MidiFile import MIDIFile from pyknon.genmidi import Midi, MidiError from pyknon.music import NoteSeq, Note class TestMidi(unittest.TestCase): def test_init(self): midi = Midi(1, tempo=120) self.assertEqual(midi.number_tracks, 1) self.assertIsInstance(midi.midi_data, MIDIFile) def test_seq_notes_with_more_tracks_than_exists(self): midi = Midi(1) with self.assertRaises(MidiError): midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_notes(self): midi = Midi(2) midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) class TestWriteMidi(unittest.TestCase): def test_write_midifile(self): notes1 = NoteSeq("D4 F#8 R A") midi = Midi(1, tempo=133) midi.seq_notes(notes1, track=0) midi.write(tempfile.TemporaryFile()) <commit_msg>Test for sequence of chords<commit_after>import unittest import tempfile from pyknon.MidiFile import MIDIFile from pyknon.genmidi import Midi, MidiError from pyknon.music import NoteSeq, Note class TestMidi(unittest.TestCase): def test_init(self): midi = Midi(1, tempo=120) self.assertEqual(midi.number_tracks, 1) self.assertIsInstance(midi.midi_data, MIDIFile) def test_seq_notes_with_more_tracks_than_exists(self): midi = Midi(1) with self.assertRaises(MidiError): midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_notes(self): midi = Midi(2) midi.seq_notes(NoteSeq("C D"), track=0) midi.seq_notes(NoteSeq("D E"), track=1) def test_seq_chords(self): chords = [NoteSeq("C E G"), NoteSeq("G B D")] midi = Midi() midi.seq_chords(chords) class TestWriteMidi(unittest.TestCase): def test_write_midifile(self): notes1 = NoteSeq("D4 F#8 R A") midi = Midi(1, tempo=133) midi.seq_notes(notes1, track=0) midi.write(tempfile.TemporaryFile())
86cf16611fe4126f4345477b24da5c15fed4c1e8
eval_kernel/eval_kernel.py
eval_kernel/eval_kernel.py
from __future__ import print_function from jupyter_kernel import MagicKernel import os class EvalKernel(MagicKernel): implementation = 'Eval' implementation_version = '1.0' language = 'python' language_version = '0.1' banner = "Eval kernel - evaluates simple Python statements and expressions" env = {} def get_usage(self): return "This is a usage statement." def set_variable(self, name, value): """ Set a variable in the kernel language. """ self.env[name] = value def get_kernel_help_on(self, expr, level): return "Sorry, no help is available on '%s'." % expr def do_execute_direct(self, code): try: return eval(code.strip(), self.env) except: try: exec code.strip() in self.env except: return "Error: " + code if __name__ == '__main__': from IPython.kernel.zmq.kernelapp import IPKernelApp IPKernelApp.launch_instance(kernel_class=EvalKernel)
from __future__ import print_function from jupyter_kernel import MagicKernel class EvalKernel(MagicKernel): implementation = 'Eval' implementation_version = '1.0' language = 'python' language_version = '0.1' banner = "Eval kernel - evaluates simple Python statements and expressions" env = {} def get_usage(self): return "This is a usage statement." def set_variable(self, name, value): """ Set a variable in the kernel language. """ self.env[name] = value def get_variable(self, name): """ Get a variable from the kernel language. """ return self.env.get(name, None) def do_execute_direct(self, code): python_magic = self.line_magics['python'] resp = python_magic.eval(code.strip()) if not resp is None: self.Print(str(resp)) def get_completions(self, token): python_magic = self.line_magics['python'] return python_magic.get_completions(token) def get_kernel_help_on(self, expr, level=0): python_magic = self.line_magics['python'] return python_magic.get_help_on(expr, level) if __name__ == '__main__': from IPython.kernel.zmq.kernelapp import IPKernelApp IPKernelApp.launch_instance(kernel_class=EvalKernel)
Update eval kernel to use python magic.
Update eval kernel to use python magic.
Python
bsd-3-clause
Calysto/metakernel
from __future__ import print_function from jupyter_kernel import MagicKernel import os class EvalKernel(MagicKernel): implementation = 'Eval' implementation_version = '1.0' language = 'python' language_version = '0.1' banner = "Eval kernel - evaluates simple Python statements and expressions" env = {} def get_usage(self): return "This is a usage statement." def set_variable(self, name, value): """ Set a variable in the kernel language. """ self.env[name] = value def get_kernel_help_on(self, expr, level): return "Sorry, no help is available on '%s'." % expr def do_execute_direct(self, code): try: return eval(code.strip(), self.env) except: try: exec code.strip() in self.env except: return "Error: " + code if __name__ == '__main__': from IPython.kernel.zmq.kernelapp import IPKernelApp IPKernelApp.launch_instance(kernel_class=EvalKernel) Update eval kernel to use python magic.
from __future__ import print_function from jupyter_kernel import MagicKernel class EvalKernel(MagicKernel): implementation = 'Eval' implementation_version = '1.0' language = 'python' language_version = '0.1' banner = "Eval kernel - evaluates simple Python statements and expressions" env = {} def get_usage(self): return "This is a usage statement." def set_variable(self, name, value): """ Set a variable in the kernel language. """ self.env[name] = value def get_variable(self, name): """ Get a variable from the kernel language. """ return self.env.get(name, None) def do_execute_direct(self, code): python_magic = self.line_magics['python'] resp = python_magic.eval(code.strip()) if not resp is None: self.Print(str(resp)) def get_completions(self, token): python_magic = self.line_magics['python'] return python_magic.get_completions(token) def get_kernel_help_on(self, expr, level=0): python_magic = self.line_magics['python'] return python_magic.get_help_on(expr, level) if __name__ == '__main__': from IPython.kernel.zmq.kernelapp import IPKernelApp IPKernelApp.launch_instance(kernel_class=EvalKernel)
<commit_before>from __future__ import print_function from jupyter_kernel import MagicKernel import os class EvalKernel(MagicKernel): implementation = 'Eval' implementation_version = '1.0' language = 'python' language_version = '0.1' banner = "Eval kernel - evaluates simple Python statements and expressions" env = {} def get_usage(self): return "This is a usage statement." def set_variable(self, name, value): """ Set a variable in the kernel language. """ self.env[name] = value def get_kernel_help_on(self, expr, level): return "Sorry, no help is available on '%s'." % expr def do_execute_direct(self, code): try: return eval(code.strip(), self.env) except: try: exec code.strip() in self.env except: return "Error: " + code if __name__ == '__main__': from IPython.kernel.zmq.kernelapp import IPKernelApp IPKernelApp.launch_instance(kernel_class=EvalKernel) <commit_msg>Update eval kernel to use python magic.<commit_after>
from __future__ import print_function from jupyter_kernel import MagicKernel class EvalKernel(MagicKernel): implementation = 'Eval' implementation_version = '1.0' language = 'python' language_version = '0.1' banner = "Eval kernel - evaluates simple Python statements and expressions" env = {} def get_usage(self): return "This is a usage statement." def set_variable(self, name, value): """ Set a variable in the kernel language. """ self.env[name] = value def get_variable(self, name): """ Get a variable from the kernel language. """ return self.env.get(name, None) def do_execute_direct(self, code): python_magic = self.line_magics['python'] resp = python_magic.eval(code.strip()) if not resp is None: self.Print(str(resp)) def get_completions(self, token): python_magic = self.line_magics['python'] return python_magic.get_completions(token) def get_kernel_help_on(self, expr, level=0): python_magic = self.line_magics['python'] return python_magic.get_help_on(expr, level) if __name__ == '__main__': from IPython.kernel.zmq.kernelapp import IPKernelApp IPKernelApp.launch_instance(kernel_class=EvalKernel)
from __future__ import print_function from jupyter_kernel import MagicKernel import os class EvalKernel(MagicKernel): implementation = 'Eval' implementation_version = '1.0' language = 'python' language_version = '0.1' banner = "Eval kernel - evaluates simple Python statements and expressions" env = {} def get_usage(self): return "This is a usage statement." def set_variable(self, name, value): """ Set a variable in the kernel language. """ self.env[name] = value def get_kernel_help_on(self, expr, level): return "Sorry, no help is available on '%s'." % expr def do_execute_direct(self, code): try: return eval(code.strip(), self.env) except: try: exec code.strip() in self.env except: return "Error: " + code if __name__ == '__main__': from IPython.kernel.zmq.kernelapp import IPKernelApp IPKernelApp.launch_instance(kernel_class=EvalKernel) Update eval kernel to use python magic.from __future__ import print_function from jupyter_kernel import MagicKernel class EvalKernel(MagicKernel): implementation = 'Eval' implementation_version = '1.0' language = 'python' language_version = '0.1' banner = "Eval kernel - evaluates simple Python statements and expressions" env = {} def get_usage(self): return "This is a usage statement." def set_variable(self, name, value): """ Set a variable in the kernel language. """ self.env[name] = value def get_variable(self, name): """ Get a variable from the kernel language. """ return self.env.get(name, None) def do_execute_direct(self, code): python_magic = self.line_magics['python'] resp = python_magic.eval(code.strip()) if not resp is None: self.Print(str(resp)) def get_completions(self, token): python_magic = self.line_magics['python'] return python_magic.get_completions(token) def get_kernel_help_on(self, expr, level=0): python_magic = self.line_magics['python'] return python_magic.get_help_on(expr, level) if __name__ == '__main__': from IPython.kernel.zmq.kernelapp import IPKernelApp IPKernelApp.launch_instance(kernel_class=EvalKernel)
<commit_before>from __future__ import print_function from jupyter_kernel import MagicKernel import os class EvalKernel(MagicKernel): implementation = 'Eval' implementation_version = '1.0' language = 'python' language_version = '0.1' banner = "Eval kernel - evaluates simple Python statements and expressions" env = {} def get_usage(self): return "This is a usage statement." def set_variable(self, name, value): """ Set a variable in the kernel language. """ self.env[name] = value def get_kernel_help_on(self, expr, level): return "Sorry, no help is available on '%s'." % expr def do_execute_direct(self, code): try: return eval(code.strip(), self.env) except: try: exec code.strip() in self.env except: return "Error: " + code if __name__ == '__main__': from IPython.kernel.zmq.kernelapp import IPKernelApp IPKernelApp.launch_instance(kernel_class=EvalKernel) <commit_msg>Update eval kernel to use python magic.<commit_after>from __future__ import print_function from jupyter_kernel import MagicKernel class EvalKernel(MagicKernel): implementation = 'Eval' implementation_version = '1.0' language = 'python' language_version = '0.1' banner = "Eval kernel - evaluates simple Python statements and expressions" env = {} def get_usage(self): return "This is a usage statement." def set_variable(self, name, value): """ Set a variable in the kernel language. """ self.env[name] = value def get_variable(self, name): """ Get a variable from the kernel language. """ return self.env.get(name, None) def do_execute_direct(self, code): python_magic = self.line_magics['python'] resp = python_magic.eval(code.strip()) if not resp is None: self.Print(str(resp)) def get_completions(self, token): python_magic = self.line_magics['python'] return python_magic.get_completions(token) def get_kernel_help_on(self, expr, level=0): python_magic = self.line_magics['python'] return python_magic.get_help_on(expr, level) if __name__ == '__main__': from IPython.kernel.zmq.kernelapp import IPKernelApp IPKernelApp.launch_instance(kernel_class=EvalKernel)
b0b4bad0ca68ebd1927229e85e7116fb63126c65
src/olympia/zadmin/helpers.py
src/olympia/zadmin/helpers.py
from jingo import register from olympia.amo.urlresolvers import reverse @register.function def admin_site_links(): return { 'addons': [ ('Search for add-ons by name or id', reverse('zadmin.addon-search')), ('Featured add-ons', reverse('zadmin.features')), ('Discovery Pane promo modules', reverse('discovery.module_admin')), ('Monthly Pick', reverse('zadmin.monthly_pick')), ('Bulk add-on validation', reverse('zadmin.validation')), ('Fake mail', reverse('zadmin.mail')), ('ACR Reports', reverse('zadmin.compat')), ('Email Add-on Developers', reverse('zadmin.email_devs')), ], 'users': [ ('Configure groups', reverse('admin:access_group_changelist')), ], 'settings': [ ('View site settings', reverse('zadmin.settings')), ('Django admin pages', reverse('zadmin.home')), ('Site Events', reverse('zadmin.site_events')), ], 'tools': [ ('View request environment', reverse('amo.env')), ('Manage elasticsearch', reverse('zadmin.elastic')), ('Purge data from memcache', reverse('zadmin.memcache')), ('View event log', reverse('admin:editors_eventlog_changelist')), ('View addon log', reverse('admin:devhub_activitylog_changelist')), ('Generate error', reverse('zadmin.generate-error')), ('Site Status', reverse('amo.monitor')), ], }
from jingo import register from olympia.amo.urlresolvers import reverse @register.function def admin_site_links(): return { 'addons': [ ('Search for add-ons by name or id', reverse('zadmin.addon-search')), ('Featured add-ons', reverse('zadmin.features')), ('Discovery Pane promo modules', reverse('discovery.module_admin')), ('Monthly Pick', reverse('zadmin.monthly_pick')), ('Bulk add-on validation', reverse('zadmin.validation')), ('Fake mail', reverse('zadmin.mail')), ('ACR Reports', reverse('zadmin.compat')), ('Email Add-on Developers', reverse('zadmin.email_devs')), ], 'users': [ ('Configure groups', reverse('admin:access_group_changelist')), ], 'settings': [ ('View site settings', reverse('zadmin.settings')), ('Django admin pages', reverse('zadmin.home')), ('Site Events', reverse('zadmin.site_events')), ], 'tools': [ ('View request environment', reverse('amo.env')), ('Manage elasticsearch', reverse('zadmin.elastic')), ('Purge data from memcache', reverse('zadmin.memcache')), ('View event log', reverse('admin:editors_eventlog_changelist')), ('View addon log', reverse('admin:devhub_activitylog_changelist')), ('Site Status', reverse('amo.monitor')), ], }
Remove generate error page from admin site
Remove generate error page from admin site
Python
bsd-3-clause
bqbn/addons-server,wagnerand/olympia,harry-7/addons-server,wagnerand/addons-server,harikishen/addons-server,psiinon/addons-server,lavish205/olympia,mstriemer/addons-server,kumar303/addons-server,Prashant-Surya/addons-server,mstriemer/olympia,mozilla/addons-server,harikishen/addons-server,Revanth47/addons-server,mstriemer/addons-server,mstriemer/olympia,lavish205/olympia,lavish205/olympia,wagnerand/olympia,diox/olympia,eviljeff/olympia,aviarypl/mozilla-l10n-addons-server,mozilla/olympia,tsl143/addons-server,Revanth47/addons-server,wagnerand/addons-server,psiinon/addons-server,eviljeff/olympia,wagnerand/addons-server,harry-7/addons-server,kumar303/addons-server,wagnerand/olympia,eviljeff/olympia,Prashant-Surya/addons-server,bqbn/addons-server,kumar303/addons-server,Revanth47/addons-server,kumar303/olympia,harry-7/addons-server,kumar303/olympia,aviarypl/mozilla-l10n-addons-server,kumar303/addons-server,mstriemer/addons-server,harikishen/addons-server,mstriemer/olympia,Prashant-Surya/addons-server,mozilla/olympia,diox/olympia,psiinon/addons-server,harry-7/addons-server,wagnerand/olympia,aviarypl/mozilla-l10n-addons-server,kumar303/olympia,mstriemer/olympia,mozilla/addons-server,bqbn/addons-server,Revanth47/addons-server,mstriemer/addons-server,diox/olympia,harikishen/addons-server,wagnerand/addons-server,diox/olympia,atiqueahmedziad/addons-server,psiinon/addons-server,eviljeff/olympia,tsl143/addons-server,mozilla/olympia,kumar303/olympia,lavish205/olympia,atiqueahmedziad/addons-server,tsl143/addons-server,tsl143/addons-server,mozilla/addons-server,mozilla/olympia,bqbn/addons-server,aviarypl/mozilla-l10n-addons-server,Prashant-Surya/addons-server,atiqueahmedziad/addons-server,mozilla/addons-server,atiqueahmedziad/addons-server
from jingo import register from olympia.amo.urlresolvers import reverse @register.function def admin_site_links(): return { 'addons': [ ('Search for add-ons by name or id', reverse('zadmin.addon-search')), ('Featured add-ons', reverse('zadmin.features')), ('Discovery Pane promo modules', reverse('discovery.module_admin')), ('Monthly Pick', reverse('zadmin.monthly_pick')), ('Bulk add-on validation', reverse('zadmin.validation')), ('Fake mail', reverse('zadmin.mail')), ('ACR Reports', reverse('zadmin.compat')), ('Email Add-on Developers', reverse('zadmin.email_devs')), ], 'users': [ ('Configure groups', reverse('admin:access_group_changelist')), ], 'settings': [ ('View site settings', reverse('zadmin.settings')), ('Django admin pages', reverse('zadmin.home')), ('Site Events', reverse('zadmin.site_events')), ], 'tools': [ ('View request environment', reverse('amo.env')), ('Manage elasticsearch', reverse('zadmin.elastic')), ('Purge data from memcache', reverse('zadmin.memcache')), ('View event log', reverse('admin:editors_eventlog_changelist')), ('View addon log', reverse('admin:devhub_activitylog_changelist')), ('Generate error', reverse('zadmin.generate-error')), ('Site Status', reverse('amo.monitor')), ], } Remove generate error page from admin site
from jingo import register from olympia.amo.urlresolvers import reverse @register.function def admin_site_links(): return { 'addons': [ ('Search for add-ons by name or id', reverse('zadmin.addon-search')), ('Featured add-ons', reverse('zadmin.features')), ('Discovery Pane promo modules', reverse('discovery.module_admin')), ('Monthly Pick', reverse('zadmin.monthly_pick')), ('Bulk add-on validation', reverse('zadmin.validation')), ('Fake mail', reverse('zadmin.mail')), ('ACR Reports', reverse('zadmin.compat')), ('Email Add-on Developers', reverse('zadmin.email_devs')), ], 'users': [ ('Configure groups', reverse('admin:access_group_changelist')), ], 'settings': [ ('View site settings', reverse('zadmin.settings')), ('Django admin pages', reverse('zadmin.home')), ('Site Events', reverse('zadmin.site_events')), ], 'tools': [ ('View request environment', reverse('amo.env')), ('Manage elasticsearch', reverse('zadmin.elastic')), ('Purge data from memcache', reverse('zadmin.memcache')), ('View event log', reverse('admin:editors_eventlog_changelist')), ('View addon log', reverse('admin:devhub_activitylog_changelist')), ('Site Status', reverse('amo.monitor')), ], }
<commit_before>from jingo import register from olympia.amo.urlresolvers import reverse @register.function def admin_site_links(): return { 'addons': [ ('Search for add-ons by name or id', reverse('zadmin.addon-search')), ('Featured add-ons', reverse('zadmin.features')), ('Discovery Pane promo modules', reverse('discovery.module_admin')), ('Monthly Pick', reverse('zadmin.monthly_pick')), ('Bulk add-on validation', reverse('zadmin.validation')), ('Fake mail', reverse('zadmin.mail')), ('ACR Reports', reverse('zadmin.compat')), ('Email Add-on Developers', reverse('zadmin.email_devs')), ], 'users': [ ('Configure groups', reverse('admin:access_group_changelist')), ], 'settings': [ ('View site settings', reverse('zadmin.settings')), ('Django admin pages', reverse('zadmin.home')), ('Site Events', reverse('zadmin.site_events')), ], 'tools': [ ('View request environment', reverse('amo.env')), ('Manage elasticsearch', reverse('zadmin.elastic')), ('Purge data from memcache', reverse('zadmin.memcache')), ('View event log', reverse('admin:editors_eventlog_changelist')), ('View addon log', reverse('admin:devhub_activitylog_changelist')), ('Generate error', reverse('zadmin.generate-error')), ('Site Status', reverse('amo.monitor')), ], } <commit_msg>Remove generate error page from admin site<commit_after>
from jingo import register from olympia.amo.urlresolvers import reverse @register.function def admin_site_links(): return { 'addons': [ ('Search for add-ons by name or id', reverse('zadmin.addon-search')), ('Featured add-ons', reverse('zadmin.features')), ('Discovery Pane promo modules', reverse('discovery.module_admin')), ('Monthly Pick', reverse('zadmin.monthly_pick')), ('Bulk add-on validation', reverse('zadmin.validation')), ('Fake mail', reverse('zadmin.mail')), ('ACR Reports', reverse('zadmin.compat')), ('Email Add-on Developers', reverse('zadmin.email_devs')), ], 'users': [ ('Configure groups', reverse('admin:access_group_changelist')), ], 'settings': [ ('View site settings', reverse('zadmin.settings')), ('Django admin pages', reverse('zadmin.home')), ('Site Events', reverse('zadmin.site_events')), ], 'tools': [ ('View request environment', reverse('amo.env')), ('Manage elasticsearch', reverse('zadmin.elastic')), ('Purge data from memcache', reverse('zadmin.memcache')), ('View event log', reverse('admin:editors_eventlog_changelist')), ('View addon log', reverse('admin:devhub_activitylog_changelist')), ('Site Status', reverse('amo.monitor')), ], }
from jingo import register from olympia.amo.urlresolvers import reverse @register.function def admin_site_links(): return { 'addons': [ ('Search for add-ons by name or id', reverse('zadmin.addon-search')), ('Featured add-ons', reverse('zadmin.features')), ('Discovery Pane promo modules', reverse('discovery.module_admin')), ('Monthly Pick', reverse('zadmin.monthly_pick')), ('Bulk add-on validation', reverse('zadmin.validation')), ('Fake mail', reverse('zadmin.mail')), ('ACR Reports', reverse('zadmin.compat')), ('Email Add-on Developers', reverse('zadmin.email_devs')), ], 'users': [ ('Configure groups', reverse('admin:access_group_changelist')), ], 'settings': [ ('View site settings', reverse('zadmin.settings')), ('Django admin pages', reverse('zadmin.home')), ('Site Events', reverse('zadmin.site_events')), ], 'tools': [ ('View request environment', reverse('amo.env')), ('Manage elasticsearch', reverse('zadmin.elastic')), ('Purge data from memcache', reverse('zadmin.memcache')), ('View event log', reverse('admin:editors_eventlog_changelist')), ('View addon log', reverse('admin:devhub_activitylog_changelist')), ('Generate error', reverse('zadmin.generate-error')), ('Site Status', reverse('amo.monitor')), ], } Remove generate error page from admin sitefrom jingo import register from olympia.amo.urlresolvers import reverse @register.function def admin_site_links(): return { 'addons': [ ('Search for add-ons by name or id', reverse('zadmin.addon-search')), ('Featured add-ons', reverse('zadmin.features')), ('Discovery Pane promo modules', reverse('discovery.module_admin')), ('Monthly Pick', reverse('zadmin.monthly_pick')), ('Bulk add-on validation', reverse('zadmin.validation')), ('Fake mail', reverse('zadmin.mail')), ('ACR Reports', reverse('zadmin.compat')), ('Email Add-on Developers', reverse('zadmin.email_devs')), ], 'users': [ ('Configure groups', reverse('admin:access_group_changelist')), ], 'settings': [ ('View site settings', reverse('zadmin.settings')), ('Django admin pages', reverse('zadmin.home')), ('Site Events', reverse('zadmin.site_events')), ], 'tools': [ ('View request environment', reverse('amo.env')), ('Manage elasticsearch', reverse('zadmin.elastic')), ('Purge data from memcache', reverse('zadmin.memcache')), ('View event log', reverse('admin:editors_eventlog_changelist')), ('View addon log', reverse('admin:devhub_activitylog_changelist')), ('Site Status', reverse('amo.monitor')), ], }
<commit_before>from jingo import register from olympia.amo.urlresolvers import reverse @register.function def admin_site_links(): return { 'addons': [ ('Search for add-ons by name or id', reverse('zadmin.addon-search')), ('Featured add-ons', reverse('zadmin.features')), ('Discovery Pane promo modules', reverse('discovery.module_admin')), ('Monthly Pick', reverse('zadmin.monthly_pick')), ('Bulk add-on validation', reverse('zadmin.validation')), ('Fake mail', reverse('zadmin.mail')), ('ACR Reports', reverse('zadmin.compat')), ('Email Add-on Developers', reverse('zadmin.email_devs')), ], 'users': [ ('Configure groups', reverse('admin:access_group_changelist')), ], 'settings': [ ('View site settings', reverse('zadmin.settings')), ('Django admin pages', reverse('zadmin.home')), ('Site Events', reverse('zadmin.site_events')), ], 'tools': [ ('View request environment', reverse('amo.env')), ('Manage elasticsearch', reverse('zadmin.elastic')), ('Purge data from memcache', reverse('zadmin.memcache')), ('View event log', reverse('admin:editors_eventlog_changelist')), ('View addon log', reverse('admin:devhub_activitylog_changelist')), ('Generate error', reverse('zadmin.generate-error')), ('Site Status', reverse('amo.monitor')), ], } <commit_msg>Remove generate error page from admin site<commit_after>from jingo import register from olympia.amo.urlresolvers import reverse @register.function def admin_site_links(): return { 'addons': [ ('Search for add-ons by name or id', reverse('zadmin.addon-search')), ('Featured add-ons', reverse('zadmin.features')), ('Discovery Pane promo modules', reverse('discovery.module_admin')), ('Monthly Pick', reverse('zadmin.monthly_pick')), ('Bulk add-on validation', reverse('zadmin.validation')), ('Fake mail', reverse('zadmin.mail')), ('ACR Reports', reverse('zadmin.compat')), ('Email Add-on Developers', reverse('zadmin.email_devs')), ], 'users': [ ('Configure groups', reverse('admin:access_group_changelist')), ], 'settings': [ ('View site settings', reverse('zadmin.settings')), ('Django admin pages', reverse('zadmin.home')), ('Site Events', reverse('zadmin.site_events')), ], 'tools': [ ('View request environment', reverse('amo.env')), ('Manage elasticsearch', reverse('zadmin.elastic')), ('Purge data from memcache', reverse('zadmin.memcache')), ('View event log', reverse('admin:editors_eventlog_changelist')), ('View addon log', reverse('admin:devhub_activitylog_changelist')), ('Site Status', reverse('amo.monitor')), ], }
be964b02036159567efcaecce5b5d905f23985af
deduper/scanfiles.py
deduper/scanfiles.py
# This file is part of the File Deduper project. It is subject to # the the revised 3-clause BSD license terms as set out in the LICENSE # file found in the top-level directory of this distribution. No part of this # project, including this file, may be copied, modified, propagated, or # distributed except according to the terms contained in the LICENSE fileself. import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close()
# This file is part of the File Deduper project. It is subject to # the the revised 3-clause BSD license terms as set out in the LICENSE # file found in the top-level directory of this distribution. No part of this # project, including this file, may be copied, modified, propagated, or # distributed except according to the terms contained in the LICENSE fileself. import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if not os.path.isfile(fullpath): continue if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close()
Check that fullpath is a regular file before continuing
Check that fullpath is a regular file before continuing
Python
bsd-3-clause
cgspeck/filededuper
# This file is part of the File Deduper project. It is subject to # the the revised 3-clause BSD license terms as set out in the LICENSE # file found in the top-level directory of this distribution. No part of this # project, including this file, may be copied, modified, propagated, or # distributed except according to the terms contained in the LICENSE fileself. import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close() Check that fullpath is a regular file before continuing
# This file is part of the File Deduper project. It is subject to # the the revised 3-clause BSD license terms as set out in the LICENSE # file found in the top-level directory of this distribution. No part of this # project, including this file, may be copied, modified, propagated, or # distributed except according to the terms contained in the LICENSE fileself. import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if not os.path.isfile(fullpath): continue if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close()
<commit_before># This file is part of the File Deduper project. It is subject to # the the revised 3-clause BSD license terms as set out in the LICENSE # file found in the top-level directory of this distribution. No part of this # project, including this file, may be copied, modified, propagated, or # distributed except according to the terms contained in the LICENSE fileself. import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close() <commit_msg>Check that fullpath is a regular file before continuing<commit_after>
# This file is part of the File Deduper project. It is subject to # the the revised 3-clause BSD license terms as set out in the LICENSE # file found in the top-level directory of this distribution. No part of this # project, including this file, may be copied, modified, propagated, or # distributed except according to the terms contained in the LICENSE fileself. import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if not os.path.isfile(fullpath): continue if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close()
# This file is part of the File Deduper project. It is subject to # the the revised 3-clause BSD license terms as set out in the LICENSE # file found in the top-level directory of this distribution. No part of this # project, including this file, may be copied, modified, propagated, or # distributed except according to the terms contained in the LICENSE fileself. import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close() Check that fullpath is a regular file before continuing# This file is part of the File Deduper project. It is subject to # the the revised 3-clause BSD license terms as set out in the LICENSE # file found in the top-level directory of this distribution. No part of this # project, including this file, may be copied, modified, propagated, or # distributed except according to the terms contained in the LICENSE fileself. import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if not os.path.isfile(fullpath): continue if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close()
<commit_before># This file is part of the File Deduper project. It is subject to # the the revised 3-clause BSD license terms as set out in the LICENSE # file found in the top-level directory of this distribution. No part of this # project, including this file, may be copied, modified, propagated, or # distributed except according to the terms contained in the LICENSE fileself. import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close() <commit_msg>Check that fullpath is a regular file before continuing<commit_after># This file is part of the File Deduper project. It is subject to # the the revised 3-clause BSD license terms as set out in the LICENSE # file found in the top-level directory of this distribution. No part of this # project, including this file, may be copied, modified, propagated, or # distributed except according to the terms contained in the LICENSE fileself. import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if not os.path.isfile(fullpath): continue if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close()
a95e891b637f0182031f229465bcded966100889
readthedocs/core/models.py
readthedocs/core/models.py
from django.db import models from django.contrib.auth.models import User STANDARD_EMAIL = "anonymous@readthedocs.org" class UserProfile (models.Model): """Additional information about a User. """ user = models.ForeignKey(User, unique=True, related_name='profile') whitelisted = models.BooleanField() homepage = models.CharField(max_length=100, blank=True) allow_email = models.BooleanField(help_text='Show your email on VCS contributions.', default=True) def get_absolute_url(self): return ('profiles_profile_detail', (), {'username': self.user.username}) get_absolute_url = models.permalink(get_absolute_url) def __unicode__(self): return "%s's profile" % self.user.username def get_contribution_details(self): """ Gets the line to put into commits to attribute the author. Returns a tuple (name, email) """ if self.user.first_name and self.user.last_name: name = '%s %s' % (self.user.first_name, self.user.last_name) else: name = self.user.username if self.allow_email: email = self.user.email else: email = STANDARD_EMAIL return (name, email)
from django.db import models from django.db.models.signals import post_save from django.dispatch import receiver from django.contrib.auth.models import User STANDARD_EMAIL = "anonymous@readthedocs.org" class UserProfile (models.Model): """Additional information about a User. """ user = models.ForeignKey(User, unique=True, related_name='profile') whitelisted = models.BooleanField() homepage = models.CharField(max_length=100, blank=True) allow_email = models.BooleanField(help_text='Show your email on VCS contributions.', default=True) def get_absolute_url(self): return ('profiles_profile_detail', (), {'username': self.user.username}) get_absolute_url = models.permalink(get_absolute_url) def __unicode__(self): return "%s's profile" % self.user.username def get_contribution_details(self): """ Gets the line to put into commits to attribute the author. Returns a tuple (name, email) """ if self.user.first_name and self.user.last_name: name = '%s %s' % (self.user.first_name, self.user.last_name) else: name = self.user.username if self.allow_email: email = self.user.email else: email = STANDARD_EMAIL return (name, email) @receiver(post_save, sender=User) def create_profile(sender, **kwargs): if kwargs['created'] is True: UserProfile.objects.create(user_id=kwargs['instance'].id, whitelisted=False)
Add post save signal for user creation
Add post save signal for user creation
Python
mit
safwanrahman/readthedocs.org,raven47git/readthedocs.org,sunnyzwh/readthedocs.org,agjohnson/readthedocs.org,d0ugal/readthedocs.org,GovReady/readthedocs.org,kenshinthebattosai/readthedocs.org,jerel/readthedocs.org,jerel/readthedocs.org,dirn/readthedocs.org,gjtorikian/readthedocs.org,titiushko/readthedocs.org,wijerasa/readthedocs.org,emawind84/readthedocs.org,asampat3090/readthedocs.org,ojii/readthedocs.org,kdkeyser/readthedocs.org,d0ugal/readthedocs.org,KamranMackey/readthedocs.org,wanghaven/readthedocs.org,kenshinthebattosai/readthedocs.org,SteveViss/readthedocs.org,mhils/readthedocs.org,takluyver/readthedocs.org,espdev/readthedocs.org,davidfischer/readthedocs.org,johncosta/private-readthedocs.org,clarkperkins/readthedocs.org,fujita-shintaro/readthedocs.org,techtonik/readthedocs.org,pombredanne/readthedocs.org,attakei/readthedocs-oauth,atsuyim/readthedocs.org,wanghaven/readthedocs.org,royalwang/readthedocs.org,VishvajitP/readthedocs.org,nyergler/pythonslides,sunnyzwh/readthedocs.org,fujita-shintaro/readthedocs.org,techtonik/readthedocs.org,atsuyim/readthedocs.org,wijerasa/readthedocs.org,attakei/readthedocs-oauth,Carreau/readthedocs.org,sils1297/readthedocs.org,soulshake/readthedocs.org,attakei/readthedocs-oauth,asampat3090/readthedocs.org,VishvajitP/readthedocs.org,atsuyim/readthedocs.org,ojii/readthedocs.org,kdkeyser/readthedocs.org,cgourlay/readthedocs.org,singingwolfboy/readthedocs.org,laplaceliu/readthedocs.org,fujita-shintaro/readthedocs.org,cgourlay/readthedocs.org,istresearch/readthedocs.org,KamranMackey/readthedocs.org,nikolas/readthedocs.org,cgourlay/readthedocs.org,ojii/readthedocs.org,sid-kap/readthedocs.org,raven47git/readthedocs.org,rtfd/readthedocs.org,mrshoki/readthedocs.org,wanghaven/readthedocs.org,safwanrahman/readthedocs.org,sils1297/readthedocs.org,raven47git/readthedocs.org,singingwolfboy/readthedocs.org,johncosta/private-readthedocs.org,takluyver/readthedocs.org,kenwang76/readthedocs.org,LukasBoersma/readthedocs.org,KamranMackey/readthedocs.org,techtonik/readthedocs.org,GovReady/readthedocs.org,mhils/readthedocs.org,CedarLogic/readthedocs.org,nyergler/pythonslides,soulshake/readthedocs.org,sid-kap/readthedocs.org,kdkeyser/readthedocs.org,asampat3090/readthedocs.org,laplaceliu/readthedocs.org,Tazer/readthedocs.org,kdkeyser/readthedocs.org,ojii/readthedocs.org,agjohnson/readthedocs.org,royalwang/readthedocs.org,hach-que/readthedocs.org,michaelmcandrew/readthedocs.org,CedarLogic/readthedocs.org,CedarLogic/readthedocs.org,laplaceliu/readthedocs.org,soulshake/readthedocs.org,dirn/readthedocs.org,gjtorikian/readthedocs.org,emawind84/readthedocs.org,takluyver/readthedocs.org,soulshake/readthedocs.org,hach-que/readthedocs.org,SteveViss/readthedocs.org,agjohnson/readthedocs.org,atsuyim/readthedocs.org,clarkperkins/readthedocs.org,clarkperkins/readthedocs.org,pombredanne/readthedocs.org,stevepiercy/readthedocs.org,titiushko/readthedocs.org,stevepiercy/readthedocs.org,nikolas/readthedocs.org,michaelmcandrew/readthedocs.org,nikolas/readthedocs.org,michaelmcandrew/readthedocs.org,SteveViss/readthedocs.org,LukasBoersma/readthedocs.org,gjtorikian/readthedocs.org,Tazer/readthedocs.org,espdev/readthedocs.org,kenwang76/readthedocs.org,wijerasa/readthedocs.org,royalwang/readthedocs.org,raven47git/readthedocs.org,mrshoki/readthedocs.org,VishvajitP/readthedocs.org,stevepiercy/readthedocs.org,espdev/readthedocs.org,pombredanne/readthedocs.org,Tazer/readthedocs.org,techtonik/readthedocs.org,tddv/readthedocs.org,istresearch/readthedocs.org,d0ugal/readthedocs.org,agjohnson/readthedocs.org,Tazer/readthedocs.org,istresearch/readthedocs.org,GovReady/readthedocs.org,nyergler/pythonslides,hach-que/readthedocs.org,stevepiercy/readthedocs.org,hach-que/readthedocs.org,kenwang76/readthedocs.org,nyergler/pythonslides,rtfd/readthedocs.org,sunnyzwh/readthedocs.org,takluyver/readthedocs.org,GovReady/readthedocs.org,safwanrahman/readthedocs.org,safwanrahman/readthedocs.org,wijerasa/readthedocs.org,emawind84/readthedocs.org,mrshoki/readthedocs.org,cgourlay/readthedocs.org,singingwolfboy/readthedocs.org,rtfd/readthedocs.org,sunnyzwh/readthedocs.org,Carreau/readthedocs.org,davidfischer/readthedocs.org,VishvajitP/readthedocs.org,espdev/readthedocs.org,royalwang/readthedocs.org,emawind84/readthedocs.org,mhils/readthedocs.org,sid-kap/readthedocs.org,sid-kap/readthedocs.org,davidfischer/readthedocs.org,tddv/readthedocs.org,LukasBoersma/readthedocs.org,istresearch/readthedocs.org,dirn/readthedocs.org,LukasBoersma/readthedocs.org,CedarLogic/readthedocs.org,kenshinthebattosai/readthedocs.org,Carreau/readthedocs.org,singingwolfboy/readthedocs.org,clarkperkins/readthedocs.org,KamranMackey/readthedocs.org,sils1297/readthedocs.org,espdev/readthedocs.org,gjtorikian/readthedocs.org,titiushko/readthedocs.org,d0ugal/readthedocs.org,kenshinthebattosai/readthedocs.org,attakei/readthedocs-oauth,dirn/readthedocs.org,jerel/readthedocs.org,mrshoki/readthedocs.org,kenwang76/readthedocs.org,sils1297/readthedocs.org,mhils/readthedocs.org,wanghaven/readthedocs.org,davidfischer/readthedocs.org,jerel/readthedocs.org,asampat3090/readthedocs.org,laplaceliu/readthedocs.org,nikolas/readthedocs.org,Carreau/readthedocs.org,johncosta/private-readthedocs.org,tddv/readthedocs.org,michaelmcandrew/readthedocs.org,fujita-shintaro/readthedocs.org,rtfd/readthedocs.org,titiushko/readthedocs.org,SteveViss/readthedocs.org
from django.db import models from django.contrib.auth.models import User STANDARD_EMAIL = "anonymous@readthedocs.org" class UserProfile (models.Model): """Additional information about a User. """ user = models.ForeignKey(User, unique=True, related_name='profile') whitelisted = models.BooleanField() homepage = models.CharField(max_length=100, blank=True) allow_email = models.BooleanField(help_text='Show your email on VCS contributions.', default=True) def get_absolute_url(self): return ('profiles_profile_detail', (), {'username': self.user.username}) get_absolute_url = models.permalink(get_absolute_url) def __unicode__(self): return "%s's profile" % self.user.username def get_contribution_details(self): """ Gets the line to put into commits to attribute the author. Returns a tuple (name, email) """ if self.user.first_name and self.user.last_name: name = '%s %s' % (self.user.first_name, self.user.last_name) else: name = self.user.username if self.allow_email: email = self.user.email else: email = STANDARD_EMAIL return (name, email) Add post save signal for user creation
from django.db import models from django.db.models.signals import post_save from django.dispatch import receiver from django.contrib.auth.models import User STANDARD_EMAIL = "anonymous@readthedocs.org" class UserProfile (models.Model): """Additional information about a User. """ user = models.ForeignKey(User, unique=True, related_name='profile') whitelisted = models.BooleanField() homepage = models.CharField(max_length=100, blank=True) allow_email = models.BooleanField(help_text='Show your email on VCS contributions.', default=True) def get_absolute_url(self): return ('profiles_profile_detail', (), {'username': self.user.username}) get_absolute_url = models.permalink(get_absolute_url) def __unicode__(self): return "%s's profile" % self.user.username def get_contribution_details(self): """ Gets the line to put into commits to attribute the author. Returns a tuple (name, email) """ if self.user.first_name and self.user.last_name: name = '%s %s' % (self.user.first_name, self.user.last_name) else: name = self.user.username if self.allow_email: email = self.user.email else: email = STANDARD_EMAIL return (name, email) @receiver(post_save, sender=User) def create_profile(sender, **kwargs): if kwargs['created'] is True: UserProfile.objects.create(user_id=kwargs['instance'].id, whitelisted=False)
<commit_before>from django.db import models from django.contrib.auth.models import User STANDARD_EMAIL = "anonymous@readthedocs.org" class UserProfile (models.Model): """Additional information about a User. """ user = models.ForeignKey(User, unique=True, related_name='profile') whitelisted = models.BooleanField() homepage = models.CharField(max_length=100, blank=True) allow_email = models.BooleanField(help_text='Show your email on VCS contributions.', default=True) def get_absolute_url(self): return ('profiles_profile_detail', (), {'username': self.user.username}) get_absolute_url = models.permalink(get_absolute_url) def __unicode__(self): return "%s's profile" % self.user.username def get_contribution_details(self): """ Gets the line to put into commits to attribute the author. Returns a tuple (name, email) """ if self.user.first_name and self.user.last_name: name = '%s %s' % (self.user.first_name, self.user.last_name) else: name = self.user.username if self.allow_email: email = self.user.email else: email = STANDARD_EMAIL return (name, email) <commit_msg>Add post save signal for user creation<commit_after>
from django.db import models from django.db.models.signals import post_save from django.dispatch import receiver from django.contrib.auth.models import User STANDARD_EMAIL = "anonymous@readthedocs.org" class UserProfile (models.Model): """Additional information about a User. """ user = models.ForeignKey(User, unique=True, related_name='profile') whitelisted = models.BooleanField() homepage = models.CharField(max_length=100, blank=True) allow_email = models.BooleanField(help_text='Show your email on VCS contributions.', default=True) def get_absolute_url(self): return ('profiles_profile_detail', (), {'username': self.user.username}) get_absolute_url = models.permalink(get_absolute_url) def __unicode__(self): return "%s's profile" % self.user.username def get_contribution_details(self): """ Gets the line to put into commits to attribute the author. Returns a tuple (name, email) """ if self.user.first_name and self.user.last_name: name = '%s %s' % (self.user.first_name, self.user.last_name) else: name = self.user.username if self.allow_email: email = self.user.email else: email = STANDARD_EMAIL return (name, email) @receiver(post_save, sender=User) def create_profile(sender, **kwargs): if kwargs['created'] is True: UserProfile.objects.create(user_id=kwargs['instance'].id, whitelisted=False)
from django.db import models from django.contrib.auth.models import User STANDARD_EMAIL = "anonymous@readthedocs.org" class UserProfile (models.Model): """Additional information about a User. """ user = models.ForeignKey(User, unique=True, related_name='profile') whitelisted = models.BooleanField() homepage = models.CharField(max_length=100, blank=True) allow_email = models.BooleanField(help_text='Show your email on VCS contributions.', default=True) def get_absolute_url(self): return ('profiles_profile_detail', (), {'username': self.user.username}) get_absolute_url = models.permalink(get_absolute_url) def __unicode__(self): return "%s's profile" % self.user.username def get_contribution_details(self): """ Gets the line to put into commits to attribute the author. Returns a tuple (name, email) """ if self.user.first_name and self.user.last_name: name = '%s %s' % (self.user.first_name, self.user.last_name) else: name = self.user.username if self.allow_email: email = self.user.email else: email = STANDARD_EMAIL return (name, email) Add post save signal for user creationfrom django.db import models from django.db.models.signals import post_save from django.dispatch import receiver from django.contrib.auth.models import User STANDARD_EMAIL = "anonymous@readthedocs.org" class UserProfile (models.Model): """Additional information about a User. """ user = models.ForeignKey(User, unique=True, related_name='profile') whitelisted = models.BooleanField() homepage = models.CharField(max_length=100, blank=True) allow_email = models.BooleanField(help_text='Show your email on VCS contributions.', default=True) def get_absolute_url(self): return ('profiles_profile_detail', (), {'username': self.user.username}) get_absolute_url = models.permalink(get_absolute_url) def __unicode__(self): return "%s's profile" % self.user.username def get_contribution_details(self): """ Gets the line to put into commits to attribute the author. Returns a tuple (name, email) """ if self.user.first_name and self.user.last_name: name = '%s %s' % (self.user.first_name, self.user.last_name) else: name = self.user.username if self.allow_email: email = self.user.email else: email = STANDARD_EMAIL return (name, email) @receiver(post_save, sender=User) def create_profile(sender, **kwargs): if kwargs['created'] is True: UserProfile.objects.create(user_id=kwargs['instance'].id, whitelisted=False)
<commit_before>from django.db import models from django.contrib.auth.models import User STANDARD_EMAIL = "anonymous@readthedocs.org" class UserProfile (models.Model): """Additional information about a User. """ user = models.ForeignKey(User, unique=True, related_name='profile') whitelisted = models.BooleanField() homepage = models.CharField(max_length=100, blank=True) allow_email = models.BooleanField(help_text='Show your email on VCS contributions.', default=True) def get_absolute_url(self): return ('profiles_profile_detail', (), {'username': self.user.username}) get_absolute_url = models.permalink(get_absolute_url) def __unicode__(self): return "%s's profile" % self.user.username def get_contribution_details(self): """ Gets the line to put into commits to attribute the author. Returns a tuple (name, email) """ if self.user.first_name and self.user.last_name: name = '%s %s' % (self.user.first_name, self.user.last_name) else: name = self.user.username if self.allow_email: email = self.user.email else: email = STANDARD_EMAIL return (name, email) <commit_msg>Add post save signal for user creation<commit_after>from django.db import models from django.db.models.signals import post_save from django.dispatch import receiver from django.contrib.auth.models import User STANDARD_EMAIL = "anonymous@readthedocs.org" class UserProfile (models.Model): """Additional information about a User. """ user = models.ForeignKey(User, unique=True, related_name='profile') whitelisted = models.BooleanField() homepage = models.CharField(max_length=100, blank=True) allow_email = models.BooleanField(help_text='Show your email on VCS contributions.', default=True) def get_absolute_url(self): return ('profiles_profile_detail', (), {'username': self.user.username}) get_absolute_url = models.permalink(get_absolute_url) def __unicode__(self): return "%s's profile" % self.user.username def get_contribution_details(self): """ Gets the line to put into commits to attribute the author. Returns a tuple (name, email) """ if self.user.first_name and self.user.last_name: name = '%s %s' % (self.user.first_name, self.user.last_name) else: name = self.user.username if self.allow_email: email = self.user.email else: email = STANDARD_EMAIL return (name, email) @receiver(post_save, sender=User) def create_profile(sender, **kwargs): if kwargs['created'] is True: UserProfile.objects.create(user_id=kwargs['instance'].id, whitelisted=False)
bf6565d3bf3c2345a7187d07585bdbf08db06f61
reddit_adzerk/adzerkads.py
reddit_adzerk/adzerkads.py
from urllib import quote from pylons import c, g from r2.lib.pages import Ads as BaseAds from r2.models.subreddit import DefaultSR FRONTPAGE_NAME = "-reddit.com" class Ads(BaseAds): def __init__(self): BaseAds.__init__(self) url_key = "adzerk_https_url" if c.secure else "adzerk_url" site_name = getattr(c.site, "analytics_name", c.site.name) # adzerk reporting is easier when not using a space in the tag if isinstance(c.site, DefaultSR): site_name = FRONTPAGE_NAME self.ad_url = g.config[url_key].format( subreddit=quote(site_name.lower()), origin=c.request_origin, ) self.frame_id = "ad_main"
from urllib import quote from pylons import c, g from r2.lib.pages import Ads as BaseAds from r2.models.subreddit import DefaultSR FRONTPAGE_NAME = "-reddit.com" class Ads(BaseAds): def __init__(self): BaseAds.__init__(self) url_key = "adzerk_https_url" if c.secure else "adzerk_url" site_name = getattr(c.site, "analytics_name", c.site.name) # adzerk reporting is easier when not using a space in the tag if isinstance(c.site, DefaultSR): site_name = FRONTPAGE_NAME self.ad_url = g.config[url_key].format( subreddit=quote(site_name.lower()), origin=c.request_origin, loggedin="loggedin" if c.user_is_loggedin else "loggedout", ) self.frame_id = "ad_main"
Add loggedin keyword to adzerk Ads.
Add loggedin keyword to adzerk Ads.
Python
bsd-3-clause
madbook/reddit-plugin-adzerk,madbook/reddit-plugin-adzerk,madbook/reddit-plugin-adzerk
from urllib import quote from pylons import c, g from r2.lib.pages import Ads as BaseAds from r2.models.subreddit import DefaultSR FRONTPAGE_NAME = "-reddit.com" class Ads(BaseAds): def __init__(self): BaseAds.__init__(self) url_key = "adzerk_https_url" if c.secure else "adzerk_url" site_name = getattr(c.site, "analytics_name", c.site.name) # adzerk reporting is easier when not using a space in the tag if isinstance(c.site, DefaultSR): site_name = FRONTPAGE_NAME self.ad_url = g.config[url_key].format( subreddit=quote(site_name.lower()), origin=c.request_origin, ) self.frame_id = "ad_main" Add loggedin keyword to adzerk Ads.
from urllib import quote from pylons import c, g from r2.lib.pages import Ads as BaseAds from r2.models.subreddit import DefaultSR FRONTPAGE_NAME = "-reddit.com" class Ads(BaseAds): def __init__(self): BaseAds.__init__(self) url_key = "adzerk_https_url" if c.secure else "adzerk_url" site_name = getattr(c.site, "analytics_name", c.site.name) # adzerk reporting is easier when not using a space in the tag if isinstance(c.site, DefaultSR): site_name = FRONTPAGE_NAME self.ad_url = g.config[url_key].format( subreddit=quote(site_name.lower()), origin=c.request_origin, loggedin="loggedin" if c.user_is_loggedin else "loggedout", ) self.frame_id = "ad_main"
<commit_before>from urllib import quote from pylons import c, g from r2.lib.pages import Ads as BaseAds from r2.models.subreddit import DefaultSR FRONTPAGE_NAME = "-reddit.com" class Ads(BaseAds): def __init__(self): BaseAds.__init__(self) url_key = "adzerk_https_url" if c.secure else "adzerk_url" site_name = getattr(c.site, "analytics_name", c.site.name) # adzerk reporting is easier when not using a space in the tag if isinstance(c.site, DefaultSR): site_name = FRONTPAGE_NAME self.ad_url = g.config[url_key].format( subreddit=quote(site_name.lower()), origin=c.request_origin, ) self.frame_id = "ad_main" <commit_msg>Add loggedin keyword to adzerk Ads.<commit_after>
from urllib import quote from pylons import c, g from r2.lib.pages import Ads as BaseAds from r2.models.subreddit import DefaultSR FRONTPAGE_NAME = "-reddit.com" class Ads(BaseAds): def __init__(self): BaseAds.__init__(self) url_key = "adzerk_https_url" if c.secure else "adzerk_url" site_name = getattr(c.site, "analytics_name", c.site.name) # adzerk reporting is easier when not using a space in the tag if isinstance(c.site, DefaultSR): site_name = FRONTPAGE_NAME self.ad_url = g.config[url_key].format( subreddit=quote(site_name.lower()), origin=c.request_origin, loggedin="loggedin" if c.user_is_loggedin else "loggedout", ) self.frame_id = "ad_main"
from urllib import quote from pylons import c, g from r2.lib.pages import Ads as BaseAds from r2.models.subreddit import DefaultSR FRONTPAGE_NAME = "-reddit.com" class Ads(BaseAds): def __init__(self): BaseAds.__init__(self) url_key = "adzerk_https_url" if c.secure else "adzerk_url" site_name = getattr(c.site, "analytics_name", c.site.name) # adzerk reporting is easier when not using a space in the tag if isinstance(c.site, DefaultSR): site_name = FRONTPAGE_NAME self.ad_url = g.config[url_key].format( subreddit=quote(site_name.lower()), origin=c.request_origin, ) self.frame_id = "ad_main" Add loggedin keyword to adzerk Ads.from urllib import quote from pylons import c, g from r2.lib.pages import Ads as BaseAds from r2.models.subreddit import DefaultSR FRONTPAGE_NAME = "-reddit.com" class Ads(BaseAds): def __init__(self): BaseAds.__init__(self) url_key = "adzerk_https_url" if c.secure else "adzerk_url" site_name = getattr(c.site, "analytics_name", c.site.name) # adzerk reporting is easier when not using a space in the tag if isinstance(c.site, DefaultSR): site_name = FRONTPAGE_NAME self.ad_url = g.config[url_key].format( subreddit=quote(site_name.lower()), origin=c.request_origin, loggedin="loggedin" if c.user_is_loggedin else "loggedout", ) self.frame_id = "ad_main"
<commit_before>from urllib import quote from pylons import c, g from r2.lib.pages import Ads as BaseAds from r2.models.subreddit import DefaultSR FRONTPAGE_NAME = "-reddit.com" class Ads(BaseAds): def __init__(self): BaseAds.__init__(self) url_key = "adzerk_https_url" if c.secure else "adzerk_url" site_name = getattr(c.site, "analytics_name", c.site.name) # adzerk reporting is easier when not using a space in the tag if isinstance(c.site, DefaultSR): site_name = FRONTPAGE_NAME self.ad_url = g.config[url_key].format( subreddit=quote(site_name.lower()), origin=c.request_origin, ) self.frame_id = "ad_main" <commit_msg>Add loggedin keyword to adzerk Ads.<commit_after>from urllib import quote from pylons import c, g from r2.lib.pages import Ads as BaseAds from r2.models.subreddit import DefaultSR FRONTPAGE_NAME = "-reddit.com" class Ads(BaseAds): def __init__(self): BaseAds.__init__(self) url_key = "adzerk_https_url" if c.secure else "adzerk_url" site_name = getattr(c.site, "analytics_name", c.site.name) # adzerk reporting is easier when not using a space in the tag if isinstance(c.site, DefaultSR): site_name = FRONTPAGE_NAME self.ad_url = g.config[url_key].format( subreddit=quote(site_name.lower()), origin=c.request_origin, loggedin="loggedin" if c.user_is_loggedin else "loggedout", ) self.frame_id = "ad_main"
fe22a0f9ca92ef0e76bb2f730a2b22da500db5dd
addons/website_sale_management/__manifest__.py
addons/website_sale_management/__manifest__.py
# -*- encoding: utf-8 -*- { 'name': 'Website Sale - Sale Management', 'version': '1.0', 'category': 'Website', 'description': """ Display orders to invoice in website dashboard. """, 'depends': [ 'sale_management', 'website_sale', ], 'installable': True, 'autoinstall': True, 'data': [ ], 'demo': [ ], 'qweb': ['static/src/xml/*.xml'], }
# -*- encoding: utf-8 -*- { 'name': 'Website Sale - Sale Management', 'version': '1.0', 'category': 'Website', 'description': """ Display orders to invoice in website dashboard. """, 'depends': [ 'sale_management', 'website_sale', ], 'installable': True, 'auto_install': True, 'data': [ ], 'demo': [ ], 'qweb': ['static/src/xml/*.xml'], }
Fix auto_install typo in manifest
[FIX] website_sale_management: Fix auto_install typo in manifest
Python
agpl-3.0
ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo
# -*- encoding: utf-8 -*- { 'name': 'Website Sale - Sale Management', 'version': '1.0', 'category': 'Website', 'description': """ Display orders to invoice in website dashboard. """, 'depends': [ 'sale_management', 'website_sale', ], 'installable': True, 'autoinstall': True, 'data': [ ], 'demo': [ ], 'qweb': ['static/src/xml/*.xml'], } [FIX] website_sale_management: Fix auto_install typo in manifest
# -*- encoding: utf-8 -*- { 'name': 'Website Sale - Sale Management', 'version': '1.0', 'category': 'Website', 'description': """ Display orders to invoice in website dashboard. """, 'depends': [ 'sale_management', 'website_sale', ], 'installable': True, 'auto_install': True, 'data': [ ], 'demo': [ ], 'qweb': ['static/src/xml/*.xml'], }
<commit_before># -*- encoding: utf-8 -*- { 'name': 'Website Sale - Sale Management', 'version': '1.0', 'category': 'Website', 'description': """ Display orders to invoice in website dashboard. """, 'depends': [ 'sale_management', 'website_sale', ], 'installable': True, 'autoinstall': True, 'data': [ ], 'demo': [ ], 'qweb': ['static/src/xml/*.xml'], } <commit_msg>[FIX] website_sale_management: Fix auto_install typo in manifest<commit_after>
# -*- encoding: utf-8 -*- { 'name': 'Website Sale - Sale Management', 'version': '1.0', 'category': 'Website', 'description': """ Display orders to invoice in website dashboard. """, 'depends': [ 'sale_management', 'website_sale', ], 'installable': True, 'auto_install': True, 'data': [ ], 'demo': [ ], 'qweb': ['static/src/xml/*.xml'], }
# -*- encoding: utf-8 -*- { 'name': 'Website Sale - Sale Management', 'version': '1.0', 'category': 'Website', 'description': """ Display orders to invoice in website dashboard. """, 'depends': [ 'sale_management', 'website_sale', ], 'installable': True, 'autoinstall': True, 'data': [ ], 'demo': [ ], 'qweb': ['static/src/xml/*.xml'], } [FIX] website_sale_management: Fix auto_install typo in manifest# -*- encoding: utf-8 -*- { 'name': 'Website Sale - Sale Management', 'version': '1.0', 'category': 'Website', 'description': """ Display orders to invoice in website dashboard. """, 'depends': [ 'sale_management', 'website_sale', ], 'installable': True, 'auto_install': True, 'data': [ ], 'demo': [ ], 'qweb': ['static/src/xml/*.xml'], }
<commit_before># -*- encoding: utf-8 -*- { 'name': 'Website Sale - Sale Management', 'version': '1.0', 'category': 'Website', 'description': """ Display orders to invoice in website dashboard. """, 'depends': [ 'sale_management', 'website_sale', ], 'installable': True, 'autoinstall': True, 'data': [ ], 'demo': [ ], 'qweb': ['static/src/xml/*.xml'], } <commit_msg>[FIX] website_sale_management: Fix auto_install typo in manifest<commit_after># -*- encoding: utf-8 -*- { 'name': 'Website Sale - Sale Management', 'version': '1.0', 'category': 'Website', 'description': """ Display orders to invoice in website dashboard. """, 'depends': [ 'sale_management', 'website_sale', ], 'installable': True, 'auto_install': True, 'data': [ ], 'demo': [ ], 'qweb': ['static/src/xml/*.xml'], }
ad17f4a48bdee7ca57e8fe66e4658821b3c8789e
corehq/apps/userreports/migrations/0018_ucrexpression.py
corehq/apps/userreports/migrations/0018_ucrexpression.py
# Generated by Django 2.2.27 on 2022-04-12 10:28 import django.contrib.postgres.fields.jsonb from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('userreports', '0017_index_cleanup'), ] operations = [ migrations.CreateModel( name='UCRExpression', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), ('domain', models.CharField(db_index=True, max_length=255)), ('description', models.TextField(blank=True, null=True)), ('expression_type', models.CharField(choices=[('named_expression', 'named_expression'), ('named_filter', 'named_filter')], db_index=True, default='named_expression', max_length=20)), ('definition', django.contrib.postgres.fields.jsonb.JSONField(null=True)), ], options={ 'unique_together': {('name', 'domain')}, }, ), ]
# Generated by Django 3.2.12 on 2022-04-12 15:07 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('userreports', '0017_index_cleanup'), ] operations = [ migrations.CreateModel( name='UCRExpression', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), ('domain', models.CharField(db_index=True, max_length=255)), ('description', models.TextField(blank=True, null=True)), ('expression_type', models.CharField(choices=[('named_expression', 'named_expression'), ('named_filter', 'named_filter')], db_index=True, default='named_expression', max_length=20)), ('definition', models.JSONField(null=True)), ], options={ 'unique_together': {('name', 'domain')}, }, ), ]
Use django 3 to make migration
Use django 3 to make migration
Python
bsd-3-clause
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
# Generated by Django 2.2.27 on 2022-04-12 10:28 import django.contrib.postgres.fields.jsonb from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('userreports', '0017_index_cleanup'), ] operations = [ migrations.CreateModel( name='UCRExpression', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), ('domain', models.CharField(db_index=True, max_length=255)), ('description', models.TextField(blank=True, null=True)), ('expression_type', models.CharField(choices=[('named_expression', 'named_expression'), ('named_filter', 'named_filter')], db_index=True, default='named_expression', max_length=20)), ('definition', django.contrib.postgres.fields.jsonb.JSONField(null=True)), ], options={ 'unique_together': {('name', 'domain')}, }, ), ] Use django 3 to make migration
# Generated by Django 3.2.12 on 2022-04-12 15:07 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('userreports', '0017_index_cleanup'), ] operations = [ migrations.CreateModel( name='UCRExpression', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), ('domain', models.CharField(db_index=True, max_length=255)), ('description', models.TextField(blank=True, null=True)), ('expression_type', models.CharField(choices=[('named_expression', 'named_expression'), ('named_filter', 'named_filter')], db_index=True, default='named_expression', max_length=20)), ('definition', models.JSONField(null=True)), ], options={ 'unique_together': {('name', 'domain')}, }, ), ]
<commit_before># Generated by Django 2.2.27 on 2022-04-12 10:28 import django.contrib.postgres.fields.jsonb from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('userreports', '0017_index_cleanup'), ] operations = [ migrations.CreateModel( name='UCRExpression', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), ('domain', models.CharField(db_index=True, max_length=255)), ('description', models.TextField(blank=True, null=True)), ('expression_type', models.CharField(choices=[('named_expression', 'named_expression'), ('named_filter', 'named_filter')], db_index=True, default='named_expression', max_length=20)), ('definition', django.contrib.postgres.fields.jsonb.JSONField(null=True)), ], options={ 'unique_together': {('name', 'domain')}, }, ), ] <commit_msg>Use django 3 to make migration<commit_after>
# Generated by Django 3.2.12 on 2022-04-12 15:07 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('userreports', '0017_index_cleanup'), ] operations = [ migrations.CreateModel( name='UCRExpression', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), ('domain', models.CharField(db_index=True, max_length=255)), ('description', models.TextField(blank=True, null=True)), ('expression_type', models.CharField(choices=[('named_expression', 'named_expression'), ('named_filter', 'named_filter')], db_index=True, default='named_expression', max_length=20)), ('definition', models.JSONField(null=True)), ], options={ 'unique_together': {('name', 'domain')}, }, ), ]
# Generated by Django 2.2.27 on 2022-04-12 10:28 import django.contrib.postgres.fields.jsonb from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('userreports', '0017_index_cleanup'), ] operations = [ migrations.CreateModel( name='UCRExpression', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), ('domain', models.CharField(db_index=True, max_length=255)), ('description', models.TextField(blank=True, null=True)), ('expression_type', models.CharField(choices=[('named_expression', 'named_expression'), ('named_filter', 'named_filter')], db_index=True, default='named_expression', max_length=20)), ('definition', django.contrib.postgres.fields.jsonb.JSONField(null=True)), ], options={ 'unique_together': {('name', 'domain')}, }, ), ] Use django 3 to make migration# Generated by Django 3.2.12 on 2022-04-12 15:07 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('userreports', '0017_index_cleanup'), ] operations = [ migrations.CreateModel( name='UCRExpression', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), ('domain', models.CharField(db_index=True, max_length=255)), ('description', models.TextField(blank=True, null=True)), ('expression_type', models.CharField(choices=[('named_expression', 'named_expression'), ('named_filter', 'named_filter')], db_index=True, default='named_expression', max_length=20)), ('definition', models.JSONField(null=True)), ], options={ 'unique_together': {('name', 'domain')}, }, ), ]
<commit_before># Generated by Django 2.2.27 on 2022-04-12 10:28 import django.contrib.postgres.fields.jsonb from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('userreports', '0017_index_cleanup'), ] operations = [ migrations.CreateModel( name='UCRExpression', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), ('domain', models.CharField(db_index=True, max_length=255)), ('description', models.TextField(blank=True, null=True)), ('expression_type', models.CharField(choices=[('named_expression', 'named_expression'), ('named_filter', 'named_filter')], db_index=True, default='named_expression', max_length=20)), ('definition', django.contrib.postgres.fields.jsonb.JSONField(null=True)), ], options={ 'unique_together': {('name', 'domain')}, }, ), ] <commit_msg>Use django 3 to make migration<commit_after># Generated by Django 3.2.12 on 2022-04-12 15:07 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('userreports', '0017_index_cleanup'), ] operations = [ migrations.CreateModel( name='UCRExpression', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), ('domain', models.CharField(db_index=True, max_length=255)), ('description', models.TextField(blank=True, null=True)), ('expression_type', models.CharField(choices=[('named_expression', 'named_expression'), ('named_filter', 'named_filter')], db_index=True, default='named_expression', max_length=20)), ('definition', models.JSONField(null=True)), ], options={ 'unique_together': {('name', 'domain')}, }, ), ]
7ee5692a98a6dfc714a05f1add8e72b09c52929e
students/psbriant/final_project/clean_data.py
students/psbriant/final_project/clean_data.py
""" Name: Paul Briant Date: 12/11/16 Class: Introduction to Python Assignment: Final Project Description: Code for Final Project """ import pandas from datetime import datetime # Change source to smaller file. data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv") # print(data["Date Text"].head()) first_date = data["Date Text"].values[0] print(first_date) # datetime.strptime(first_date, "%Y-%m-%d") # datetime(2012, 3, 10, 0, 0) # data.date = data.date.apply(lambda d: datetime.strptime(d, "%Y-%m-%d")) # print(data.date.head()) # data.index = data.date # print(data) # print(data.ix[datetime(2012, 8, 19)]) # Remove date column # data = data.drop(["date"], axis=1) # print(data.columns) # Determine what values are missing # empty = data.apply(lambda col: pandas.isnull(col))
""" Name: Paul Briant Date: 12/11/16 Class: Introduction to Python Assignment: Final Project Description: Code for Final Project """ import pandas from datetime import datetime # Change source to smaller file. data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv", parse_dates=[0], infer_datetime_format=True) temp = pandas.DatetimeIndex(data["Date_Text"]) data["Month"] = temp.month data["Year"] = temp.year print(data) # print(data["Date Text"].head()) # first_date = data["Date Text"].values[0] # print(first_date) # datetime.strptime(first_date, "%b-%Y") # datetime(2012, 3, 10, 0, 0) # data.date = data["Date Text"].apply(lambda d: datetime.strptime(d, "%b-%Y")) # print(data.date.head()) # data.index = data.date # print(data) # print(data.ix[datetime(2012, 8, 19)]) # Remove date column # data = data.drop(["date"], axis=1) # print(data.columns) # Determine what values are missing # empty = data.apply(lambda col: pandas.isnull(col))
Add code to split month and year into new columns.
Add code to split month and year into new columns.
Python
unlicense
UWPCE-PythonCert/IntroPython2016,weidnem/IntroPython2016,UWPCE-PythonCert/IntroPython2016,weidnem/IntroPython2016,weidnem/IntroPython2016,UWPCE-PythonCert/IntroPython2016
""" Name: Paul Briant Date: 12/11/16 Class: Introduction to Python Assignment: Final Project Description: Code for Final Project """ import pandas from datetime import datetime # Change source to smaller file. data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv") # print(data["Date Text"].head()) first_date = data["Date Text"].values[0] print(first_date) # datetime.strptime(first_date, "%Y-%m-%d") # datetime(2012, 3, 10, 0, 0) # data.date = data.date.apply(lambda d: datetime.strptime(d, "%Y-%m-%d")) # print(data.date.head()) # data.index = data.date # print(data) # print(data.ix[datetime(2012, 8, 19)]) # Remove date column # data = data.drop(["date"], axis=1) # print(data.columns) # Determine what values are missing # empty = data.apply(lambda col: pandas.isnull(col)) Add code to split month and year into new columns.
""" Name: Paul Briant Date: 12/11/16 Class: Introduction to Python Assignment: Final Project Description: Code for Final Project """ import pandas from datetime import datetime # Change source to smaller file. data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv", parse_dates=[0], infer_datetime_format=True) temp = pandas.DatetimeIndex(data["Date_Text"]) data["Month"] = temp.month data["Year"] = temp.year print(data) # print(data["Date Text"].head()) # first_date = data["Date Text"].values[0] # print(first_date) # datetime.strptime(first_date, "%b-%Y") # datetime(2012, 3, 10, 0, 0) # data.date = data["Date Text"].apply(lambda d: datetime.strptime(d, "%b-%Y")) # print(data.date.head()) # data.index = data.date # print(data) # print(data.ix[datetime(2012, 8, 19)]) # Remove date column # data = data.drop(["date"], axis=1) # print(data.columns) # Determine what values are missing # empty = data.apply(lambda col: pandas.isnull(col))
<commit_before>""" Name: Paul Briant Date: 12/11/16 Class: Introduction to Python Assignment: Final Project Description: Code for Final Project """ import pandas from datetime import datetime # Change source to smaller file. data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv") # print(data["Date Text"].head()) first_date = data["Date Text"].values[0] print(first_date) # datetime.strptime(first_date, "%Y-%m-%d") # datetime(2012, 3, 10, 0, 0) # data.date = data.date.apply(lambda d: datetime.strptime(d, "%Y-%m-%d")) # print(data.date.head()) # data.index = data.date # print(data) # print(data.ix[datetime(2012, 8, 19)]) # Remove date column # data = data.drop(["date"], axis=1) # print(data.columns) # Determine what values are missing # empty = data.apply(lambda col: pandas.isnull(col)) <commit_msg>Add code to split month and year into new columns.<commit_after>
""" Name: Paul Briant Date: 12/11/16 Class: Introduction to Python Assignment: Final Project Description: Code for Final Project """ import pandas from datetime import datetime # Change source to smaller file. data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv", parse_dates=[0], infer_datetime_format=True) temp = pandas.DatetimeIndex(data["Date_Text"]) data["Month"] = temp.month data["Year"] = temp.year print(data) # print(data["Date Text"].head()) # first_date = data["Date Text"].values[0] # print(first_date) # datetime.strptime(first_date, "%b-%Y") # datetime(2012, 3, 10, 0, 0) # data.date = data["Date Text"].apply(lambda d: datetime.strptime(d, "%b-%Y")) # print(data.date.head()) # data.index = data.date # print(data) # print(data.ix[datetime(2012, 8, 19)]) # Remove date column # data = data.drop(["date"], axis=1) # print(data.columns) # Determine what values are missing # empty = data.apply(lambda col: pandas.isnull(col))
""" Name: Paul Briant Date: 12/11/16 Class: Introduction to Python Assignment: Final Project Description: Code for Final Project """ import pandas from datetime import datetime # Change source to smaller file. data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv") # print(data["Date Text"].head()) first_date = data["Date Text"].values[0] print(first_date) # datetime.strptime(first_date, "%Y-%m-%d") # datetime(2012, 3, 10, 0, 0) # data.date = data.date.apply(lambda d: datetime.strptime(d, "%Y-%m-%d")) # print(data.date.head()) # data.index = data.date # print(data) # print(data.ix[datetime(2012, 8, 19)]) # Remove date column # data = data.drop(["date"], axis=1) # print(data.columns) # Determine what values are missing # empty = data.apply(lambda col: pandas.isnull(col)) Add code to split month and year into new columns.""" Name: Paul Briant Date: 12/11/16 Class: Introduction to Python Assignment: Final Project Description: Code for Final Project """ import pandas from datetime import datetime # Change source to smaller file. data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv", parse_dates=[0], infer_datetime_format=True) temp = pandas.DatetimeIndex(data["Date_Text"]) data["Month"] = temp.month data["Year"] = temp.year print(data) # print(data["Date Text"].head()) # first_date = data["Date Text"].values[0] # print(first_date) # datetime.strptime(first_date, "%b-%Y") # datetime(2012, 3, 10, 0, 0) # data.date = data["Date Text"].apply(lambda d: datetime.strptime(d, "%b-%Y")) # print(data.date.head()) # data.index = data.date # print(data) # print(data.ix[datetime(2012, 8, 19)]) # Remove date column # data = data.drop(["date"], axis=1) # print(data.columns) # Determine what values are missing # empty = data.apply(lambda col: pandas.isnull(col))
<commit_before>""" Name: Paul Briant Date: 12/11/16 Class: Introduction to Python Assignment: Final Project Description: Code for Final Project """ import pandas from datetime import datetime # Change source to smaller file. data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv") # print(data["Date Text"].head()) first_date = data["Date Text"].values[0] print(first_date) # datetime.strptime(first_date, "%Y-%m-%d") # datetime(2012, 3, 10, 0, 0) # data.date = data.date.apply(lambda d: datetime.strptime(d, "%Y-%m-%d")) # print(data.date.head()) # data.index = data.date # print(data) # print(data.ix[datetime(2012, 8, 19)]) # Remove date column # data = data.drop(["date"], axis=1) # print(data.columns) # Determine what values are missing # empty = data.apply(lambda col: pandas.isnull(col)) <commit_msg>Add code to split month and year into new columns.<commit_after>""" Name: Paul Briant Date: 12/11/16 Class: Introduction to Python Assignment: Final Project Description: Code for Final Project """ import pandas from datetime import datetime # Change source to smaller file. data = pandas.read_csv("data/Residential_Water_Usage_Zip_Code_on_Top.csv", parse_dates=[0], infer_datetime_format=True) temp = pandas.DatetimeIndex(data["Date_Text"]) data["Month"] = temp.month data["Year"] = temp.year print(data) # print(data["Date Text"].head()) # first_date = data["Date Text"].values[0] # print(first_date) # datetime.strptime(first_date, "%b-%Y") # datetime(2012, 3, 10, 0, 0) # data.date = data["Date Text"].apply(lambda d: datetime.strptime(d, "%b-%Y")) # print(data.date.head()) # data.index = data.date # print(data) # print(data.ix[datetime(2012, 8, 19)]) # Remove date column # data = data.drop(["date"], axis=1) # print(data.columns) # Determine what values are missing # empty = data.apply(lambda col: pandas.isnull(col))
f334c9770dd7d9fb38fd5c118df806dbf532d17a
wsgi.py
wsgi.py
# -*- coding: utf-8 -*- """ wsgi ~~~~ ieee wsgi module """ from werkzeug.serving import run_simple from werkzeug.wsgi import DispatcherMiddleware from myip.factory import create_app if __name__ == "__main__": run_simple('localhost', 5000, create_app(), use_reloader=True, use_debugger=True, use_evalex=True)
# -*- coding: utf-8 -*- """ wsgi ~~~~ ieee wsgi module """ from werkzeug.serving import run_simple from werkzeug.wsgi import DispatcherMiddleware from myip.factory import create_app application = create_app() if __name__ == "__main__": run_simple('localhost', 5000, application, use_reloader=True, use_debugger=True, use_evalex=True)
Add application variable for the procfile.
Add application variable for the procfile.
Python
mit
brotatos/myip,brotatos/myip
# -*- coding: utf-8 -*- """ wsgi ~~~~ ieee wsgi module """ from werkzeug.serving import run_simple from werkzeug.wsgi import DispatcherMiddleware from myip.factory import create_app if __name__ == "__main__": run_simple('localhost', 5000, create_app(), use_reloader=True, use_debugger=True, use_evalex=True) Add application variable for the procfile.
# -*- coding: utf-8 -*- """ wsgi ~~~~ ieee wsgi module """ from werkzeug.serving import run_simple from werkzeug.wsgi import DispatcherMiddleware from myip.factory import create_app application = create_app() if __name__ == "__main__": run_simple('localhost', 5000, application, use_reloader=True, use_debugger=True, use_evalex=True)
<commit_before># -*- coding: utf-8 -*- """ wsgi ~~~~ ieee wsgi module """ from werkzeug.serving import run_simple from werkzeug.wsgi import DispatcherMiddleware from myip.factory import create_app if __name__ == "__main__": run_simple('localhost', 5000, create_app(), use_reloader=True, use_debugger=True, use_evalex=True) <commit_msg>Add application variable for the procfile.<commit_after>
# -*- coding: utf-8 -*- """ wsgi ~~~~ ieee wsgi module """ from werkzeug.serving import run_simple from werkzeug.wsgi import DispatcherMiddleware from myip.factory import create_app application = create_app() if __name__ == "__main__": run_simple('localhost', 5000, application, use_reloader=True, use_debugger=True, use_evalex=True)
# -*- coding: utf-8 -*- """ wsgi ~~~~ ieee wsgi module """ from werkzeug.serving import run_simple from werkzeug.wsgi import DispatcherMiddleware from myip.factory import create_app if __name__ == "__main__": run_simple('localhost', 5000, create_app(), use_reloader=True, use_debugger=True, use_evalex=True) Add application variable for the procfile.# -*- coding: utf-8 -*- """ wsgi ~~~~ ieee wsgi module """ from werkzeug.serving import run_simple from werkzeug.wsgi import DispatcherMiddleware from myip.factory import create_app application = create_app() if __name__ == "__main__": run_simple('localhost', 5000, application, use_reloader=True, use_debugger=True, use_evalex=True)
<commit_before># -*- coding: utf-8 -*- """ wsgi ~~~~ ieee wsgi module """ from werkzeug.serving import run_simple from werkzeug.wsgi import DispatcherMiddleware from myip.factory import create_app if __name__ == "__main__": run_simple('localhost', 5000, create_app(), use_reloader=True, use_debugger=True, use_evalex=True) <commit_msg>Add application variable for the procfile.<commit_after># -*- coding: utf-8 -*- """ wsgi ~~~~ ieee wsgi module """ from werkzeug.serving import run_simple from werkzeug.wsgi import DispatcherMiddleware from myip.factory import create_app application = create_app() if __name__ == "__main__": run_simple('localhost', 5000, application, use_reloader=True, use_debugger=True, use_evalex=True)
d52a7b19f7b5596e88d7233dfea35a70b2645385
osmaxx-py/excerptconverter/converter_manager.py
osmaxx-py/excerptconverter/converter_manager.py
from excerptconverter.baseexcerptconverter import BaseExcerptConverter class ConverterManager: @staticmethod def converter_configuration(): export_options = {} for Converter in BaseExcerptConverter.available_converters: export_options[Converter.__name__] = Converter.converter_configuration() return export_options def __init__(self, extraction_order, available_converters=BaseExcerptConverter.available_converters, run_as_celery_tasks=True): """" :param execution_configuration example: { 'gis': { 'formats': ['txt', 'file_gdb'], 'options': { 'coordinate_reference_system': 'wgs72', 'detail_level': 'verbatim' } }, 'routing': { ... } } """ self.extraction_order = extraction_order self.available_converters = available_converters self.run_as_celery_tasks = run_as_celery_tasks def execute_converters(self): for Converter in self.available_converters: if Converter.__name__ in self.extraction_order.extraction_configuration: Converter.execute( self.extraction_order, self.extraction_order.extraction_configuration[Converter.__name__], self.run_as_celery_tasks )
from excerptconverter.baseexcerptconverter import BaseExcerptConverter class ConverterManager: @staticmethod def converter_configuration(): return {Converter.__name__: Converter.converter_configuration() for Converter in BaseExcerptConverter.available_converters} def __init__(self, extraction_order, available_converters=BaseExcerptConverter.available_converters, run_as_celery_tasks=True): """" :param execution_configuration example: { 'gis': { 'formats': ['txt', 'file_gdb'], 'options': { 'coordinate_reference_system': 'wgs72', 'detail_level': 'verbatim' } }, 'routing': { ... } } """ self.extraction_order = extraction_order self.available_converters = available_converters self.run_as_celery_tasks = run_as_celery_tasks def execute_converters(self): for Converter in self.available_converters: if Converter.__name__ in self.extraction_order.extraction_configuration: Converter.execute( self.extraction_order, self.extraction_order.extraction_configuration[Converter.__name__], self.run_as_celery_tasks )
Replace loop by dictionary comprehension
Refactoring: Replace loop by dictionary comprehension
Python
mit
geometalab/osmaxx,geometalab/drf-utm-zone-info,geometalab/drf-utm-zone-info,geometalab/osmaxx-frontend,geometalab/osmaxx-frontend,geometalab/osmaxx-frontend,geometalab/osmaxx,geometalab/osmaxx,geometalab/osmaxx,geometalab/osmaxx-frontend
from excerptconverter.baseexcerptconverter import BaseExcerptConverter class ConverterManager: @staticmethod def converter_configuration(): export_options = {} for Converter in BaseExcerptConverter.available_converters: export_options[Converter.__name__] = Converter.converter_configuration() return export_options def __init__(self, extraction_order, available_converters=BaseExcerptConverter.available_converters, run_as_celery_tasks=True): """" :param execution_configuration example: { 'gis': { 'formats': ['txt', 'file_gdb'], 'options': { 'coordinate_reference_system': 'wgs72', 'detail_level': 'verbatim' } }, 'routing': { ... } } """ self.extraction_order = extraction_order self.available_converters = available_converters self.run_as_celery_tasks = run_as_celery_tasks def execute_converters(self): for Converter in self.available_converters: if Converter.__name__ in self.extraction_order.extraction_configuration: Converter.execute( self.extraction_order, self.extraction_order.extraction_configuration[Converter.__name__], self.run_as_celery_tasks ) Refactoring: Replace loop by dictionary comprehension
from excerptconverter.baseexcerptconverter import BaseExcerptConverter class ConverterManager: @staticmethod def converter_configuration(): return {Converter.__name__: Converter.converter_configuration() for Converter in BaseExcerptConverter.available_converters} def __init__(self, extraction_order, available_converters=BaseExcerptConverter.available_converters, run_as_celery_tasks=True): """" :param execution_configuration example: { 'gis': { 'formats': ['txt', 'file_gdb'], 'options': { 'coordinate_reference_system': 'wgs72', 'detail_level': 'verbatim' } }, 'routing': { ... } } """ self.extraction_order = extraction_order self.available_converters = available_converters self.run_as_celery_tasks = run_as_celery_tasks def execute_converters(self): for Converter in self.available_converters: if Converter.__name__ in self.extraction_order.extraction_configuration: Converter.execute( self.extraction_order, self.extraction_order.extraction_configuration[Converter.__name__], self.run_as_celery_tasks )
<commit_before>from excerptconverter.baseexcerptconverter import BaseExcerptConverter class ConverterManager: @staticmethod def converter_configuration(): export_options = {} for Converter in BaseExcerptConverter.available_converters: export_options[Converter.__name__] = Converter.converter_configuration() return export_options def __init__(self, extraction_order, available_converters=BaseExcerptConverter.available_converters, run_as_celery_tasks=True): """" :param execution_configuration example: { 'gis': { 'formats': ['txt', 'file_gdb'], 'options': { 'coordinate_reference_system': 'wgs72', 'detail_level': 'verbatim' } }, 'routing': { ... } } """ self.extraction_order = extraction_order self.available_converters = available_converters self.run_as_celery_tasks = run_as_celery_tasks def execute_converters(self): for Converter in self.available_converters: if Converter.__name__ in self.extraction_order.extraction_configuration: Converter.execute( self.extraction_order, self.extraction_order.extraction_configuration[Converter.__name__], self.run_as_celery_tasks ) <commit_msg>Refactoring: Replace loop by dictionary comprehension<commit_after>
from excerptconverter.baseexcerptconverter import BaseExcerptConverter class ConverterManager: @staticmethod def converter_configuration(): return {Converter.__name__: Converter.converter_configuration() for Converter in BaseExcerptConverter.available_converters} def __init__(self, extraction_order, available_converters=BaseExcerptConverter.available_converters, run_as_celery_tasks=True): """" :param execution_configuration example: { 'gis': { 'formats': ['txt', 'file_gdb'], 'options': { 'coordinate_reference_system': 'wgs72', 'detail_level': 'verbatim' } }, 'routing': { ... } } """ self.extraction_order = extraction_order self.available_converters = available_converters self.run_as_celery_tasks = run_as_celery_tasks def execute_converters(self): for Converter in self.available_converters: if Converter.__name__ in self.extraction_order.extraction_configuration: Converter.execute( self.extraction_order, self.extraction_order.extraction_configuration[Converter.__name__], self.run_as_celery_tasks )
from excerptconverter.baseexcerptconverter import BaseExcerptConverter class ConverterManager: @staticmethod def converter_configuration(): export_options = {} for Converter in BaseExcerptConverter.available_converters: export_options[Converter.__name__] = Converter.converter_configuration() return export_options def __init__(self, extraction_order, available_converters=BaseExcerptConverter.available_converters, run_as_celery_tasks=True): """" :param execution_configuration example: { 'gis': { 'formats': ['txt', 'file_gdb'], 'options': { 'coordinate_reference_system': 'wgs72', 'detail_level': 'verbatim' } }, 'routing': { ... } } """ self.extraction_order = extraction_order self.available_converters = available_converters self.run_as_celery_tasks = run_as_celery_tasks def execute_converters(self): for Converter in self.available_converters: if Converter.__name__ in self.extraction_order.extraction_configuration: Converter.execute( self.extraction_order, self.extraction_order.extraction_configuration[Converter.__name__], self.run_as_celery_tasks ) Refactoring: Replace loop by dictionary comprehensionfrom excerptconverter.baseexcerptconverter import BaseExcerptConverter class ConverterManager: @staticmethod def converter_configuration(): return {Converter.__name__: Converter.converter_configuration() for Converter in BaseExcerptConverter.available_converters} def __init__(self, extraction_order, available_converters=BaseExcerptConverter.available_converters, run_as_celery_tasks=True): """" :param execution_configuration example: { 'gis': { 'formats': ['txt', 'file_gdb'], 'options': { 'coordinate_reference_system': 'wgs72', 'detail_level': 'verbatim' } }, 'routing': { ... } } """ self.extraction_order = extraction_order self.available_converters = available_converters self.run_as_celery_tasks = run_as_celery_tasks def execute_converters(self): for Converter in self.available_converters: if Converter.__name__ in self.extraction_order.extraction_configuration: Converter.execute( self.extraction_order, self.extraction_order.extraction_configuration[Converter.__name__], self.run_as_celery_tasks )
<commit_before>from excerptconverter.baseexcerptconverter import BaseExcerptConverter class ConverterManager: @staticmethod def converter_configuration(): export_options = {} for Converter in BaseExcerptConverter.available_converters: export_options[Converter.__name__] = Converter.converter_configuration() return export_options def __init__(self, extraction_order, available_converters=BaseExcerptConverter.available_converters, run_as_celery_tasks=True): """" :param execution_configuration example: { 'gis': { 'formats': ['txt', 'file_gdb'], 'options': { 'coordinate_reference_system': 'wgs72', 'detail_level': 'verbatim' } }, 'routing': { ... } } """ self.extraction_order = extraction_order self.available_converters = available_converters self.run_as_celery_tasks = run_as_celery_tasks def execute_converters(self): for Converter in self.available_converters: if Converter.__name__ in self.extraction_order.extraction_configuration: Converter.execute( self.extraction_order, self.extraction_order.extraction_configuration[Converter.__name__], self.run_as_celery_tasks ) <commit_msg>Refactoring: Replace loop by dictionary comprehension<commit_after>from excerptconverter.baseexcerptconverter import BaseExcerptConverter class ConverterManager: @staticmethod def converter_configuration(): return {Converter.__name__: Converter.converter_configuration() for Converter in BaseExcerptConverter.available_converters} def __init__(self, extraction_order, available_converters=BaseExcerptConverter.available_converters, run_as_celery_tasks=True): """" :param execution_configuration example: { 'gis': { 'formats': ['txt', 'file_gdb'], 'options': { 'coordinate_reference_system': 'wgs72', 'detail_level': 'verbatim' } }, 'routing': { ... } } """ self.extraction_order = extraction_order self.available_converters = available_converters self.run_as_celery_tasks = run_as_celery_tasks def execute_converters(self): for Converter in self.available_converters: if Converter.__name__ in self.extraction_order.extraction_configuration: Converter.execute( self.extraction_order, self.extraction_order.extraction_configuration[Converter.__name__], self.run_as_celery_tasks )
d9da661c9493c0d1b54eeda8bb416ad98ca07a33
acapi/resources/environmentlist.py
acapi/resources/environmentlist.py
""" Acquia Cloud API server list resource. """ from .acquialist import AcquiaList class EnvironmentList(AcquiaList): """Dict of Acquia Cloud API Server resources keyed by hostname.""" def set_base_uri(self, base_uri): """ Set the base URI for server resources. Parameters ---------- base_uri : str The base URI to use for generating the new URI. """ uri = '{}/servers'.format(base_uri) self.uri = uri
""" Acquia Cloud API server list resource. """ from .acquialist import AcquiaList class EnvironmentList(AcquiaList): """Dict of Acquia Cloud API Server resources keyed by hostname.""" def set_base_uri(self, base_uri): """ Set the base URI for server resources. Parameters ---------- base_uri : str The base URI to use for generating the new URI. """ uri = '{}/envs'.format(base_uri) self.uri = uri
Use correct URI for looking up environment resources
Use correct URI for looking up environment resources
Python
mit
skwashd/python-acquia-cloud
""" Acquia Cloud API server list resource. """ from .acquialist import AcquiaList class EnvironmentList(AcquiaList): """Dict of Acquia Cloud API Server resources keyed by hostname.""" def set_base_uri(self, base_uri): """ Set the base URI for server resources. Parameters ---------- base_uri : str The base URI to use for generating the new URI. """ uri = '{}/servers'.format(base_uri) self.uri = uri Use correct URI for looking up environment resources
""" Acquia Cloud API server list resource. """ from .acquialist import AcquiaList class EnvironmentList(AcquiaList): """Dict of Acquia Cloud API Server resources keyed by hostname.""" def set_base_uri(self, base_uri): """ Set the base URI for server resources. Parameters ---------- base_uri : str The base URI to use for generating the new URI. """ uri = '{}/envs'.format(base_uri) self.uri = uri
<commit_before>""" Acquia Cloud API server list resource. """ from .acquialist import AcquiaList class EnvironmentList(AcquiaList): """Dict of Acquia Cloud API Server resources keyed by hostname.""" def set_base_uri(self, base_uri): """ Set the base URI for server resources. Parameters ---------- base_uri : str The base URI to use for generating the new URI. """ uri = '{}/servers'.format(base_uri) self.uri = uri <commit_msg>Use correct URI for looking up environment resources<commit_after>
""" Acquia Cloud API server list resource. """ from .acquialist import AcquiaList class EnvironmentList(AcquiaList): """Dict of Acquia Cloud API Server resources keyed by hostname.""" def set_base_uri(self, base_uri): """ Set the base URI for server resources. Parameters ---------- base_uri : str The base URI to use for generating the new URI. """ uri = '{}/envs'.format(base_uri) self.uri = uri
""" Acquia Cloud API server list resource. """ from .acquialist import AcquiaList class EnvironmentList(AcquiaList): """Dict of Acquia Cloud API Server resources keyed by hostname.""" def set_base_uri(self, base_uri): """ Set the base URI for server resources. Parameters ---------- base_uri : str The base URI to use for generating the new URI. """ uri = '{}/servers'.format(base_uri) self.uri = uri Use correct URI for looking up environment resources""" Acquia Cloud API server list resource. """ from .acquialist import AcquiaList class EnvironmentList(AcquiaList): """Dict of Acquia Cloud API Server resources keyed by hostname.""" def set_base_uri(self, base_uri): """ Set the base URI for server resources. Parameters ---------- base_uri : str The base URI to use for generating the new URI. """ uri = '{}/envs'.format(base_uri) self.uri = uri
<commit_before>""" Acquia Cloud API server list resource. """ from .acquialist import AcquiaList class EnvironmentList(AcquiaList): """Dict of Acquia Cloud API Server resources keyed by hostname.""" def set_base_uri(self, base_uri): """ Set the base URI for server resources. Parameters ---------- base_uri : str The base URI to use for generating the new URI. """ uri = '{}/servers'.format(base_uri) self.uri = uri <commit_msg>Use correct URI for looking up environment resources<commit_after>""" Acquia Cloud API server list resource. """ from .acquialist import AcquiaList class EnvironmentList(AcquiaList): """Dict of Acquia Cloud API Server resources keyed by hostname.""" def set_base_uri(self, base_uri): """ Set the base URI for server resources. Parameters ---------- base_uri : str The base URI to use for generating the new URI. """ uri = '{}/envs'.format(base_uri) self.uri = uri
f5d864c2a5c9b4d2ea1ff95e59d60adf2ebd176e
recipes/sos-bash/run_test.py
recipes/sos-bash/run_test.py
import unittest import sys from sos_notebook.test_utils import sos_kernel from ipykernel.tests.utils import execute, wait_for_idle, assemble_output @unittest.skipIf(sys.platform == 'win32', 'bash does not exist on win32') class TestSoSKernel(unittest.TestCase): def testKernel(self): with sos_kernel() as kc: execute(kc=kc, code='a = 1') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stdout.strip(), '', f'Stdout is not empty, "{stdout}" received') self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') execute(kc=kc, code='%use Bash\n%get a\necho $a') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') self.assertEqual(stdout.strip(), '1', f'Stdout should be 1, "{stdout}" received') if __name__ == '__main__': unittest.main()
import unittest import sys from sos_notebook.test_utils import sos_kernel from ipykernel.tests.utils import execute, wait_for_idle, assemble_output class TestSoSKernel(unittest.TestCase): def testKernel(self): with sos_kernel() as kc: execute(kc=kc, code='a = 1') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stdout.strip(), '', f'Stdout is not empty, "{stdout}" received') self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') execute(kc=kc, code='%use Bash\n%get a\necho $a') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') self.assertEqual(stdout.strip(), '1', f'Stdout should be 1, "{stdout}" received') if __name__ == '__main__': unittest.main()
Test does not have to fail under windows
Test does not have to fail under windows
Python
bsd-3-clause
birdsarah/staged-recipes,synapticarbors/staged-recipes,SylvainCorlay/staged-recipes,kwilcox/staged-recipes,johanneskoester/staged-recipes,synapticarbors/staged-recipes,asmeurer/staged-recipes,SylvainCorlay/staged-recipes,scopatz/staged-recipes,patricksnape/staged-recipes,goanpeca/staged-recipes,hadim/staged-recipes,hadim/staged-recipes,ReimarBauer/staged-recipes,chrisburr/staged-recipes,conda-forge/staged-recipes,chrisburr/staged-recipes,dschreij/staged-recipes,stuertz/staged-recipes,ReimarBauer/staged-recipes,scopatz/staged-recipes,stuertz/staged-recipes,kwilcox/staged-recipes,jakirkham/staged-recipes,goanpeca/staged-recipes,jochym/staged-recipes,Juanlu001/staged-recipes,jochym/staged-recipes,ocefpaf/staged-recipes,ocefpaf/staged-recipes,Juanlu001/staged-recipes,asmeurer/staged-recipes,johanneskoester/staged-recipes,mariusvniekerk/staged-recipes,mcs07/staged-recipes,isuruf/staged-recipes,petrushy/staged-recipes,conda-forge/staged-recipes,mariusvniekerk/staged-recipes,dschreij/staged-recipes,petrushy/staged-recipes,birdsarah/staged-recipes,isuruf/staged-recipes,igortg/staged-recipes,jakirkham/staged-recipes,patricksnape/staged-recipes,igortg/staged-recipes,mcs07/staged-recipes
import unittest import sys from sos_notebook.test_utils import sos_kernel from ipykernel.tests.utils import execute, wait_for_idle, assemble_output @unittest.skipIf(sys.platform == 'win32', 'bash does not exist on win32') class TestSoSKernel(unittest.TestCase): def testKernel(self): with sos_kernel() as kc: execute(kc=kc, code='a = 1') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stdout.strip(), '', f'Stdout is not empty, "{stdout}" received') self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') execute(kc=kc, code='%use Bash\n%get a\necho $a') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') self.assertEqual(stdout.strip(), '1', f'Stdout should be 1, "{stdout}" received') if __name__ == '__main__': unittest.main() Test does not have to fail under windows
import unittest import sys from sos_notebook.test_utils import sos_kernel from ipykernel.tests.utils import execute, wait_for_idle, assemble_output class TestSoSKernel(unittest.TestCase): def testKernel(self): with sos_kernel() as kc: execute(kc=kc, code='a = 1') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stdout.strip(), '', f'Stdout is not empty, "{stdout}" received') self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') execute(kc=kc, code='%use Bash\n%get a\necho $a') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') self.assertEqual(stdout.strip(), '1', f'Stdout should be 1, "{stdout}" received') if __name__ == '__main__': unittest.main()
<commit_before>import unittest import sys from sos_notebook.test_utils import sos_kernel from ipykernel.tests.utils import execute, wait_for_idle, assemble_output @unittest.skipIf(sys.platform == 'win32', 'bash does not exist on win32') class TestSoSKernel(unittest.TestCase): def testKernel(self): with sos_kernel() as kc: execute(kc=kc, code='a = 1') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stdout.strip(), '', f'Stdout is not empty, "{stdout}" received') self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') execute(kc=kc, code='%use Bash\n%get a\necho $a') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') self.assertEqual(stdout.strip(), '1', f'Stdout should be 1, "{stdout}" received') if __name__ == '__main__': unittest.main() <commit_msg>Test does not have to fail under windows<commit_after>
import unittest import sys from sos_notebook.test_utils import sos_kernel from ipykernel.tests.utils import execute, wait_for_idle, assemble_output class TestSoSKernel(unittest.TestCase): def testKernel(self): with sos_kernel() as kc: execute(kc=kc, code='a = 1') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stdout.strip(), '', f'Stdout is not empty, "{stdout}" received') self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') execute(kc=kc, code='%use Bash\n%get a\necho $a') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') self.assertEqual(stdout.strip(), '1', f'Stdout should be 1, "{stdout}" received') if __name__ == '__main__': unittest.main()
import unittest import sys from sos_notebook.test_utils import sos_kernel from ipykernel.tests.utils import execute, wait_for_idle, assemble_output @unittest.skipIf(sys.platform == 'win32', 'bash does not exist on win32') class TestSoSKernel(unittest.TestCase): def testKernel(self): with sos_kernel() as kc: execute(kc=kc, code='a = 1') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stdout.strip(), '', f'Stdout is not empty, "{stdout}" received') self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') execute(kc=kc, code='%use Bash\n%get a\necho $a') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') self.assertEqual(stdout.strip(), '1', f'Stdout should be 1, "{stdout}" received') if __name__ == '__main__': unittest.main() Test does not have to fail under windowsimport unittest import sys from sos_notebook.test_utils import sos_kernel from ipykernel.tests.utils import execute, wait_for_idle, assemble_output class TestSoSKernel(unittest.TestCase): def testKernel(self): with sos_kernel() as kc: execute(kc=kc, code='a = 1') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stdout.strip(), '', f'Stdout is not empty, "{stdout}" received') self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') execute(kc=kc, code='%use Bash\n%get a\necho $a') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') self.assertEqual(stdout.strip(), '1', f'Stdout should be 1, "{stdout}" received') if __name__ == '__main__': unittest.main()
<commit_before>import unittest import sys from sos_notebook.test_utils import sos_kernel from ipykernel.tests.utils import execute, wait_for_idle, assemble_output @unittest.skipIf(sys.platform == 'win32', 'bash does not exist on win32') class TestSoSKernel(unittest.TestCase): def testKernel(self): with sos_kernel() as kc: execute(kc=kc, code='a = 1') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stdout.strip(), '', f'Stdout is not empty, "{stdout}" received') self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') execute(kc=kc, code='%use Bash\n%get a\necho $a') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') self.assertEqual(stdout.strip(), '1', f'Stdout should be 1, "{stdout}" received') if __name__ == '__main__': unittest.main() <commit_msg>Test does not have to fail under windows<commit_after>import unittest import sys from sos_notebook.test_utils import sos_kernel from ipykernel.tests.utils import execute, wait_for_idle, assemble_output class TestSoSKernel(unittest.TestCase): def testKernel(self): with sos_kernel() as kc: execute(kc=kc, code='a = 1') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stdout.strip(), '', f'Stdout is not empty, "{stdout}" received') self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') execute(kc=kc, code='%use Bash\n%get a\necho $a') stdout, stderr = assemble_output(kc.iopub_channel) self.assertEqual(stderr.strip(), '', f'Stderr is not empty, "{stderr}" received') self.assertEqual(stdout.strip(), '1', f'Stdout should be 1, "{stdout}" received') if __name__ == '__main__': unittest.main()
c7bc130efaaba1e079bc8a0f39f37ef3c2534413
Yank/tests/test_utils.py
Yank/tests/test_utils.py
#!/usr/local/bin/env python """ Test various utility functions. """ #============================================================================================= # GLOBAL IMPORTS #============================================================================================= from yank.utils import YankOptions #============================================================================================= # TESTING FUNCTIONS #============================================================================================= def test_yank_options(): """Test option priorities and handling.""" cl_opt = {'option1': 1} yaml_opt = {'option1': 2, 'option2': 'test'} yank_opt = YankOptions(cl_opt=cl_opt, yaml_opt=yaml_opt) assert yank_opt['option2'] == 'test' assert yank_opt['option1'] == 1 # command line > yaml assert len(yank_opt) == 2 # runtime > command line yank_opt['option1'] = 0 assert yank_opt['option1'] == 0 # restore old option when deleted at runtime del yank_opt['option1'] assert yank_opt['option1'] == 1 # modify specific priority level yank_opt.default = {'option3': -2} assert len(yank_opt) == 3 assert yank_opt['option3'] == -2 # test iteration interface assert yank_opt.items() == [('option1', 1), ('option2', 'test'), ('option3', -2)] assert yank_opt.keys() == ['option1', 'option2', 'option3']
#!/usr/local/bin/env python """ Test various utility functions. """ #============================================================================================= # GLOBAL IMPORTS #============================================================================================= from yank.utils import YankOptions #============================================================================================= # TESTING FUNCTIONS #============================================================================================= def test_yank_options(): """Test option priorities and handling.""" cl_opt = {'option1': 1} yaml_opt = {'option1': 2, 'option2': 'test'} yank_opt = YankOptions(cl_opt=cl_opt, yaml_opt=yaml_opt) assert yank_opt['option2'] == 'test' assert yank_opt['option1'] == 1 # command line > yaml assert len(yank_opt) == 2, "Excepted two options, found: %s" % str([x for x in yank_opt]) # runtime > command line yank_opt['option1'] = 0 assert yank_opt['option1'] == 0 # restore old option when deleted at runtime del yank_opt['option1'] assert yank_opt['option1'] == 1 # modify specific priority level yank_opt.default = {'option3': -2} assert len(yank_opt) == 3 assert yank_opt['option3'] == -2 # test iteration interface assert yank_opt.items() == [('option1', 1), ('option2', 'test'), ('option3', -2)] assert yank_opt.keys() == ['option1', 'option2', 'option3']
Add more informative error message for yank options assertion test.
Add more informative error message for yank options assertion test.
Python
mit
andrrizzi/yank,andrrizzi/yank,choderalab/yank,andrrizzi/yank,choderalab/yank
#!/usr/local/bin/env python """ Test various utility functions. """ #============================================================================================= # GLOBAL IMPORTS #============================================================================================= from yank.utils import YankOptions #============================================================================================= # TESTING FUNCTIONS #============================================================================================= def test_yank_options(): """Test option priorities and handling.""" cl_opt = {'option1': 1} yaml_opt = {'option1': 2, 'option2': 'test'} yank_opt = YankOptions(cl_opt=cl_opt, yaml_opt=yaml_opt) assert yank_opt['option2'] == 'test' assert yank_opt['option1'] == 1 # command line > yaml assert len(yank_opt) == 2 # runtime > command line yank_opt['option1'] = 0 assert yank_opt['option1'] == 0 # restore old option when deleted at runtime del yank_opt['option1'] assert yank_opt['option1'] == 1 # modify specific priority level yank_opt.default = {'option3': -2} assert len(yank_opt) == 3 assert yank_opt['option3'] == -2 # test iteration interface assert yank_opt.items() == [('option1', 1), ('option2', 'test'), ('option3', -2)] assert yank_opt.keys() == ['option1', 'option2', 'option3'] Add more informative error message for yank options assertion test.
#!/usr/local/bin/env python """ Test various utility functions. """ #============================================================================================= # GLOBAL IMPORTS #============================================================================================= from yank.utils import YankOptions #============================================================================================= # TESTING FUNCTIONS #============================================================================================= def test_yank_options(): """Test option priorities and handling.""" cl_opt = {'option1': 1} yaml_opt = {'option1': 2, 'option2': 'test'} yank_opt = YankOptions(cl_opt=cl_opt, yaml_opt=yaml_opt) assert yank_opt['option2'] == 'test' assert yank_opt['option1'] == 1 # command line > yaml assert len(yank_opt) == 2, "Excepted two options, found: %s" % str([x for x in yank_opt]) # runtime > command line yank_opt['option1'] = 0 assert yank_opt['option1'] == 0 # restore old option when deleted at runtime del yank_opt['option1'] assert yank_opt['option1'] == 1 # modify specific priority level yank_opt.default = {'option3': -2} assert len(yank_opt) == 3 assert yank_opt['option3'] == -2 # test iteration interface assert yank_opt.items() == [('option1', 1), ('option2', 'test'), ('option3', -2)] assert yank_opt.keys() == ['option1', 'option2', 'option3']
<commit_before>#!/usr/local/bin/env python """ Test various utility functions. """ #============================================================================================= # GLOBAL IMPORTS #============================================================================================= from yank.utils import YankOptions #============================================================================================= # TESTING FUNCTIONS #============================================================================================= def test_yank_options(): """Test option priorities and handling.""" cl_opt = {'option1': 1} yaml_opt = {'option1': 2, 'option2': 'test'} yank_opt = YankOptions(cl_opt=cl_opt, yaml_opt=yaml_opt) assert yank_opt['option2'] == 'test' assert yank_opt['option1'] == 1 # command line > yaml assert len(yank_opt) == 2 # runtime > command line yank_opt['option1'] = 0 assert yank_opt['option1'] == 0 # restore old option when deleted at runtime del yank_opt['option1'] assert yank_opt['option1'] == 1 # modify specific priority level yank_opt.default = {'option3': -2} assert len(yank_opt) == 3 assert yank_opt['option3'] == -2 # test iteration interface assert yank_opt.items() == [('option1', 1), ('option2', 'test'), ('option3', -2)] assert yank_opt.keys() == ['option1', 'option2', 'option3'] <commit_msg>Add more informative error message for yank options assertion test.<commit_after>
#!/usr/local/bin/env python """ Test various utility functions. """ #============================================================================================= # GLOBAL IMPORTS #============================================================================================= from yank.utils import YankOptions #============================================================================================= # TESTING FUNCTIONS #============================================================================================= def test_yank_options(): """Test option priorities and handling.""" cl_opt = {'option1': 1} yaml_opt = {'option1': 2, 'option2': 'test'} yank_opt = YankOptions(cl_opt=cl_opt, yaml_opt=yaml_opt) assert yank_opt['option2'] == 'test' assert yank_opt['option1'] == 1 # command line > yaml assert len(yank_opt) == 2, "Excepted two options, found: %s" % str([x for x in yank_opt]) # runtime > command line yank_opt['option1'] = 0 assert yank_opt['option1'] == 0 # restore old option when deleted at runtime del yank_opt['option1'] assert yank_opt['option1'] == 1 # modify specific priority level yank_opt.default = {'option3': -2} assert len(yank_opt) == 3 assert yank_opt['option3'] == -2 # test iteration interface assert yank_opt.items() == [('option1', 1), ('option2', 'test'), ('option3', -2)] assert yank_opt.keys() == ['option1', 'option2', 'option3']
#!/usr/local/bin/env python """ Test various utility functions. """ #============================================================================================= # GLOBAL IMPORTS #============================================================================================= from yank.utils import YankOptions #============================================================================================= # TESTING FUNCTIONS #============================================================================================= def test_yank_options(): """Test option priorities and handling.""" cl_opt = {'option1': 1} yaml_opt = {'option1': 2, 'option2': 'test'} yank_opt = YankOptions(cl_opt=cl_opt, yaml_opt=yaml_opt) assert yank_opt['option2'] == 'test' assert yank_opt['option1'] == 1 # command line > yaml assert len(yank_opt) == 2 # runtime > command line yank_opt['option1'] = 0 assert yank_opt['option1'] == 0 # restore old option when deleted at runtime del yank_opt['option1'] assert yank_opt['option1'] == 1 # modify specific priority level yank_opt.default = {'option3': -2} assert len(yank_opt) == 3 assert yank_opt['option3'] == -2 # test iteration interface assert yank_opt.items() == [('option1', 1), ('option2', 'test'), ('option3', -2)] assert yank_opt.keys() == ['option1', 'option2', 'option3'] Add more informative error message for yank options assertion test.#!/usr/local/bin/env python """ Test various utility functions. """ #============================================================================================= # GLOBAL IMPORTS #============================================================================================= from yank.utils import YankOptions #============================================================================================= # TESTING FUNCTIONS #============================================================================================= def test_yank_options(): """Test option priorities and handling.""" cl_opt = {'option1': 1} yaml_opt = {'option1': 2, 'option2': 'test'} yank_opt = YankOptions(cl_opt=cl_opt, yaml_opt=yaml_opt) assert yank_opt['option2'] == 'test' assert yank_opt['option1'] == 1 # command line > yaml assert len(yank_opt) == 2, "Excepted two options, found: %s" % str([x for x in yank_opt]) # runtime > command line yank_opt['option1'] = 0 assert yank_opt['option1'] == 0 # restore old option when deleted at runtime del yank_opt['option1'] assert yank_opt['option1'] == 1 # modify specific priority level yank_opt.default = {'option3': -2} assert len(yank_opt) == 3 assert yank_opt['option3'] == -2 # test iteration interface assert yank_opt.items() == [('option1', 1), ('option2', 'test'), ('option3', -2)] assert yank_opt.keys() == ['option1', 'option2', 'option3']
<commit_before>#!/usr/local/bin/env python """ Test various utility functions. """ #============================================================================================= # GLOBAL IMPORTS #============================================================================================= from yank.utils import YankOptions #============================================================================================= # TESTING FUNCTIONS #============================================================================================= def test_yank_options(): """Test option priorities and handling.""" cl_opt = {'option1': 1} yaml_opt = {'option1': 2, 'option2': 'test'} yank_opt = YankOptions(cl_opt=cl_opt, yaml_opt=yaml_opt) assert yank_opt['option2'] == 'test' assert yank_opt['option1'] == 1 # command line > yaml assert len(yank_opt) == 2 # runtime > command line yank_opt['option1'] = 0 assert yank_opt['option1'] == 0 # restore old option when deleted at runtime del yank_opt['option1'] assert yank_opt['option1'] == 1 # modify specific priority level yank_opt.default = {'option3': -2} assert len(yank_opt) == 3 assert yank_opt['option3'] == -2 # test iteration interface assert yank_opt.items() == [('option1', 1), ('option2', 'test'), ('option3', -2)] assert yank_opt.keys() == ['option1', 'option2', 'option3'] <commit_msg>Add more informative error message for yank options assertion test.<commit_after>#!/usr/local/bin/env python """ Test various utility functions. """ #============================================================================================= # GLOBAL IMPORTS #============================================================================================= from yank.utils import YankOptions #============================================================================================= # TESTING FUNCTIONS #============================================================================================= def test_yank_options(): """Test option priorities and handling.""" cl_opt = {'option1': 1} yaml_opt = {'option1': 2, 'option2': 'test'} yank_opt = YankOptions(cl_opt=cl_opt, yaml_opt=yaml_opt) assert yank_opt['option2'] == 'test' assert yank_opt['option1'] == 1 # command line > yaml assert len(yank_opt) == 2, "Excepted two options, found: %s" % str([x for x in yank_opt]) # runtime > command line yank_opt['option1'] = 0 assert yank_opt['option1'] == 0 # restore old option when deleted at runtime del yank_opt['option1'] assert yank_opt['option1'] == 1 # modify specific priority level yank_opt.default = {'option3': -2} assert len(yank_opt) == 3 assert yank_opt['option3'] == -2 # test iteration interface assert yank_opt.items() == [('option1', 1), ('option2', 'test'), ('option3', -2)] assert yank_opt.keys() == ['option1', 'option2', 'option3']
81dc92b3c2875b6775d33321b1bcd9f994be8a10
txircd/modules/extra/snotice_remoteconnect.py
txircd/modules/extra/snotice_remoteconnect.py
from twisted.plugin import IPlugin from txircd.module_interface import IModuleData, ModuleData from zope.interface import implements class SnoRemoteConnect(ModuleData): implements(IPlugin, IModuleData) name = "ServerNoticeRemoteConnect" def __init__(self): self.burstingServer = None def actions(self): return [ ("remoteregister", 1, self.sendRemoteConnectNotice), ("servernoticetype", 1, self.checkSnoType), ("startburstcommand", 1, self.markStartBurst), ("endburstcommand", 1, self.markEndBurst) ] def sendRemoteConnectNotice(self, user, *params): server = self.ircd.servers[user.uuid[:3]].name if server == self.burstingServer: return message = "Client connected on {}: {} ({}) [{}]".format(server, user.hostmaskWithRealHost(), user.ip, user.gecos) snodata = { "mask": "connect", "message": message } self.ircd.runActionProcessing("sendservernotice", snodata) def checkSnoType(self, user, typename): return typename == "remoteconnect" def markStartBurst(self, server, command): self.burstingServer = server def markEndBurst(self, server, command): self.burstingServer = None snoRemoteConnect = SnoRemoteConnect()
from twisted.plugin import IPlugin from txircd.module_interface import IModuleData, ModuleData from zope.interface import implements class SnoRemoteConnect(ModuleData): implements(IPlugin, IModuleData) name = "ServerNoticeRemoteConnect" def __init__(self): self.burstingServer = None def actions(self): return [ ("remoteregister", 1, self.sendRemoteConnectNotice), ("servernoticetype", 1, self.checkSnoType), ("startburstcommand", 1, self.markStartBurst), ("endburstcommand", 1, self.markEndBurst) ] def sendRemoteConnectNotice(self, user, *params): server = self.ircd.servers[user.uuid[:3]] if server == self.burstingServer: return server = server.name message = "Client connected on {}: {} ({}) [{}]".format(server, user.hostmaskWithRealHost(), user.ip, user.gecos) snodata = { "mask": "connect", "message": message } self.ircd.runActionProcessing("sendservernotice", snodata) def checkSnoType(self, user, typename): return typename == "remoteconnect" def markStartBurst(self, server, command): self.burstingServer = server def markEndBurst(self, server, command): self.burstingServer = None snoRemoteConnect = SnoRemoteConnect()
Fix checking a name against an object
Fix checking a name against an object
Python
bsd-3-clause
Heufneutje/txircd
from twisted.plugin import IPlugin from txircd.module_interface import IModuleData, ModuleData from zope.interface import implements class SnoRemoteConnect(ModuleData): implements(IPlugin, IModuleData) name = "ServerNoticeRemoteConnect" def __init__(self): self.burstingServer = None def actions(self): return [ ("remoteregister", 1, self.sendRemoteConnectNotice), ("servernoticetype", 1, self.checkSnoType), ("startburstcommand", 1, self.markStartBurst), ("endburstcommand", 1, self.markEndBurst) ] def sendRemoteConnectNotice(self, user, *params): server = self.ircd.servers[user.uuid[:3]].name if server == self.burstingServer: return message = "Client connected on {}: {} ({}) [{}]".format(server, user.hostmaskWithRealHost(), user.ip, user.gecos) snodata = { "mask": "connect", "message": message } self.ircd.runActionProcessing("sendservernotice", snodata) def checkSnoType(self, user, typename): return typename == "remoteconnect" def markStartBurst(self, server, command): self.burstingServer = server def markEndBurst(self, server, command): self.burstingServer = None snoRemoteConnect = SnoRemoteConnect()Fix checking a name against an object
from twisted.plugin import IPlugin from txircd.module_interface import IModuleData, ModuleData from zope.interface import implements class SnoRemoteConnect(ModuleData): implements(IPlugin, IModuleData) name = "ServerNoticeRemoteConnect" def __init__(self): self.burstingServer = None def actions(self): return [ ("remoteregister", 1, self.sendRemoteConnectNotice), ("servernoticetype", 1, self.checkSnoType), ("startburstcommand", 1, self.markStartBurst), ("endburstcommand", 1, self.markEndBurst) ] def sendRemoteConnectNotice(self, user, *params): server = self.ircd.servers[user.uuid[:3]] if server == self.burstingServer: return server = server.name message = "Client connected on {}: {} ({}) [{}]".format(server, user.hostmaskWithRealHost(), user.ip, user.gecos) snodata = { "mask": "connect", "message": message } self.ircd.runActionProcessing("sendservernotice", snodata) def checkSnoType(self, user, typename): return typename == "remoteconnect" def markStartBurst(self, server, command): self.burstingServer = server def markEndBurst(self, server, command): self.burstingServer = None snoRemoteConnect = SnoRemoteConnect()
<commit_before>from twisted.plugin import IPlugin from txircd.module_interface import IModuleData, ModuleData from zope.interface import implements class SnoRemoteConnect(ModuleData): implements(IPlugin, IModuleData) name = "ServerNoticeRemoteConnect" def __init__(self): self.burstingServer = None def actions(self): return [ ("remoteregister", 1, self.sendRemoteConnectNotice), ("servernoticetype", 1, self.checkSnoType), ("startburstcommand", 1, self.markStartBurst), ("endburstcommand", 1, self.markEndBurst) ] def sendRemoteConnectNotice(self, user, *params): server = self.ircd.servers[user.uuid[:3]].name if server == self.burstingServer: return message = "Client connected on {}: {} ({}) [{}]".format(server, user.hostmaskWithRealHost(), user.ip, user.gecos) snodata = { "mask": "connect", "message": message } self.ircd.runActionProcessing("sendservernotice", snodata) def checkSnoType(self, user, typename): return typename == "remoteconnect" def markStartBurst(self, server, command): self.burstingServer = server def markEndBurst(self, server, command): self.burstingServer = None snoRemoteConnect = SnoRemoteConnect()<commit_msg>Fix checking a name against an object<commit_after>
from twisted.plugin import IPlugin from txircd.module_interface import IModuleData, ModuleData from zope.interface import implements class SnoRemoteConnect(ModuleData): implements(IPlugin, IModuleData) name = "ServerNoticeRemoteConnect" def __init__(self): self.burstingServer = None def actions(self): return [ ("remoteregister", 1, self.sendRemoteConnectNotice), ("servernoticetype", 1, self.checkSnoType), ("startburstcommand", 1, self.markStartBurst), ("endburstcommand", 1, self.markEndBurst) ] def sendRemoteConnectNotice(self, user, *params): server = self.ircd.servers[user.uuid[:3]] if server == self.burstingServer: return server = server.name message = "Client connected on {}: {} ({}) [{}]".format(server, user.hostmaskWithRealHost(), user.ip, user.gecos) snodata = { "mask": "connect", "message": message } self.ircd.runActionProcessing("sendservernotice", snodata) def checkSnoType(self, user, typename): return typename == "remoteconnect" def markStartBurst(self, server, command): self.burstingServer = server def markEndBurst(self, server, command): self.burstingServer = None snoRemoteConnect = SnoRemoteConnect()
from twisted.plugin import IPlugin from txircd.module_interface import IModuleData, ModuleData from zope.interface import implements class SnoRemoteConnect(ModuleData): implements(IPlugin, IModuleData) name = "ServerNoticeRemoteConnect" def __init__(self): self.burstingServer = None def actions(self): return [ ("remoteregister", 1, self.sendRemoteConnectNotice), ("servernoticetype", 1, self.checkSnoType), ("startburstcommand", 1, self.markStartBurst), ("endburstcommand", 1, self.markEndBurst) ] def sendRemoteConnectNotice(self, user, *params): server = self.ircd.servers[user.uuid[:3]].name if server == self.burstingServer: return message = "Client connected on {}: {} ({}) [{}]".format(server, user.hostmaskWithRealHost(), user.ip, user.gecos) snodata = { "mask": "connect", "message": message } self.ircd.runActionProcessing("sendservernotice", snodata) def checkSnoType(self, user, typename): return typename == "remoteconnect" def markStartBurst(self, server, command): self.burstingServer = server def markEndBurst(self, server, command): self.burstingServer = None snoRemoteConnect = SnoRemoteConnect()Fix checking a name against an objectfrom twisted.plugin import IPlugin from txircd.module_interface import IModuleData, ModuleData from zope.interface import implements class SnoRemoteConnect(ModuleData): implements(IPlugin, IModuleData) name = "ServerNoticeRemoteConnect" def __init__(self): self.burstingServer = None def actions(self): return [ ("remoteregister", 1, self.sendRemoteConnectNotice), ("servernoticetype", 1, self.checkSnoType), ("startburstcommand", 1, self.markStartBurst), ("endburstcommand", 1, self.markEndBurst) ] def sendRemoteConnectNotice(self, user, *params): server = self.ircd.servers[user.uuid[:3]] if server == self.burstingServer: return server = server.name message = "Client connected on {}: {} ({}) [{}]".format(server, user.hostmaskWithRealHost(), user.ip, user.gecos) snodata = { "mask": "connect", "message": message } self.ircd.runActionProcessing("sendservernotice", snodata) def checkSnoType(self, user, typename): return typename == "remoteconnect" def markStartBurst(self, server, command): self.burstingServer = server def markEndBurst(self, server, command): self.burstingServer = None snoRemoteConnect = SnoRemoteConnect()
<commit_before>from twisted.plugin import IPlugin from txircd.module_interface import IModuleData, ModuleData from zope.interface import implements class SnoRemoteConnect(ModuleData): implements(IPlugin, IModuleData) name = "ServerNoticeRemoteConnect" def __init__(self): self.burstingServer = None def actions(self): return [ ("remoteregister", 1, self.sendRemoteConnectNotice), ("servernoticetype", 1, self.checkSnoType), ("startburstcommand", 1, self.markStartBurst), ("endburstcommand", 1, self.markEndBurst) ] def sendRemoteConnectNotice(self, user, *params): server = self.ircd.servers[user.uuid[:3]].name if server == self.burstingServer: return message = "Client connected on {}: {} ({}) [{}]".format(server, user.hostmaskWithRealHost(), user.ip, user.gecos) snodata = { "mask": "connect", "message": message } self.ircd.runActionProcessing("sendservernotice", snodata) def checkSnoType(self, user, typename): return typename == "remoteconnect" def markStartBurst(self, server, command): self.burstingServer = server def markEndBurst(self, server, command): self.burstingServer = None snoRemoteConnect = SnoRemoteConnect()<commit_msg>Fix checking a name against an object<commit_after>from twisted.plugin import IPlugin from txircd.module_interface import IModuleData, ModuleData from zope.interface import implements class SnoRemoteConnect(ModuleData): implements(IPlugin, IModuleData) name = "ServerNoticeRemoteConnect" def __init__(self): self.burstingServer = None def actions(self): return [ ("remoteregister", 1, self.sendRemoteConnectNotice), ("servernoticetype", 1, self.checkSnoType), ("startburstcommand", 1, self.markStartBurst), ("endburstcommand", 1, self.markEndBurst) ] def sendRemoteConnectNotice(self, user, *params): server = self.ircd.servers[user.uuid[:3]] if server == self.burstingServer: return server = server.name message = "Client connected on {}: {} ({}) [{}]".format(server, user.hostmaskWithRealHost(), user.ip, user.gecos) snodata = { "mask": "connect", "message": message } self.ircd.runActionProcessing("sendservernotice", snodata) def checkSnoType(self, user, typename): return typename == "remoteconnect" def markStartBurst(self, server, command): self.burstingServer = server def markEndBurst(self, server, command): self.burstingServer = None snoRemoteConnect = SnoRemoteConnect()
8ecf97b338dd37eaf5a4e2672e33e27cc40d215d
sacred/observers/__init__.py
sacred/observers/__init__.py
#!/usr/bin/env python # coding=utf-8 from __future__ import division, print_function, unicode_literals from sacred.commandline_options import CommandLineOption from sacred.observers.base import RunObserver from sacred.observers.file_storage import FileStorageObserver import sacred.optional as opt from sacred.observers.tinydb_hashfs import TinyDbObserver if opt.has_pymongo: from sacred.observers.mongo import MongoObserver else: MongoObserver = opt.MissingDependencyMock('pymongo') class MongoDbOption(CommandLineOption): """To use the MongoObserver you need to install pymongo first.""" arg = 'DB' @classmethod def apply(cls, args, run): raise ImportError('cannot use -m/--mongo_db flag: ' 'missing pymongo dependency') if opt.has_sqlalchemy: from sacred.observers.sql import SqlObserver else: SqlObserver = opt.MissingDependencyMock('sqlalchemy') class SqlOption(CommandLineOption): """To use the SqlObserver you need to install sqlalchemy first.""" arg = 'DB_URL' @classmethod def apply(cls, args, run): raise ImportError('cannot use -s/--sql flag: ' 'missing sqlalchemy dependency') __all__ = ('FileStorageObserver', 'RunObserver', 'MongoObserver', 'SqlObserver', 'TinyDbObserver')
#!/usr/bin/env python # coding=utf-8 from __future__ import division, print_function, unicode_literals from sacred.commandline_options import CommandLineOption from sacred.observers.base import RunObserver from sacred.observers.file_storage import FileStorageObserver import sacred.optional as opt from sacred.observers.tinydb_hashfs import TinyDbObserver, TinyDbReader if opt.has_pymongo: from sacred.observers.mongo import MongoObserver else: MongoObserver = opt.MissingDependencyMock('pymongo') class MongoDbOption(CommandLineOption): """To use the MongoObserver you need to install pymongo first.""" arg = 'DB' @classmethod def apply(cls, args, run): raise ImportError('cannot use -m/--mongo_db flag: ' 'missing pymongo dependency') if opt.has_sqlalchemy: from sacred.observers.sql import SqlObserver else: SqlObserver = opt.MissingDependencyMock('sqlalchemy') class SqlOption(CommandLineOption): """To use the SqlObserver you need to install sqlalchemy first.""" arg = 'DB_URL' @classmethod def apply(cls, args, run): raise ImportError('cannot use -s/--sql flag: ' 'missing sqlalchemy dependency') __all__ = ('FileStorageObserver', 'RunObserver', 'MongoObserver', 'SqlObserver', 'TinyDbObserver', 'TinyDbReader')
Add TinyDbReader to observers init
Add TinyDbReader to observers init
Python
mit
IDSIA/sacred,IDSIA/sacred
#!/usr/bin/env python # coding=utf-8 from __future__ import division, print_function, unicode_literals from sacred.commandline_options import CommandLineOption from sacred.observers.base import RunObserver from sacred.observers.file_storage import FileStorageObserver import sacred.optional as opt from sacred.observers.tinydb_hashfs import TinyDbObserver if opt.has_pymongo: from sacred.observers.mongo import MongoObserver else: MongoObserver = opt.MissingDependencyMock('pymongo') class MongoDbOption(CommandLineOption): """To use the MongoObserver you need to install pymongo first.""" arg = 'DB' @classmethod def apply(cls, args, run): raise ImportError('cannot use -m/--mongo_db flag: ' 'missing pymongo dependency') if opt.has_sqlalchemy: from sacred.observers.sql import SqlObserver else: SqlObserver = opt.MissingDependencyMock('sqlalchemy') class SqlOption(CommandLineOption): """To use the SqlObserver you need to install sqlalchemy first.""" arg = 'DB_URL' @classmethod def apply(cls, args, run): raise ImportError('cannot use -s/--sql flag: ' 'missing sqlalchemy dependency') __all__ = ('FileStorageObserver', 'RunObserver', 'MongoObserver', 'SqlObserver', 'TinyDbObserver') Add TinyDbReader to observers init
#!/usr/bin/env python # coding=utf-8 from __future__ import division, print_function, unicode_literals from sacred.commandline_options import CommandLineOption from sacred.observers.base import RunObserver from sacred.observers.file_storage import FileStorageObserver import sacred.optional as opt from sacred.observers.tinydb_hashfs import TinyDbObserver, TinyDbReader if opt.has_pymongo: from sacred.observers.mongo import MongoObserver else: MongoObserver = opt.MissingDependencyMock('pymongo') class MongoDbOption(CommandLineOption): """To use the MongoObserver you need to install pymongo first.""" arg = 'DB' @classmethod def apply(cls, args, run): raise ImportError('cannot use -m/--mongo_db flag: ' 'missing pymongo dependency') if opt.has_sqlalchemy: from sacred.observers.sql import SqlObserver else: SqlObserver = opt.MissingDependencyMock('sqlalchemy') class SqlOption(CommandLineOption): """To use the SqlObserver you need to install sqlalchemy first.""" arg = 'DB_URL' @classmethod def apply(cls, args, run): raise ImportError('cannot use -s/--sql flag: ' 'missing sqlalchemy dependency') __all__ = ('FileStorageObserver', 'RunObserver', 'MongoObserver', 'SqlObserver', 'TinyDbObserver', 'TinyDbReader')
<commit_before>#!/usr/bin/env python # coding=utf-8 from __future__ import division, print_function, unicode_literals from sacred.commandline_options import CommandLineOption from sacred.observers.base import RunObserver from sacred.observers.file_storage import FileStorageObserver import sacred.optional as opt from sacred.observers.tinydb_hashfs import TinyDbObserver if opt.has_pymongo: from sacred.observers.mongo import MongoObserver else: MongoObserver = opt.MissingDependencyMock('pymongo') class MongoDbOption(CommandLineOption): """To use the MongoObserver you need to install pymongo first.""" arg = 'DB' @classmethod def apply(cls, args, run): raise ImportError('cannot use -m/--mongo_db flag: ' 'missing pymongo dependency') if opt.has_sqlalchemy: from sacred.observers.sql import SqlObserver else: SqlObserver = opt.MissingDependencyMock('sqlalchemy') class SqlOption(CommandLineOption): """To use the SqlObserver you need to install sqlalchemy first.""" arg = 'DB_URL' @classmethod def apply(cls, args, run): raise ImportError('cannot use -s/--sql flag: ' 'missing sqlalchemy dependency') __all__ = ('FileStorageObserver', 'RunObserver', 'MongoObserver', 'SqlObserver', 'TinyDbObserver') <commit_msg>Add TinyDbReader to observers init<commit_after>
#!/usr/bin/env python # coding=utf-8 from __future__ import division, print_function, unicode_literals from sacred.commandline_options import CommandLineOption from sacred.observers.base import RunObserver from sacred.observers.file_storage import FileStorageObserver import sacred.optional as opt from sacred.observers.tinydb_hashfs import TinyDbObserver, TinyDbReader if opt.has_pymongo: from sacred.observers.mongo import MongoObserver else: MongoObserver = opt.MissingDependencyMock('pymongo') class MongoDbOption(CommandLineOption): """To use the MongoObserver you need to install pymongo first.""" arg = 'DB' @classmethod def apply(cls, args, run): raise ImportError('cannot use -m/--mongo_db flag: ' 'missing pymongo dependency') if opt.has_sqlalchemy: from sacred.observers.sql import SqlObserver else: SqlObserver = opt.MissingDependencyMock('sqlalchemy') class SqlOption(CommandLineOption): """To use the SqlObserver you need to install sqlalchemy first.""" arg = 'DB_URL' @classmethod def apply(cls, args, run): raise ImportError('cannot use -s/--sql flag: ' 'missing sqlalchemy dependency') __all__ = ('FileStorageObserver', 'RunObserver', 'MongoObserver', 'SqlObserver', 'TinyDbObserver', 'TinyDbReader')
#!/usr/bin/env python # coding=utf-8 from __future__ import division, print_function, unicode_literals from sacred.commandline_options import CommandLineOption from sacred.observers.base import RunObserver from sacred.observers.file_storage import FileStorageObserver import sacred.optional as opt from sacred.observers.tinydb_hashfs import TinyDbObserver if opt.has_pymongo: from sacred.observers.mongo import MongoObserver else: MongoObserver = opt.MissingDependencyMock('pymongo') class MongoDbOption(CommandLineOption): """To use the MongoObserver you need to install pymongo first.""" arg = 'DB' @classmethod def apply(cls, args, run): raise ImportError('cannot use -m/--mongo_db flag: ' 'missing pymongo dependency') if opt.has_sqlalchemy: from sacred.observers.sql import SqlObserver else: SqlObserver = opt.MissingDependencyMock('sqlalchemy') class SqlOption(CommandLineOption): """To use the SqlObserver you need to install sqlalchemy first.""" arg = 'DB_URL' @classmethod def apply(cls, args, run): raise ImportError('cannot use -s/--sql flag: ' 'missing sqlalchemy dependency') __all__ = ('FileStorageObserver', 'RunObserver', 'MongoObserver', 'SqlObserver', 'TinyDbObserver') Add TinyDbReader to observers init#!/usr/bin/env python # coding=utf-8 from __future__ import division, print_function, unicode_literals from sacred.commandline_options import CommandLineOption from sacred.observers.base import RunObserver from sacred.observers.file_storage import FileStorageObserver import sacred.optional as opt from sacred.observers.tinydb_hashfs import TinyDbObserver, TinyDbReader if opt.has_pymongo: from sacred.observers.mongo import MongoObserver else: MongoObserver = opt.MissingDependencyMock('pymongo') class MongoDbOption(CommandLineOption): """To use the MongoObserver you need to install pymongo first.""" arg = 'DB' @classmethod def apply(cls, args, run): raise ImportError('cannot use -m/--mongo_db flag: ' 'missing pymongo dependency') if opt.has_sqlalchemy: from sacred.observers.sql import SqlObserver else: SqlObserver = opt.MissingDependencyMock('sqlalchemy') class SqlOption(CommandLineOption): """To use the SqlObserver you need to install sqlalchemy first.""" arg = 'DB_URL' @classmethod def apply(cls, args, run): raise ImportError('cannot use -s/--sql flag: ' 'missing sqlalchemy dependency') __all__ = ('FileStorageObserver', 'RunObserver', 'MongoObserver', 'SqlObserver', 'TinyDbObserver', 'TinyDbReader')
<commit_before>#!/usr/bin/env python # coding=utf-8 from __future__ import division, print_function, unicode_literals from sacred.commandline_options import CommandLineOption from sacred.observers.base import RunObserver from sacred.observers.file_storage import FileStorageObserver import sacred.optional as opt from sacred.observers.tinydb_hashfs import TinyDbObserver if opt.has_pymongo: from sacred.observers.mongo import MongoObserver else: MongoObserver = opt.MissingDependencyMock('pymongo') class MongoDbOption(CommandLineOption): """To use the MongoObserver you need to install pymongo first.""" arg = 'DB' @classmethod def apply(cls, args, run): raise ImportError('cannot use -m/--mongo_db flag: ' 'missing pymongo dependency') if opt.has_sqlalchemy: from sacred.observers.sql import SqlObserver else: SqlObserver = opt.MissingDependencyMock('sqlalchemy') class SqlOption(CommandLineOption): """To use the SqlObserver you need to install sqlalchemy first.""" arg = 'DB_URL' @classmethod def apply(cls, args, run): raise ImportError('cannot use -s/--sql flag: ' 'missing sqlalchemy dependency') __all__ = ('FileStorageObserver', 'RunObserver', 'MongoObserver', 'SqlObserver', 'TinyDbObserver') <commit_msg>Add TinyDbReader to observers init<commit_after>#!/usr/bin/env python # coding=utf-8 from __future__ import division, print_function, unicode_literals from sacred.commandline_options import CommandLineOption from sacred.observers.base import RunObserver from sacred.observers.file_storage import FileStorageObserver import sacred.optional as opt from sacred.observers.tinydb_hashfs import TinyDbObserver, TinyDbReader if opt.has_pymongo: from sacred.observers.mongo import MongoObserver else: MongoObserver = opt.MissingDependencyMock('pymongo') class MongoDbOption(CommandLineOption): """To use the MongoObserver you need to install pymongo first.""" arg = 'DB' @classmethod def apply(cls, args, run): raise ImportError('cannot use -m/--mongo_db flag: ' 'missing pymongo dependency') if opt.has_sqlalchemy: from sacred.observers.sql import SqlObserver else: SqlObserver = opt.MissingDependencyMock('sqlalchemy') class SqlOption(CommandLineOption): """To use the SqlObserver you need to install sqlalchemy first.""" arg = 'DB_URL' @classmethod def apply(cls, args, run): raise ImportError('cannot use -s/--sql flag: ' 'missing sqlalchemy dependency') __all__ = ('FileStorageObserver', 'RunObserver', 'MongoObserver', 'SqlObserver', 'TinyDbObserver', 'TinyDbReader')
04fec1c50ac81c1b80c22f37bd43845a0e08c1a3
fancypages/assets/models.py
fancypages/assets/models.py
from django.db import models from django.utils.translation import ugettext as _ class AbstractAsset(models.Model): name = models.CharField(_("Name"), max_length=255) date_created = models.DateTimeField(_("Date created"), auto_now_add=True) date_modified = models.DateTimeField(_("Date modified"), auto_now=True) description = models.TextField(_("Description"), default="") creator = models.ForeignKey('auth.User', verbose_name=_("Creator")) def __unicode__(self): return self.name class Meta: abstract = True class ImageAsset(AbstractAsset): image = models.ImageField( upload_to='asset/images/%Y/%m', width_field='width', height_field='height', verbose_name=_("Image") ) width = models.IntegerField(_("Width"), blank=True) height = models.IntegerField(_("Height"), blank=True) size = models.IntegerField(_("Size"), blank=True, null=True) # Bytes @property def asset_type(self): return self._meta.object_name.lower() def get_absolute_url(self): return self.image.url
from django.db import models from django.utils.translation import ugettext as _ try: from django.contrib.auth import get_user_model User = get_user_model() except ImportError: from django.contrib.auth.models import User class AbstractAsset(models.Model): name = models.CharField(_("Name"), max_length=255) date_created = models.DateTimeField(_("Date created"), auto_now_add=True) date_modified = models.DateTimeField(_("Date modified"), auto_now=True) description = models.TextField(_("Description"), default="") creator = models.ForeignKey(User, verbose_name=_("Creator")) def __unicode__(self): return self.name class Meta: abstract = True class ImageAsset(AbstractAsset): image = models.ImageField( upload_to='asset/images/%Y/%m', width_field='width', height_field='height', verbose_name=_("Image") ) width = models.IntegerField(_("Width"), blank=True) height = models.IntegerField(_("Height"), blank=True) size = models.IntegerField(_("Size"), blank=True, null=True) # Bytes @property def asset_type(self): return self._meta.object_name.lower() def get_absolute_url(self): return self.image.url
Add support for custom user model (Django 1.5+)
Add support for custom user model (Django 1.5+)
Python
bsd-3-clause
socradev/django-fancypages,tangentlabs/django-fancypages,socradev/django-fancypages,tangentlabs/django-fancypages,tangentlabs/django-fancypages,socradev/django-fancypages
from django.db import models from django.utils.translation import ugettext as _ class AbstractAsset(models.Model): name = models.CharField(_("Name"), max_length=255) date_created = models.DateTimeField(_("Date created"), auto_now_add=True) date_modified = models.DateTimeField(_("Date modified"), auto_now=True) description = models.TextField(_("Description"), default="") creator = models.ForeignKey('auth.User', verbose_name=_("Creator")) def __unicode__(self): return self.name class Meta: abstract = True class ImageAsset(AbstractAsset): image = models.ImageField( upload_to='asset/images/%Y/%m', width_field='width', height_field='height', verbose_name=_("Image") ) width = models.IntegerField(_("Width"), blank=True) height = models.IntegerField(_("Height"), blank=True) size = models.IntegerField(_("Size"), blank=True, null=True) # Bytes @property def asset_type(self): return self._meta.object_name.lower() def get_absolute_url(self): return self.image.url Add support for custom user model (Django 1.5+)
from django.db import models from django.utils.translation import ugettext as _ try: from django.contrib.auth import get_user_model User = get_user_model() except ImportError: from django.contrib.auth.models import User class AbstractAsset(models.Model): name = models.CharField(_("Name"), max_length=255) date_created = models.DateTimeField(_("Date created"), auto_now_add=True) date_modified = models.DateTimeField(_("Date modified"), auto_now=True) description = models.TextField(_("Description"), default="") creator = models.ForeignKey(User, verbose_name=_("Creator")) def __unicode__(self): return self.name class Meta: abstract = True class ImageAsset(AbstractAsset): image = models.ImageField( upload_to='asset/images/%Y/%m', width_field='width', height_field='height', verbose_name=_("Image") ) width = models.IntegerField(_("Width"), blank=True) height = models.IntegerField(_("Height"), blank=True) size = models.IntegerField(_("Size"), blank=True, null=True) # Bytes @property def asset_type(self): return self._meta.object_name.lower() def get_absolute_url(self): return self.image.url
<commit_before>from django.db import models from django.utils.translation import ugettext as _ class AbstractAsset(models.Model): name = models.CharField(_("Name"), max_length=255) date_created = models.DateTimeField(_("Date created"), auto_now_add=True) date_modified = models.DateTimeField(_("Date modified"), auto_now=True) description = models.TextField(_("Description"), default="") creator = models.ForeignKey('auth.User', verbose_name=_("Creator")) def __unicode__(self): return self.name class Meta: abstract = True class ImageAsset(AbstractAsset): image = models.ImageField( upload_to='asset/images/%Y/%m', width_field='width', height_field='height', verbose_name=_("Image") ) width = models.IntegerField(_("Width"), blank=True) height = models.IntegerField(_("Height"), blank=True) size = models.IntegerField(_("Size"), blank=True, null=True) # Bytes @property def asset_type(self): return self._meta.object_name.lower() def get_absolute_url(self): return self.image.url <commit_msg>Add support for custom user model (Django 1.5+)<commit_after>
from django.db import models from django.utils.translation import ugettext as _ try: from django.contrib.auth import get_user_model User = get_user_model() except ImportError: from django.contrib.auth.models import User class AbstractAsset(models.Model): name = models.CharField(_("Name"), max_length=255) date_created = models.DateTimeField(_("Date created"), auto_now_add=True) date_modified = models.DateTimeField(_("Date modified"), auto_now=True) description = models.TextField(_("Description"), default="") creator = models.ForeignKey(User, verbose_name=_("Creator")) def __unicode__(self): return self.name class Meta: abstract = True class ImageAsset(AbstractAsset): image = models.ImageField( upload_to='asset/images/%Y/%m', width_field='width', height_field='height', verbose_name=_("Image") ) width = models.IntegerField(_("Width"), blank=True) height = models.IntegerField(_("Height"), blank=True) size = models.IntegerField(_("Size"), blank=True, null=True) # Bytes @property def asset_type(self): return self._meta.object_name.lower() def get_absolute_url(self): return self.image.url
from django.db import models from django.utils.translation import ugettext as _ class AbstractAsset(models.Model): name = models.CharField(_("Name"), max_length=255) date_created = models.DateTimeField(_("Date created"), auto_now_add=True) date_modified = models.DateTimeField(_("Date modified"), auto_now=True) description = models.TextField(_("Description"), default="") creator = models.ForeignKey('auth.User', verbose_name=_("Creator")) def __unicode__(self): return self.name class Meta: abstract = True class ImageAsset(AbstractAsset): image = models.ImageField( upload_to='asset/images/%Y/%m', width_field='width', height_field='height', verbose_name=_("Image") ) width = models.IntegerField(_("Width"), blank=True) height = models.IntegerField(_("Height"), blank=True) size = models.IntegerField(_("Size"), blank=True, null=True) # Bytes @property def asset_type(self): return self._meta.object_name.lower() def get_absolute_url(self): return self.image.url Add support for custom user model (Django 1.5+)from django.db import models from django.utils.translation import ugettext as _ try: from django.contrib.auth import get_user_model User = get_user_model() except ImportError: from django.contrib.auth.models import User class AbstractAsset(models.Model): name = models.CharField(_("Name"), max_length=255) date_created = models.DateTimeField(_("Date created"), auto_now_add=True) date_modified = models.DateTimeField(_("Date modified"), auto_now=True) description = models.TextField(_("Description"), default="") creator = models.ForeignKey(User, verbose_name=_("Creator")) def __unicode__(self): return self.name class Meta: abstract = True class ImageAsset(AbstractAsset): image = models.ImageField( upload_to='asset/images/%Y/%m', width_field='width', height_field='height', verbose_name=_("Image") ) width = models.IntegerField(_("Width"), blank=True) height = models.IntegerField(_("Height"), blank=True) size = models.IntegerField(_("Size"), blank=True, null=True) # Bytes @property def asset_type(self): return self._meta.object_name.lower() def get_absolute_url(self): return self.image.url
<commit_before>from django.db import models from django.utils.translation import ugettext as _ class AbstractAsset(models.Model): name = models.CharField(_("Name"), max_length=255) date_created = models.DateTimeField(_("Date created"), auto_now_add=True) date_modified = models.DateTimeField(_("Date modified"), auto_now=True) description = models.TextField(_("Description"), default="") creator = models.ForeignKey('auth.User', verbose_name=_("Creator")) def __unicode__(self): return self.name class Meta: abstract = True class ImageAsset(AbstractAsset): image = models.ImageField( upload_to='asset/images/%Y/%m', width_field='width', height_field='height', verbose_name=_("Image") ) width = models.IntegerField(_("Width"), blank=True) height = models.IntegerField(_("Height"), blank=True) size = models.IntegerField(_("Size"), blank=True, null=True) # Bytes @property def asset_type(self): return self._meta.object_name.lower() def get_absolute_url(self): return self.image.url <commit_msg>Add support for custom user model (Django 1.5+)<commit_after>from django.db import models from django.utils.translation import ugettext as _ try: from django.contrib.auth import get_user_model User = get_user_model() except ImportError: from django.contrib.auth.models import User class AbstractAsset(models.Model): name = models.CharField(_("Name"), max_length=255) date_created = models.DateTimeField(_("Date created"), auto_now_add=True) date_modified = models.DateTimeField(_("Date modified"), auto_now=True) description = models.TextField(_("Description"), default="") creator = models.ForeignKey(User, verbose_name=_("Creator")) def __unicode__(self): return self.name class Meta: abstract = True class ImageAsset(AbstractAsset): image = models.ImageField( upload_to='asset/images/%Y/%m', width_field='width', height_field='height', verbose_name=_("Image") ) width = models.IntegerField(_("Width"), blank=True) height = models.IntegerField(_("Height"), blank=True) size = models.IntegerField(_("Size"), blank=True, null=True) # Bytes @property def asset_type(self): return self._meta.object_name.lower() def get_absolute_url(self): return self.image.url
7629afde2627457b4f4b19e1542a87e695c1837d
tests/events/test_models.py
tests/events/test_models.py
"""Unit tests for events models.""" import datetime from app.events.factories import EventFactory from app.events.models import Event def test_event_factory(db): # noqa: D103 # GIVEN an empty database assert Event.objects.count() == 0 # WHEN saving a new event instance to the database EventFactory.create(title='five') # THEN it's there assert Event.objects.count() == 1 def test_event_has_all_the_attributes(): # noqa: D103 # GIVEN an event e = EventFactory.build() # THEN it has … assert e.title assert e.date assert e.venue assert e.description assert e.fb_event_url def test_event_has_slug(db): # noqa: D103 # GIVEN an event e = EventFactory.build( title='One Happy Family', date=datetime.date(2018, 1, 1), venue=None, ) assert e.slug == '' # WHEN saving the event e.save() # THEN it gets a slug generated from its date and title assert e.slug == '2018-01-01-one-happy-family'
"""Unit tests for events models.""" import datetime from app.events.factories import EventFactory from app.events.models import Event def test_event_factory(db): # noqa: D103 # GIVEN an empty database assert Event.objects.count() == 0 # WHEN saving a new event instance to the database EventFactory.create(title='five') # THEN it's there assert Event.objects.count() == 1 def test_event_has_all_the_attributes(): # noqa: D103 # GIVEN an event e = EventFactory.build() # THEN it has … assert e.title assert e.date assert e.venue assert e.description assert e.fb_event_url def test_event_has_slug(db): # noqa: D103 # GIVEN an event e = EventFactory.build( title='One Happy Family', date=datetime.date(2018, 1, 1), venue=None, ) assert e.slug == '' # WHEN saving the event e.save() # THEN it gets a slug generated from its date and title assert e.slug == '2018-01-01-one-happy-family' def test_event_slug_gets_updated_on_date_change(db): # noqa: D103 # GIVEN an event e = EventFactory.create( date=datetime.date(2018, 1, 1), venue=None, ) # WHEN changing the date assert e.slug.startswith('2018-01-01') e.date = datetime.date(2018, 1, 2) e.save() # THEN the slug changes to reflect the new date assert e.slug.startswith('2018-01-02')
Make sure slug gets updated on date change
Make sure slug gets updated on date change
Python
mit
FlowFX/reggae-cdmx,FlowFX/reggae-cdmx
"""Unit tests for events models.""" import datetime from app.events.factories import EventFactory from app.events.models import Event def test_event_factory(db): # noqa: D103 # GIVEN an empty database assert Event.objects.count() == 0 # WHEN saving a new event instance to the database EventFactory.create(title='five') # THEN it's there assert Event.objects.count() == 1 def test_event_has_all_the_attributes(): # noqa: D103 # GIVEN an event e = EventFactory.build() # THEN it has … assert e.title assert e.date assert e.venue assert e.description assert e.fb_event_url def test_event_has_slug(db): # noqa: D103 # GIVEN an event e = EventFactory.build( title='One Happy Family', date=datetime.date(2018, 1, 1), venue=None, ) assert e.slug == '' # WHEN saving the event e.save() # THEN it gets a slug generated from its date and title assert e.slug == '2018-01-01-one-happy-family' Make sure slug gets updated on date change
"""Unit tests for events models.""" import datetime from app.events.factories import EventFactory from app.events.models import Event def test_event_factory(db): # noqa: D103 # GIVEN an empty database assert Event.objects.count() == 0 # WHEN saving a new event instance to the database EventFactory.create(title='five') # THEN it's there assert Event.objects.count() == 1 def test_event_has_all_the_attributes(): # noqa: D103 # GIVEN an event e = EventFactory.build() # THEN it has … assert e.title assert e.date assert e.venue assert e.description assert e.fb_event_url def test_event_has_slug(db): # noqa: D103 # GIVEN an event e = EventFactory.build( title='One Happy Family', date=datetime.date(2018, 1, 1), venue=None, ) assert e.slug == '' # WHEN saving the event e.save() # THEN it gets a slug generated from its date and title assert e.slug == '2018-01-01-one-happy-family' def test_event_slug_gets_updated_on_date_change(db): # noqa: D103 # GIVEN an event e = EventFactory.create( date=datetime.date(2018, 1, 1), venue=None, ) # WHEN changing the date assert e.slug.startswith('2018-01-01') e.date = datetime.date(2018, 1, 2) e.save() # THEN the slug changes to reflect the new date assert e.slug.startswith('2018-01-02')
<commit_before>"""Unit tests for events models.""" import datetime from app.events.factories import EventFactory from app.events.models import Event def test_event_factory(db): # noqa: D103 # GIVEN an empty database assert Event.objects.count() == 0 # WHEN saving a new event instance to the database EventFactory.create(title='five') # THEN it's there assert Event.objects.count() == 1 def test_event_has_all_the_attributes(): # noqa: D103 # GIVEN an event e = EventFactory.build() # THEN it has … assert e.title assert e.date assert e.venue assert e.description assert e.fb_event_url def test_event_has_slug(db): # noqa: D103 # GIVEN an event e = EventFactory.build( title='One Happy Family', date=datetime.date(2018, 1, 1), venue=None, ) assert e.slug == '' # WHEN saving the event e.save() # THEN it gets a slug generated from its date and title assert e.slug == '2018-01-01-one-happy-family' <commit_msg>Make sure slug gets updated on date change<commit_after>
"""Unit tests for events models.""" import datetime from app.events.factories import EventFactory from app.events.models import Event def test_event_factory(db): # noqa: D103 # GIVEN an empty database assert Event.objects.count() == 0 # WHEN saving a new event instance to the database EventFactory.create(title='five') # THEN it's there assert Event.objects.count() == 1 def test_event_has_all_the_attributes(): # noqa: D103 # GIVEN an event e = EventFactory.build() # THEN it has … assert e.title assert e.date assert e.venue assert e.description assert e.fb_event_url def test_event_has_slug(db): # noqa: D103 # GIVEN an event e = EventFactory.build( title='One Happy Family', date=datetime.date(2018, 1, 1), venue=None, ) assert e.slug == '' # WHEN saving the event e.save() # THEN it gets a slug generated from its date and title assert e.slug == '2018-01-01-one-happy-family' def test_event_slug_gets_updated_on_date_change(db): # noqa: D103 # GIVEN an event e = EventFactory.create( date=datetime.date(2018, 1, 1), venue=None, ) # WHEN changing the date assert e.slug.startswith('2018-01-01') e.date = datetime.date(2018, 1, 2) e.save() # THEN the slug changes to reflect the new date assert e.slug.startswith('2018-01-02')
"""Unit tests for events models.""" import datetime from app.events.factories import EventFactory from app.events.models import Event def test_event_factory(db): # noqa: D103 # GIVEN an empty database assert Event.objects.count() == 0 # WHEN saving a new event instance to the database EventFactory.create(title='five') # THEN it's there assert Event.objects.count() == 1 def test_event_has_all_the_attributes(): # noqa: D103 # GIVEN an event e = EventFactory.build() # THEN it has … assert e.title assert e.date assert e.venue assert e.description assert e.fb_event_url def test_event_has_slug(db): # noqa: D103 # GIVEN an event e = EventFactory.build( title='One Happy Family', date=datetime.date(2018, 1, 1), venue=None, ) assert e.slug == '' # WHEN saving the event e.save() # THEN it gets a slug generated from its date and title assert e.slug == '2018-01-01-one-happy-family' Make sure slug gets updated on date change"""Unit tests for events models.""" import datetime from app.events.factories import EventFactory from app.events.models import Event def test_event_factory(db): # noqa: D103 # GIVEN an empty database assert Event.objects.count() == 0 # WHEN saving a new event instance to the database EventFactory.create(title='five') # THEN it's there assert Event.objects.count() == 1 def test_event_has_all_the_attributes(): # noqa: D103 # GIVEN an event e = EventFactory.build() # THEN it has … assert e.title assert e.date assert e.venue assert e.description assert e.fb_event_url def test_event_has_slug(db): # noqa: D103 # GIVEN an event e = EventFactory.build( title='One Happy Family', date=datetime.date(2018, 1, 1), venue=None, ) assert e.slug == '' # WHEN saving the event e.save() # THEN it gets a slug generated from its date and title assert e.slug == '2018-01-01-one-happy-family' def test_event_slug_gets_updated_on_date_change(db): # noqa: D103 # GIVEN an event e = EventFactory.create( date=datetime.date(2018, 1, 1), venue=None, ) # WHEN changing the date assert e.slug.startswith('2018-01-01') e.date = datetime.date(2018, 1, 2) e.save() # THEN the slug changes to reflect the new date assert e.slug.startswith('2018-01-02')
<commit_before>"""Unit tests for events models.""" import datetime from app.events.factories import EventFactory from app.events.models import Event def test_event_factory(db): # noqa: D103 # GIVEN an empty database assert Event.objects.count() == 0 # WHEN saving a new event instance to the database EventFactory.create(title='five') # THEN it's there assert Event.objects.count() == 1 def test_event_has_all_the_attributes(): # noqa: D103 # GIVEN an event e = EventFactory.build() # THEN it has … assert e.title assert e.date assert e.venue assert e.description assert e.fb_event_url def test_event_has_slug(db): # noqa: D103 # GIVEN an event e = EventFactory.build( title='One Happy Family', date=datetime.date(2018, 1, 1), venue=None, ) assert e.slug == '' # WHEN saving the event e.save() # THEN it gets a slug generated from its date and title assert e.slug == '2018-01-01-one-happy-family' <commit_msg>Make sure slug gets updated on date change<commit_after>"""Unit tests for events models.""" import datetime from app.events.factories import EventFactory from app.events.models import Event def test_event_factory(db): # noqa: D103 # GIVEN an empty database assert Event.objects.count() == 0 # WHEN saving a new event instance to the database EventFactory.create(title='five') # THEN it's there assert Event.objects.count() == 1 def test_event_has_all_the_attributes(): # noqa: D103 # GIVEN an event e = EventFactory.build() # THEN it has … assert e.title assert e.date assert e.venue assert e.description assert e.fb_event_url def test_event_has_slug(db): # noqa: D103 # GIVEN an event e = EventFactory.build( title='One Happy Family', date=datetime.date(2018, 1, 1), venue=None, ) assert e.slug == '' # WHEN saving the event e.save() # THEN it gets a slug generated from its date and title assert e.slug == '2018-01-01-one-happy-family' def test_event_slug_gets_updated_on_date_change(db): # noqa: D103 # GIVEN an event e = EventFactory.create( date=datetime.date(2018, 1, 1), venue=None, ) # WHEN changing the date assert e.slug.startswith('2018-01-01') e.date = datetime.date(2018, 1, 2) e.save() # THEN the slug changes to reflect the new date assert e.slug.startswith('2018-01-02')
8209b77a16c899436418dbc85dc891f671949bfc
bot/logger/message_sender/asynchronous.py
bot/logger/message_sender/asynchronous.py
from bot.logger.message_sender import IntermediateMessageSender, MessageSender from bot.multithreading.work import Work from bot.multithreading.worker import Worker class AsynchronousMessageSender(IntermediateMessageSender): def __init__(self, sender: MessageSender, worker: Worker): super().__init__(sender) self.worker = worker def send(self, text): self.worker.post(Work(lambda: self.sender.send(text), "async_message_send"))
from bot.logger.message_sender import IntermediateMessageSender, MessageSender from bot.multithreading.work import Work from bot.multithreading.worker import Worker class AsynchronousMessageSender(IntermediateMessageSender): def __init__(self, sender: MessageSender, worker: Worker): super().__init__(sender) self.worker = worker def send(self, text): self.worker.post(Work(lambda: self.sender.send(text), "asynchronous_message_sender:send"))
Clarify work action in AsynchronousMessageSender
Clarify work action in AsynchronousMessageSender
Python
agpl-3.0
alvarogzp/telegram-bot,alvarogzp/telegram-bot
from bot.logger.message_sender import IntermediateMessageSender, MessageSender from bot.multithreading.work import Work from bot.multithreading.worker import Worker class AsynchronousMessageSender(IntermediateMessageSender): def __init__(self, sender: MessageSender, worker: Worker): super().__init__(sender) self.worker = worker def send(self, text): self.worker.post(Work(lambda: self.sender.send(text), "async_message_send")) Clarify work action in AsynchronousMessageSender
from bot.logger.message_sender import IntermediateMessageSender, MessageSender from bot.multithreading.work import Work from bot.multithreading.worker import Worker class AsynchronousMessageSender(IntermediateMessageSender): def __init__(self, sender: MessageSender, worker: Worker): super().__init__(sender) self.worker = worker def send(self, text): self.worker.post(Work(lambda: self.sender.send(text), "asynchronous_message_sender:send"))
<commit_before>from bot.logger.message_sender import IntermediateMessageSender, MessageSender from bot.multithreading.work import Work from bot.multithreading.worker import Worker class AsynchronousMessageSender(IntermediateMessageSender): def __init__(self, sender: MessageSender, worker: Worker): super().__init__(sender) self.worker = worker def send(self, text): self.worker.post(Work(lambda: self.sender.send(text), "async_message_send")) <commit_msg>Clarify work action in AsynchronousMessageSender<commit_after>
from bot.logger.message_sender import IntermediateMessageSender, MessageSender from bot.multithreading.work import Work from bot.multithreading.worker import Worker class AsynchronousMessageSender(IntermediateMessageSender): def __init__(self, sender: MessageSender, worker: Worker): super().__init__(sender) self.worker = worker def send(self, text): self.worker.post(Work(lambda: self.sender.send(text), "asynchronous_message_sender:send"))
from bot.logger.message_sender import IntermediateMessageSender, MessageSender from bot.multithreading.work import Work from bot.multithreading.worker import Worker class AsynchronousMessageSender(IntermediateMessageSender): def __init__(self, sender: MessageSender, worker: Worker): super().__init__(sender) self.worker = worker def send(self, text): self.worker.post(Work(lambda: self.sender.send(text), "async_message_send")) Clarify work action in AsynchronousMessageSenderfrom bot.logger.message_sender import IntermediateMessageSender, MessageSender from bot.multithreading.work import Work from bot.multithreading.worker import Worker class AsynchronousMessageSender(IntermediateMessageSender): def __init__(self, sender: MessageSender, worker: Worker): super().__init__(sender) self.worker = worker def send(self, text): self.worker.post(Work(lambda: self.sender.send(text), "asynchronous_message_sender:send"))
<commit_before>from bot.logger.message_sender import IntermediateMessageSender, MessageSender from bot.multithreading.work import Work from bot.multithreading.worker import Worker class AsynchronousMessageSender(IntermediateMessageSender): def __init__(self, sender: MessageSender, worker: Worker): super().__init__(sender) self.worker = worker def send(self, text): self.worker.post(Work(lambda: self.sender.send(text), "async_message_send")) <commit_msg>Clarify work action in AsynchronousMessageSender<commit_after>from bot.logger.message_sender import IntermediateMessageSender, MessageSender from bot.multithreading.work import Work from bot.multithreading.worker import Worker class AsynchronousMessageSender(IntermediateMessageSender): def __init__(self, sender: MessageSender, worker: Worker): super().__init__(sender) self.worker = worker def send(self, text): self.worker.post(Work(lambda: self.sender.send(text), "asynchronous_message_sender:send"))
ccf3c508bd6750073ea3bbaefff567b92880df73
rest_framework/authtoken/serializers.py
rest_framework/authtoken/serializers.py
from django.contrib.auth import authenticate from django.utils.translation import ugettext_lazy as _ from rest_framework import serializers class AuthTokenSerializer(serializers.Serializer): username = serializers.CharField() password = serializers.CharField() def validate(self, attrs): username = attrs.get('username') password = attrs.get('password') if username and password: user = authenticate(username=username, password=password) if user: if not user.is_active: msg = _('User account is disabled.') raise serializers.ValidationError() attrs['user'] = user return attrs else: msg = _('Unable to login with provided credentials.') raise serializers.ValidationError(msg) else: msg = _('Must include "username" and "password"') raise serializers.ValidationError(msg)
from django.contrib.auth import authenticate from django.utils.translation import ugettext_lazy as _ from rest_framework import serializers class AuthTokenSerializer(serializers.Serializer): username = serializers.CharField() password = serializers.CharField() def validate(self, attrs): username = attrs.get('username') password = attrs.get('password') if username and password: user = authenticate(username=username, password=password) if user: if not user.is_active: msg = _('User account is disabled.') raise serializers.ValidationError(msg) attrs['user'] = user return attrs else: msg = _('Unable to login with provided credentials.') raise serializers.ValidationError(msg) else: msg = _('Must include "username" and "password"') raise serializers.ValidationError(msg)
Fix missing message in ValidationError
Fix missing message in ValidationError
Python
bsd-2-clause
jtiai/django-rest-framework,sbellem/django-rest-framework,kylefox/django-rest-framework,jerryhebert/django-rest-framework,zeldalink0515/django-rest-framework,adambain-vokal/django-rest-framework,tcroiset/django-rest-framework,alacritythief/django-rest-framework,johnraz/django-rest-framework,rhblind/django-rest-framework,kezabelle/django-rest-framework,cyberj/django-rest-framework,bluedazzle/django-rest-framework,andriy-s/django-rest-framework,alacritythief/django-rest-framework,antonyc/django-rest-framework,wedaly/django-rest-framework,paolopaolopaolo/django-rest-framework,paolopaolopaolo/django-rest-framework,wedaly/django-rest-framework,AlexandreProenca/django-rest-framework,werthen/django-rest-framework,rafaelang/django-rest-framework,tomchristie/django-rest-framework,tomchristie/django-rest-framework,rubendura/django-rest-framework,kgeorgy/django-rest-framework,uploadcare/django-rest-framework,waytai/django-rest-framework,antonyc/django-rest-framework,ajaali/django-rest-framework,jpadilla/django-rest-framework,uruz/django-rest-framework,potpath/django-rest-framework,d0ugal/django-rest-framework,brandoncazander/django-rest-framework,wangpanjun/django-rest-framework,maryokhin/django-rest-framework,elim/django-rest-framework,wedaly/django-rest-framework,hnakamur/django-rest-framework,davesque/django-rest-framework,kennydude/django-rest-framework,ajaali/django-rest-framework,HireAnEsquire/django-rest-framework,kezabelle/django-rest-framework,fishky/django-rest-framework,arpheno/django-rest-framework,James1345/django-rest-framework,rubendura/django-rest-framework,iheitlager/django-rest-framework,damycra/django-rest-framework,YBJAY00000/django-rest-framework,pombredanne/django-rest-framework,HireAnEsquire/django-rest-framework,HireAnEsquire/django-rest-framework,waytai/django-rest-framework,ticosax/django-rest-framework,davesque/django-rest-framework,hnakamur/django-rest-framework,MJafarMashhadi/django-rest-framework,damycra/django-rest-framework,justanr/django-rest-framework,ashishfinoit/django-rest-framework,qsorix/django-rest-framework,johnraz/django-rest-framework,fishky/django-rest-framework,gregmuellegger/django-rest-framework,cheif/django-rest-framework,xiaotangyuan/django-rest-framework,maryokhin/django-rest-framework,douwevandermeij/django-rest-framework,canassa/django-rest-framework,sehmaschine/django-rest-framework,tcroiset/django-rest-framework,ticosax/django-rest-framework,justanr/django-rest-framework,mgaitan/django-rest-framework,buptlsl/django-rest-framework,wzbozon/django-rest-framework,thedrow/django-rest-framework-1,cyberj/django-rest-framework,adambain-vokal/django-rest-framework,edx/django-rest-framework,vstoykov/django-rest-framework,zeldalink0515/django-rest-framework,kezabelle/django-rest-framework,xiaotangyuan/django-rest-framework,ambivalentno/django-rest-framework,andriy-s/django-rest-framework,nhorelik/django-rest-framework,hunter007/django-rest-framework,delinhabit/django-rest-framework,wzbozon/django-rest-framework,sbellem/django-rest-framework,jtiai/django-rest-framework,kgeorgy/django-rest-framework,thedrow/django-rest-framework-1,abdulhaq-e/django-rest-framework,krinart/django-rest-framework,maryokhin/django-rest-framework,jtiai/django-rest-framework,adambain-vokal/django-rest-framework,rafaelcaricio/django-rest-framework,iheitlager/django-rest-framework,ambivalentno/django-rest-framework,VishvajitP/django-rest-framework,ebsaral/django-rest-framework,sheppard/django-rest-framework,sheppard/django-rest-framework,abdulhaq-e/django-rest-framework,tigeraniya/django-rest-framework,lubomir/django-rest-framework,hnakamur/django-rest-framework,brandoncazander/django-rest-framework,ebsaral/django-rest-framework,delinhabit/django-rest-framework,mgaitan/django-rest-framework,aericson/django-rest-framework,elim/django-rest-framework,potpath/django-rest-framework,jness/django-rest-framework,kgeorgy/django-rest-framework,ashishfinoit/django-rest-framework,callorico/django-rest-framework,rafaelang/django-rest-framework,wangpanjun/django-rest-framework,jpadilla/django-rest-framework,James1345/django-rest-framework,antonyc/django-rest-framework,ajaali/django-rest-framework,aericson/django-rest-framework,nryoung/django-rest-framework,alacritythief/django-rest-framework,kennydude/django-rest-framework,dmwyatt/django-rest-framework,delinhabit/django-rest-framework,uploadcare/django-rest-framework,linovia/django-rest-framework,uruz/django-rest-framework,tomchristie/django-rest-framework,ezheidtmann/django-rest-framework,rafaelang/django-rest-framework,akalipetis/django-rest-framework,sheppard/django-rest-framework,raphaelmerx/django-rest-framework,leeahoward/django-rest-framework,potpath/django-rest-framework,pombredanne/django-rest-framework,nryoung/django-rest-framework,jpulec/django-rest-framework,tigeraniya/django-rest-framework,dmwyatt/django-rest-framework,d0ugal/django-rest-framework,cheif/django-rest-framework,jness/django-rest-framework,ossanna16/django-rest-framework,MJafarMashhadi/django-rest-framework,callorico/django-rest-framework,rafaelcaricio/django-rest-framework,aericson/django-rest-framework,cyberj/django-rest-framework,simudream/django-rest-framework,rhblind/django-rest-framework,wangpanjun/django-rest-framework,tcroiset/django-rest-framework,yiyocx/django-rest-framework,raphaelmerx/django-rest-framework,andriy-s/django-rest-framework,wwj718/django-rest-framework,damycra/django-rest-framework,tigeraniya/django-rest-framework,canassa/django-rest-framework,canassa/django-rest-framework,werthen/django-rest-framework,kylefox/django-rest-framework,linovia/django-rest-framework,agconti/django-rest-framework,nhorelik/django-rest-framework,vstoykov/django-rest-framework,sehmaschine/django-rest-framework,sehmaschine/django-rest-framework,MJafarMashhadi/django-rest-framework,werthen/django-rest-framework,buptlsl/django-rest-framework,krinart/django-rest-framework,zeldalink0515/django-rest-framework,AlexandreProenca/django-rest-framework,davesque/django-rest-framework,VishvajitP/django-rest-framework,ticosax/django-rest-framework,brandoncazander/django-rest-framework,akalipetis/django-rest-framework,leeahoward/django-rest-framework,kylefox/django-rest-framework,nhorelik/django-rest-framework,linovia/django-rest-framework,wzbozon/django-rest-framework,agconti/django-rest-framework,krinart/django-rest-framework,qsorix/django-rest-framework,leeahoward/django-rest-framework,rubendura/django-rest-framework,abdulhaq-e/django-rest-framework,jness/django-rest-framework,YBJAY00000/django-rest-framework,gregmuellegger/django-rest-framework,iheitlager/django-rest-framework,pombredanne/django-rest-framework,ossanna16/django-rest-framework,jerryhebert/django-rest-framework,paolopaolopaolo/django-rest-framework,AlexandreProenca/django-rest-framework,rafaelcaricio/django-rest-framework,raphaelmerx/django-rest-framework,uruz/django-rest-framework,agconti/django-rest-framework,buptlsl/django-rest-framework,cheif/django-rest-framework,bluedazzle/django-rest-framework,jerryhebert/django-rest-framework,edx/django-rest-framework,douwevandermeij/django-rest-framework,simudream/django-rest-framework,atombrella/django-rest-framework,jpulec/django-rest-framework,wwj718/django-rest-framework,douwevandermeij/django-rest-framework,thedrow/django-rest-framework-1,simudream/django-rest-framework,bluedazzle/django-rest-framework,nryoung/django-rest-framework,qsorix/django-rest-framework,ossanna16/django-rest-framework,johnraz/django-rest-framework,sbellem/django-rest-framework,waytai/django-rest-framework,gregmuellegger/django-rest-framework,d0ugal/django-rest-framework,arpheno/django-rest-framework,akalipetis/django-rest-framework,rhblind/django-rest-framework,lubomir/django-rest-framework,justanr/django-rest-framework,atombrella/django-rest-framework,edx/django-rest-framework,ashishfinoit/django-rest-framework,xiaotangyuan/django-rest-framework,arpheno/django-rest-framework,yiyocx/django-rest-framework,mgaitan/django-rest-framework,hunter007/django-rest-framework,uploadcare/django-rest-framework,lubomir/django-rest-framework,hunter007/django-rest-framework,vstoykov/django-rest-framework,James1345/django-rest-framework,ezheidtmann/django-rest-framework,atombrella/django-rest-framework,kennydude/django-rest-framework,ambivalentno/django-rest-framework,fishky/django-rest-framework,jpadilla/django-rest-framework,VishvajitP/django-rest-framework,hnarayanan/django-rest-framework,ebsaral/django-rest-framework,yiyocx/django-rest-framework,hnarayanan/django-rest-framework,hnarayanan/django-rest-framework,wwj718/django-rest-framework,dmwyatt/django-rest-framework,YBJAY00000/django-rest-framework,elim/django-rest-framework,jpulec/django-rest-framework,ezheidtmann/django-rest-framework,callorico/django-rest-framework
from django.contrib.auth import authenticate from django.utils.translation import ugettext_lazy as _ from rest_framework import serializers class AuthTokenSerializer(serializers.Serializer): username = serializers.CharField() password = serializers.CharField() def validate(self, attrs): username = attrs.get('username') password = attrs.get('password') if username and password: user = authenticate(username=username, password=password) if user: if not user.is_active: msg = _('User account is disabled.') raise serializers.ValidationError() attrs['user'] = user return attrs else: msg = _('Unable to login with provided credentials.') raise serializers.ValidationError(msg) else: msg = _('Must include "username" and "password"') raise serializers.ValidationError(msg) Fix missing message in ValidationError
from django.contrib.auth import authenticate from django.utils.translation import ugettext_lazy as _ from rest_framework import serializers class AuthTokenSerializer(serializers.Serializer): username = serializers.CharField() password = serializers.CharField() def validate(self, attrs): username = attrs.get('username') password = attrs.get('password') if username and password: user = authenticate(username=username, password=password) if user: if not user.is_active: msg = _('User account is disabled.') raise serializers.ValidationError(msg) attrs['user'] = user return attrs else: msg = _('Unable to login with provided credentials.') raise serializers.ValidationError(msg) else: msg = _('Must include "username" and "password"') raise serializers.ValidationError(msg)
<commit_before>from django.contrib.auth import authenticate from django.utils.translation import ugettext_lazy as _ from rest_framework import serializers class AuthTokenSerializer(serializers.Serializer): username = serializers.CharField() password = serializers.CharField() def validate(self, attrs): username = attrs.get('username') password = attrs.get('password') if username and password: user = authenticate(username=username, password=password) if user: if not user.is_active: msg = _('User account is disabled.') raise serializers.ValidationError() attrs['user'] = user return attrs else: msg = _('Unable to login with provided credentials.') raise serializers.ValidationError(msg) else: msg = _('Must include "username" and "password"') raise serializers.ValidationError(msg) <commit_msg>Fix missing message in ValidationError<commit_after>
from django.contrib.auth import authenticate from django.utils.translation import ugettext_lazy as _ from rest_framework import serializers class AuthTokenSerializer(serializers.Serializer): username = serializers.CharField() password = serializers.CharField() def validate(self, attrs): username = attrs.get('username') password = attrs.get('password') if username and password: user = authenticate(username=username, password=password) if user: if not user.is_active: msg = _('User account is disabled.') raise serializers.ValidationError(msg) attrs['user'] = user return attrs else: msg = _('Unable to login with provided credentials.') raise serializers.ValidationError(msg) else: msg = _('Must include "username" and "password"') raise serializers.ValidationError(msg)
from django.contrib.auth import authenticate from django.utils.translation import ugettext_lazy as _ from rest_framework import serializers class AuthTokenSerializer(serializers.Serializer): username = serializers.CharField() password = serializers.CharField() def validate(self, attrs): username = attrs.get('username') password = attrs.get('password') if username and password: user = authenticate(username=username, password=password) if user: if not user.is_active: msg = _('User account is disabled.') raise serializers.ValidationError() attrs['user'] = user return attrs else: msg = _('Unable to login with provided credentials.') raise serializers.ValidationError(msg) else: msg = _('Must include "username" and "password"') raise serializers.ValidationError(msg) Fix missing message in ValidationErrorfrom django.contrib.auth import authenticate from django.utils.translation import ugettext_lazy as _ from rest_framework import serializers class AuthTokenSerializer(serializers.Serializer): username = serializers.CharField() password = serializers.CharField() def validate(self, attrs): username = attrs.get('username') password = attrs.get('password') if username and password: user = authenticate(username=username, password=password) if user: if not user.is_active: msg = _('User account is disabled.') raise serializers.ValidationError(msg) attrs['user'] = user return attrs else: msg = _('Unable to login with provided credentials.') raise serializers.ValidationError(msg) else: msg = _('Must include "username" and "password"') raise serializers.ValidationError(msg)
<commit_before>from django.contrib.auth import authenticate from django.utils.translation import ugettext_lazy as _ from rest_framework import serializers class AuthTokenSerializer(serializers.Serializer): username = serializers.CharField() password = serializers.CharField() def validate(self, attrs): username = attrs.get('username') password = attrs.get('password') if username and password: user = authenticate(username=username, password=password) if user: if not user.is_active: msg = _('User account is disabled.') raise serializers.ValidationError() attrs['user'] = user return attrs else: msg = _('Unable to login with provided credentials.') raise serializers.ValidationError(msg) else: msg = _('Must include "username" and "password"') raise serializers.ValidationError(msg) <commit_msg>Fix missing message in ValidationError<commit_after>from django.contrib.auth import authenticate from django.utils.translation import ugettext_lazy as _ from rest_framework import serializers class AuthTokenSerializer(serializers.Serializer): username = serializers.CharField() password = serializers.CharField() def validate(self, attrs): username = attrs.get('username') password = attrs.get('password') if username and password: user = authenticate(username=username, password=password) if user: if not user.is_active: msg = _('User account is disabled.') raise serializers.ValidationError(msg) attrs['user'] = user return attrs else: msg = _('Unable to login with provided credentials.') raise serializers.ValidationError(msg) else: msg = _('Must include "username" and "password"') raise serializers.ValidationError(msg)
ac1a2191b7923a21b621c5e2ea6465e79a36e579
doc2dash/__init__.py
doc2dash/__init__.py
""" Convert docs to Dash.app's docset format. """ from __future__ import absolute_import, division, print_function __author__ = 'Hynek Schlawack' __version__ = '2.0.0' __license__ = 'MIT'
""" Convert docs to Dash.app's docset format. """ from __future__ import absolute_import, division, print_function __author__ = 'Hynek Schlawack' __version__ = '2.0.0-dev' __license__ = 'MIT'
Use -dev tag for in-dev versions
Use -dev tag for in-dev versions
Python
mit
hynek/doc2dash,hynek/doc2dash
""" Convert docs to Dash.app's docset format. """ from __future__ import absolute_import, division, print_function __author__ = 'Hynek Schlawack' __version__ = '2.0.0' __license__ = 'MIT' Use -dev tag for in-dev versions
""" Convert docs to Dash.app's docset format. """ from __future__ import absolute_import, division, print_function __author__ = 'Hynek Schlawack' __version__ = '2.0.0-dev' __license__ = 'MIT'
<commit_before>""" Convert docs to Dash.app's docset format. """ from __future__ import absolute_import, division, print_function __author__ = 'Hynek Schlawack' __version__ = '2.0.0' __license__ = 'MIT' <commit_msg>Use -dev tag for in-dev versions<commit_after>
""" Convert docs to Dash.app's docset format. """ from __future__ import absolute_import, division, print_function __author__ = 'Hynek Schlawack' __version__ = '2.0.0-dev' __license__ = 'MIT'
""" Convert docs to Dash.app's docset format. """ from __future__ import absolute_import, division, print_function __author__ = 'Hynek Schlawack' __version__ = '2.0.0' __license__ = 'MIT' Use -dev tag for in-dev versions""" Convert docs to Dash.app's docset format. """ from __future__ import absolute_import, division, print_function __author__ = 'Hynek Schlawack' __version__ = '2.0.0-dev' __license__ = 'MIT'
<commit_before>""" Convert docs to Dash.app's docset format. """ from __future__ import absolute_import, division, print_function __author__ = 'Hynek Schlawack' __version__ = '2.0.0' __license__ = 'MIT' <commit_msg>Use -dev tag for in-dev versions<commit_after>""" Convert docs to Dash.app's docset format. """ from __future__ import absolute_import, division, print_function __author__ = 'Hynek Schlawack' __version__ = '2.0.0-dev' __license__ = 'MIT'
5f6fb5866ca74793b05308ac27c4698033068cfe
tvtk/tests/test_garbage_collection.py
tvtk/tests/test_garbage_collection.py
""" Tests for the garbage collection of objects in tvtk package. """ # Authors: Deepak Surti, Ioannis Tziakos # Copyright (c) 2015, Enthought, Inc. # License: BSD Style. import unittest from traits.etsconfig.api import ETSConfig from tvtk.pyface.scene import Scene from tvtk.pyface.scene_model import SceneModel from tvtk.tests.common import TestGarbageCollection class TestTVTKGarbageCollection(TestGarbageCollection): """ See: tvtk.tests.common.TestGarbageCollection """ @unittest.skipIf( ETSConfig.toolkit=='wx', 'Test segfaults using WX (issue #216)') def test_scene(self): """ Tests if Scene can be garbage collected.""" def obj_fn(): return Scene() def close_fn(o): o.closed = True self.check_object_garbage_collected(obj_fn, close_fn) def test_scene_model(self): """ Tests if SceneModel can be garbage collected.""" def create_fn(): return SceneModel() def close_fn(obj): obj.closed = True self.check_object_garbage_collected(create_fn, close_fn)
""" Tests for the garbage collection of objects in tvtk package. """ # Authors: Deepak Surti, Ioannis Tziakos # Copyright (c) 2015, Enthought, Inc. # License: BSD Style. import unittest from traits.etsconfig.api import ETSConfig from tvtk.pyface.tvtk_scene import TVTKScene from tvtk.pyface.scene import Scene from tvtk.pyface.scene_model import SceneModel from tvtk.tests.common import TestGarbageCollection class TestTVTKGarbageCollection(TestGarbageCollection): """ See: tvtk.tests.common.TestGarbageCollection """ def test_tvtk_scene(self): """ Tests if TVTK scene can be garbage collected.""" def create_fn(): return TVTKScene() def close_fn(o): o.closed = True self.check_object_garbage_collected(create_fn, close_fn) @unittest.skipIf( ETSConfig.toolkit=='wx', 'Test segfaults using WX (issue #216)') def test_scene(self): """ Tests if Scene can be garbage collected.""" def create_fn(): return Scene() def close_fn(o): o.closed = True self.check_object_garbage_collected(create_fn, close_fn) def test_scene_model(self): """ Tests if SceneModel can be garbage collected.""" def create_fn(): return SceneModel() def close_fn(obj): obj.closed = True self.check_object_garbage_collected(create_fn, close_fn)
Test TVTKScene garbage collection and renaming
Test TVTKScene garbage collection and renaming
Python
bsd-3-clause
alexandreleroux/mayavi,alexandreleroux/mayavi,liulion/mayavi,liulion/mayavi,dmsurti/mayavi,dmsurti/mayavi
""" Tests for the garbage collection of objects in tvtk package. """ # Authors: Deepak Surti, Ioannis Tziakos # Copyright (c) 2015, Enthought, Inc. # License: BSD Style. import unittest from traits.etsconfig.api import ETSConfig from tvtk.pyface.scene import Scene from tvtk.pyface.scene_model import SceneModel from tvtk.tests.common import TestGarbageCollection class TestTVTKGarbageCollection(TestGarbageCollection): """ See: tvtk.tests.common.TestGarbageCollection """ @unittest.skipIf( ETSConfig.toolkit=='wx', 'Test segfaults using WX (issue #216)') def test_scene(self): """ Tests if Scene can be garbage collected.""" def obj_fn(): return Scene() def close_fn(o): o.closed = True self.check_object_garbage_collected(obj_fn, close_fn) def test_scene_model(self): """ Tests if SceneModel can be garbage collected.""" def create_fn(): return SceneModel() def close_fn(obj): obj.closed = True self.check_object_garbage_collected(create_fn, close_fn) Test TVTKScene garbage collection and renaming
""" Tests for the garbage collection of objects in tvtk package. """ # Authors: Deepak Surti, Ioannis Tziakos # Copyright (c) 2015, Enthought, Inc. # License: BSD Style. import unittest from traits.etsconfig.api import ETSConfig from tvtk.pyface.tvtk_scene import TVTKScene from tvtk.pyface.scene import Scene from tvtk.pyface.scene_model import SceneModel from tvtk.tests.common import TestGarbageCollection class TestTVTKGarbageCollection(TestGarbageCollection): """ See: tvtk.tests.common.TestGarbageCollection """ def test_tvtk_scene(self): """ Tests if TVTK scene can be garbage collected.""" def create_fn(): return TVTKScene() def close_fn(o): o.closed = True self.check_object_garbage_collected(create_fn, close_fn) @unittest.skipIf( ETSConfig.toolkit=='wx', 'Test segfaults using WX (issue #216)') def test_scene(self): """ Tests if Scene can be garbage collected.""" def create_fn(): return Scene() def close_fn(o): o.closed = True self.check_object_garbage_collected(create_fn, close_fn) def test_scene_model(self): """ Tests if SceneModel can be garbage collected.""" def create_fn(): return SceneModel() def close_fn(obj): obj.closed = True self.check_object_garbage_collected(create_fn, close_fn)
<commit_before>""" Tests for the garbage collection of objects in tvtk package. """ # Authors: Deepak Surti, Ioannis Tziakos # Copyright (c) 2015, Enthought, Inc. # License: BSD Style. import unittest from traits.etsconfig.api import ETSConfig from tvtk.pyface.scene import Scene from tvtk.pyface.scene_model import SceneModel from tvtk.tests.common import TestGarbageCollection class TestTVTKGarbageCollection(TestGarbageCollection): """ See: tvtk.tests.common.TestGarbageCollection """ @unittest.skipIf( ETSConfig.toolkit=='wx', 'Test segfaults using WX (issue #216)') def test_scene(self): """ Tests if Scene can be garbage collected.""" def obj_fn(): return Scene() def close_fn(o): o.closed = True self.check_object_garbage_collected(obj_fn, close_fn) def test_scene_model(self): """ Tests if SceneModel can be garbage collected.""" def create_fn(): return SceneModel() def close_fn(obj): obj.closed = True self.check_object_garbage_collected(create_fn, close_fn) <commit_msg>Test TVTKScene garbage collection and renaming<commit_after>
""" Tests for the garbage collection of objects in tvtk package. """ # Authors: Deepak Surti, Ioannis Tziakos # Copyright (c) 2015, Enthought, Inc. # License: BSD Style. import unittest from traits.etsconfig.api import ETSConfig from tvtk.pyface.tvtk_scene import TVTKScene from tvtk.pyface.scene import Scene from tvtk.pyface.scene_model import SceneModel from tvtk.tests.common import TestGarbageCollection class TestTVTKGarbageCollection(TestGarbageCollection): """ See: tvtk.tests.common.TestGarbageCollection """ def test_tvtk_scene(self): """ Tests if TVTK scene can be garbage collected.""" def create_fn(): return TVTKScene() def close_fn(o): o.closed = True self.check_object_garbage_collected(create_fn, close_fn) @unittest.skipIf( ETSConfig.toolkit=='wx', 'Test segfaults using WX (issue #216)') def test_scene(self): """ Tests if Scene can be garbage collected.""" def create_fn(): return Scene() def close_fn(o): o.closed = True self.check_object_garbage_collected(create_fn, close_fn) def test_scene_model(self): """ Tests if SceneModel can be garbage collected.""" def create_fn(): return SceneModel() def close_fn(obj): obj.closed = True self.check_object_garbage_collected(create_fn, close_fn)
""" Tests for the garbage collection of objects in tvtk package. """ # Authors: Deepak Surti, Ioannis Tziakos # Copyright (c) 2015, Enthought, Inc. # License: BSD Style. import unittest from traits.etsconfig.api import ETSConfig from tvtk.pyface.scene import Scene from tvtk.pyface.scene_model import SceneModel from tvtk.tests.common import TestGarbageCollection class TestTVTKGarbageCollection(TestGarbageCollection): """ See: tvtk.tests.common.TestGarbageCollection """ @unittest.skipIf( ETSConfig.toolkit=='wx', 'Test segfaults using WX (issue #216)') def test_scene(self): """ Tests if Scene can be garbage collected.""" def obj_fn(): return Scene() def close_fn(o): o.closed = True self.check_object_garbage_collected(obj_fn, close_fn) def test_scene_model(self): """ Tests if SceneModel can be garbage collected.""" def create_fn(): return SceneModel() def close_fn(obj): obj.closed = True self.check_object_garbage_collected(create_fn, close_fn) Test TVTKScene garbage collection and renaming""" Tests for the garbage collection of objects in tvtk package. """ # Authors: Deepak Surti, Ioannis Tziakos # Copyright (c) 2015, Enthought, Inc. # License: BSD Style. import unittest from traits.etsconfig.api import ETSConfig from tvtk.pyface.tvtk_scene import TVTKScene from tvtk.pyface.scene import Scene from tvtk.pyface.scene_model import SceneModel from tvtk.tests.common import TestGarbageCollection class TestTVTKGarbageCollection(TestGarbageCollection): """ See: tvtk.tests.common.TestGarbageCollection """ def test_tvtk_scene(self): """ Tests if TVTK scene can be garbage collected.""" def create_fn(): return TVTKScene() def close_fn(o): o.closed = True self.check_object_garbage_collected(create_fn, close_fn) @unittest.skipIf( ETSConfig.toolkit=='wx', 'Test segfaults using WX (issue #216)') def test_scene(self): """ Tests if Scene can be garbage collected.""" def create_fn(): return Scene() def close_fn(o): o.closed = True self.check_object_garbage_collected(create_fn, close_fn) def test_scene_model(self): """ Tests if SceneModel can be garbage collected.""" def create_fn(): return SceneModel() def close_fn(obj): obj.closed = True self.check_object_garbage_collected(create_fn, close_fn)
<commit_before>""" Tests for the garbage collection of objects in tvtk package. """ # Authors: Deepak Surti, Ioannis Tziakos # Copyright (c) 2015, Enthought, Inc. # License: BSD Style. import unittest from traits.etsconfig.api import ETSConfig from tvtk.pyface.scene import Scene from tvtk.pyface.scene_model import SceneModel from tvtk.tests.common import TestGarbageCollection class TestTVTKGarbageCollection(TestGarbageCollection): """ See: tvtk.tests.common.TestGarbageCollection """ @unittest.skipIf( ETSConfig.toolkit=='wx', 'Test segfaults using WX (issue #216)') def test_scene(self): """ Tests if Scene can be garbage collected.""" def obj_fn(): return Scene() def close_fn(o): o.closed = True self.check_object_garbage_collected(obj_fn, close_fn) def test_scene_model(self): """ Tests if SceneModel can be garbage collected.""" def create_fn(): return SceneModel() def close_fn(obj): obj.closed = True self.check_object_garbage_collected(create_fn, close_fn) <commit_msg>Test TVTKScene garbage collection and renaming<commit_after>""" Tests for the garbage collection of objects in tvtk package. """ # Authors: Deepak Surti, Ioannis Tziakos # Copyright (c) 2015, Enthought, Inc. # License: BSD Style. import unittest from traits.etsconfig.api import ETSConfig from tvtk.pyface.tvtk_scene import TVTKScene from tvtk.pyface.scene import Scene from tvtk.pyface.scene_model import SceneModel from tvtk.tests.common import TestGarbageCollection class TestTVTKGarbageCollection(TestGarbageCollection): """ See: tvtk.tests.common.TestGarbageCollection """ def test_tvtk_scene(self): """ Tests if TVTK scene can be garbage collected.""" def create_fn(): return TVTKScene() def close_fn(o): o.closed = True self.check_object_garbage_collected(create_fn, close_fn) @unittest.skipIf( ETSConfig.toolkit=='wx', 'Test segfaults using WX (issue #216)') def test_scene(self): """ Tests if Scene can be garbage collected.""" def create_fn(): return Scene() def close_fn(o): o.closed = True self.check_object_garbage_collected(create_fn, close_fn) def test_scene_model(self): """ Tests if SceneModel can be garbage collected.""" def create_fn(): return SceneModel() def close_fn(obj): obj.closed = True self.check_object_garbage_collected(create_fn, close_fn)
6af828b3f541adf0c42e73d76d7998bc21a072cb
sample_proj/polls/filters.py
sample_proj/polls/filters.py
from datafilters.filterform import FilterForm from datafilters.specs import (GenericSpec, DateFieldFilterSpec, GreaterThanFilterSpec, ContainsFilterSpec, GreaterThanZeroFilterSpec) class PollsFilterForm(FilterForm): has_exact_votes = GenericSpec('choice__votes') has_choice_with_votes = GreaterThanZeroFilterSpec('choice__votes') pub_date = DateFieldFilterSpec('pub_date', label='Date of publishing') has_major_choice = GreaterThanFilterSpec('choice__votes', value=50) question_contains = ContainsFilterSpec('question') choice_contains = ContainsFilterSpec('choice__choice_text')
from datafilters.filterform import FilterForm from datafilters.filterspec import FilterSpec from datafilters.specs import (DateFieldFilterSpec, GreaterThanFilterSpec, ContainsFilterSpec, GreaterThanZeroFilterSpec) class PollsFilterForm(FilterForm): has_exact_votes = FilterSpec('choice__votes') has_choice_with_votes = GreaterThanZeroFilterSpec('choice__votes') pub_date = DateFieldFilterSpec('pub_date', label='Date of publishing') has_major_choice = GreaterThanFilterSpec('choice__votes', value=50) question_contains = ContainsFilterSpec('question') choice_contains = ContainsFilterSpec('choice__choice_text')
Use FilterSpec instead of GenericSpec in sample_proj
Use FilterSpec instead of GenericSpec in sample_proj
Python
mit
zorainc/django-datafilters,freevoid/django-datafilters,zorainc/django-datafilters
from datafilters.filterform import FilterForm from datafilters.specs import (GenericSpec, DateFieldFilterSpec, GreaterThanFilterSpec, ContainsFilterSpec, GreaterThanZeroFilterSpec) class PollsFilterForm(FilterForm): has_exact_votes = GenericSpec('choice__votes') has_choice_with_votes = GreaterThanZeroFilterSpec('choice__votes') pub_date = DateFieldFilterSpec('pub_date', label='Date of publishing') has_major_choice = GreaterThanFilterSpec('choice__votes', value=50) question_contains = ContainsFilterSpec('question') choice_contains = ContainsFilterSpec('choice__choice_text') Use FilterSpec instead of GenericSpec in sample_proj
from datafilters.filterform import FilterForm from datafilters.filterspec import FilterSpec from datafilters.specs import (DateFieldFilterSpec, GreaterThanFilterSpec, ContainsFilterSpec, GreaterThanZeroFilterSpec) class PollsFilterForm(FilterForm): has_exact_votes = FilterSpec('choice__votes') has_choice_with_votes = GreaterThanZeroFilterSpec('choice__votes') pub_date = DateFieldFilterSpec('pub_date', label='Date of publishing') has_major_choice = GreaterThanFilterSpec('choice__votes', value=50) question_contains = ContainsFilterSpec('question') choice_contains = ContainsFilterSpec('choice__choice_text')
<commit_before>from datafilters.filterform import FilterForm from datafilters.specs import (GenericSpec, DateFieldFilterSpec, GreaterThanFilterSpec, ContainsFilterSpec, GreaterThanZeroFilterSpec) class PollsFilterForm(FilterForm): has_exact_votes = GenericSpec('choice__votes') has_choice_with_votes = GreaterThanZeroFilterSpec('choice__votes') pub_date = DateFieldFilterSpec('pub_date', label='Date of publishing') has_major_choice = GreaterThanFilterSpec('choice__votes', value=50) question_contains = ContainsFilterSpec('question') choice_contains = ContainsFilterSpec('choice__choice_text') <commit_msg>Use FilterSpec instead of GenericSpec in sample_proj<commit_after>
from datafilters.filterform import FilterForm from datafilters.filterspec import FilterSpec from datafilters.specs import (DateFieldFilterSpec, GreaterThanFilterSpec, ContainsFilterSpec, GreaterThanZeroFilterSpec) class PollsFilterForm(FilterForm): has_exact_votes = FilterSpec('choice__votes') has_choice_with_votes = GreaterThanZeroFilterSpec('choice__votes') pub_date = DateFieldFilterSpec('pub_date', label='Date of publishing') has_major_choice = GreaterThanFilterSpec('choice__votes', value=50) question_contains = ContainsFilterSpec('question') choice_contains = ContainsFilterSpec('choice__choice_text')
from datafilters.filterform import FilterForm from datafilters.specs import (GenericSpec, DateFieldFilterSpec, GreaterThanFilterSpec, ContainsFilterSpec, GreaterThanZeroFilterSpec) class PollsFilterForm(FilterForm): has_exact_votes = GenericSpec('choice__votes') has_choice_with_votes = GreaterThanZeroFilterSpec('choice__votes') pub_date = DateFieldFilterSpec('pub_date', label='Date of publishing') has_major_choice = GreaterThanFilterSpec('choice__votes', value=50) question_contains = ContainsFilterSpec('question') choice_contains = ContainsFilterSpec('choice__choice_text') Use FilterSpec instead of GenericSpec in sample_projfrom datafilters.filterform import FilterForm from datafilters.filterspec import FilterSpec from datafilters.specs import (DateFieldFilterSpec, GreaterThanFilterSpec, ContainsFilterSpec, GreaterThanZeroFilterSpec) class PollsFilterForm(FilterForm): has_exact_votes = FilterSpec('choice__votes') has_choice_with_votes = GreaterThanZeroFilterSpec('choice__votes') pub_date = DateFieldFilterSpec('pub_date', label='Date of publishing') has_major_choice = GreaterThanFilterSpec('choice__votes', value=50) question_contains = ContainsFilterSpec('question') choice_contains = ContainsFilterSpec('choice__choice_text')
<commit_before>from datafilters.filterform import FilterForm from datafilters.specs import (GenericSpec, DateFieldFilterSpec, GreaterThanFilterSpec, ContainsFilterSpec, GreaterThanZeroFilterSpec) class PollsFilterForm(FilterForm): has_exact_votes = GenericSpec('choice__votes') has_choice_with_votes = GreaterThanZeroFilterSpec('choice__votes') pub_date = DateFieldFilterSpec('pub_date', label='Date of publishing') has_major_choice = GreaterThanFilterSpec('choice__votes', value=50) question_contains = ContainsFilterSpec('question') choice_contains = ContainsFilterSpec('choice__choice_text') <commit_msg>Use FilterSpec instead of GenericSpec in sample_proj<commit_after>from datafilters.filterform import FilterForm from datafilters.filterspec import FilterSpec from datafilters.specs import (DateFieldFilterSpec, GreaterThanFilterSpec, ContainsFilterSpec, GreaterThanZeroFilterSpec) class PollsFilterForm(FilterForm): has_exact_votes = FilterSpec('choice__votes') has_choice_with_votes = GreaterThanZeroFilterSpec('choice__votes') pub_date = DateFieldFilterSpec('pub_date', label='Date of publishing') has_major_choice = GreaterThanFilterSpec('choice__votes', value=50) question_contains = ContainsFilterSpec('question') choice_contains = ContainsFilterSpec('choice__choice_text')
d611eb18b234c7d85371d38d8c875aaae447c231
testing/urls.py
testing/urls.py
# -*- coding: utf-8 -*- from django.conf.urls import patterns, include, url from django.views.generic import View urlpatterns = patterns('', url(r'foo$', View.as_view(), name="foo"), )
# -*- coding: utf-8 -*- from django.conf.urls import url from django.views.generic import View urlpatterns = [ url(r'foo$', View.as_view(), name="foo"), ]
Fix tests remove unneccessary imports
Fix tests remove unneccessary imports
Python
bsd-3-clause
niwinz/django-sites,niwinz/django-sites
# -*- coding: utf-8 -*- from django.conf.urls import patterns, include, url from django.views.generic import View urlpatterns = patterns('', url(r'foo$', View.as_view(), name="foo"), ) Fix tests remove unneccessary imports
# -*- coding: utf-8 -*- from django.conf.urls import url from django.views.generic import View urlpatterns = [ url(r'foo$', View.as_view(), name="foo"), ]
<commit_before># -*- coding: utf-8 -*- from django.conf.urls import patterns, include, url from django.views.generic import View urlpatterns = patterns('', url(r'foo$', View.as_view(), name="foo"), ) <commit_msg>Fix tests remove unneccessary imports<commit_after>
# -*- coding: utf-8 -*- from django.conf.urls import url from django.views.generic import View urlpatterns = [ url(r'foo$', View.as_view(), name="foo"), ]
# -*- coding: utf-8 -*- from django.conf.urls import patterns, include, url from django.views.generic import View urlpatterns = patterns('', url(r'foo$', View.as_view(), name="foo"), ) Fix tests remove unneccessary imports# -*- coding: utf-8 -*- from django.conf.urls import url from django.views.generic import View urlpatterns = [ url(r'foo$', View.as_view(), name="foo"), ]
<commit_before># -*- coding: utf-8 -*- from django.conf.urls import patterns, include, url from django.views.generic import View urlpatterns = patterns('', url(r'foo$', View.as_view(), name="foo"), ) <commit_msg>Fix tests remove unneccessary imports<commit_after># -*- coding: utf-8 -*- from django.conf.urls import url from django.views.generic import View urlpatterns = [ url(r'foo$', View.as_view(), name="foo"), ]
69ce0c86e2eb5be9b4977cc6033c5aa330c88122
tfx/utils/dsl_utils_test.py
tfx/utils/dsl_utils_test.py
# Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.utils.dsl_utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.utils import dsl_utils class DslUtilsTest(tf.test.TestCase): def csv_input(self): [csv] = dsl_utils.csv_input(uri='path') self.assertEqual('ExternalPath', csv.type_name) self.assertEqual('path', csv.uri) if __name__ == '__main__': tf.test.main()
# Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.utils.dsl_utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.utils import dsl_utils class DslUtilsTest(tf.test.TestCase): def testCsvInput(self): [csv] = dsl_utils.csv_input(uri='path') self.assertEqual('ExternalPath', csv.type_name) self.assertEqual('path', csv.uri) if __name__ == '__main__': tf.test.main()
Change a test name to CamelCase instead of snake, to conform with TFX convention.
Change a test name to CamelCase instead of snake, to conform with TFX convention. PiperOrigin-RevId: 246340014
Python
apache-2.0
tensorflow/tfx,tensorflow/tfx
# Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.utils.dsl_utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.utils import dsl_utils class DslUtilsTest(tf.test.TestCase): def csv_input(self): [csv] = dsl_utils.csv_input(uri='path') self.assertEqual('ExternalPath', csv.type_name) self.assertEqual('path', csv.uri) if __name__ == '__main__': tf.test.main() Change a test name to CamelCase instead of snake, to conform with TFX convention. PiperOrigin-RevId: 246340014
# Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.utils.dsl_utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.utils import dsl_utils class DslUtilsTest(tf.test.TestCase): def testCsvInput(self): [csv] = dsl_utils.csv_input(uri='path') self.assertEqual('ExternalPath', csv.type_name) self.assertEqual('path', csv.uri) if __name__ == '__main__': tf.test.main()
<commit_before># Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.utils.dsl_utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.utils import dsl_utils class DslUtilsTest(tf.test.TestCase): def csv_input(self): [csv] = dsl_utils.csv_input(uri='path') self.assertEqual('ExternalPath', csv.type_name) self.assertEqual('path', csv.uri) if __name__ == '__main__': tf.test.main() <commit_msg>Change a test name to CamelCase instead of snake, to conform with TFX convention. PiperOrigin-RevId: 246340014<commit_after>
# Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.utils.dsl_utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.utils import dsl_utils class DslUtilsTest(tf.test.TestCase): def testCsvInput(self): [csv] = dsl_utils.csv_input(uri='path') self.assertEqual('ExternalPath', csv.type_name) self.assertEqual('path', csv.uri) if __name__ == '__main__': tf.test.main()
# Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.utils.dsl_utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.utils import dsl_utils class DslUtilsTest(tf.test.TestCase): def csv_input(self): [csv] = dsl_utils.csv_input(uri='path') self.assertEqual('ExternalPath', csv.type_name) self.assertEqual('path', csv.uri) if __name__ == '__main__': tf.test.main() Change a test name to CamelCase instead of snake, to conform with TFX convention. PiperOrigin-RevId: 246340014# Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.utils.dsl_utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.utils import dsl_utils class DslUtilsTest(tf.test.TestCase): def testCsvInput(self): [csv] = dsl_utils.csv_input(uri='path') self.assertEqual('ExternalPath', csv.type_name) self.assertEqual('path', csv.uri) if __name__ == '__main__': tf.test.main()
<commit_before># Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.utils.dsl_utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.utils import dsl_utils class DslUtilsTest(tf.test.TestCase): def csv_input(self): [csv] = dsl_utils.csv_input(uri='path') self.assertEqual('ExternalPath', csv.type_name) self.assertEqual('path', csv.uri) if __name__ == '__main__': tf.test.main() <commit_msg>Change a test name to CamelCase instead of snake, to conform with TFX convention. PiperOrigin-RevId: 246340014<commit_after># Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.utils.dsl_utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.utils import dsl_utils class DslUtilsTest(tf.test.TestCase): def testCsvInput(self): [csv] = dsl_utils.csv_input(uri='path') self.assertEqual('ExternalPath', csv.type_name) self.assertEqual('path', csv.uri) if __name__ == '__main__': tf.test.main()
72dd3f7245f07f7c596da707dffbb67ae6aa7cd0
tohu/custom_generator_v2.py
tohu/custom_generator_v2.py
import logging logger = logging.getLogger('tohu') class CustomGeneratorMetaV2(type): def __new__(metacls, cg_name, bases, clsdict): logger.debug('[DDD] CustomGeneratorMetaV2.__new__') logger.debug(f' - metacls={metacls}') logger.debug(f' - cg_name={cg_name}') logger.debug(f' - bases={bases}') logger.debug(f' - clsdict={clsdict}') new_obj = super(CustomGeneratorMetaV2, metacls).__new__(metacls, cg_name, bases, clsdict) logger.debug(f' -- new_obj={new_obj}') def reset(self, seed=None): logger.debug(f'[EEE] Inside automatically generated reset() method for {self} (seed={seed})') logger.debug(f' TODO: reset internal seed generator and call reset() on each child generator') new_obj.reset = reset return new_obj
import logging logger = logging.getLogger('tohu') class CustomGeneratorMetaV2(type): def __new__(metacls, cg_name, bases, clsdict): logger.debug('[DDD]') logger.debug('CustomGeneratorMetaV2.__new__') logger.debug(f' - metacls={metacls}') logger.debug(f' - cg_name={cg_name}') logger.debug(f' - bases={bases}') logger.debug(f' - clsdict={clsdict}') # # Create new custom generator object # new_obj = super(CustomGeneratorMetaV2, metacls).__new__(metacls, cg_name, bases, clsdict) logger.debug(f' - new_obj={new_obj}') # # Create and assign automatically generated reset() method # def reset(self, seed=None): logger.debug(f'[EEE] Inside automatically generated reset() method for {self} (seed={seed})') logger.debug(f' TODO: reset internal seed generator and call reset() on each child generator') new_obj.reset = reset return new_obj
Tweak debugging messages, add comments
Tweak debugging messages, add comments
Python
mit
maxalbert/tohu
import logging logger = logging.getLogger('tohu') class CustomGeneratorMetaV2(type): def __new__(metacls, cg_name, bases, clsdict): logger.debug('[DDD] CustomGeneratorMetaV2.__new__') logger.debug(f' - metacls={metacls}') logger.debug(f' - cg_name={cg_name}') logger.debug(f' - bases={bases}') logger.debug(f' - clsdict={clsdict}') new_obj = super(CustomGeneratorMetaV2, metacls).__new__(metacls, cg_name, bases, clsdict) logger.debug(f' -- new_obj={new_obj}') def reset(self, seed=None): logger.debug(f'[EEE] Inside automatically generated reset() method for {self} (seed={seed})') logger.debug(f' TODO: reset internal seed generator and call reset() on each child generator') new_obj.reset = reset return new_objTweak debugging messages, add comments
import logging logger = logging.getLogger('tohu') class CustomGeneratorMetaV2(type): def __new__(metacls, cg_name, bases, clsdict): logger.debug('[DDD]') logger.debug('CustomGeneratorMetaV2.__new__') logger.debug(f' - metacls={metacls}') logger.debug(f' - cg_name={cg_name}') logger.debug(f' - bases={bases}') logger.debug(f' - clsdict={clsdict}') # # Create new custom generator object # new_obj = super(CustomGeneratorMetaV2, metacls).__new__(metacls, cg_name, bases, clsdict) logger.debug(f' - new_obj={new_obj}') # # Create and assign automatically generated reset() method # def reset(self, seed=None): logger.debug(f'[EEE] Inside automatically generated reset() method for {self} (seed={seed})') logger.debug(f' TODO: reset internal seed generator and call reset() on each child generator') new_obj.reset = reset return new_obj
<commit_before>import logging logger = logging.getLogger('tohu') class CustomGeneratorMetaV2(type): def __new__(metacls, cg_name, bases, clsdict): logger.debug('[DDD] CustomGeneratorMetaV2.__new__') logger.debug(f' - metacls={metacls}') logger.debug(f' - cg_name={cg_name}') logger.debug(f' - bases={bases}') logger.debug(f' - clsdict={clsdict}') new_obj = super(CustomGeneratorMetaV2, metacls).__new__(metacls, cg_name, bases, clsdict) logger.debug(f' -- new_obj={new_obj}') def reset(self, seed=None): logger.debug(f'[EEE] Inside automatically generated reset() method for {self} (seed={seed})') logger.debug(f' TODO: reset internal seed generator and call reset() on each child generator') new_obj.reset = reset return new_obj<commit_msg>Tweak debugging messages, add comments<commit_after>
import logging logger = logging.getLogger('tohu') class CustomGeneratorMetaV2(type): def __new__(metacls, cg_name, bases, clsdict): logger.debug('[DDD]') logger.debug('CustomGeneratorMetaV2.__new__') logger.debug(f' - metacls={metacls}') logger.debug(f' - cg_name={cg_name}') logger.debug(f' - bases={bases}') logger.debug(f' - clsdict={clsdict}') # # Create new custom generator object # new_obj = super(CustomGeneratorMetaV2, metacls).__new__(metacls, cg_name, bases, clsdict) logger.debug(f' - new_obj={new_obj}') # # Create and assign automatically generated reset() method # def reset(self, seed=None): logger.debug(f'[EEE] Inside automatically generated reset() method for {self} (seed={seed})') logger.debug(f' TODO: reset internal seed generator and call reset() on each child generator') new_obj.reset = reset return new_obj
import logging logger = logging.getLogger('tohu') class CustomGeneratorMetaV2(type): def __new__(metacls, cg_name, bases, clsdict): logger.debug('[DDD] CustomGeneratorMetaV2.__new__') logger.debug(f' - metacls={metacls}') logger.debug(f' - cg_name={cg_name}') logger.debug(f' - bases={bases}') logger.debug(f' - clsdict={clsdict}') new_obj = super(CustomGeneratorMetaV2, metacls).__new__(metacls, cg_name, bases, clsdict) logger.debug(f' -- new_obj={new_obj}') def reset(self, seed=None): logger.debug(f'[EEE] Inside automatically generated reset() method for {self} (seed={seed})') logger.debug(f' TODO: reset internal seed generator and call reset() on each child generator') new_obj.reset = reset return new_objTweak debugging messages, add commentsimport logging logger = logging.getLogger('tohu') class CustomGeneratorMetaV2(type): def __new__(metacls, cg_name, bases, clsdict): logger.debug('[DDD]') logger.debug('CustomGeneratorMetaV2.__new__') logger.debug(f' - metacls={metacls}') logger.debug(f' - cg_name={cg_name}') logger.debug(f' - bases={bases}') logger.debug(f' - clsdict={clsdict}') # # Create new custom generator object # new_obj = super(CustomGeneratorMetaV2, metacls).__new__(metacls, cg_name, bases, clsdict) logger.debug(f' - new_obj={new_obj}') # # Create and assign automatically generated reset() method # def reset(self, seed=None): logger.debug(f'[EEE] Inside automatically generated reset() method for {self} (seed={seed})') logger.debug(f' TODO: reset internal seed generator and call reset() on each child generator') new_obj.reset = reset return new_obj
<commit_before>import logging logger = logging.getLogger('tohu') class CustomGeneratorMetaV2(type): def __new__(metacls, cg_name, bases, clsdict): logger.debug('[DDD] CustomGeneratorMetaV2.__new__') logger.debug(f' - metacls={metacls}') logger.debug(f' - cg_name={cg_name}') logger.debug(f' - bases={bases}') logger.debug(f' - clsdict={clsdict}') new_obj = super(CustomGeneratorMetaV2, metacls).__new__(metacls, cg_name, bases, clsdict) logger.debug(f' -- new_obj={new_obj}') def reset(self, seed=None): logger.debug(f'[EEE] Inside automatically generated reset() method for {self} (seed={seed})') logger.debug(f' TODO: reset internal seed generator and call reset() on each child generator') new_obj.reset = reset return new_obj<commit_msg>Tweak debugging messages, add comments<commit_after>import logging logger = logging.getLogger('tohu') class CustomGeneratorMetaV2(type): def __new__(metacls, cg_name, bases, clsdict): logger.debug('[DDD]') logger.debug('CustomGeneratorMetaV2.__new__') logger.debug(f' - metacls={metacls}') logger.debug(f' - cg_name={cg_name}') logger.debug(f' - bases={bases}') logger.debug(f' - clsdict={clsdict}') # # Create new custom generator object # new_obj = super(CustomGeneratorMetaV2, metacls).__new__(metacls, cg_name, bases, clsdict) logger.debug(f' - new_obj={new_obj}') # # Create and assign automatically generated reset() method # def reset(self, seed=None): logger.debug(f'[EEE] Inside automatically generated reset() method for {self} (seed={seed})') logger.debug(f' TODO: reset internal seed generator and call reset() on each child generator') new_obj.reset = reset return new_obj
af2b561cd1a25fc4abd7c7948e5ff8ceb507a497
tests/cli/test_rasa_shell.py
tests/cli/test_rasa_shell.py
from typing import Callable from _pytest.pytester import RunResult def test_shell_help(run: Callable[..., RunResult]): output = run("shell", "--help") help_text = """usage: rasa shell [-h] [-v] [-vv] [--quiet] [-m MODEL] [--log-file LOG_FILE] [--endpoints ENDPOINTS] [-p PORT] [-t AUTH_TOKEN] [--cors [CORS [CORS ...]]] [--enable-api] [--remote-storage REMOTE_STORAGE] [--ssl-certificate SSL_CERTIFICATE] [--ssl-keyfile SSL_KEYFILE] [--ssl-ca-file SSL_CA_FILE] [--ssl-password SSL_PASSWORD] [--credentials CREDENTIALS] [--connector CONNECTOR] [--jwt-secret JWT_SECRET] [--jwt-method JWT_METHOD] {nlu} ... [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line def test_shell_nlu_help(run: Callable[..., RunResult]): output = run("shell", "nlu", "--help") help_text = """usage: rasa shell nlu [-h] [-v] [-vv] [--quiet] [-m MODEL] [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line
from typing import Callable from _pytest.pytester import RunResult def test_shell_help(run: Callable[..., RunResult]): output = run("shell", "--help") help_text = """usage: rasa shell [-h] [-v] [-vv] [--quiet] [--conversation-id CONVERSATION_ID] [-m MODEL] [--log-file LOG_FILE] [--endpoints ENDPOINTS] [-p PORT] [-t AUTH_TOKEN] [--cors [CORS [CORS ...]]] [--enable-api] [--remote-storage REMOTE_STORAGE] [--ssl-certificate SSL_CERTIFICATE] [--ssl-keyfile SSL_KEYFILE] [--ssl-ca-file SSL_CA_FILE] [--ssl-password SSL_PASSWORD] [--credentials CREDENTIALS] [--connector CONNECTOR] [--jwt-secret JWT_SECRET] [--jwt-method JWT_METHOD] {nlu} ... [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line def test_shell_nlu_help(run: Callable[..., RunResult]): output = run("shell", "nlu", "--help") help_text = """usage: rasa shell nlu [-h] [-v] [-vv] [--quiet] [-m MODEL] [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line
Adjust rasa shell help test to changes.
Adjust rasa shell help test to changes.
Python
apache-2.0
RasaHQ/rasa_nlu,RasaHQ/rasa_nlu,RasaHQ/rasa_nlu
from typing import Callable from _pytest.pytester import RunResult def test_shell_help(run: Callable[..., RunResult]): output = run("shell", "--help") help_text = """usage: rasa shell [-h] [-v] [-vv] [--quiet] [-m MODEL] [--log-file LOG_FILE] [--endpoints ENDPOINTS] [-p PORT] [-t AUTH_TOKEN] [--cors [CORS [CORS ...]]] [--enable-api] [--remote-storage REMOTE_STORAGE] [--ssl-certificate SSL_CERTIFICATE] [--ssl-keyfile SSL_KEYFILE] [--ssl-ca-file SSL_CA_FILE] [--ssl-password SSL_PASSWORD] [--credentials CREDENTIALS] [--connector CONNECTOR] [--jwt-secret JWT_SECRET] [--jwt-method JWT_METHOD] {nlu} ... [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line def test_shell_nlu_help(run: Callable[..., RunResult]): output = run("shell", "nlu", "--help") help_text = """usage: rasa shell nlu [-h] [-v] [-vv] [--quiet] [-m MODEL] [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line Adjust rasa shell help test to changes.
from typing import Callable from _pytest.pytester import RunResult def test_shell_help(run: Callable[..., RunResult]): output = run("shell", "--help") help_text = """usage: rasa shell [-h] [-v] [-vv] [--quiet] [--conversation-id CONVERSATION_ID] [-m MODEL] [--log-file LOG_FILE] [--endpoints ENDPOINTS] [-p PORT] [-t AUTH_TOKEN] [--cors [CORS [CORS ...]]] [--enable-api] [--remote-storage REMOTE_STORAGE] [--ssl-certificate SSL_CERTIFICATE] [--ssl-keyfile SSL_KEYFILE] [--ssl-ca-file SSL_CA_FILE] [--ssl-password SSL_PASSWORD] [--credentials CREDENTIALS] [--connector CONNECTOR] [--jwt-secret JWT_SECRET] [--jwt-method JWT_METHOD] {nlu} ... [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line def test_shell_nlu_help(run: Callable[..., RunResult]): output = run("shell", "nlu", "--help") help_text = """usage: rasa shell nlu [-h] [-v] [-vv] [--quiet] [-m MODEL] [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line
<commit_before>from typing import Callable from _pytest.pytester import RunResult def test_shell_help(run: Callable[..., RunResult]): output = run("shell", "--help") help_text = """usage: rasa shell [-h] [-v] [-vv] [--quiet] [-m MODEL] [--log-file LOG_FILE] [--endpoints ENDPOINTS] [-p PORT] [-t AUTH_TOKEN] [--cors [CORS [CORS ...]]] [--enable-api] [--remote-storage REMOTE_STORAGE] [--ssl-certificate SSL_CERTIFICATE] [--ssl-keyfile SSL_KEYFILE] [--ssl-ca-file SSL_CA_FILE] [--ssl-password SSL_PASSWORD] [--credentials CREDENTIALS] [--connector CONNECTOR] [--jwt-secret JWT_SECRET] [--jwt-method JWT_METHOD] {nlu} ... [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line def test_shell_nlu_help(run: Callable[..., RunResult]): output = run("shell", "nlu", "--help") help_text = """usage: rasa shell nlu [-h] [-v] [-vv] [--quiet] [-m MODEL] [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line <commit_msg>Adjust rasa shell help test to changes.<commit_after>
from typing import Callable from _pytest.pytester import RunResult def test_shell_help(run: Callable[..., RunResult]): output = run("shell", "--help") help_text = """usage: rasa shell [-h] [-v] [-vv] [--quiet] [--conversation-id CONVERSATION_ID] [-m MODEL] [--log-file LOG_FILE] [--endpoints ENDPOINTS] [-p PORT] [-t AUTH_TOKEN] [--cors [CORS [CORS ...]]] [--enable-api] [--remote-storage REMOTE_STORAGE] [--ssl-certificate SSL_CERTIFICATE] [--ssl-keyfile SSL_KEYFILE] [--ssl-ca-file SSL_CA_FILE] [--ssl-password SSL_PASSWORD] [--credentials CREDENTIALS] [--connector CONNECTOR] [--jwt-secret JWT_SECRET] [--jwt-method JWT_METHOD] {nlu} ... [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line def test_shell_nlu_help(run: Callable[..., RunResult]): output = run("shell", "nlu", "--help") help_text = """usage: rasa shell nlu [-h] [-v] [-vv] [--quiet] [-m MODEL] [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line
from typing import Callable from _pytest.pytester import RunResult def test_shell_help(run: Callable[..., RunResult]): output = run("shell", "--help") help_text = """usage: rasa shell [-h] [-v] [-vv] [--quiet] [-m MODEL] [--log-file LOG_FILE] [--endpoints ENDPOINTS] [-p PORT] [-t AUTH_TOKEN] [--cors [CORS [CORS ...]]] [--enable-api] [--remote-storage REMOTE_STORAGE] [--ssl-certificate SSL_CERTIFICATE] [--ssl-keyfile SSL_KEYFILE] [--ssl-ca-file SSL_CA_FILE] [--ssl-password SSL_PASSWORD] [--credentials CREDENTIALS] [--connector CONNECTOR] [--jwt-secret JWT_SECRET] [--jwt-method JWT_METHOD] {nlu} ... [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line def test_shell_nlu_help(run: Callable[..., RunResult]): output = run("shell", "nlu", "--help") help_text = """usage: rasa shell nlu [-h] [-v] [-vv] [--quiet] [-m MODEL] [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line Adjust rasa shell help test to changes.from typing import Callable from _pytest.pytester import RunResult def test_shell_help(run: Callable[..., RunResult]): output = run("shell", "--help") help_text = """usage: rasa shell [-h] [-v] [-vv] [--quiet] [--conversation-id CONVERSATION_ID] [-m MODEL] [--log-file LOG_FILE] [--endpoints ENDPOINTS] [-p PORT] [-t AUTH_TOKEN] [--cors [CORS [CORS ...]]] [--enable-api] [--remote-storage REMOTE_STORAGE] [--ssl-certificate SSL_CERTIFICATE] [--ssl-keyfile SSL_KEYFILE] [--ssl-ca-file SSL_CA_FILE] [--ssl-password SSL_PASSWORD] [--credentials CREDENTIALS] [--connector CONNECTOR] [--jwt-secret JWT_SECRET] [--jwt-method JWT_METHOD] {nlu} ... [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line def test_shell_nlu_help(run: Callable[..., RunResult]): output = run("shell", "nlu", "--help") help_text = """usage: rasa shell nlu [-h] [-v] [-vv] [--quiet] [-m MODEL] [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line
<commit_before>from typing import Callable from _pytest.pytester import RunResult def test_shell_help(run: Callable[..., RunResult]): output = run("shell", "--help") help_text = """usage: rasa shell [-h] [-v] [-vv] [--quiet] [-m MODEL] [--log-file LOG_FILE] [--endpoints ENDPOINTS] [-p PORT] [-t AUTH_TOKEN] [--cors [CORS [CORS ...]]] [--enable-api] [--remote-storage REMOTE_STORAGE] [--ssl-certificate SSL_CERTIFICATE] [--ssl-keyfile SSL_KEYFILE] [--ssl-ca-file SSL_CA_FILE] [--ssl-password SSL_PASSWORD] [--credentials CREDENTIALS] [--connector CONNECTOR] [--jwt-secret JWT_SECRET] [--jwt-method JWT_METHOD] {nlu} ... [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line def test_shell_nlu_help(run: Callable[..., RunResult]): output = run("shell", "nlu", "--help") help_text = """usage: rasa shell nlu [-h] [-v] [-vv] [--quiet] [-m MODEL] [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line <commit_msg>Adjust rasa shell help test to changes.<commit_after>from typing import Callable from _pytest.pytester import RunResult def test_shell_help(run: Callable[..., RunResult]): output = run("shell", "--help") help_text = """usage: rasa shell [-h] [-v] [-vv] [--quiet] [--conversation-id CONVERSATION_ID] [-m MODEL] [--log-file LOG_FILE] [--endpoints ENDPOINTS] [-p PORT] [-t AUTH_TOKEN] [--cors [CORS [CORS ...]]] [--enable-api] [--remote-storage REMOTE_STORAGE] [--ssl-certificate SSL_CERTIFICATE] [--ssl-keyfile SSL_KEYFILE] [--ssl-ca-file SSL_CA_FILE] [--ssl-password SSL_PASSWORD] [--credentials CREDENTIALS] [--connector CONNECTOR] [--jwt-secret JWT_SECRET] [--jwt-method JWT_METHOD] {nlu} ... [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line def test_shell_nlu_help(run: Callable[..., RunResult]): output = run("shell", "nlu", "--help") help_text = """usage: rasa shell nlu [-h] [-v] [-vv] [--quiet] [-m MODEL] [model-as-positional-argument]""" lines = help_text.split("\n") for i, line in enumerate(lines): assert output.outlines[i] == line
43ca4bf6a9c30055385519bb4c4472a530cce4dd
site_scons/site_tools/build_and_compress.py
site_scons/site_tools/build_and_compress.py
def build_and_compress(env, source, target): env.Tool('compactor') uncompressed = env.Program('uncompressed', source) compressed = env.Compact( target = target, source = uncompressed, ) return compressed def generate(env): env.AddMethod(build_and_compress, 'BuildAndCompress') def exists(env): return True
def build_and_compress(env, target, source): env.Tool('compactor') uncompressed = env.Program('uncompressed', source) compressed = env.Compact( target = target, source = uncompressed, ) return compressed def generate(env): env.AddMethod(build_and_compress, 'BuildAndCompress') def exists(env): return True
Fix argument order in BuildAndCompress pseudo-builder
Fix argument order in BuildAndCompress pseudo-builder
Python
mit
robobrobro/terminus
def build_and_compress(env, source, target): env.Tool('compactor') uncompressed = env.Program('uncompressed', source) compressed = env.Compact( target = target, source = uncompressed, ) return compressed def generate(env): env.AddMethod(build_and_compress, 'BuildAndCompress') def exists(env): return True Fix argument order in BuildAndCompress pseudo-builder
def build_and_compress(env, target, source): env.Tool('compactor') uncompressed = env.Program('uncompressed', source) compressed = env.Compact( target = target, source = uncompressed, ) return compressed def generate(env): env.AddMethod(build_and_compress, 'BuildAndCompress') def exists(env): return True
<commit_before>def build_and_compress(env, source, target): env.Tool('compactor') uncompressed = env.Program('uncompressed', source) compressed = env.Compact( target = target, source = uncompressed, ) return compressed def generate(env): env.AddMethod(build_and_compress, 'BuildAndCompress') def exists(env): return True <commit_msg>Fix argument order in BuildAndCompress pseudo-builder<commit_after>
def build_and_compress(env, target, source): env.Tool('compactor') uncompressed = env.Program('uncompressed', source) compressed = env.Compact( target = target, source = uncompressed, ) return compressed def generate(env): env.AddMethod(build_and_compress, 'BuildAndCompress') def exists(env): return True
def build_and_compress(env, source, target): env.Tool('compactor') uncompressed = env.Program('uncompressed', source) compressed = env.Compact( target = target, source = uncompressed, ) return compressed def generate(env): env.AddMethod(build_and_compress, 'BuildAndCompress') def exists(env): return True Fix argument order in BuildAndCompress pseudo-builderdef build_and_compress(env, target, source): env.Tool('compactor') uncompressed = env.Program('uncompressed', source) compressed = env.Compact( target = target, source = uncompressed, ) return compressed def generate(env): env.AddMethod(build_and_compress, 'BuildAndCompress') def exists(env): return True
<commit_before>def build_and_compress(env, source, target): env.Tool('compactor') uncompressed = env.Program('uncompressed', source) compressed = env.Compact( target = target, source = uncompressed, ) return compressed def generate(env): env.AddMethod(build_and_compress, 'BuildAndCompress') def exists(env): return True <commit_msg>Fix argument order in BuildAndCompress pseudo-builder<commit_after>def build_and_compress(env, target, source): env.Tool('compactor') uncompressed = env.Program('uncompressed', source) compressed = env.Compact( target = target, source = uncompressed, ) return compressed def generate(env): env.AddMethod(build_and_compress, 'BuildAndCompress') def exists(env): return True
fb7c10fd96747285e13f89a5415da8e09c8beae6
setup.py
setup.py
from setuptools import find_packages, setup setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
from setuptools import find_packages, setup setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
Update Development Status from Planning to Alpha
Update Development Status from Planning to Alpha
Python
bsd-2-clause
incuna/incuna-pigeon
from setuptools import find_packages, setup setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), ) Update Development Status from Planning to Alpha
from setuptools import find_packages, setup setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
<commit_before>from setuptools import find_packages, setup setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), ) <commit_msg>Update Development Status from Planning to Alpha<commit_after>
from setuptools import find_packages, setup setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
from setuptools import find_packages, setup setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), ) Update Development Status from Planning to Alphafrom setuptools import find_packages, setup setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
<commit_before>from setuptools import find_packages, setup setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), ) <commit_msg>Update Development Status from Planning to Alpha<commit_after>from setuptools import find_packages, setup setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
cc7da0c83a27e42619f0d800b11ef7bf7f82e4bb
setup.py
setup.py
from distutils.core import setup setup( name = 'sdnotify', packages = ['sdnotify'], version = '0.3.0', description = 'A pure Python implementation of systemd\'s service notification protocol (sd_notify)', author = 'Brett Bethke', author_email = 'bbethke@gmail.com', url = 'https://github.com/bb4242/sdnotify', download_url = 'https://github.com/bb4242/sdnotify/tarball/v0.3.0', keywords = ['systemd'], classifiers = [ "Programming Language :: Python", "Programming Language :: Python :: 3", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description = """\ systemd Service Notification This is a pure Python implementation of the systemd sd_notify protocol. This protocol can be used to inform systemd about service start-up completion, watchdog events, and other service status changes. Thus, this package can be used to write system services in Python that play nicely with systemd. sdnotify is compatible with both Python 2 and Python 3. """ )
from distutils.core import setup VERSION='0.3.1' setup( name = 'sdnotify', packages = ['sdnotify'], version = VERSION, description = 'A pure Python implementation of systemd\'s service notification protocol (sd_notify)', author = 'Brett Bethke', author_email = 'bbethke@gmail.com', url = 'https://github.com/bb4242/sdnotify', download_url = 'https://github.com/bb4242/sdnotify/tarball/v{}'.format(VERSION), keywords = ['systemd'], classifiers = [ "Programming Language :: Python", "Programming Language :: Python :: 3", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Topic :: Software Development :: Libraries :: Python Modules", ], data_files = [("", ["LICENSE.txt"])], long_description = """\ systemd Service Notification This is a pure Python implementation of the systemd sd_notify protocol. This protocol can be used to inform systemd about service start-up completion, watchdog events, and other service status changes. Thus, this package can be used to write system services in Python that play nicely with systemd. sdnotify is compatible with both Python 2 and Python 3. """ )
Include LICENSE.txt in the package tarball
Include LICENSE.txt in the package tarball
Python
mit
bb4242/sdnotify
from distutils.core import setup setup( name = 'sdnotify', packages = ['sdnotify'], version = '0.3.0', description = 'A pure Python implementation of systemd\'s service notification protocol (sd_notify)', author = 'Brett Bethke', author_email = 'bbethke@gmail.com', url = 'https://github.com/bb4242/sdnotify', download_url = 'https://github.com/bb4242/sdnotify/tarball/v0.3.0', keywords = ['systemd'], classifiers = [ "Programming Language :: Python", "Programming Language :: Python :: 3", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description = """\ systemd Service Notification This is a pure Python implementation of the systemd sd_notify protocol. This protocol can be used to inform systemd about service start-up completion, watchdog events, and other service status changes. Thus, this package can be used to write system services in Python that play nicely with systemd. sdnotify is compatible with both Python 2 and Python 3. """ ) Include LICENSE.txt in the package tarball
from distutils.core import setup VERSION='0.3.1' setup( name = 'sdnotify', packages = ['sdnotify'], version = VERSION, description = 'A pure Python implementation of systemd\'s service notification protocol (sd_notify)', author = 'Brett Bethke', author_email = 'bbethke@gmail.com', url = 'https://github.com/bb4242/sdnotify', download_url = 'https://github.com/bb4242/sdnotify/tarball/v{}'.format(VERSION), keywords = ['systemd'], classifiers = [ "Programming Language :: Python", "Programming Language :: Python :: 3", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Topic :: Software Development :: Libraries :: Python Modules", ], data_files = [("", ["LICENSE.txt"])], long_description = """\ systemd Service Notification This is a pure Python implementation of the systemd sd_notify protocol. This protocol can be used to inform systemd about service start-up completion, watchdog events, and other service status changes. Thus, this package can be used to write system services in Python that play nicely with systemd. sdnotify is compatible with both Python 2 and Python 3. """ )
<commit_before>from distutils.core import setup setup( name = 'sdnotify', packages = ['sdnotify'], version = '0.3.0', description = 'A pure Python implementation of systemd\'s service notification protocol (sd_notify)', author = 'Brett Bethke', author_email = 'bbethke@gmail.com', url = 'https://github.com/bb4242/sdnotify', download_url = 'https://github.com/bb4242/sdnotify/tarball/v0.3.0', keywords = ['systemd'], classifiers = [ "Programming Language :: Python", "Programming Language :: Python :: 3", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description = """\ systemd Service Notification This is a pure Python implementation of the systemd sd_notify protocol. This protocol can be used to inform systemd about service start-up completion, watchdog events, and other service status changes. Thus, this package can be used to write system services in Python that play nicely with systemd. sdnotify is compatible with both Python 2 and Python 3. """ ) <commit_msg>Include LICENSE.txt in the package tarball<commit_after>
from distutils.core import setup VERSION='0.3.1' setup( name = 'sdnotify', packages = ['sdnotify'], version = VERSION, description = 'A pure Python implementation of systemd\'s service notification protocol (sd_notify)', author = 'Brett Bethke', author_email = 'bbethke@gmail.com', url = 'https://github.com/bb4242/sdnotify', download_url = 'https://github.com/bb4242/sdnotify/tarball/v{}'.format(VERSION), keywords = ['systemd'], classifiers = [ "Programming Language :: Python", "Programming Language :: Python :: 3", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Topic :: Software Development :: Libraries :: Python Modules", ], data_files = [("", ["LICENSE.txt"])], long_description = """\ systemd Service Notification This is a pure Python implementation of the systemd sd_notify protocol. This protocol can be used to inform systemd about service start-up completion, watchdog events, and other service status changes. Thus, this package can be used to write system services in Python that play nicely with systemd. sdnotify is compatible with both Python 2 and Python 3. """ )
from distutils.core import setup setup( name = 'sdnotify', packages = ['sdnotify'], version = '0.3.0', description = 'A pure Python implementation of systemd\'s service notification protocol (sd_notify)', author = 'Brett Bethke', author_email = 'bbethke@gmail.com', url = 'https://github.com/bb4242/sdnotify', download_url = 'https://github.com/bb4242/sdnotify/tarball/v0.3.0', keywords = ['systemd'], classifiers = [ "Programming Language :: Python", "Programming Language :: Python :: 3", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description = """\ systemd Service Notification This is a pure Python implementation of the systemd sd_notify protocol. This protocol can be used to inform systemd about service start-up completion, watchdog events, and other service status changes. Thus, this package can be used to write system services in Python that play nicely with systemd. sdnotify is compatible with both Python 2 and Python 3. """ ) Include LICENSE.txt in the package tarballfrom distutils.core import setup VERSION='0.3.1' setup( name = 'sdnotify', packages = ['sdnotify'], version = VERSION, description = 'A pure Python implementation of systemd\'s service notification protocol (sd_notify)', author = 'Brett Bethke', author_email = 'bbethke@gmail.com', url = 'https://github.com/bb4242/sdnotify', download_url = 'https://github.com/bb4242/sdnotify/tarball/v{}'.format(VERSION), keywords = ['systemd'], classifiers = [ "Programming Language :: Python", "Programming Language :: Python :: 3", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Topic :: Software Development :: Libraries :: Python Modules", ], data_files = [("", ["LICENSE.txt"])], long_description = """\ systemd Service Notification This is a pure Python implementation of the systemd sd_notify protocol. This protocol can be used to inform systemd about service start-up completion, watchdog events, and other service status changes. Thus, this package can be used to write system services in Python that play nicely with systemd. sdnotify is compatible with both Python 2 and Python 3. """ )
<commit_before>from distutils.core import setup setup( name = 'sdnotify', packages = ['sdnotify'], version = '0.3.0', description = 'A pure Python implementation of systemd\'s service notification protocol (sd_notify)', author = 'Brett Bethke', author_email = 'bbethke@gmail.com', url = 'https://github.com/bb4242/sdnotify', download_url = 'https://github.com/bb4242/sdnotify/tarball/v0.3.0', keywords = ['systemd'], classifiers = [ "Programming Language :: Python", "Programming Language :: Python :: 3", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description = """\ systemd Service Notification This is a pure Python implementation of the systemd sd_notify protocol. This protocol can be used to inform systemd about service start-up completion, watchdog events, and other service status changes. Thus, this package can be used to write system services in Python that play nicely with systemd. sdnotify is compatible with both Python 2 and Python 3. """ ) <commit_msg>Include LICENSE.txt in the package tarball<commit_after>from distutils.core import setup VERSION='0.3.1' setup( name = 'sdnotify', packages = ['sdnotify'], version = VERSION, description = 'A pure Python implementation of systemd\'s service notification protocol (sd_notify)', author = 'Brett Bethke', author_email = 'bbethke@gmail.com', url = 'https://github.com/bb4242/sdnotify', download_url = 'https://github.com/bb4242/sdnotify/tarball/v{}'.format(VERSION), keywords = ['systemd'], classifiers = [ "Programming Language :: Python", "Programming Language :: Python :: 3", "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX :: Linux", "Topic :: Software Development :: Libraries :: Python Modules", ], data_files = [("", ["LICENSE.txt"])], long_description = """\ systemd Service Notification This is a pure Python implementation of the systemd sd_notify protocol. This protocol can be used to inform systemd about service start-up completion, watchdog events, and other service status changes. Thus, this package can be used to write system services in Python that play nicely with systemd. sdnotify is compatible with both Python 2 and Python 3. """ )
9c9a461cf85d123a249460aa306d8300c515ac81
setup.py
setup.py
from setuptools import setup setup(name='tfr', version='0.1', description='Time-frequency reassigned spectrograms', url='http://github.com/bzamecnik/tfr', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', packages=['tfr'], zip_safe=False, install_requires=[ 'numpy', 'scikit-learn', 'scipy', 'soundfile', ], # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Science/Research', 'Topic :: Multimedia :: Sound/Audio :: Analysis', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Operating System :: POSIX :: Linux', 'Operating System :: MacOS :: MacOS X', ])
from setuptools import setup setup(name='tfr', version='0.1', description='Time-frequency reassigned spectrograms', url='http://github.com/bzamecnik/tfr', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', packages=['tfr'], zip_safe=False, install_requires=[ 'numpy', 'scikit-learn', 'scipy', 'soundfile', ], setup_requires=['setuptools-markdown'], long_description_markdown_filename='README.md', # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Science/Research', 'Topic :: Multimedia :: Sound/Audio :: Analysis', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Operating System :: POSIX :: Linux', 'Operating System :: MacOS :: MacOS X', ])
Add the markdown readme to the pip package.
Add the markdown readme to the pip package.
Python
mit
bzamecnik/tfr,bzamecnik/tfr
from setuptools import setup setup(name='tfr', version='0.1', description='Time-frequency reassigned spectrograms', url='http://github.com/bzamecnik/tfr', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', packages=['tfr'], zip_safe=False, install_requires=[ 'numpy', 'scikit-learn', 'scipy', 'soundfile', ], # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Science/Research', 'Topic :: Multimedia :: Sound/Audio :: Analysis', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Operating System :: POSIX :: Linux', 'Operating System :: MacOS :: MacOS X', ]) Add the markdown readme to the pip package.
from setuptools import setup setup(name='tfr', version='0.1', description='Time-frequency reassigned spectrograms', url='http://github.com/bzamecnik/tfr', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', packages=['tfr'], zip_safe=False, install_requires=[ 'numpy', 'scikit-learn', 'scipy', 'soundfile', ], setup_requires=['setuptools-markdown'], long_description_markdown_filename='README.md', # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Science/Research', 'Topic :: Multimedia :: Sound/Audio :: Analysis', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Operating System :: POSIX :: Linux', 'Operating System :: MacOS :: MacOS X', ])
<commit_before>from setuptools import setup setup(name='tfr', version='0.1', description='Time-frequency reassigned spectrograms', url='http://github.com/bzamecnik/tfr', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', packages=['tfr'], zip_safe=False, install_requires=[ 'numpy', 'scikit-learn', 'scipy', 'soundfile', ], # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Science/Research', 'Topic :: Multimedia :: Sound/Audio :: Analysis', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Operating System :: POSIX :: Linux', 'Operating System :: MacOS :: MacOS X', ]) <commit_msg>Add the markdown readme to the pip package.<commit_after>
from setuptools import setup setup(name='tfr', version='0.1', description='Time-frequency reassigned spectrograms', url='http://github.com/bzamecnik/tfr', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', packages=['tfr'], zip_safe=False, install_requires=[ 'numpy', 'scikit-learn', 'scipy', 'soundfile', ], setup_requires=['setuptools-markdown'], long_description_markdown_filename='README.md', # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Science/Research', 'Topic :: Multimedia :: Sound/Audio :: Analysis', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Operating System :: POSIX :: Linux', 'Operating System :: MacOS :: MacOS X', ])
from setuptools import setup setup(name='tfr', version='0.1', description='Time-frequency reassigned spectrograms', url='http://github.com/bzamecnik/tfr', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', packages=['tfr'], zip_safe=False, install_requires=[ 'numpy', 'scikit-learn', 'scipy', 'soundfile', ], # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Science/Research', 'Topic :: Multimedia :: Sound/Audio :: Analysis', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Operating System :: POSIX :: Linux', 'Operating System :: MacOS :: MacOS X', ]) Add the markdown readme to the pip package.from setuptools import setup setup(name='tfr', version='0.1', description='Time-frequency reassigned spectrograms', url='http://github.com/bzamecnik/tfr', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', packages=['tfr'], zip_safe=False, install_requires=[ 'numpy', 'scikit-learn', 'scipy', 'soundfile', ], setup_requires=['setuptools-markdown'], long_description_markdown_filename='README.md', # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Science/Research', 'Topic :: Multimedia :: Sound/Audio :: Analysis', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Operating System :: POSIX :: Linux', 'Operating System :: MacOS :: MacOS X', ])
<commit_before>from setuptools import setup setup(name='tfr', version='0.1', description='Time-frequency reassigned spectrograms', url='http://github.com/bzamecnik/tfr', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', packages=['tfr'], zip_safe=False, install_requires=[ 'numpy', 'scikit-learn', 'scipy', 'soundfile', ], # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Science/Research', 'Topic :: Multimedia :: Sound/Audio :: Analysis', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Operating System :: POSIX :: Linux', 'Operating System :: MacOS :: MacOS X', ]) <commit_msg>Add the markdown readme to the pip package.<commit_after>from setuptools import setup setup(name='tfr', version='0.1', description='Time-frequency reassigned spectrograms', url='http://github.com/bzamecnik/tfr', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', packages=['tfr'], zip_safe=False, install_requires=[ 'numpy', 'scikit-learn', 'scipy', 'soundfile', ], setup_requires=['setuptools-markdown'], long_description_markdown_filename='README.md', # See https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ # How mature is this project? Common values are # 3 - Alpha # 4 - Beta # 5 - Production/Stable 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Intended Audience :: Science/Research', 'Topic :: Multimedia :: Sound/Audio :: Analysis', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', 'Operating System :: POSIX :: Linux', 'Operating System :: MacOS :: MacOS X', ])
935c1bfa6fedea6ee3f1d4329c3632fecb7a4bdc
setup.py
setup.py
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-sentry', version='.'.join(map(str, __import__('sentry').__version__)), author='David Cramer', author_email='dcramer@gmail.com', url='http://github.com/dcramer/django-sentry', description = 'Exception Logging to a Database in Django', packages=find_packages(), zip_safe=False, install_requires=[ # Optional 'django-paging>=0.2.2', 'django-indexer==0.2', ], include_package_data=True, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-sentry', version='.'.join(map(str, __import__('sentry').__version__)), author='David Cramer', author_email='dcramer@gmail.com', url='http://github.com/dcramer/django-sentry', description = 'Exception Logging to a Database in Django', packages=find_packages(), zip_safe=False, install_requires=[ 'django-paging>=0.2.2', 'django-indexer==0.2', ], include_package_data=True, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
Remove all info pertaining to pygooglechart
Remove all info pertaining to pygooglechart
Python
bsd-3-clause
primepix/django-sentry,kevinlondon/sentry,JTCunning/sentry,Goldmund-Wyldebeast-Wunderliebe/raven-python,imankulov/sentry,zenefits/sentry,arthurlogilab/raven-python,johansteffner/raven-python,Kryz/sentry,smarkets/raven-python,ewdurbin/raven-python,kevinastone/sentry,songyi199111/sentry,rdio/sentry,mvaled/sentry,beniwohli/apm-agent-python,dirtycoder/opbeat_python,SilentCircle/sentry,icereval/raven-python,wong2/sentry,arthurlogilab/raven-python,jbarbuto/raven-python,SilentCircle/sentry,JTCunning/sentry,fuziontech/sentry,jbarbuto/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,percipient/raven-python,ticosax/opbeat_python,mitsuhiko/raven,patrys/opbeat_python,jean/sentry,WoLpH/django-sentry,wujuguang/sentry,alex/sentry,looker/sentry,arthurlogilab/raven-python,jokey2k/sentry,ifduyue/sentry,1tush/sentry,icereval/raven-python,jmagnusson/raven-python,BayanGroup/sentry,beeftornado/sentry,arthurlogilab/raven-python,pauloschilling/sentry,alex/sentry,tbarbugli/sentry_fork,akalipetis/raven-python,TedaLIEz/sentry,chayapan/django-sentry,Goldmund-Wyldebeast-Wunderliebe/raven-python,percipient/raven-python,danriti/raven-python,SilentCircle/sentry,mvaled/sentry,dbravender/raven-python,gg7/sentry,JamesMura/sentry,gg7/sentry,mvaled/sentry,gencer/sentry,drcapulet/sentry,ronaldevers/raven-python,ifduyue/sentry,ngonzalvez/sentry,jean/sentry,SilentCircle/sentry,akheron/raven-python,ronaldevers/raven-python,wujuguang/sentry,nicholasserra/sentry,dbravender/raven-python,johansteffner/raven-python,mitsuhiko/sentry,1tush/sentry,Goldmund-Wyldebeast-Wunderliebe/raven-python,NickPresta/sentry,jean/sentry,smarkets/raven-python,pauloschilling/sentry,rdio/sentry,BuildingLink/sentry,tbarbugli/sentry_fork,fuziontech/sentry,tbarbugli/sentry_fork,getsentry/raven-python,Photonomie/raven-python,beni55/sentry,TedaLIEz/sentry,Photonomie/raven-python,argonemyth/sentry,songyi199111/sentry,fotinakis/sentry,boneyao/sentry,nikolas/raven-python,jmagnusson/raven-python,fotinakis/sentry,jmagnusson/raven-python,someonehan/raven-python,BuildingLink/sentry,felixbuenemann/sentry,Natim/sentry,nikolas/raven-python,jean/sentry,Photonomie/raven-python,imankulov/sentry,alexm92/sentry,hzy/raven-python,akheron/raven-python,hongliang5623/sentry,tarkatronic/opbeat_python,JamesMura/sentry,recht/raven-python,Kronuz/django-sentry,smarkets/raven-python,beeftornado/sentry,gg7/sentry,drcapulet/sentry,gencer/sentry,akalipetis/raven-python,primepix/django-sentry,recht/raven-python,BayanGroup/sentry,camilonova/sentry,beni55/sentry,TedaLIEz/sentry,JackDanger/sentry,gencer/sentry,beniwohli/apm-agent-python,WoLpH/django-sentry,BuildingLink/sentry,mvaled/sentry,korealerts1/sentry,tarkatronic/opbeat_python,camilonova/sentry,jokey2k/sentry,inspirehep/raven-python,primepix/django-sentry,songyi199111/sentry,kevinlondon/sentry,drcapulet/sentry,ifduyue/sentry,akheron/raven-python,zenefits/sentry,daikeren/opbeat_python,nicholasserra/sentry,llonchj/sentry,JackDanger/sentry,fotinakis/sentry,jmp0xf/raven-python,tarkatronic/opbeat_python,ticosax/opbeat_python,JackDanger/sentry,alex/raven,fuziontech/sentry,JamesMura/sentry,felixbuenemann/sentry,daevaorn/sentry,smarkets/raven-python,mitsuhiko/sentry,Kronuz/django-sentry,lepture/raven-python,akalipetis/raven-python,dcramer/sentry-old,inspirehep/raven-python,ifduyue/sentry,wujuguang/sentry,beniwohli/apm-agent-python,felixbuenemann/sentry,jbarbuto/raven-python,gencer/sentry,ngonzalvez/sentry,ewdurbin/raven-python,patrys/opbeat_python,patrys/opbeat_python,mvaled/sentry,ticosax/opbeat_python,hongliang5623/sentry,chayapan/django-sentry,jokey2k/sentry,icereval/raven-python,lopter/raven-python-old,BuildingLink/sentry,korealerts1/sentry,beeftornado/sentry,nicholasserra/sentry,pauloschilling/sentry,wong2/sentry,boneyao/sentry,rdio/sentry,dirtycoder/opbeat_python,daevaorn/sentry,ngonzalvez/sentry,hzy/raven-python,wong2/sentry,someonehan/raven-python,zenefits/sentry,dcramer/sentry-old,JamesMura/sentry,recht/raven-python,looker/sentry,getsentry/raven-python,BuildingLink/sentry,ewdurbin/sentry,argonemyth/sentry,inspirehep/raven-python,JamesMura/sentry,Natim/sentry,hongliang5623/sentry,llonchj/sentry,imankulov/sentry,beni55/sentry,vperron/sentry,fotinakis/sentry,zenefits/sentry,1tush/sentry,ewdurbin/raven-python,NickPresta/sentry,Kronuz/django-sentry,JTCunning/sentry,kevinlondon/sentry,rdio/sentry,daevaorn/sentry,hzy/raven-python,dirtycoder/opbeat_python,danriti/raven-python,beniwohli/apm-agent-python,vperron/sentry,NickPresta/sentry,ronaldevers/raven-python,mitsuhiko/raven,chayapan/django-sentry,getsentry/raven-python,kevinastone/sentry,danriti/raven-python,ewdurbin/sentry,zenefits/sentry,lepture/raven-python,nikolas/raven-python,Kryz/sentry,patrys/opbeat_python,jmp0xf/raven-python,jmp0xf/raven-python,someonehan/raven-python,BayanGroup/sentry,inspirehep/raven-python,looker/sentry,ewdurbin/sentry,Natim/sentry,kevinastone/sentry,NickPresta/sentry,camilonova/sentry,boneyao/sentry,nikolas/raven-python,korealerts1/sentry,ifduyue/sentry,dbravender/raven-python,alexm92/sentry,Kryz/sentry,percipient/raven-python,collective/mr.poe,johansteffner/raven-python,mvaled/sentry,looker/sentry,icereval/raven-python,WoLpH/django-sentry,argonemyth/sentry,lepture/raven-python,openlabs/raven,alexm92/sentry,alex/sentry,llonchj/sentry,vperron/sentry,gencer/sentry,jbarbuto/raven-python,daikeren/opbeat_python,jean/sentry,looker/sentry,daikeren/opbeat_python,dcramer/sentry-old,daevaorn/sentry
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-sentry', version='.'.join(map(str, __import__('sentry').__version__)), author='David Cramer', author_email='dcramer@gmail.com', url='http://github.com/dcramer/django-sentry', description = 'Exception Logging to a Database in Django', packages=find_packages(), zip_safe=False, install_requires=[ # Optional 'django-paging>=0.2.2', 'django-indexer==0.2', ], include_package_data=True, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )Remove all info pertaining to pygooglechart
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-sentry', version='.'.join(map(str, __import__('sentry').__version__)), author='David Cramer', author_email='dcramer@gmail.com', url='http://github.com/dcramer/django-sentry', description = 'Exception Logging to a Database in Django', packages=find_packages(), zip_safe=False, install_requires=[ 'django-paging>=0.2.2', 'django-indexer==0.2', ], include_package_data=True, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-sentry', version='.'.join(map(str, __import__('sentry').__version__)), author='David Cramer', author_email='dcramer@gmail.com', url='http://github.com/dcramer/django-sentry', description = 'Exception Logging to a Database in Django', packages=find_packages(), zip_safe=False, install_requires=[ # Optional 'django-paging>=0.2.2', 'django-indexer==0.2', ], include_package_data=True, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )<commit_msg>Remove all info pertaining to pygooglechart<commit_after>
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-sentry', version='.'.join(map(str, __import__('sentry').__version__)), author='David Cramer', author_email='dcramer@gmail.com', url='http://github.com/dcramer/django-sentry', description = 'Exception Logging to a Database in Django', packages=find_packages(), zip_safe=False, install_requires=[ 'django-paging>=0.2.2', 'django-indexer==0.2', ], include_package_data=True, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-sentry', version='.'.join(map(str, __import__('sentry').__version__)), author='David Cramer', author_email='dcramer@gmail.com', url='http://github.com/dcramer/django-sentry', description = 'Exception Logging to a Database in Django', packages=find_packages(), zip_safe=False, install_requires=[ # Optional 'django-paging>=0.2.2', 'django-indexer==0.2', ], include_package_data=True, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )Remove all info pertaining to pygooglechart#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-sentry', version='.'.join(map(str, __import__('sentry').__version__)), author='David Cramer', author_email='dcramer@gmail.com', url='http://github.com/dcramer/django-sentry', description = 'Exception Logging to a Database in Django', packages=find_packages(), zip_safe=False, install_requires=[ 'django-paging>=0.2.2', 'django-indexer==0.2', ], include_package_data=True, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-sentry', version='.'.join(map(str, __import__('sentry').__version__)), author='David Cramer', author_email='dcramer@gmail.com', url='http://github.com/dcramer/django-sentry', description = 'Exception Logging to a Database in Django', packages=find_packages(), zip_safe=False, install_requires=[ # Optional 'django-paging>=0.2.2', 'django-indexer==0.2', ], include_package_data=True, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )<commit_msg>Remove all info pertaining to pygooglechart<commit_after>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-sentry', version='.'.join(map(str, __import__('sentry').__version__)), author='David Cramer', author_email='dcramer@gmail.com', url='http://github.com/dcramer/django-sentry', description = 'Exception Logging to a Database in Django', packages=find_packages(), zip_safe=False, install_requires=[ 'django-paging>=0.2.2', 'django-indexer==0.2', ], include_package_data=True, classifiers=[ 'Framework :: Django', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Operating System :: OS Independent', 'Topic :: Software Development' ], )
c486d44cbb6007d5a89f36746822e68ea8cb0afa
setup.py
setup.py
from setuptools import setup, find_packages setup( name="cumulus", version="0.1.0", description="Girder API endpoints for interacting with cloud providers.", author="Chris Haris", author_email="chris.harris@kitware.com", url="https://github.com/Kitware/cumulus", packages=find_packages(), package_data={ "": ["*.json"], "cumulus": ["conf/*.json"], })
from setuptools import setup, find_packages setup( name="cumulus", version="0.1.0", description="Girder API endpoints for interacting with cloud providers.", author="Chris Haris", author_email="chris.harris@kitware.com", url="https://github.com/Kitware/cumulus", packages=find_packages(exclude=["*.tests", "*.tests.*", "tests.*", "tests"]), package_data={ "": ["*.json", "*.sh"], "cumulus": ["conf/*.json"], })
Exclude unit tests from install
Exclude unit tests from install
Python
apache-2.0
Kitware/cumulus,Kitware/cumulus,cjh1/cumulus,cjh1/cumulus
from setuptools import setup, find_packages setup( name="cumulus", version="0.1.0", description="Girder API endpoints for interacting with cloud providers.", author="Chris Haris", author_email="chris.harris@kitware.com", url="https://github.com/Kitware/cumulus", packages=find_packages(), package_data={ "": ["*.json"], "cumulus": ["conf/*.json"], }) Exclude unit tests from install
from setuptools import setup, find_packages setup( name="cumulus", version="0.1.0", description="Girder API endpoints for interacting with cloud providers.", author="Chris Haris", author_email="chris.harris@kitware.com", url="https://github.com/Kitware/cumulus", packages=find_packages(exclude=["*.tests", "*.tests.*", "tests.*", "tests"]), package_data={ "": ["*.json", "*.sh"], "cumulus": ["conf/*.json"], })
<commit_before>from setuptools import setup, find_packages setup( name="cumulus", version="0.1.0", description="Girder API endpoints for interacting with cloud providers.", author="Chris Haris", author_email="chris.harris@kitware.com", url="https://github.com/Kitware/cumulus", packages=find_packages(), package_data={ "": ["*.json"], "cumulus": ["conf/*.json"], }) <commit_msg>Exclude unit tests from install<commit_after>
from setuptools import setup, find_packages setup( name="cumulus", version="0.1.0", description="Girder API endpoints for interacting with cloud providers.", author="Chris Haris", author_email="chris.harris@kitware.com", url="https://github.com/Kitware/cumulus", packages=find_packages(exclude=["*.tests", "*.tests.*", "tests.*", "tests"]), package_data={ "": ["*.json", "*.sh"], "cumulus": ["conf/*.json"], })
from setuptools import setup, find_packages setup( name="cumulus", version="0.1.0", description="Girder API endpoints for interacting with cloud providers.", author="Chris Haris", author_email="chris.harris@kitware.com", url="https://github.com/Kitware/cumulus", packages=find_packages(), package_data={ "": ["*.json"], "cumulus": ["conf/*.json"], }) Exclude unit tests from installfrom setuptools import setup, find_packages setup( name="cumulus", version="0.1.0", description="Girder API endpoints for interacting with cloud providers.", author="Chris Haris", author_email="chris.harris@kitware.com", url="https://github.com/Kitware/cumulus", packages=find_packages(exclude=["*.tests", "*.tests.*", "tests.*", "tests"]), package_data={ "": ["*.json", "*.sh"], "cumulus": ["conf/*.json"], })
<commit_before>from setuptools import setup, find_packages setup( name="cumulus", version="0.1.0", description="Girder API endpoints for interacting with cloud providers.", author="Chris Haris", author_email="chris.harris@kitware.com", url="https://github.com/Kitware/cumulus", packages=find_packages(), package_data={ "": ["*.json"], "cumulus": ["conf/*.json"], }) <commit_msg>Exclude unit tests from install<commit_after>from setuptools import setup, find_packages setup( name="cumulus", version="0.1.0", description="Girder API endpoints for interacting with cloud providers.", author="Chris Haris", author_email="chris.harris@kitware.com", url="https://github.com/Kitware/cumulus", packages=find_packages(exclude=["*.tests", "*.tests.*", "tests.*", "tests"]), package_data={ "": ["*.json", "*.sh"], "cumulus": ["conf/*.json"], })
208f31883a42bc1407f2f911f627b44fd23bf8b1
setup.py
setup.py
#!/usr/bin/env python import io from setuptools import find_packages, setup, Extension with io.open('README.rst', encoding='utf8') as readme: long_description = readme.read() setup( name="pyspamsum", version="1.0.5", description="A Python wrapper for Andrew Tridgell's spamsum algorithm", long_description=long_description, author="Russell Keith-Magee", author_email="russell@keith-magee.com", url='http://github.com/freakboy3742/pyspamsum/', license="New BSD", classifiers=[ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Text Processing', 'Topic :: Utilities', ], ext_modules=[ Extension( "spamsum", [ "pyspamsum.c", "spamsum.c", "edit_dist.c", ] ) ], test_suite='tests', )
#!/usr/bin/env python import io from setuptools import find_packages, setup, Extension with io.open('README.rst', encoding='utf8') as readme: long_description = readme.read() setup( name="pyspamsum", version="1.0.5", description="A Python wrapper for Andrew Tridgell's spamsum algorithm", long_description=long_description, author="Russell Keith-Magee", author_email="russell@keith-magee.com", url='http://github.com/freakboy3742/pyspamsum/', license="New BSD", classifiers=[ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Text Processing', 'Topic :: Utilities', ], ext_modules=[ Extension( "spamsum", [ "pyspamsum.c", "spamsum.c", "edit_dist.c", ] ) ], test_suite='tests', )
Remove support for deprecated Python versions.
Remove support for deprecated Python versions.
Python
bsd-3-clause
freakboy3742/pyspamsum,freakboy3742/pyspamsum
#!/usr/bin/env python import io from setuptools import find_packages, setup, Extension with io.open('README.rst', encoding='utf8') as readme: long_description = readme.read() setup( name="pyspamsum", version="1.0.5", description="A Python wrapper for Andrew Tridgell's spamsum algorithm", long_description=long_description, author="Russell Keith-Magee", author_email="russell@keith-magee.com", url='http://github.com/freakboy3742/pyspamsum/', license="New BSD", classifiers=[ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Text Processing', 'Topic :: Utilities', ], ext_modules=[ Extension( "spamsum", [ "pyspamsum.c", "spamsum.c", "edit_dist.c", ] ) ], test_suite='tests', ) Remove support for deprecated Python versions.
#!/usr/bin/env python import io from setuptools import find_packages, setup, Extension with io.open('README.rst', encoding='utf8') as readme: long_description = readme.read() setup( name="pyspamsum", version="1.0.5", description="A Python wrapper for Andrew Tridgell's spamsum algorithm", long_description=long_description, author="Russell Keith-Magee", author_email="russell@keith-magee.com", url='http://github.com/freakboy3742/pyspamsum/', license="New BSD", classifiers=[ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Text Processing', 'Topic :: Utilities', ], ext_modules=[ Extension( "spamsum", [ "pyspamsum.c", "spamsum.c", "edit_dist.c", ] ) ], test_suite='tests', )
<commit_before>#!/usr/bin/env python import io from setuptools import find_packages, setup, Extension with io.open('README.rst', encoding='utf8') as readme: long_description = readme.read() setup( name="pyspamsum", version="1.0.5", description="A Python wrapper for Andrew Tridgell's spamsum algorithm", long_description=long_description, author="Russell Keith-Magee", author_email="russell@keith-magee.com", url='http://github.com/freakboy3742/pyspamsum/', license="New BSD", classifiers=[ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Text Processing', 'Topic :: Utilities', ], ext_modules=[ Extension( "spamsum", [ "pyspamsum.c", "spamsum.c", "edit_dist.c", ] ) ], test_suite='tests', ) <commit_msg>Remove support for deprecated Python versions.<commit_after>
#!/usr/bin/env python import io from setuptools import find_packages, setup, Extension with io.open('README.rst', encoding='utf8') as readme: long_description = readme.read() setup( name="pyspamsum", version="1.0.5", description="A Python wrapper for Andrew Tridgell's spamsum algorithm", long_description=long_description, author="Russell Keith-Magee", author_email="russell@keith-magee.com", url='http://github.com/freakboy3742/pyspamsum/', license="New BSD", classifiers=[ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Text Processing', 'Topic :: Utilities', ], ext_modules=[ Extension( "spamsum", [ "pyspamsum.c", "spamsum.c", "edit_dist.c", ] ) ], test_suite='tests', )
#!/usr/bin/env python import io from setuptools import find_packages, setup, Extension with io.open('README.rst', encoding='utf8') as readme: long_description = readme.read() setup( name="pyspamsum", version="1.0.5", description="A Python wrapper for Andrew Tridgell's spamsum algorithm", long_description=long_description, author="Russell Keith-Magee", author_email="russell@keith-magee.com", url='http://github.com/freakboy3742/pyspamsum/', license="New BSD", classifiers=[ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Text Processing', 'Topic :: Utilities', ], ext_modules=[ Extension( "spamsum", [ "pyspamsum.c", "spamsum.c", "edit_dist.c", ] ) ], test_suite='tests', ) Remove support for deprecated Python versions.#!/usr/bin/env python import io from setuptools import find_packages, setup, Extension with io.open('README.rst', encoding='utf8') as readme: long_description = readme.read() setup( name="pyspamsum", version="1.0.5", description="A Python wrapper for Andrew Tridgell's spamsum algorithm", long_description=long_description, author="Russell Keith-Magee", author_email="russell@keith-magee.com", url='http://github.com/freakboy3742/pyspamsum/', license="New BSD", classifiers=[ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Text Processing', 'Topic :: Utilities', ], ext_modules=[ Extension( "spamsum", [ "pyspamsum.c", "spamsum.c", "edit_dist.c", ] ) ], test_suite='tests', )
<commit_before>#!/usr/bin/env python import io from setuptools import find_packages, setup, Extension with io.open('README.rst', encoding='utf8') as readme: long_description = readme.read() setup( name="pyspamsum", version="1.0.5", description="A Python wrapper for Andrew Tridgell's spamsum algorithm", long_description=long_description, author="Russell Keith-Magee", author_email="russell@keith-magee.com", url='http://github.com/freakboy3742/pyspamsum/', license="New BSD", classifiers=[ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Text Processing', 'Topic :: Utilities', ], ext_modules=[ Extension( "spamsum", [ "pyspamsum.c", "spamsum.c", "edit_dist.c", ] ) ], test_suite='tests', ) <commit_msg>Remove support for deprecated Python versions.<commit_after>#!/usr/bin/env python import io from setuptools import find_packages, setup, Extension with io.open('README.rst', encoding='utf8') as readme: long_description = readme.read() setup( name="pyspamsum", version="1.0.5", description="A Python wrapper for Andrew Tridgell's spamsum algorithm", long_description=long_description, author="Russell Keith-Magee", author_email="russell@keith-magee.com", url='http://github.com/freakboy3742/pyspamsum/', license="New BSD", classifiers=[ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Text Processing', 'Topic :: Utilities', ], ext_modules=[ Extension( "spamsum", [ "pyspamsum.c", "spamsum.c", "edit_dist.c", ] ) ], test_suite='tests', )
1fd8ea9c746bc1c3fa0dc95a3b00244f97373976
setup.py
setup.py
from distutils.core import setup setup( name='django-robots', version=__import__('robots').__version__, description='Robots exclusion application for Django, complementing Sitemaps.', long_description=open('docs/overview.txt').read(), author='Jannis Leidel', author_email='jannis@leidel.info', url='http://code.google.com/p/django-robots/', download_url='http://github.com/jezdez/django-dbtemplates/zipball/0.5.4', packages=['robots'], package_dir={'dbtemplates': 'dbtemplates'}, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
from distutils.core import setup setup( name='django-robots', version=__import__('robots').__version__, description='Robots exclusion application for Django, complementing Sitemaps.', long_description=open('docs/overview.txt').read(), author='Jannis Leidel', author_email='jannis@leidel.info', url='http://code.google.com/p/django-robots/', packages=['robots'], package_dir={'dbtemplates': 'dbtemplates'}, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
Remove download URL since Github doesn't get his act together. Damnit
Remove download URL since Github doesn't get his act together. Damnit git-svn-id: https://django-robots.googlecode.com/svn/trunk@36 12edf5ea-513a-0410-8a8c-37067077e60f committer: leidel <48255722c72bb2e7f7bec0fff5f60dc7fb5357fb@12edf5ea-513a-0410-8a8c-37067077e60f>
Python
bsd-3-clause
Jythoner/django-robots,Jythoner/django-robots
from distutils.core import setup setup( name='django-robots', version=__import__('robots').__version__, description='Robots exclusion application for Django, complementing Sitemaps.', long_description=open('docs/overview.txt').read(), author='Jannis Leidel', author_email='jannis@leidel.info', url='http://code.google.com/p/django-robots/', download_url='http://github.com/jezdez/django-dbtemplates/zipball/0.5.4', packages=['robots'], package_dir={'dbtemplates': 'dbtemplates'}, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] ) Remove download URL since Github doesn't get his act together. Damnit git-svn-id: https://django-robots.googlecode.com/svn/trunk@36 12edf5ea-513a-0410-8a8c-37067077e60f committer: leidel <48255722c72bb2e7f7bec0fff5f60dc7fb5357fb@12edf5ea-513a-0410-8a8c-37067077e60f>
from distutils.core import setup setup( name='django-robots', version=__import__('robots').__version__, description='Robots exclusion application for Django, complementing Sitemaps.', long_description=open('docs/overview.txt').read(), author='Jannis Leidel', author_email='jannis@leidel.info', url='http://code.google.com/p/django-robots/', packages=['robots'], package_dir={'dbtemplates': 'dbtemplates'}, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
<commit_before>from distutils.core import setup setup( name='django-robots', version=__import__('robots').__version__, description='Robots exclusion application for Django, complementing Sitemaps.', long_description=open('docs/overview.txt').read(), author='Jannis Leidel', author_email='jannis@leidel.info', url='http://code.google.com/p/django-robots/', download_url='http://github.com/jezdez/django-dbtemplates/zipball/0.5.4', packages=['robots'], package_dir={'dbtemplates': 'dbtemplates'}, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] ) <commit_msg>Remove download URL since Github doesn't get his act together. Damnit git-svn-id: https://django-robots.googlecode.com/svn/trunk@36 12edf5ea-513a-0410-8a8c-37067077e60f committer: leidel <48255722c72bb2e7f7bec0fff5f60dc7fb5357fb@12edf5ea-513a-0410-8a8c-37067077e60f><commit_after>
from distutils.core import setup setup( name='django-robots', version=__import__('robots').__version__, description='Robots exclusion application for Django, complementing Sitemaps.', long_description=open('docs/overview.txt').read(), author='Jannis Leidel', author_email='jannis@leidel.info', url='http://code.google.com/p/django-robots/', packages=['robots'], package_dir={'dbtemplates': 'dbtemplates'}, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
from distutils.core import setup setup( name='django-robots', version=__import__('robots').__version__, description='Robots exclusion application for Django, complementing Sitemaps.', long_description=open('docs/overview.txt').read(), author='Jannis Leidel', author_email='jannis@leidel.info', url='http://code.google.com/p/django-robots/', download_url='http://github.com/jezdez/django-dbtemplates/zipball/0.5.4', packages=['robots'], package_dir={'dbtemplates': 'dbtemplates'}, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] ) Remove download URL since Github doesn't get his act together. Damnit git-svn-id: https://django-robots.googlecode.com/svn/trunk@36 12edf5ea-513a-0410-8a8c-37067077e60f committer: leidel <48255722c72bb2e7f7bec0fff5f60dc7fb5357fb@12edf5ea-513a-0410-8a8c-37067077e60f>from distutils.core import setup setup( name='django-robots', version=__import__('robots').__version__, description='Robots exclusion application for Django, complementing Sitemaps.', long_description=open('docs/overview.txt').read(), author='Jannis Leidel', author_email='jannis@leidel.info', url='http://code.google.com/p/django-robots/', packages=['robots'], package_dir={'dbtemplates': 'dbtemplates'}, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
<commit_before>from distutils.core import setup setup( name='django-robots', version=__import__('robots').__version__, description='Robots exclusion application for Django, complementing Sitemaps.', long_description=open('docs/overview.txt').read(), author='Jannis Leidel', author_email='jannis@leidel.info', url='http://code.google.com/p/django-robots/', download_url='http://github.com/jezdez/django-dbtemplates/zipball/0.5.4', packages=['robots'], package_dir={'dbtemplates': 'dbtemplates'}, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] ) <commit_msg>Remove download URL since Github doesn't get his act together. Damnit git-svn-id: https://django-robots.googlecode.com/svn/trunk@36 12edf5ea-513a-0410-8a8c-37067077e60f committer: leidel <48255722c72bb2e7f7bec0fff5f60dc7fb5357fb@12edf5ea-513a-0410-8a8c-37067077e60f><commit_after>from distutils.core import setup setup( name='django-robots', version=__import__('robots').__version__, description='Robots exclusion application for Django, complementing Sitemaps.', long_description=open('docs/overview.txt').read(), author='Jannis Leidel', author_email='jannis@leidel.info', url='http://code.google.com/p/django-robots/', packages=['robots'], package_dir={'dbtemplates': 'dbtemplates'}, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
743fcf5266235b47cfec7f9e12b7df0378f4f9c4
solar/solar/extensions/modules/discovery.py
solar/solar/extensions/modules/discovery.py
import io import os import yaml from solar.extensions import base class Discovery(base.BaseExtension): VERSION = '1.0.0' ID = 'discovery' PROVIDES = ['nodes_resources'] COLLECTION_NAME = 'nodes' FILE_PATH = os.path.join( # TODO(pkaminski): no way we need '..' here... os.path.dirname(__file__), '..', '..', '..', '..', 'examples', 'nodes_list.yaml') def discover(self): with io.open(self.FILE_PATH) as f: nodes = yaml.load(f) for node in nodes: node['tags'] = [] self.db.store_list(self.COLLECTION_NAME, nodes) return nodes def nodes_resources(self): nodes_list = self.db.get_list(self.COLLECTION_NAME) nodes_resources = [] for node in nodes_list: node_resource = {} node_resource['id'] = node['id'] node_resource['name'] = node['id'] node_resource['handler'] = 'data' node_resource['type'] = 'resource' node_resource['version'] = self.VERSION node_resource['tags'] = node['tags'] node_resource['output'] = node node_resource['ssh_host'] = node['ip'] # TODO replace it with ssh type node_resource['connection_type'] = 'local' nodes_resources.append(node_resource) return nodes_resources
import io import os import yaml from solar.extensions import base class Discovery(base.BaseExtension): VERSION = '1.0.0' ID = 'discovery' PROVIDES = ['nodes_resources'] COLLECTION_NAME = 'nodes' FILE_PATH = os.path.join( # TODO(pkaminski): no way we need '..' here... os.path.dirname(__file__), '..', '..', '..', '..', 'examples', 'nodes_list.yaml') def discover(self): with io.open(self.FILE_PATH) as f: nodes = yaml.load(f) for node in nodes: node['tags'] = ['node/{0}'.format(node['id'])] self.db.store_list(self.COLLECTION_NAME, nodes) return nodes def nodes_resources(self): nodes_list = self.db.get_list(self.COLLECTION_NAME) nodes_resources = [] for node in nodes_list: node_resource = {} node_resource['id'] = node['id'] node_resource['name'] = node['id'] node_resource['handler'] = 'data' node_resource['type'] = 'resource' node_resource['version'] = self.VERSION node_resource['tags'] = node['tags'] node_resource['output'] = node node_resource['ssh_host'] = node['ip'] # TODO replace it with ssh type node_resource['connection_type'] = 'local' nodes_resources.append(node_resource) return nodes_resources
Add node specific tag generation
Add node specific tag generation
Python
apache-2.0
pigmej/solar,pigmej/solar,torgartor21/solar,torgartor21/solar,zen/solar,loles/solar,Mirantis/solar,CGenie/solar,pigmej/solar,zen/solar,dshulyak/solar,openstack/solar,Mirantis/solar,loles/solar,openstack/solar,Mirantis/solar,dshulyak/solar,Mirantis/solar,CGenie/solar,loles/solar,loles/solar,openstack/solar,zen/solar,zen/solar
import io import os import yaml from solar.extensions import base class Discovery(base.BaseExtension): VERSION = '1.0.0' ID = 'discovery' PROVIDES = ['nodes_resources'] COLLECTION_NAME = 'nodes' FILE_PATH = os.path.join( # TODO(pkaminski): no way we need '..' here... os.path.dirname(__file__), '..', '..', '..', '..', 'examples', 'nodes_list.yaml') def discover(self): with io.open(self.FILE_PATH) as f: nodes = yaml.load(f) for node in nodes: node['tags'] = [] self.db.store_list(self.COLLECTION_NAME, nodes) return nodes def nodes_resources(self): nodes_list = self.db.get_list(self.COLLECTION_NAME) nodes_resources = [] for node in nodes_list: node_resource = {} node_resource['id'] = node['id'] node_resource['name'] = node['id'] node_resource['handler'] = 'data' node_resource['type'] = 'resource' node_resource['version'] = self.VERSION node_resource['tags'] = node['tags'] node_resource['output'] = node node_resource['ssh_host'] = node['ip'] # TODO replace it with ssh type node_resource['connection_type'] = 'local' nodes_resources.append(node_resource) return nodes_resources Add node specific tag generation
import io import os import yaml from solar.extensions import base class Discovery(base.BaseExtension): VERSION = '1.0.0' ID = 'discovery' PROVIDES = ['nodes_resources'] COLLECTION_NAME = 'nodes' FILE_PATH = os.path.join( # TODO(pkaminski): no way we need '..' here... os.path.dirname(__file__), '..', '..', '..', '..', 'examples', 'nodes_list.yaml') def discover(self): with io.open(self.FILE_PATH) as f: nodes = yaml.load(f) for node in nodes: node['tags'] = ['node/{0}'.format(node['id'])] self.db.store_list(self.COLLECTION_NAME, nodes) return nodes def nodes_resources(self): nodes_list = self.db.get_list(self.COLLECTION_NAME) nodes_resources = [] for node in nodes_list: node_resource = {} node_resource['id'] = node['id'] node_resource['name'] = node['id'] node_resource['handler'] = 'data' node_resource['type'] = 'resource' node_resource['version'] = self.VERSION node_resource['tags'] = node['tags'] node_resource['output'] = node node_resource['ssh_host'] = node['ip'] # TODO replace it with ssh type node_resource['connection_type'] = 'local' nodes_resources.append(node_resource) return nodes_resources
<commit_before>import io import os import yaml from solar.extensions import base class Discovery(base.BaseExtension): VERSION = '1.0.0' ID = 'discovery' PROVIDES = ['nodes_resources'] COLLECTION_NAME = 'nodes' FILE_PATH = os.path.join( # TODO(pkaminski): no way we need '..' here... os.path.dirname(__file__), '..', '..', '..', '..', 'examples', 'nodes_list.yaml') def discover(self): with io.open(self.FILE_PATH) as f: nodes = yaml.load(f) for node in nodes: node['tags'] = [] self.db.store_list(self.COLLECTION_NAME, nodes) return nodes def nodes_resources(self): nodes_list = self.db.get_list(self.COLLECTION_NAME) nodes_resources = [] for node in nodes_list: node_resource = {} node_resource['id'] = node['id'] node_resource['name'] = node['id'] node_resource['handler'] = 'data' node_resource['type'] = 'resource' node_resource['version'] = self.VERSION node_resource['tags'] = node['tags'] node_resource['output'] = node node_resource['ssh_host'] = node['ip'] # TODO replace it with ssh type node_resource['connection_type'] = 'local' nodes_resources.append(node_resource) return nodes_resources <commit_msg>Add node specific tag generation<commit_after>
import io import os import yaml from solar.extensions import base class Discovery(base.BaseExtension): VERSION = '1.0.0' ID = 'discovery' PROVIDES = ['nodes_resources'] COLLECTION_NAME = 'nodes' FILE_PATH = os.path.join( # TODO(pkaminski): no way we need '..' here... os.path.dirname(__file__), '..', '..', '..', '..', 'examples', 'nodes_list.yaml') def discover(self): with io.open(self.FILE_PATH) as f: nodes = yaml.load(f) for node in nodes: node['tags'] = ['node/{0}'.format(node['id'])] self.db.store_list(self.COLLECTION_NAME, nodes) return nodes def nodes_resources(self): nodes_list = self.db.get_list(self.COLLECTION_NAME) nodes_resources = [] for node in nodes_list: node_resource = {} node_resource['id'] = node['id'] node_resource['name'] = node['id'] node_resource['handler'] = 'data' node_resource['type'] = 'resource' node_resource['version'] = self.VERSION node_resource['tags'] = node['tags'] node_resource['output'] = node node_resource['ssh_host'] = node['ip'] # TODO replace it with ssh type node_resource['connection_type'] = 'local' nodes_resources.append(node_resource) return nodes_resources
import io import os import yaml from solar.extensions import base class Discovery(base.BaseExtension): VERSION = '1.0.0' ID = 'discovery' PROVIDES = ['nodes_resources'] COLLECTION_NAME = 'nodes' FILE_PATH = os.path.join( # TODO(pkaminski): no way we need '..' here... os.path.dirname(__file__), '..', '..', '..', '..', 'examples', 'nodes_list.yaml') def discover(self): with io.open(self.FILE_PATH) as f: nodes = yaml.load(f) for node in nodes: node['tags'] = [] self.db.store_list(self.COLLECTION_NAME, nodes) return nodes def nodes_resources(self): nodes_list = self.db.get_list(self.COLLECTION_NAME) nodes_resources = [] for node in nodes_list: node_resource = {} node_resource['id'] = node['id'] node_resource['name'] = node['id'] node_resource['handler'] = 'data' node_resource['type'] = 'resource' node_resource['version'] = self.VERSION node_resource['tags'] = node['tags'] node_resource['output'] = node node_resource['ssh_host'] = node['ip'] # TODO replace it with ssh type node_resource['connection_type'] = 'local' nodes_resources.append(node_resource) return nodes_resources Add node specific tag generationimport io import os import yaml from solar.extensions import base class Discovery(base.BaseExtension): VERSION = '1.0.0' ID = 'discovery' PROVIDES = ['nodes_resources'] COLLECTION_NAME = 'nodes' FILE_PATH = os.path.join( # TODO(pkaminski): no way we need '..' here... os.path.dirname(__file__), '..', '..', '..', '..', 'examples', 'nodes_list.yaml') def discover(self): with io.open(self.FILE_PATH) as f: nodes = yaml.load(f) for node in nodes: node['tags'] = ['node/{0}'.format(node['id'])] self.db.store_list(self.COLLECTION_NAME, nodes) return nodes def nodes_resources(self): nodes_list = self.db.get_list(self.COLLECTION_NAME) nodes_resources = [] for node in nodes_list: node_resource = {} node_resource['id'] = node['id'] node_resource['name'] = node['id'] node_resource['handler'] = 'data' node_resource['type'] = 'resource' node_resource['version'] = self.VERSION node_resource['tags'] = node['tags'] node_resource['output'] = node node_resource['ssh_host'] = node['ip'] # TODO replace it with ssh type node_resource['connection_type'] = 'local' nodes_resources.append(node_resource) return nodes_resources
<commit_before>import io import os import yaml from solar.extensions import base class Discovery(base.BaseExtension): VERSION = '1.0.0' ID = 'discovery' PROVIDES = ['nodes_resources'] COLLECTION_NAME = 'nodes' FILE_PATH = os.path.join( # TODO(pkaminski): no way we need '..' here... os.path.dirname(__file__), '..', '..', '..', '..', 'examples', 'nodes_list.yaml') def discover(self): with io.open(self.FILE_PATH) as f: nodes = yaml.load(f) for node in nodes: node['tags'] = [] self.db.store_list(self.COLLECTION_NAME, nodes) return nodes def nodes_resources(self): nodes_list = self.db.get_list(self.COLLECTION_NAME) nodes_resources = [] for node in nodes_list: node_resource = {} node_resource['id'] = node['id'] node_resource['name'] = node['id'] node_resource['handler'] = 'data' node_resource['type'] = 'resource' node_resource['version'] = self.VERSION node_resource['tags'] = node['tags'] node_resource['output'] = node node_resource['ssh_host'] = node['ip'] # TODO replace it with ssh type node_resource['connection_type'] = 'local' nodes_resources.append(node_resource) return nodes_resources <commit_msg>Add node specific tag generation<commit_after>import io import os import yaml from solar.extensions import base class Discovery(base.BaseExtension): VERSION = '1.0.0' ID = 'discovery' PROVIDES = ['nodes_resources'] COLLECTION_NAME = 'nodes' FILE_PATH = os.path.join( # TODO(pkaminski): no way we need '..' here... os.path.dirname(__file__), '..', '..', '..', '..', 'examples', 'nodes_list.yaml') def discover(self): with io.open(self.FILE_PATH) as f: nodes = yaml.load(f) for node in nodes: node['tags'] = ['node/{0}'.format(node['id'])] self.db.store_list(self.COLLECTION_NAME, nodes) return nodes def nodes_resources(self): nodes_list = self.db.get_list(self.COLLECTION_NAME) nodes_resources = [] for node in nodes_list: node_resource = {} node_resource['id'] = node['id'] node_resource['name'] = node['id'] node_resource['handler'] = 'data' node_resource['type'] = 'resource' node_resource['version'] = self.VERSION node_resource['tags'] = node['tags'] node_resource['output'] = node node_resource['ssh_host'] = node['ip'] # TODO replace it with ssh type node_resource['connection_type'] = 'local' nodes_resources.append(node_resource) return nodes_resources
d5e05aa44998e6f2cebc41b0c945d234eeb6317d
gcloud/search/connection.py
gcloud/search/connection.py
# Copyright 2015 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Create / interact with gcloud search connections.""" from gcloud import connection as base_connection class Connection(base_connection.JSONConnection): """A connection to Google Cloud Search via the JSON REST API.""" API_BASE_URL = 'https://cloudsearch.googleapis.com' """The base of the API call URL.""" API_VERSION = 'v1' """The version of the API, used in building the API call's URL.""" API_URL_TEMPLATE = '{api_base_url}/{api_version}{path}' """A template for the URL of a particular API call.""" SCOPE = ('https://www.googleapis.com/auth/ndev.cloudsearch',) """The scopes required for authenticating as a Cloud Search consumer."""
# Copyright 2015 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Create / interact with gcloud search connections.""" from gcloud import connection as base_connection class Connection(base_connection.JSONConnection): """A connection to Google Cloud Search via the JSON REST API.""" API_BASE_URL = 'https://cloudsearch.googleapis.com' """The base of the API call URL.""" API_VERSION = 'v1' """The version of the API, used in building the API call's URL.""" API_URL_TEMPLATE = '{api_base_url}/{api_version}{path}' """A template for the URL of a particular API call.""" SCOPE = ('https://www.googleapis.com/auth/cloudsearch',) """The scopes required for authenticating as a Cloud Search consumer."""
Fix scope used for Cloud Search.
Fix scope used for Cloud Search.
Python
apache-2.0
googleapis/google-cloud-python,VitalLabs/gcloud-python,jgeewax/gcloud-python,waprin/google-cloud-python,tseaver/gcloud-python,jgeewax/gcloud-python,jonparrott/google-cloud-python,tswast/google-cloud-python,dhermes/google-cloud-python,tseaver/google-cloud-python,jonparrott/gcloud-python,dhermes/google-cloud-python,jonparrott/google-cloud-python,Fkawala/gcloud-python,quom/google-cloud-python,daspecster/google-cloud-python,waprin/gcloud-python,waprin/gcloud-python,dhermes/gcloud-python,Fkawala/gcloud-python,GoogleCloudPlatform/gcloud-python,tartavull/google-cloud-python,jonparrott/gcloud-python,tseaver/google-cloud-python,tseaver/google-cloud-python,tswast/google-cloud-python,dhermes/gcloud-python,dhermes/google-cloud-python,quom/google-cloud-python,elibixby/gcloud-python,tseaver/gcloud-python,VitalLabs/gcloud-python,elibixby/gcloud-python,tartavull/google-cloud-python,tswast/google-cloud-python,calpeyser/google-cloud-python,googleapis/google-cloud-python,calpeyser/google-cloud-python,GoogleCloudPlatform/gcloud-python,daspecster/google-cloud-python,waprin/google-cloud-python
# Copyright 2015 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Create / interact with gcloud search connections.""" from gcloud import connection as base_connection class Connection(base_connection.JSONConnection): """A connection to Google Cloud Search via the JSON REST API.""" API_BASE_URL = 'https://cloudsearch.googleapis.com' """The base of the API call URL.""" API_VERSION = 'v1' """The version of the API, used in building the API call's URL.""" API_URL_TEMPLATE = '{api_base_url}/{api_version}{path}' """A template for the URL of a particular API call.""" SCOPE = ('https://www.googleapis.com/auth/ndev.cloudsearch',) """The scopes required for authenticating as a Cloud Search consumer.""" Fix scope used for Cloud Search.
# Copyright 2015 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Create / interact with gcloud search connections.""" from gcloud import connection as base_connection class Connection(base_connection.JSONConnection): """A connection to Google Cloud Search via the JSON REST API.""" API_BASE_URL = 'https://cloudsearch.googleapis.com' """The base of the API call URL.""" API_VERSION = 'v1' """The version of the API, used in building the API call's URL.""" API_URL_TEMPLATE = '{api_base_url}/{api_version}{path}' """A template for the URL of a particular API call.""" SCOPE = ('https://www.googleapis.com/auth/cloudsearch',) """The scopes required for authenticating as a Cloud Search consumer."""
<commit_before># Copyright 2015 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Create / interact with gcloud search connections.""" from gcloud import connection as base_connection class Connection(base_connection.JSONConnection): """A connection to Google Cloud Search via the JSON REST API.""" API_BASE_URL = 'https://cloudsearch.googleapis.com' """The base of the API call URL.""" API_VERSION = 'v1' """The version of the API, used in building the API call's URL.""" API_URL_TEMPLATE = '{api_base_url}/{api_version}{path}' """A template for the URL of a particular API call.""" SCOPE = ('https://www.googleapis.com/auth/ndev.cloudsearch',) """The scopes required for authenticating as a Cloud Search consumer.""" <commit_msg>Fix scope used for Cloud Search.<commit_after>
# Copyright 2015 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Create / interact with gcloud search connections.""" from gcloud import connection as base_connection class Connection(base_connection.JSONConnection): """A connection to Google Cloud Search via the JSON REST API.""" API_BASE_URL = 'https://cloudsearch.googleapis.com' """The base of the API call URL.""" API_VERSION = 'v1' """The version of the API, used in building the API call's URL.""" API_URL_TEMPLATE = '{api_base_url}/{api_version}{path}' """A template for the URL of a particular API call.""" SCOPE = ('https://www.googleapis.com/auth/cloudsearch',) """The scopes required for authenticating as a Cloud Search consumer."""
# Copyright 2015 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Create / interact with gcloud search connections.""" from gcloud import connection as base_connection class Connection(base_connection.JSONConnection): """A connection to Google Cloud Search via the JSON REST API.""" API_BASE_URL = 'https://cloudsearch.googleapis.com' """The base of the API call URL.""" API_VERSION = 'v1' """The version of the API, used in building the API call's URL.""" API_URL_TEMPLATE = '{api_base_url}/{api_version}{path}' """A template for the URL of a particular API call.""" SCOPE = ('https://www.googleapis.com/auth/ndev.cloudsearch',) """The scopes required for authenticating as a Cloud Search consumer.""" Fix scope used for Cloud Search.# Copyright 2015 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Create / interact with gcloud search connections.""" from gcloud import connection as base_connection class Connection(base_connection.JSONConnection): """A connection to Google Cloud Search via the JSON REST API.""" API_BASE_URL = 'https://cloudsearch.googleapis.com' """The base of the API call URL.""" API_VERSION = 'v1' """The version of the API, used in building the API call's URL.""" API_URL_TEMPLATE = '{api_base_url}/{api_version}{path}' """A template for the URL of a particular API call.""" SCOPE = ('https://www.googleapis.com/auth/cloudsearch',) """The scopes required for authenticating as a Cloud Search consumer."""
<commit_before># Copyright 2015 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Create / interact with gcloud search connections.""" from gcloud import connection as base_connection class Connection(base_connection.JSONConnection): """A connection to Google Cloud Search via the JSON REST API.""" API_BASE_URL = 'https://cloudsearch.googleapis.com' """The base of the API call URL.""" API_VERSION = 'v1' """The version of the API, used in building the API call's URL.""" API_URL_TEMPLATE = '{api_base_url}/{api_version}{path}' """A template for the URL of a particular API call.""" SCOPE = ('https://www.googleapis.com/auth/ndev.cloudsearch',) """The scopes required for authenticating as a Cloud Search consumer.""" <commit_msg>Fix scope used for Cloud Search.<commit_after># Copyright 2015 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Create / interact with gcloud search connections.""" from gcloud import connection as base_connection class Connection(base_connection.JSONConnection): """A connection to Google Cloud Search via the JSON REST API.""" API_BASE_URL = 'https://cloudsearch.googleapis.com' """The base of the API call URL.""" API_VERSION = 'v1' """The version of the API, used in building the API call's URL.""" API_URL_TEMPLATE = '{api_base_url}/{api_version}{path}' """A template for the URL of a particular API call.""" SCOPE = ('https://www.googleapis.com/auth/cloudsearch',) """The scopes required for authenticating as a Cloud Search consumer."""
f4fb4e90725cec378634c22ee8803eeb29ccf143
conveyor/processor.py
conveyor/processor.py
from __future__ import absolute_import from __future__ import division from xmlrpc2 import client as xmlrpc2 class BaseProcessor(object): def __init__(self, index, *args, **kwargs): super(BaseProcessor, self).__init__(*args, **kwargs) self.index = index self.client = xmlrpc2.Client(self.index) def process(self): raise NotImplementedError def get_releases(self, name, version=None): if version is None: return set(self.client.package_releases(name, True)) else: return set([version]) class BulkProcessor(BaseProcessor): def process(self): pass
from __future__ import absolute_import from __future__ import division from xmlrpc2 import client as xmlrpc2 class BaseProcessor(object): def __init__(self, index, *args, **kwargs): super(BaseProcessor, self).__init__(*args, **kwargs) self.index = index self.client = xmlrpc2.Client(self.index) def process(self): raise NotImplementedError def get_releases(self, name, version=None): if version is None: return self.client.package_releases(name, True) else: return [version] class BulkProcessor(BaseProcessor): def process(self): pass
Switch from returning a set to returning a list
Switch from returning a set to returning a list
Python
bsd-2-clause
crateio/carrier
from __future__ import absolute_import from __future__ import division from xmlrpc2 import client as xmlrpc2 class BaseProcessor(object): def __init__(self, index, *args, **kwargs): super(BaseProcessor, self).__init__(*args, **kwargs) self.index = index self.client = xmlrpc2.Client(self.index) def process(self): raise NotImplementedError def get_releases(self, name, version=None): if version is None: return set(self.client.package_releases(name, True)) else: return set([version]) class BulkProcessor(BaseProcessor): def process(self): pass Switch from returning a set to returning a list
from __future__ import absolute_import from __future__ import division from xmlrpc2 import client as xmlrpc2 class BaseProcessor(object): def __init__(self, index, *args, **kwargs): super(BaseProcessor, self).__init__(*args, **kwargs) self.index = index self.client = xmlrpc2.Client(self.index) def process(self): raise NotImplementedError def get_releases(self, name, version=None): if version is None: return self.client.package_releases(name, True) else: return [version] class BulkProcessor(BaseProcessor): def process(self): pass
<commit_before>from __future__ import absolute_import from __future__ import division from xmlrpc2 import client as xmlrpc2 class BaseProcessor(object): def __init__(self, index, *args, **kwargs): super(BaseProcessor, self).__init__(*args, **kwargs) self.index = index self.client = xmlrpc2.Client(self.index) def process(self): raise NotImplementedError def get_releases(self, name, version=None): if version is None: return set(self.client.package_releases(name, True)) else: return set([version]) class BulkProcessor(BaseProcessor): def process(self): pass <commit_msg>Switch from returning a set to returning a list<commit_after>
from __future__ import absolute_import from __future__ import division from xmlrpc2 import client as xmlrpc2 class BaseProcessor(object): def __init__(self, index, *args, **kwargs): super(BaseProcessor, self).__init__(*args, **kwargs) self.index = index self.client = xmlrpc2.Client(self.index) def process(self): raise NotImplementedError def get_releases(self, name, version=None): if version is None: return self.client.package_releases(name, True) else: return [version] class BulkProcessor(BaseProcessor): def process(self): pass
from __future__ import absolute_import from __future__ import division from xmlrpc2 import client as xmlrpc2 class BaseProcessor(object): def __init__(self, index, *args, **kwargs): super(BaseProcessor, self).__init__(*args, **kwargs) self.index = index self.client = xmlrpc2.Client(self.index) def process(self): raise NotImplementedError def get_releases(self, name, version=None): if version is None: return set(self.client.package_releases(name, True)) else: return set([version]) class BulkProcessor(BaseProcessor): def process(self): pass Switch from returning a set to returning a listfrom __future__ import absolute_import from __future__ import division from xmlrpc2 import client as xmlrpc2 class BaseProcessor(object): def __init__(self, index, *args, **kwargs): super(BaseProcessor, self).__init__(*args, **kwargs) self.index = index self.client = xmlrpc2.Client(self.index) def process(self): raise NotImplementedError def get_releases(self, name, version=None): if version is None: return self.client.package_releases(name, True) else: return [version] class BulkProcessor(BaseProcessor): def process(self): pass
<commit_before>from __future__ import absolute_import from __future__ import division from xmlrpc2 import client as xmlrpc2 class BaseProcessor(object): def __init__(self, index, *args, **kwargs): super(BaseProcessor, self).__init__(*args, **kwargs) self.index = index self.client = xmlrpc2.Client(self.index) def process(self): raise NotImplementedError def get_releases(self, name, version=None): if version is None: return set(self.client.package_releases(name, True)) else: return set([version]) class BulkProcessor(BaseProcessor): def process(self): pass <commit_msg>Switch from returning a set to returning a list<commit_after>from __future__ import absolute_import from __future__ import division from xmlrpc2 import client as xmlrpc2 class BaseProcessor(object): def __init__(self, index, *args, **kwargs): super(BaseProcessor, self).__init__(*args, **kwargs) self.index = index self.client = xmlrpc2.Client(self.index) def process(self): raise NotImplementedError def get_releases(self, name, version=None): if version is None: return self.client.package_releases(name, True) else: return [version] class BulkProcessor(BaseProcessor): def process(self): pass
22ab7f94c3b8b04a50cd11e73b2963bddb0470a5
poradnia/contrib/sites/migrations/0002_set_site_domain_and_name.py
poradnia/contrib/sites/migrations/0002_set_site_domain_and_name.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.db import models, migrations def update_site_forward(apps, schema_editor): """Set site domain and name.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "poradnia.informacjapubliczna.org.pl", "name": "poradnia" } ) def update_site_backward(apps, schema_editor): """Revert site domain and name to default.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "example.com", "name": "example.com" } ) class Migration(migrations.Migration): dependencies = [ ('sites', '0001_initial'), ] operations = [ migrations.RunPython(update_site_forward, update_site_backward), ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.db import models, migrations def update_site_forward(apps, schema_editor): """Set site domain and name.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "porady.siecobywatelska.pl", "name": "poradnia" } ) def update_site_backward(apps, schema_editor): """Revert site domain and name to default.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "example.com", "name": "example.com" } ) class Migration(migrations.Migration): dependencies = [ ('sites', '0001_initial'), ] operations = [ migrations.RunPython(update_site_forward, update_site_backward), ]
Fix migrations - drop old domain
Fix migrations - drop old domain
Python
mit
rwakulszowa/poradnia,rwakulszowa/poradnia,watchdogpolska/poradnia.siecobywatelska.pl,watchdogpolska/poradnia,rwakulszowa/poradnia,watchdogpolska/poradnia,watchdogpolska/poradnia.siecobywatelska.pl,watchdogpolska/poradnia.siecobywatelska.pl,rwakulszowa/poradnia,watchdogpolska/poradnia,watchdogpolska/poradnia
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.db import models, migrations def update_site_forward(apps, schema_editor): """Set site domain and name.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "poradnia.informacjapubliczna.org.pl", "name": "poradnia" } ) def update_site_backward(apps, schema_editor): """Revert site domain and name to default.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "example.com", "name": "example.com" } ) class Migration(migrations.Migration): dependencies = [ ('sites', '0001_initial'), ] operations = [ migrations.RunPython(update_site_forward, update_site_backward), ] Fix migrations - drop old domain
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.db import models, migrations def update_site_forward(apps, schema_editor): """Set site domain and name.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "porady.siecobywatelska.pl", "name": "poradnia" } ) def update_site_backward(apps, schema_editor): """Revert site domain and name to default.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "example.com", "name": "example.com" } ) class Migration(migrations.Migration): dependencies = [ ('sites', '0001_initial'), ] operations = [ migrations.RunPython(update_site_forward, update_site_backward), ]
<commit_before># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.db import models, migrations def update_site_forward(apps, schema_editor): """Set site domain and name.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "poradnia.informacjapubliczna.org.pl", "name": "poradnia" } ) def update_site_backward(apps, schema_editor): """Revert site domain and name to default.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "example.com", "name": "example.com" } ) class Migration(migrations.Migration): dependencies = [ ('sites', '0001_initial'), ] operations = [ migrations.RunPython(update_site_forward, update_site_backward), ] <commit_msg>Fix migrations - drop old domain<commit_after>
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.db import models, migrations def update_site_forward(apps, schema_editor): """Set site domain and name.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "porady.siecobywatelska.pl", "name": "poradnia" } ) def update_site_backward(apps, schema_editor): """Revert site domain and name to default.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "example.com", "name": "example.com" } ) class Migration(migrations.Migration): dependencies = [ ('sites', '0001_initial'), ] operations = [ migrations.RunPython(update_site_forward, update_site_backward), ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.db import models, migrations def update_site_forward(apps, schema_editor): """Set site domain and name.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "poradnia.informacjapubliczna.org.pl", "name": "poradnia" } ) def update_site_backward(apps, schema_editor): """Revert site domain and name to default.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "example.com", "name": "example.com" } ) class Migration(migrations.Migration): dependencies = [ ('sites', '0001_initial'), ] operations = [ migrations.RunPython(update_site_forward, update_site_backward), ] Fix migrations - drop old domain# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.db import models, migrations def update_site_forward(apps, schema_editor): """Set site domain and name.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "porady.siecobywatelska.pl", "name": "poradnia" } ) def update_site_backward(apps, schema_editor): """Revert site domain and name to default.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "example.com", "name": "example.com" } ) class Migration(migrations.Migration): dependencies = [ ('sites', '0001_initial'), ] operations = [ migrations.RunPython(update_site_forward, update_site_backward), ]
<commit_before># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.db import models, migrations def update_site_forward(apps, schema_editor): """Set site domain and name.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "poradnia.informacjapubliczna.org.pl", "name": "poradnia" } ) def update_site_backward(apps, schema_editor): """Revert site domain and name to default.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "example.com", "name": "example.com" } ) class Migration(migrations.Migration): dependencies = [ ('sites', '0001_initial'), ] operations = [ migrations.RunPython(update_site_forward, update_site_backward), ] <commit_msg>Fix migrations - drop old domain<commit_after># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.conf import settings from django.db import models, migrations def update_site_forward(apps, schema_editor): """Set site domain and name.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "porady.siecobywatelska.pl", "name": "poradnia" } ) def update_site_backward(apps, schema_editor): """Revert site domain and name to default.""" Site = apps.get_model("sites", "Site") Site.objects.update_or_create( id=settings.SITE_ID, defaults={ "domain": "example.com", "name": "example.com" } ) class Migration(migrations.Migration): dependencies = [ ('sites', '0001_initial'), ] operations = [ migrations.RunPython(update_site_forward, update_site_backward), ]
73feddb22ad3a2543ad4f8047061d909c64fd75d
server/system/CommonMySQL.py
server/system/CommonMySQL.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from system.DBMysql import connect from system.ConfigLoader import getCfg def loginUser(login, password): ''' Try to login a user regarding login/password ''' userContent = None table = getCfg('MYSQL', 'table') tableId = getCfg('MYSQL', 'idField') tableLogin = getCfg('MYSQL', 'loginField') tablePassword = getCfg('MYSQL', 'passwordField') try: # Starting con = connect() cur = con.cursor() cur.execute( 'SELECT ' + tableId + ' FROM ' + table + ' WHERE ' + tableLogin + '=%s AND ' + tablePassword + '=%s', ( login, password ) ) userContent = cur.fetchone() if userContent is not None: userContent = userContent[0] except db.Error as e: logging.error('loginUser: Error from MySQL => %s' % e) finally: if con: con.close() return userContent
#!/usr/bin/env python # -*- coding: utf-8 -*- from system.DBMysql import connect from system.ConfigLoader import getCfg import logging def loginUser(login, password): ''' Try to login a user regarding login/password ''' userContent = None table = getCfg('MYSQL', 'table') tableId = getCfg('MYSQL', 'idField') tableLogin = getCfg('MYSQL', 'loginField') tablePassword = getCfg('MYSQL', 'passwordField') con = None try: # Starting con = connect() cur = con.cursor() cur.execute( 'SELECT ' + tableId + ' FROM ' + table + ' WHERE ' + tableLogin + '=%s AND ' + tablePassword + '=%s', ( login, password ) ) userContent = cur.fetchone() if userContent is not None: userContent = userContent[0] except Exception as e: logging.error('loginUser: Error from MySQL => %s' % e) finally: if con: con.close() return userContent
Remove some bug on mysql elements.
Remove some bug on mysql elements.
Python
mit
Deisss/webservice-notification,Deisss/webservice-notification
#!/usr/bin/env python # -*- coding: utf-8 -*- from system.DBMysql import connect from system.ConfigLoader import getCfg def loginUser(login, password): ''' Try to login a user regarding login/password ''' userContent = None table = getCfg('MYSQL', 'table') tableId = getCfg('MYSQL', 'idField') tableLogin = getCfg('MYSQL', 'loginField') tablePassword = getCfg('MYSQL', 'passwordField') try: # Starting con = connect() cur = con.cursor() cur.execute( 'SELECT ' + tableId + ' FROM ' + table + ' WHERE ' + tableLogin + '=%s AND ' + tablePassword + '=%s', ( login, password ) ) userContent = cur.fetchone() if userContent is not None: userContent = userContent[0] except db.Error as e: logging.error('loginUser: Error from MySQL => %s' % e) finally: if con: con.close() return userContentRemove some bug on mysql elements.
#!/usr/bin/env python # -*- coding: utf-8 -*- from system.DBMysql import connect from system.ConfigLoader import getCfg import logging def loginUser(login, password): ''' Try to login a user regarding login/password ''' userContent = None table = getCfg('MYSQL', 'table') tableId = getCfg('MYSQL', 'idField') tableLogin = getCfg('MYSQL', 'loginField') tablePassword = getCfg('MYSQL', 'passwordField') con = None try: # Starting con = connect() cur = con.cursor() cur.execute( 'SELECT ' + tableId + ' FROM ' + table + ' WHERE ' + tableLogin + '=%s AND ' + tablePassword + '=%s', ( login, password ) ) userContent = cur.fetchone() if userContent is not None: userContent = userContent[0] except Exception as e: logging.error('loginUser: Error from MySQL => %s' % e) finally: if con: con.close() return userContent
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from system.DBMysql import connect from system.ConfigLoader import getCfg def loginUser(login, password): ''' Try to login a user regarding login/password ''' userContent = None table = getCfg('MYSQL', 'table') tableId = getCfg('MYSQL', 'idField') tableLogin = getCfg('MYSQL', 'loginField') tablePassword = getCfg('MYSQL', 'passwordField') try: # Starting con = connect() cur = con.cursor() cur.execute( 'SELECT ' + tableId + ' FROM ' + table + ' WHERE ' + tableLogin + '=%s AND ' + tablePassword + '=%s', ( login, password ) ) userContent = cur.fetchone() if userContent is not None: userContent = userContent[0] except db.Error as e: logging.error('loginUser: Error from MySQL => %s' % e) finally: if con: con.close() return userContent<commit_msg>Remove some bug on mysql elements.<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from system.DBMysql import connect from system.ConfigLoader import getCfg import logging def loginUser(login, password): ''' Try to login a user regarding login/password ''' userContent = None table = getCfg('MYSQL', 'table') tableId = getCfg('MYSQL', 'idField') tableLogin = getCfg('MYSQL', 'loginField') tablePassword = getCfg('MYSQL', 'passwordField') con = None try: # Starting con = connect() cur = con.cursor() cur.execute( 'SELECT ' + tableId + ' FROM ' + table + ' WHERE ' + tableLogin + '=%s AND ' + tablePassword + '=%s', ( login, password ) ) userContent = cur.fetchone() if userContent is not None: userContent = userContent[0] except Exception as e: logging.error('loginUser: Error from MySQL => %s' % e) finally: if con: con.close() return userContent
#!/usr/bin/env python # -*- coding: utf-8 -*- from system.DBMysql import connect from system.ConfigLoader import getCfg def loginUser(login, password): ''' Try to login a user regarding login/password ''' userContent = None table = getCfg('MYSQL', 'table') tableId = getCfg('MYSQL', 'idField') tableLogin = getCfg('MYSQL', 'loginField') tablePassword = getCfg('MYSQL', 'passwordField') try: # Starting con = connect() cur = con.cursor() cur.execute( 'SELECT ' + tableId + ' FROM ' + table + ' WHERE ' + tableLogin + '=%s AND ' + tablePassword + '=%s', ( login, password ) ) userContent = cur.fetchone() if userContent is not None: userContent = userContent[0] except db.Error as e: logging.error('loginUser: Error from MySQL => %s' % e) finally: if con: con.close() return userContentRemove some bug on mysql elements.#!/usr/bin/env python # -*- coding: utf-8 -*- from system.DBMysql import connect from system.ConfigLoader import getCfg import logging def loginUser(login, password): ''' Try to login a user regarding login/password ''' userContent = None table = getCfg('MYSQL', 'table') tableId = getCfg('MYSQL', 'idField') tableLogin = getCfg('MYSQL', 'loginField') tablePassword = getCfg('MYSQL', 'passwordField') con = None try: # Starting con = connect() cur = con.cursor() cur.execute( 'SELECT ' + tableId + ' FROM ' + table + ' WHERE ' + tableLogin + '=%s AND ' + tablePassword + '=%s', ( login, password ) ) userContent = cur.fetchone() if userContent is not None: userContent = userContent[0] except Exception as e: logging.error('loginUser: Error from MySQL => %s' % e) finally: if con: con.close() return userContent
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from system.DBMysql import connect from system.ConfigLoader import getCfg def loginUser(login, password): ''' Try to login a user regarding login/password ''' userContent = None table = getCfg('MYSQL', 'table') tableId = getCfg('MYSQL', 'idField') tableLogin = getCfg('MYSQL', 'loginField') tablePassword = getCfg('MYSQL', 'passwordField') try: # Starting con = connect() cur = con.cursor() cur.execute( 'SELECT ' + tableId + ' FROM ' + table + ' WHERE ' + tableLogin + '=%s AND ' + tablePassword + '=%s', ( login, password ) ) userContent = cur.fetchone() if userContent is not None: userContent = userContent[0] except db.Error as e: logging.error('loginUser: Error from MySQL => %s' % e) finally: if con: con.close() return userContent<commit_msg>Remove some bug on mysql elements.<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from system.DBMysql import connect from system.ConfigLoader import getCfg import logging def loginUser(login, password): ''' Try to login a user regarding login/password ''' userContent = None table = getCfg('MYSQL', 'table') tableId = getCfg('MYSQL', 'idField') tableLogin = getCfg('MYSQL', 'loginField') tablePassword = getCfg('MYSQL', 'passwordField') con = None try: # Starting con = connect() cur = con.cursor() cur.execute( 'SELECT ' + tableId + ' FROM ' + table + ' WHERE ' + tableLogin + '=%s AND ' + tablePassword + '=%s', ( login, password ) ) userContent = cur.fetchone() if userContent is not None: userContent = userContent[0] except Exception as e: logging.error('loginUser: Error from MySQL => %s' % e) finally: if con: con.close() return userContent
9ffb43c969288bdb06de20aded660bd7e4e5c337
plugins/witai.py
plugins/witai.py
import os from slackbot.bot import respond_to from slackbot.bot import listen_to from wit import Wit from wit_actions import actions wit_client = Wit( access_token=os.environ['WIT_API_TOKEN'], actions=actions ) @respond_to('') def wit(message): message.react('+1') message.reply('You said: ' + message.body['text']) wit_client.run_actions('asdf', message.body['text'])
import os import random import string from slackbot.bot import respond_to from slackbot.bot import listen_to from wit import Wit from wit_actions import actions wit_client = Wit( access_token=os.environ['WIT_API_TOKEN'], actions=actions ) wit_context = {} def random_word(length): return ''.join(random.choice(string.lowercase) for i in range(length)) @respond_to('') def wit(message): global wit_context message.react('+1') wit_context = wit_client.run_actions( random_word(10), message.body['text'], wit_context)
Create new session on restart
Create new session on restart
Python
mit
sanjaybv/netbot
import os from slackbot.bot import respond_to from slackbot.bot import listen_to from wit import Wit from wit_actions import actions wit_client = Wit( access_token=os.environ['WIT_API_TOKEN'], actions=actions ) @respond_to('') def wit(message): message.react('+1') message.reply('You said: ' + message.body['text']) wit_client.run_actions('asdf', message.body['text']) Create new session on restart
import os import random import string from slackbot.bot import respond_to from slackbot.bot import listen_to from wit import Wit from wit_actions import actions wit_client = Wit( access_token=os.environ['WIT_API_TOKEN'], actions=actions ) wit_context = {} def random_word(length): return ''.join(random.choice(string.lowercase) for i in range(length)) @respond_to('') def wit(message): global wit_context message.react('+1') wit_context = wit_client.run_actions( random_word(10), message.body['text'], wit_context)
<commit_before>import os from slackbot.bot import respond_to from slackbot.bot import listen_to from wit import Wit from wit_actions import actions wit_client = Wit( access_token=os.environ['WIT_API_TOKEN'], actions=actions ) @respond_to('') def wit(message): message.react('+1') message.reply('You said: ' + message.body['text']) wit_client.run_actions('asdf', message.body['text']) <commit_msg>Create new session on restart<commit_after>
import os import random import string from slackbot.bot import respond_to from slackbot.bot import listen_to from wit import Wit from wit_actions import actions wit_client = Wit( access_token=os.environ['WIT_API_TOKEN'], actions=actions ) wit_context = {} def random_word(length): return ''.join(random.choice(string.lowercase) for i in range(length)) @respond_to('') def wit(message): global wit_context message.react('+1') wit_context = wit_client.run_actions( random_word(10), message.body['text'], wit_context)
import os from slackbot.bot import respond_to from slackbot.bot import listen_to from wit import Wit from wit_actions import actions wit_client = Wit( access_token=os.environ['WIT_API_TOKEN'], actions=actions ) @respond_to('') def wit(message): message.react('+1') message.reply('You said: ' + message.body['text']) wit_client.run_actions('asdf', message.body['text']) Create new session on restartimport os import random import string from slackbot.bot import respond_to from slackbot.bot import listen_to from wit import Wit from wit_actions import actions wit_client = Wit( access_token=os.environ['WIT_API_TOKEN'], actions=actions ) wit_context = {} def random_word(length): return ''.join(random.choice(string.lowercase) for i in range(length)) @respond_to('') def wit(message): global wit_context message.react('+1') wit_context = wit_client.run_actions( random_word(10), message.body['text'], wit_context)
<commit_before>import os from slackbot.bot import respond_to from slackbot.bot import listen_to from wit import Wit from wit_actions import actions wit_client = Wit( access_token=os.environ['WIT_API_TOKEN'], actions=actions ) @respond_to('') def wit(message): message.react('+1') message.reply('You said: ' + message.body['text']) wit_client.run_actions('asdf', message.body['text']) <commit_msg>Create new session on restart<commit_after>import os import random import string from slackbot.bot import respond_to from slackbot.bot import listen_to from wit import Wit from wit_actions import actions wit_client = Wit( access_token=os.environ['WIT_API_TOKEN'], actions=actions ) wit_context = {} def random_word(length): return ''.join(random.choice(string.lowercase) for i in range(length)) @respond_to('') def wit(message): global wit_context message.react('+1') wit_context = wit_client.run_actions( random_word(10), message.body['text'], wit_context)
667fbe85d0e0ce60a696e63aa705dcb6b59c1214
geokey_wegovnow/__init__.py
geokey_wegovnow/__init__.py
"""Main initialization for the WeGovNow extension.""" VERSION = (3, 1, 1) __version__ = '.'.join(map(str, VERSION)) try: from geokey.extensions.base import register register( 'geokey_wegovnow', 'WeGovNow', display_admin=False, superuser=False, version=__version__ ) except BaseException: print 'Please install GeoKey first'
"""Main initialization for the WeGovNow extension.""" VERSION = (3, 1, 2) __version__ = '.'.join(map(str, VERSION)) try: from geokey.extensions.base import register register( 'geokey_wegovnow', 'WeGovNow', display_admin=False, superuser=False, version=__version__ ) except BaseException: print 'Please install GeoKey first'
Increment patch number following bugfix.
Increment patch number following bugfix.
Python
mit
ExCiteS/geokey-wegovnow,ExCiteS/geokey-wegovnow
"""Main initialization for the WeGovNow extension.""" VERSION = (3, 1, 1) __version__ = '.'.join(map(str, VERSION)) try: from geokey.extensions.base import register register( 'geokey_wegovnow', 'WeGovNow', display_admin=False, superuser=False, version=__version__ ) except BaseException: print 'Please install GeoKey first' Increment patch number following bugfix.
"""Main initialization for the WeGovNow extension.""" VERSION = (3, 1, 2) __version__ = '.'.join(map(str, VERSION)) try: from geokey.extensions.base import register register( 'geokey_wegovnow', 'WeGovNow', display_admin=False, superuser=False, version=__version__ ) except BaseException: print 'Please install GeoKey first'
<commit_before>"""Main initialization for the WeGovNow extension.""" VERSION = (3, 1, 1) __version__ = '.'.join(map(str, VERSION)) try: from geokey.extensions.base import register register( 'geokey_wegovnow', 'WeGovNow', display_admin=False, superuser=False, version=__version__ ) except BaseException: print 'Please install GeoKey first' <commit_msg>Increment patch number following bugfix.<commit_after>
"""Main initialization for the WeGovNow extension.""" VERSION = (3, 1, 2) __version__ = '.'.join(map(str, VERSION)) try: from geokey.extensions.base import register register( 'geokey_wegovnow', 'WeGovNow', display_admin=False, superuser=False, version=__version__ ) except BaseException: print 'Please install GeoKey first'
"""Main initialization for the WeGovNow extension.""" VERSION = (3, 1, 1) __version__ = '.'.join(map(str, VERSION)) try: from geokey.extensions.base import register register( 'geokey_wegovnow', 'WeGovNow', display_admin=False, superuser=False, version=__version__ ) except BaseException: print 'Please install GeoKey first' Increment patch number following bugfix."""Main initialization for the WeGovNow extension.""" VERSION = (3, 1, 2) __version__ = '.'.join(map(str, VERSION)) try: from geokey.extensions.base import register register( 'geokey_wegovnow', 'WeGovNow', display_admin=False, superuser=False, version=__version__ ) except BaseException: print 'Please install GeoKey first'
<commit_before>"""Main initialization for the WeGovNow extension.""" VERSION = (3, 1, 1) __version__ = '.'.join(map(str, VERSION)) try: from geokey.extensions.base import register register( 'geokey_wegovnow', 'WeGovNow', display_admin=False, superuser=False, version=__version__ ) except BaseException: print 'Please install GeoKey first' <commit_msg>Increment patch number following bugfix.<commit_after>"""Main initialization for the WeGovNow extension.""" VERSION = (3, 1, 2) __version__ = '.'.join(map(str, VERSION)) try: from geokey.extensions.base import register register( 'geokey_wegovnow', 'WeGovNow', display_admin=False, superuser=False, version=__version__ ) except BaseException: print 'Please install GeoKey first'
25351cd6b9119ea27123a2fddbbcc274c3620886
examples/examples.py
examples/examples.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """Usage examples """ from __future__ import print_function, division import numpy as np import matplotlib.pylab as plt import seaborn as sns from multidensity import MultiDensity from skewstudent import SkewStudent def estimate_bivariate_mle(): ndim = 2 size = (1000, ndim) data = np.random.normal(size=size) eta, lam = 4, -.9 skst = SkewStudent(eta=eta, lam=lam) data = skst.rvs(size=size) out = MultiDensity.fit_mle(data=data) print(out) mdens = MultiDensity() mdens.from_theta(out.x) fig, axes = plt.subplots(nrows=size[1], ncols=1) for innov, ax in zip(data.T, axes): sns.kdeplot(innov, ax=ax) lines = [ax.get_lines()[0].get_xdata() for ax in axes] lines = np.vstack(lines).T marginals = mdens.marginals(lines) for line, margin, ax in zip(lines.T, marginals.T, axes): ax.plot(line, margin) plt.show() if __name__ == '__main__': estimate_bivariate_mle()
#!/usr/bin/env python # -*- coding: utf-8 -*- """Usage examples """ from __future__ import print_function, division import numpy as np import matplotlib.pylab as plt import seaborn as sns from multidensity import MultiDensity from skewstudent import SkewStudent def estimate_bivariate_mle(): ndim = 2 size = (1000, ndim) data = np.random.normal(size=size) eta, lam = 4, -.9 skst = SkewStudent(eta=eta, lam=lam) data = skst.rvs(size=size) out = MultiDensity.fit_mle(data=data) print(out) mdens = MultiDensity() mdens.from_theta(out.x) fig, axes = plt.subplots(nrows=size[1], ncols=1) for innov, ax in zip(data.T, axes): sns.kdeplot(innov, ax=ax, label='data') lines = [ax.get_lines()[0].get_xdata() for ax in axes] lines = np.vstack(lines).T marginals = mdens.marginals(lines) for line, margin, ax in zip(lines.T, marginals.T, axes): ax.plot(line, margin, label='fitted') ax.legend() plt.show() if __name__ == '__main__': estimate_bivariate_mle()
Add plot legend in the example
Add plot legend in the example
Python
mit
khrapovs/multidensity
#!/usr/bin/env python # -*- coding: utf-8 -*- """Usage examples """ from __future__ import print_function, division import numpy as np import matplotlib.pylab as plt import seaborn as sns from multidensity import MultiDensity from skewstudent import SkewStudent def estimate_bivariate_mle(): ndim = 2 size = (1000, ndim) data = np.random.normal(size=size) eta, lam = 4, -.9 skst = SkewStudent(eta=eta, lam=lam) data = skst.rvs(size=size) out = MultiDensity.fit_mle(data=data) print(out) mdens = MultiDensity() mdens.from_theta(out.x) fig, axes = plt.subplots(nrows=size[1], ncols=1) for innov, ax in zip(data.T, axes): sns.kdeplot(innov, ax=ax) lines = [ax.get_lines()[0].get_xdata() for ax in axes] lines = np.vstack(lines).T marginals = mdens.marginals(lines) for line, margin, ax in zip(lines.T, marginals.T, axes): ax.plot(line, margin) plt.show() if __name__ == '__main__': estimate_bivariate_mle() Add plot legend in the example
#!/usr/bin/env python # -*- coding: utf-8 -*- """Usage examples """ from __future__ import print_function, division import numpy as np import matplotlib.pylab as plt import seaborn as sns from multidensity import MultiDensity from skewstudent import SkewStudent def estimate_bivariate_mle(): ndim = 2 size = (1000, ndim) data = np.random.normal(size=size) eta, lam = 4, -.9 skst = SkewStudent(eta=eta, lam=lam) data = skst.rvs(size=size) out = MultiDensity.fit_mle(data=data) print(out) mdens = MultiDensity() mdens.from_theta(out.x) fig, axes = plt.subplots(nrows=size[1], ncols=1) for innov, ax in zip(data.T, axes): sns.kdeplot(innov, ax=ax, label='data') lines = [ax.get_lines()[0].get_xdata() for ax in axes] lines = np.vstack(lines).T marginals = mdens.marginals(lines) for line, margin, ax in zip(lines.T, marginals.T, axes): ax.plot(line, margin, label='fitted') ax.legend() plt.show() if __name__ == '__main__': estimate_bivariate_mle()
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """Usage examples """ from __future__ import print_function, division import numpy as np import matplotlib.pylab as plt import seaborn as sns from multidensity import MultiDensity from skewstudent import SkewStudent def estimate_bivariate_mle(): ndim = 2 size = (1000, ndim) data = np.random.normal(size=size) eta, lam = 4, -.9 skst = SkewStudent(eta=eta, lam=lam) data = skst.rvs(size=size) out = MultiDensity.fit_mle(data=data) print(out) mdens = MultiDensity() mdens.from_theta(out.x) fig, axes = plt.subplots(nrows=size[1], ncols=1) for innov, ax in zip(data.T, axes): sns.kdeplot(innov, ax=ax) lines = [ax.get_lines()[0].get_xdata() for ax in axes] lines = np.vstack(lines).T marginals = mdens.marginals(lines) for line, margin, ax in zip(lines.T, marginals.T, axes): ax.plot(line, margin) plt.show() if __name__ == '__main__': estimate_bivariate_mle() <commit_msg>Add plot legend in the example<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- """Usage examples """ from __future__ import print_function, division import numpy as np import matplotlib.pylab as plt import seaborn as sns from multidensity import MultiDensity from skewstudent import SkewStudent def estimate_bivariate_mle(): ndim = 2 size = (1000, ndim) data = np.random.normal(size=size) eta, lam = 4, -.9 skst = SkewStudent(eta=eta, lam=lam) data = skst.rvs(size=size) out = MultiDensity.fit_mle(data=data) print(out) mdens = MultiDensity() mdens.from_theta(out.x) fig, axes = plt.subplots(nrows=size[1], ncols=1) for innov, ax in zip(data.T, axes): sns.kdeplot(innov, ax=ax, label='data') lines = [ax.get_lines()[0].get_xdata() for ax in axes] lines = np.vstack(lines).T marginals = mdens.marginals(lines) for line, margin, ax in zip(lines.T, marginals.T, axes): ax.plot(line, margin, label='fitted') ax.legend() plt.show() if __name__ == '__main__': estimate_bivariate_mle()
#!/usr/bin/env python # -*- coding: utf-8 -*- """Usage examples """ from __future__ import print_function, division import numpy as np import matplotlib.pylab as plt import seaborn as sns from multidensity import MultiDensity from skewstudent import SkewStudent def estimate_bivariate_mle(): ndim = 2 size = (1000, ndim) data = np.random.normal(size=size) eta, lam = 4, -.9 skst = SkewStudent(eta=eta, lam=lam) data = skst.rvs(size=size) out = MultiDensity.fit_mle(data=data) print(out) mdens = MultiDensity() mdens.from_theta(out.x) fig, axes = plt.subplots(nrows=size[1], ncols=1) for innov, ax in zip(data.T, axes): sns.kdeplot(innov, ax=ax) lines = [ax.get_lines()[0].get_xdata() for ax in axes] lines = np.vstack(lines).T marginals = mdens.marginals(lines) for line, margin, ax in zip(lines.T, marginals.T, axes): ax.plot(line, margin) plt.show() if __name__ == '__main__': estimate_bivariate_mle() Add plot legend in the example#!/usr/bin/env python # -*- coding: utf-8 -*- """Usage examples """ from __future__ import print_function, division import numpy as np import matplotlib.pylab as plt import seaborn as sns from multidensity import MultiDensity from skewstudent import SkewStudent def estimate_bivariate_mle(): ndim = 2 size = (1000, ndim) data = np.random.normal(size=size) eta, lam = 4, -.9 skst = SkewStudent(eta=eta, lam=lam) data = skst.rvs(size=size) out = MultiDensity.fit_mle(data=data) print(out) mdens = MultiDensity() mdens.from_theta(out.x) fig, axes = plt.subplots(nrows=size[1], ncols=1) for innov, ax in zip(data.T, axes): sns.kdeplot(innov, ax=ax, label='data') lines = [ax.get_lines()[0].get_xdata() for ax in axes] lines = np.vstack(lines).T marginals = mdens.marginals(lines) for line, margin, ax in zip(lines.T, marginals.T, axes): ax.plot(line, margin, label='fitted') ax.legend() plt.show() if __name__ == '__main__': estimate_bivariate_mle()
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """Usage examples """ from __future__ import print_function, division import numpy as np import matplotlib.pylab as plt import seaborn as sns from multidensity import MultiDensity from skewstudent import SkewStudent def estimate_bivariate_mle(): ndim = 2 size = (1000, ndim) data = np.random.normal(size=size) eta, lam = 4, -.9 skst = SkewStudent(eta=eta, lam=lam) data = skst.rvs(size=size) out = MultiDensity.fit_mle(data=data) print(out) mdens = MultiDensity() mdens.from_theta(out.x) fig, axes = plt.subplots(nrows=size[1], ncols=1) for innov, ax in zip(data.T, axes): sns.kdeplot(innov, ax=ax) lines = [ax.get_lines()[0].get_xdata() for ax in axes] lines = np.vstack(lines).T marginals = mdens.marginals(lines) for line, margin, ax in zip(lines.T, marginals.T, axes): ax.plot(line, margin) plt.show() if __name__ == '__main__': estimate_bivariate_mle() <commit_msg>Add plot legend in the example<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- """Usage examples """ from __future__ import print_function, division import numpy as np import matplotlib.pylab as plt import seaborn as sns from multidensity import MultiDensity from skewstudent import SkewStudent def estimate_bivariate_mle(): ndim = 2 size = (1000, ndim) data = np.random.normal(size=size) eta, lam = 4, -.9 skst = SkewStudent(eta=eta, lam=lam) data = skst.rvs(size=size) out = MultiDensity.fit_mle(data=data) print(out) mdens = MultiDensity() mdens.from_theta(out.x) fig, axes = plt.subplots(nrows=size[1], ncols=1) for innov, ax in zip(data.T, axes): sns.kdeplot(innov, ax=ax, label='data') lines = [ax.get_lines()[0].get_xdata() for ax in axes] lines = np.vstack(lines).T marginals = mdens.marginals(lines) for line, margin, ax in zip(lines.T, marginals.T, axes): ax.plot(line, margin, label='fitted') ax.legend() plt.show() if __name__ == '__main__': estimate_bivariate_mle()
cde9dd479b2974f26f2e50b3611bfd0756f86c2b
game_of_thrones/__init__.py
game_of_thrones/__init__.py
class MarkovChain: """ Entity which contains a chunk of text and a Markov chain generated from it. """ def __init__(self, text): self.text = text def pair_symbols(self, text): return [pair for pair in zip(text[0::1], text[1::1])]
class MarkovChain: """ Entity which contains a chunk of text and a Markov chain generated from it. """ def __init__(self, text): self.text = text def pair_symbols(self, text): """ Takes an string and returns a list of tuples. For example: >>> pair_symbols('Arya') [('A', 'r'), ('r', 'y'), ('y', 'a')] """ return [pair for pair in zip(text[0::1], text[1::1])]
Add docstring to the `pair_symbols` method
Add docstring to the `pair_symbols` method
Python
mit
Matt-Deacalion/Name-of-Thrones
class MarkovChain: """ Entity which contains a chunk of text and a Markov chain generated from it. """ def __init__(self, text): self.text = text def pair_symbols(self, text): return [pair for pair in zip(text[0::1], text[1::1])] Add docstring to the `pair_symbols` method
class MarkovChain: """ Entity which contains a chunk of text and a Markov chain generated from it. """ def __init__(self, text): self.text = text def pair_symbols(self, text): """ Takes an string and returns a list of tuples. For example: >>> pair_symbols('Arya') [('A', 'r'), ('r', 'y'), ('y', 'a')] """ return [pair for pair in zip(text[0::1], text[1::1])]
<commit_before>class MarkovChain: """ Entity which contains a chunk of text and a Markov chain generated from it. """ def __init__(self, text): self.text = text def pair_symbols(self, text): return [pair for pair in zip(text[0::1], text[1::1])] <commit_msg>Add docstring to the `pair_symbols` method<commit_after>
class MarkovChain: """ Entity which contains a chunk of text and a Markov chain generated from it. """ def __init__(self, text): self.text = text def pair_symbols(self, text): """ Takes an string and returns a list of tuples. For example: >>> pair_symbols('Arya') [('A', 'r'), ('r', 'y'), ('y', 'a')] """ return [pair for pair in zip(text[0::1], text[1::1])]
class MarkovChain: """ Entity which contains a chunk of text and a Markov chain generated from it. """ def __init__(self, text): self.text = text def pair_symbols(self, text): return [pair for pair in zip(text[0::1], text[1::1])] Add docstring to the `pair_symbols` methodclass MarkovChain: """ Entity which contains a chunk of text and a Markov chain generated from it. """ def __init__(self, text): self.text = text def pair_symbols(self, text): """ Takes an string and returns a list of tuples. For example: >>> pair_symbols('Arya') [('A', 'r'), ('r', 'y'), ('y', 'a')] """ return [pair for pair in zip(text[0::1], text[1::1])]
<commit_before>class MarkovChain: """ Entity which contains a chunk of text and a Markov chain generated from it. """ def __init__(self, text): self.text = text def pair_symbols(self, text): return [pair for pair in zip(text[0::1], text[1::1])] <commit_msg>Add docstring to the `pair_symbols` method<commit_after>class MarkovChain: """ Entity which contains a chunk of text and a Markov chain generated from it. """ def __init__(self, text): self.text = text def pair_symbols(self, text): """ Takes an string and returns a list of tuples. For example: >>> pair_symbols('Arya') [('A', 'r'), ('r', 'y'), ('y', 'a')] """ return [pair for pair in zip(text[0::1], text[1::1])]
c8cbd2c660a2e10cc021bdc0a2fead872f77967d
ynr/apps/uk_results/migrations/0054_update_is_winner_to_elected.py
ynr/apps/uk_results/migrations/0054_update_is_winner_to_elected.py
# Generated by Django 2.2.18 on 2021-05-13 11:27 from django.db import migrations def update_versions(versions, current_key, new_key): for version in versions: for result in version["candidate_results"]: try: result[new_key] = result.pop(current_key) except KeyError: continue def forwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all(): update_versions( versions=result.versions, current_key="is_winner", new_key="elected" ) result.save() def backwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all(): update_versions( versions=result.versions, current_key="elected", new_key="is_winner" ) result.save() class Migration(migrations.Migration): dependencies = [("uk_results", "0053_auto_20210928_1007")] operations = [migrations.RunPython(code=forwards, reverse_code=backwards)]
# Generated by Django 2.2.18 on 2021-05-13 11:27 from django.db import migrations def update_versions(versions, current_key, new_key): for version in versions: for result in version["candidate_results"]: try: result[new_key] = result.pop(current_key) except KeyError: continue def forwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all().iterator(): update_versions( versions=result.versions, current_key="is_winner", new_key="elected" ) result.save() def backwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.iterator(): update_versions( versions=result.versions, current_key="elected", new_key="is_winner" ) result.save() class Migration(migrations.Migration): dependencies = [("uk_results", "0053_auto_20210928_1007")] operations = [migrations.RunPython(code=forwards, reverse_code=backwards)]
Use iterator in resultset data migration
Use iterator in resultset data migration
Python
agpl-3.0
DemocracyClub/yournextrepresentative,DemocracyClub/yournextrepresentative,DemocracyClub/yournextrepresentative
# Generated by Django 2.2.18 on 2021-05-13 11:27 from django.db import migrations def update_versions(versions, current_key, new_key): for version in versions: for result in version["candidate_results"]: try: result[new_key] = result.pop(current_key) except KeyError: continue def forwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all(): update_versions( versions=result.versions, current_key="is_winner", new_key="elected" ) result.save() def backwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all(): update_versions( versions=result.versions, current_key="elected", new_key="is_winner" ) result.save() class Migration(migrations.Migration): dependencies = [("uk_results", "0053_auto_20210928_1007")] operations = [migrations.RunPython(code=forwards, reverse_code=backwards)] Use iterator in resultset data migration
# Generated by Django 2.2.18 on 2021-05-13 11:27 from django.db import migrations def update_versions(versions, current_key, new_key): for version in versions: for result in version["candidate_results"]: try: result[new_key] = result.pop(current_key) except KeyError: continue def forwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all().iterator(): update_versions( versions=result.versions, current_key="is_winner", new_key="elected" ) result.save() def backwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.iterator(): update_versions( versions=result.versions, current_key="elected", new_key="is_winner" ) result.save() class Migration(migrations.Migration): dependencies = [("uk_results", "0053_auto_20210928_1007")] operations = [migrations.RunPython(code=forwards, reverse_code=backwards)]
<commit_before># Generated by Django 2.2.18 on 2021-05-13 11:27 from django.db import migrations def update_versions(versions, current_key, new_key): for version in versions: for result in version["candidate_results"]: try: result[new_key] = result.pop(current_key) except KeyError: continue def forwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all(): update_versions( versions=result.versions, current_key="is_winner", new_key="elected" ) result.save() def backwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all(): update_versions( versions=result.versions, current_key="elected", new_key="is_winner" ) result.save() class Migration(migrations.Migration): dependencies = [("uk_results", "0053_auto_20210928_1007")] operations = [migrations.RunPython(code=forwards, reverse_code=backwards)] <commit_msg>Use iterator in resultset data migration<commit_after>
# Generated by Django 2.2.18 on 2021-05-13 11:27 from django.db import migrations def update_versions(versions, current_key, new_key): for version in versions: for result in version["candidate_results"]: try: result[new_key] = result.pop(current_key) except KeyError: continue def forwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all().iterator(): update_versions( versions=result.versions, current_key="is_winner", new_key="elected" ) result.save() def backwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.iterator(): update_versions( versions=result.versions, current_key="elected", new_key="is_winner" ) result.save() class Migration(migrations.Migration): dependencies = [("uk_results", "0053_auto_20210928_1007")] operations = [migrations.RunPython(code=forwards, reverse_code=backwards)]
# Generated by Django 2.2.18 on 2021-05-13 11:27 from django.db import migrations def update_versions(versions, current_key, new_key): for version in versions: for result in version["candidate_results"]: try: result[new_key] = result.pop(current_key) except KeyError: continue def forwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all(): update_versions( versions=result.versions, current_key="is_winner", new_key="elected" ) result.save() def backwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all(): update_versions( versions=result.versions, current_key="elected", new_key="is_winner" ) result.save() class Migration(migrations.Migration): dependencies = [("uk_results", "0053_auto_20210928_1007")] operations = [migrations.RunPython(code=forwards, reverse_code=backwards)] Use iterator in resultset data migration# Generated by Django 2.2.18 on 2021-05-13 11:27 from django.db import migrations def update_versions(versions, current_key, new_key): for version in versions: for result in version["candidate_results"]: try: result[new_key] = result.pop(current_key) except KeyError: continue def forwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all().iterator(): update_versions( versions=result.versions, current_key="is_winner", new_key="elected" ) result.save() def backwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.iterator(): update_versions( versions=result.versions, current_key="elected", new_key="is_winner" ) result.save() class Migration(migrations.Migration): dependencies = [("uk_results", "0053_auto_20210928_1007")] operations = [migrations.RunPython(code=forwards, reverse_code=backwards)]
<commit_before># Generated by Django 2.2.18 on 2021-05-13 11:27 from django.db import migrations def update_versions(versions, current_key, new_key): for version in versions: for result in version["candidate_results"]: try: result[new_key] = result.pop(current_key) except KeyError: continue def forwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all(): update_versions( versions=result.versions, current_key="is_winner", new_key="elected" ) result.save() def backwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all(): update_versions( versions=result.versions, current_key="elected", new_key="is_winner" ) result.save() class Migration(migrations.Migration): dependencies = [("uk_results", "0053_auto_20210928_1007")] operations = [migrations.RunPython(code=forwards, reverse_code=backwards)] <commit_msg>Use iterator in resultset data migration<commit_after># Generated by Django 2.2.18 on 2021-05-13 11:27 from django.db import migrations def update_versions(versions, current_key, new_key): for version in versions: for result in version["candidate_results"]: try: result[new_key] = result.pop(current_key) except KeyError: continue def forwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.all().iterator(): update_versions( versions=result.versions, current_key="is_winner", new_key="elected" ) result.save() def backwards(apps, schema_editor): ResultSet = apps.get_model("uk_results", "ResultSet") for result in ResultSet.objects.iterator(): update_versions( versions=result.versions, current_key="elected", new_key="is_winner" ) result.save() class Migration(migrations.Migration): dependencies = [("uk_results", "0053_auto_20210928_1007")] operations = [migrations.RunPython(code=forwards, reverse_code=backwards)]
32d2d6069191cd2f7e98d1dedce13ddc1a6b1f32
config.py
config.py
dbdir = "/usr/local/finance/db" imgdir = "/usr/local/finance/img" apiurl = "https://sixpak.org/finance/api.py" sessiondir = "/usr/local/finance/session" sessiontimeout = (60*15) # 15 minutes banks = [ "bankofamerica", "citicards", "wellsfargo", "chase", "paypal", "gap", "fecu", "csvurl" ]
dbdir = "/usr/local/finance/db" imgdir = "/usr/local/finance/img" apiurl = "https://example.com/pywebfinance/api.py" sessiondir = "/usr/local/finance/session" sessiontimeout = (60*15) # 15 minutes banks = [ "bankofamerica", "citicards", "wellsfargo", "chase", "paypal", "gap", "fecu", "csvurl" ]
Use example.com for API URL.
Use example.com for API URL.
Python
agpl-3.0
vincebusam/pyWebCash,vincebusam/pyWebCash,vincebusam/pyWebCash
dbdir = "/usr/local/finance/db" imgdir = "/usr/local/finance/img" apiurl = "https://sixpak.org/finance/api.py" sessiondir = "/usr/local/finance/session" sessiontimeout = (60*15) # 15 minutes banks = [ "bankofamerica", "citicards", "wellsfargo", "chase", "paypal", "gap", "fecu", "csvurl" ] Use example.com for API URL.
dbdir = "/usr/local/finance/db" imgdir = "/usr/local/finance/img" apiurl = "https://example.com/pywebfinance/api.py" sessiondir = "/usr/local/finance/session" sessiontimeout = (60*15) # 15 minutes banks = [ "bankofamerica", "citicards", "wellsfargo", "chase", "paypal", "gap", "fecu", "csvurl" ]
<commit_before>dbdir = "/usr/local/finance/db" imgdir = "/usr/local/finance/img" apiurl = "https://sixpak.org/finance/api.py" sessiondir = "/usr/local/finance/session" sessiontimeout = (60*15) # 15 minutes banks = [ "bankofamerica", "citicards", "wellsfargo", "chase", "paypal", "gap", "fecu", "csvurl" ] <commit_msg>Use example.com for API URL.<commit_after>
dbdir = "/usr/local/finance/db" imgdir = "/usr/local/finance/img" apiurl = "https://example.com/pywebfinance/api.py" sessiondir = "/usr/local/finance/session" sessiontimeout = (60*15) # 15 minutes banks = [ "bankofamerica", "citicards", "wellsfargo", "chase", "paypal", "gap", "fecu", "csvurl" ]
dbdir = "/usr/local/finance/db" imgdir = "/usr/local/finance/img" apiurl = "https://sixpak.org/finance/api.py" sessiondir = "/usr/local/finance/session" sessiontimeout = (60*15) # 15 minutes banks = [ "bankofamerica", "citicards", "wellsfargo", "chase", "paypal", "gap", "fecu", "csvurl" ] Use example.com for API URL.dbdir = "/usr/local/finance/db" imgdir = "/usr/local/finance/img" apiurl = "https://example.com/pywebfinance/api.py" sessiondir = "/usr/local/finance/session" sessiontimeout = (60*15) # 15 minutes banks = [ "bankofamerica", "citicards", "wellsfargo", "chase", "paypal", "gap", "fecu", "csvurl" ]
<commit_before>dbdir = "/usr/local/finance/db" imgdir = "/usr/local/finance/img" apiurl = "https://sixpak.org/finance/api.py" sessiondir = "/usr/local/finance/session" sessiontimeout = (60*15) # 15 minutes banks = [ "bankofamerica", "citicards", "wellsfargo", "chase", "paypal", "gap", "fecu", "csvurl" ] <commit_msg>Use example.com for API URL.<commit_after>dbdir = "/usr/local/finance/db" imgdir = "/usr/local/finance/img" apiurl = "https://example.com/pywebfinance/api.py" sessiondir = "/usr/local/finance/session" sessiontimeout = (60*15) # 15 minutes banks = [ "bankofamerica", "citicards", "wellsfargo", "chase", "paypal", "gap", "fecu", "csvurl" ]