commit stringlengths 40 40 | old_file stringlengths 4 118 | new_file stringlengths 4 118 | old_contents stringlengths 0 2.94k | new_contents stringlengths 1 4.43k | subject stringlengths 15 444 | message stringlengths 16 3.45k | lang stringclasses 1 value | license stringclasses 13 values | repos stringlengths 5 43.2k | prompt stringlengths 17 4.58k | response stringlengths 1 4.43k | prompt_tagged stringlengths 58 4.62k | response_tagged stringlengths 1 4.43k | text stringlengths 132 7.29k | text_tagged stringlengths 173 7.33k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
7d756efb7361c13b0db0f37ead9668351c3a6887 | unitTestUtils/parseXML.py | unitTestUtils/parseXML.py | #!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
| #!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
print(infile)
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
| Add a print with file where mistake is | Add a print with file where mistake is
| Python | apache-2.0 | wkrzemien/j-pet-framework,wkrzemien/j-pet-framework,wkrzemien/j-pet-framework,wkrzemien/j-pet-framework,wkrzemien/j-pet-framework | #!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
Add a print with file where mistake is | #!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
print(infile)
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
| <commit_before>#!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
<commit_msg>Add a print with file where mistake is<commit_after> | #!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
print(infile)
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
| #!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
Add a print with file where mistake is#!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
print(infile)
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
| <commit_before>#!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
<commit_msg>Add a print with file where mistake is<commit_after>#!/usr/bin/python
# -*- coding: utf-8 -*-
from __future__ import print_function
from xml.etree.ElementTree import ParseError
import xml.etree.ElementTree as ET
import glob
import sys
def eprint(*args, **kwargs):
print(*args, file=sys.stderr, **kwargs)
def parse():
for infile in glob.glob('*.xml'):
try:
tree = ET.parse(infile)
root = tree.getroot()
if root.findall('.//FatalError'):
eprint("Error detected")
print(infile)
sys.exit(1)
except ParseError:
eprint("The file xml isn't correct. There were some mistakes in the tests ")
sys.exit(1)
def main():
parse()
if __name__ == '__main__':
main()
|
817de880b1bce7b16348ae3e0b1494753e7fe6b8 | emission/net/ext_service/push/notify_queries.py | emission/net/ext_service/push/notify_queries.py | # Standard imports
import json
import logging
import uuid
# Our imports
import emission.core.get_database as edb
def get_platform_query(platform):
return {"curr_platform": platform}
def get_sync_interval_query(interval):
return {"curr_sync_interval": interval}
def get_user_query(user_id_list):
return {"user_id": {"$in": user_id_list}}
def combine_queries(query_list):
combined_query = {}
for query in query_list:
combined_query.update(query)
return combined_query
def get_matching_tokens(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "device_token": True})
mapped_list = map(lambda e: e.get("device_token"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
| # Standard imports
import json
import logging
import uuid
# Our imports
import emission.core.get_database as edb
def get_platform_query(platform):
return {"curr_platform": platform}
def get_sync_interval_query(interval):
return {"curr_sync_interval": interval}
def get_user_query(user_id_list):
return {"user_id": {"$in": user_id_list}}
def combine_queries(query_list):
combined_query = {}
for query in query_list:
combined_query.update(query)
return combined_query
def get_matching_tokens(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "device_token": True})
mapped_list = map(lambda e: e.get("device_token"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
def get_matching_user_ids(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "user_id": True})
mapped_list = map(lambda e: e.get("user_id"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
| Add support for returning the uuids for a profile query | Add support for returning the uuids for a profile query
in addition to the tokens. This allows us to link profile query results to the
survey/push notification feature.
| Python | bsd-3-clause | e-mission/e-mission-server,e-mission/e-mission-server,shankari/e-mission-server,e-mission/e-mission-server,sunil07t/e-mission-server,shankari/e-mission-server,shankari/e-mission-server,sunil07t/e-mission-server,sunil07t/e-mission-server,shankari/e-mission-server,e-mission/e-mission-server,sunil07t/e-mission-server | # Standard imports
import json
import logging
import uuid
# Our imports
import emission.core.get_database as edb
def get_platform_query(platform):
return {"curr_platform": platform}
def get_sync_interval_query(interval):
return {"curr_sync_interval": interval}
def get_user_query(user_id_list):
return {"user_id": {"$in": user_id_list}}
def combine_queries(query_list):
combined_query = {}
for query in query_list:
combined_query.update(query)
return combined_query
def get_matching_tokens(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "device_token": True})
mapped_list = map(lambda e: e.get("device_token"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
Add support for returning the uuids for a profile query
in addition to the tokens. This allows us to link profile query results to the
survey/push notification feature. | # Standard imports
import json
import logging
import uuid
# Our imports
import emission.core.get_database as edb
def get_platform_query(platform):
return {"curr_platform": platform}
def get_sync_interval_query(interval):
return {"curr_sync_interval": interval}
def get_user_query(user_id_list):
return {"user_id": {"$in": user_id_list}}
def combine_queries(query_list):
combined_query = {}
for query in query_list:
combined_query.update(query)
return combined_query
def get_matching_tokens(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "device_token": True})
mapped_list = map(lambda e: e.get("device_token"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
def get_matching_user_ids(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "user_id": True})
mapped_list = map(lambda e: e.get("user_id"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
| <commit_before># Standard imports
import json
import logging
import uuid
# Our imports
import emission.core.get_database as edb
def get_platform_query(platform):
return {"curr_platform": platform}
def get_sync_interval_query(interval):
return {"curr_sync_interval": interval}
def get_user_query(user_id_list):
return {"user_id": {"$in": user_id_list}}
def combine_queries(query_list):
combined_query = {}
for query in query_list:
combined_query.update(query)
return combined_query
def get_matching_tokens(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "device_token": True})
mapped_list = map(lambda e: e.get("device_token"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
<commit_msg>Add support for returning the uuids for a profile query
in addition to the tokens. This allows us to link profile query results to the
survey/push notification feature.<commit_after> | # Standard imports
import json
import logging
import uuid
# Our imports
import emission.core.get_database as edb
def get_platform_query(platform):
return {"curr_platform": platform}
def get_sync_interval_query(interval):
return {"curr_sync_interval": interval}
def get_user_query(user_id_list):
return {"user_id": {"$in": user_id_list}}
def combine_queries(query_list):
combined_query = {}
for query in query_list:
combined_query.update(query)
return combined_query
def get_matching_tokens(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "device_token": True})
mapped_list = map(lambda e: e.get("device_token"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
def get_matching_user_ids(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "user_id": True})
mapped_list = map(lambda e: e.get("user_id"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
| # Standard imports
import json
import logging
import uuid
# Our imports
import emission.core.get_database as edb
def get_platform_query(platform):
return {"curr_platform": platform}
def get_sync_interval_query(interval):
return {"curr_sync_interval": interval}
def get_user_query(user_id_list):
return {"user_id": {"$in": user_id_list}}
def combine_queries(query_list):
combined_query = {}
for query in query_list:
combined_query.update(query)
return combined_query
def get_matching_tokens(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "device_token": True})
mapped_list = map(lambda e: e.get("device_token"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
Add support for returning the uuids for a profile query
in addition to the tokens. This allows us to link profile query results to the
survey/push notification feature.# Standard imports
import json
import logging
import uuid
# Our imports
import emission.core.get_database as edb
def get_platform_query(platform):
return {"curr_platform": platform}
def get_sync_interval_query(interval):
return {"curr_sync_interval": interval}
def get_user_query(user_id_list):
return {"user_id": {"$in": user_id_list}}
def combine_queries(query_list):
combined_query = {}
for query in query_list:
combined_query.update(query)
return combined_query
def get_matching_tokens(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "device_token": True})
mapped_list = map(lambda e: e.get("device_token"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
def get_matching_user_ids(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "user_id": True})
mapped_list = map(lambda e: e.get("user_id"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
| <commit_before># Standard imports
import json
import logging
import uuid
# Our imports
import emission.core.get_database as edb
def get_platform_query(platform):
return {"curr_platform": platform}
def get_sync_interval_query(interval):
return {"curr_sync_interval": interval}
def get_user_query(user_id_list):
return {"user_id": {"$in": user_id_list}}
def combine_queries(query_list):
combined_query = {}
for query in query_list:
combined_query.update(query)
return combined_query
def get_matching_tokens(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "device_token": True})
mapped_list = map(lambda e: e.get("device_token"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
<commit_msg>Add support for returning the uuids for a profile query
in addition to the tokens. This allows us to link profile query results to the
survey/push notification feature.<commit_after># Standard imports
import json
import logging
import uuid
# Our imports
import emission.core.get_database as edb
def get_platform_query(platform):
return {"curr_platform": platform}
def get_sync_interval_query(interval):
return {"curr_sync_interval": interval}
def get_user_query(user_id_list):
return {"user_id": {"$in": user_id_list}}
def combine_queries(query_list):
combined_query = {}
for query in query_list:
combined_query.update(query)
return combined_query
def get_matching_tokens(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "device_token": True})
mapped_list = map(lambda e: e.get("device_token"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
def get_matching_user_ids(query):
logging.debug("Getting tokens matching query %s" % query)
ret_cursor = edb.get_profile_db().find(query, {"_id": False, "user_id": True})
mapped_list = map(lambda e: e.get("user_id"), ret_cursor)
ret_list = [item for item in mapped_list if item is not None]
return ret_list
|
e885701a12fcb2d2557c975fadbabc7ee28ebf8b | djoauth2/helpers.py | djoauth2/helpers.py | # coding: utf-8
import random
from string import ascii_letters, digits
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
| # coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib2 import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse(url)
query_parameters = urlparse.parse_qsl(parsed_url.query)
parsed_url.query = urlencode(query_parameters + parameters.items())
return urlparse.urlunparse(parsed_url)
| Add helper for updating URL GET parameters. | Add helper for updating URL GET parameters.
| Python | mit | vden/djoauth2-ng,vden/djoauth2-ng,Locu/djoauth2,seler/djoauth2,Locu/djoauth2,seler/djoauth2 | # coding: utf-8
import random
from string import ascii_letters, digits
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
Add helper for updating URL GET parameters. | # coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib2 import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse(url)
query_parameters = urlparse.parse_qsl(parsed_url.query)
parsed_url.query = urlencode(query_parameters + parameters.items())
return urlparse.urlunparse(parsed_url)
| <commit_before># coding: utf-8
import random
from string import ascii_letters, digits
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
<commit_msg>Add helper for updating URL GET parameters.<commit_after> | # coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib2 import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse(url)
query_parameters = urlparse.parse_qsl(parsed_url.query)
parsed_url.query = urlencode(query_parameters + parameters.items())
return urlparse.urlunparse(parsed_url)
| # coding: utf-8
import random
from string import ascii_letters, digits
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
Add helper for updating URL GET parameters.# coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib2 import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse(url)
query_parameters = urlparse.parse_qsl(parsed_url.query)
parsed_url.query = urlencode(query_parameters + parameters.items())
return urlparse.urlunparse(parsed_url)
| <commit_before># coding: utf-8
import random
from string import ascii_letters, digits
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
<commit_msg>Add helper for updating URL GET parameters.<commit_after># coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib2 import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse(url)
query_parameters = urlparse.parse_qsl(parsed_url.query)
parsed_url.query = urlencode(query_parameters + parameters.items())
return urlparse.urlunparse(parsed_url)
|
e2722385831a0930765d2c4bb78a582d41f4b64b | src/sentry/replays.py | src/sentry/replays.py | from __future__ import absolute_import
import socket
from httplib import HTTPConnection, HTTPSConnection
from urllib import urlencode
from urlparse import urlparse
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
urlparts = urlparse(self.url)
if urlparts.scheme == 'http':
conn_cls = HTTPConnection
elif urlparts.scheme == 'https':
conn_cls = HTTPSConnection
else:
raise ValueError(self.url)
data = self.data
if isinstance(data, dict):
data = urlencode(data)
if urlparts.query:
full_url = urlparts.path + '?' + urlparts.query
else:
full_url = urlparts.path
conn = conn_cls(urlparts.netloc)
try:
conn.request(self.method, full_url, data, self.headers or {})
response = conn.getresponse()
except socket.error as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status,
'reason': response.reason,
'headers': response.getheaders(),
'body': response.read(),
}
| from __future__ import absolute_import
import requests
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
try:
response = requests.request(
self.method,
self.url,
data=self.data,
headers=self.headers or {}
)
except requests.RequestException as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status_code,
'reason': response.reason,
'headers': response.headers,
'body': response.content,
}
| Use requests instead of httplib to do replay | Use requests instead of httplib to do replay
| Python | bsd-3-clause | beeftornado/sentry,nicholasserra/sentry,Kryz/sentry,JackDanger/sentry,imankulov/sentry,JamesMura/sentry,zenefits/sentry,kevinlondon/sentry,mvaled/sentry,JamesMura/sentry,ifduyue/sentry,looker/sentry,daevaorn/sentry,fotinakis/sentry,gencer/sentry,looker/sentry,JackDanger/sentry,mvaled/sentry,Natim/sentry,beeftornado/sentry,korealerts1/sentry,imankulov/sentry,zenefits/sentry,jean/sentry,alexm92/sentry,fotinakis/sentry,daevaorn/sentry,beeftornado/sentry,BuildingLink/sentry,ngonzalvez/sentry,BayanGroup/sentry,mvaled/sentry,mitsuhiko/sentry,daevaorn/sentry,ngonzalvez/sentry,looker/sentry,korealerts1/sentry,mvaled/sentry,gencer/sentry,imankulov/sentry,Kryz/sentry,looker/sentry,felixbuenemann/sentry,jean/sentry,mitsuhiko/sentry,fotinakis/sentry,ifduyue/sentry,mvaled/sentry,korealerts1/sentry,kevinlondon/sentry,Natim/sentry,alexm92/sentry,zenefits/sentry,zenefits/sentry,BayanGroup/sentry,jean/sentry,mvaled/sentry,ifduyue/sentry,JamesMura/sentry,kevinlondon/sentry,BuildingLink/sentry,BayanGroup/sentry,JamesMura/sentry,nicholasserra/sentry,felixbuenemann/sentry,Kryz/sentry,BuildingLink/sentry,BuildingLink/sentry,Natim/sentry,BuildingLink/sentry,jean/sentry,JackDanger/sentry,felixbuenemann/sentry,fotinakis/sentry,nicholasserra/sentry,daevaorn/sentry,ngonzalvez/sentry,looker/sentry,alexm92/sentry,gencer/sentry,gencer/sentry,gencer/sentry,zenefits/sentry,JamesMura/sentry,ifduyue/sentry,jean/sentry,ifduyue/sentry | from __future__ import absolute_import
import socket
from httplib import HTTPConnection, HTTPSConnection
from urllib import urlencode
from urlparse import urlparse
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
urlparts = urlparse(self.url)
if urlparts.scheme == 'http':
conn_cls = HTTPConnection
elif urlparts.scheme == 'https':
conn_cls = HTTPSConnection
else:
raise ValueError(self.url)
data = self.data
if isinstance(data, dict):
data = urlencode(data)
if urlparts.query:
full_url = urlparts.path + '?' + urlparts.query
else:
full_url = urlparts.path
conn = conn_cls(urlparts.netloc)
try:
conn.request(self.method, full_url, data, self.headers or {})
response = conn.getresponse()
except socket.error as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status,
'reason': response.reason,
'headers': response.getheaders(),
'body': response.read(),
}
Use requests instead of httplib to do replay | from __future__ import absolute_import
import requests
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
try:
response = requests.request(
self.method,
self.url,
data=self.data,
headers=self.headers or {}
)
except requests.RequestException as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status_code,
'reason': response.reason,
'headers': response.headers,
'body': response.content,
}
| <commit_before>from __future__ import absolute_import
import socket
from httplib import HTTPConnection, HTTPSConnection
from urllib import urlencode
from urlparse import urlparse
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
urlparts = urlparse(self.url)
if urlparts.scheme == 'http':
conn_cls = HTTPConnection
elif urlparts.scheme == 'https':
conn_cls = HTTPSConnection
else:
raise ValueError(self.url)
data = self.data
if isinstance(data, dict):
data = urlencode(data)
if urlparts.query:
full_url = urlparts.path + '?' + urlparts.query
else:
full_url = urlparts.path
conn = conn_cls(urlparts.netloc)
try:
conn.request(self.method, full_url, data, self.headers or {})
response = conn.getresponse()
except socket.error as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status,
'reason': response.reason,
'headers': response.getheaders(),
'body': response.read(),
}
<commit_msg>Use requests instead of httplib to do replay<commit_after> | from __future__ import absolute_import
import requests
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
try:
response = requests.request(
self.method,
self.url,
data=self.data,
headers=self.headers or {}
)
except requests.RequestException as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status_code,
'reason': response.reason,
'headers': response.headers,
'body': response.content,
}
| from __future__ import absolute_import
import socket
from httplib import HTTPConnection, HTTPSConnection
from urllib import urlencode
from urlparse import urlparse
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
urlparts = urlparse(self.url)
if urlparts.scheme == 'http':
conn_cls = HTTPConnection
elif urlparts.scheme == 'https':
conn_cls = HTTPSConnection
else:
raise ValueError(self.url)
data = self.data
if isinstance(data, dict):
data = urlencode(data)
if urlparts.query:
full_url = urlparts.path + '?' + urlparts.query
else:
full_url = urlparts.path
conn = conn_cls(urlparts.netloc)
try:
conn.request(self.method, full_url, data, self.headers or {})
response = conn.getresponse()
except socket.error as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status,
'reason': response.reason,
'headers': response.getheaders(),
'body': response.read(),
}
Use requests instead of httplib to do replayfrom __future__ import absolute_import
import requests
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
try:
response = requests.request(
self.method,
self.url,
data=self.data,
headers=self.headers or {}
)
except requests.RequestException as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status_code,
'reason': response.reason,
'headers': response.headers,
'body': response.content,
}
| <commit_before>from __future__ import absolute_import
import socket
from httplib import HTTPConnection, HTTPSConnection
from urllib import urlencode
from urlparse import urlparse
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
urlparts = urlparse(self.url)
if urlparts.scheme == 'http':
conn_cls = HTTPConnection
elif urlparts.scheme == 'https':
conn_cls = HTTPSConnection
else:
raise ValueError(self.url)
data = self.data
if isinstance(data, dict):
data = urlencode(data)
if urlparts.query:
full_url = urlparts.path + '?' + urlparts.query
else:
full_url = urlparts.path
conn = conn_cls(urlparts.netloc)
try:
conn.request(self.method, full_url, data, self.headers or {})
response = conn.getresponse()
except socket.error as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status,
'reason': response.reason,
'headers': response.getheaders(),
'body': response.read(),
}
<commit_msg>Use requests instead of httplib to do replay<commit_after>from __future__ import absolute_import
import requests
class Replayer(object):
def __init__(self, url, method, data=None, headers=None):
self.url = url
self.method = method
self.data = data
self.headers = headers
def replay(self):
try:
response = requests.request(
self.method,
self.url,
data=self.data,
headers=self.headers or {}
)
except requests.RequestException as e:
return {
'status': 'error',
'reason': str(e),
}
return {
'status': response.status_code,
'reason': response.reason,
'headers': response.headers,
'body': response.content,
}
|
133617660fe96a817b47d4d0fba4cfa7567dcafb | exceptional.py | exceptional.py | """A module to demonstrate exceptions."""
import sys
def convert(item):
'''
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
'''
try:
x = int(item)
print(str.format('Conversion succeeded! x= {}', x))
except ValueError:
print('Conversion Failed')
x = -1
return x
if __name__ == '__main__':
print(convert(sys.argv[1])) | """A module to demonstrate exceptions."""
import sys
def convert(item):
"""
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
"""
try:
return int(item)
except (ValueError, TypeError):
return -1
if __name__ == '__main__':
print(convert(sys.argv[1])) | Use two return statements and remove printing | Use two return statements and remove printing
| Python | mit | kentoj/python-fundamentals | """A module to demonstrate exceptions."""
import sys
def convert(item):
'''
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
'''
try:
x = int(item)
print(str.format('Conversion succeeded! x= {}', x))
except ValueError:
print('Conversion Failed')
x = -1
return x
if __name__ == '__main__':
print(convert(sys.argv[1]))Use two return statements and remove printing | """A module to demonstrate exceptions."""
import sys
def convert(item):
"""
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
"""
try:
return int(item)
except (ValueError, TypeError):
return -1
if __name__ == '__main__':
print(convert(sys.argv[1])) | <commit_before>"""A module to demonstrate exceptions."""
import sys
def convert(item):
'''
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
'''
try:
x = int(item)
print(str.format('Conversion succeeded! x= {}', x))
except ValueError:
print('Conversion Failed')
x = -1
return x
if __name__ == '__main__':
print(convert(sys.argv[1]))<commit_msg>Use two return statements and remove printing<commit_after> | """A module to demonstrate exceptions."""
import sys
def convert(item):
"""
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
"""
try:
return int(item)
except (ValueError, TypeError):
return -1
if __name__ == '__main__':
print(convert(sys.argv[1])) | """A module to demonstrate exceptions."""
import sys
def convert(item):
'''
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
'''
try:
x = int(item)
print(str.format('Conversion succeeded! x= {}', x))
except ValueError:
print('Conversion Failed')
x = -1
return x
if __name__ == '__main__':
print(convert(sys.argv[1]))Use two return statements and remove printing"""A module to demonstrate exceptions."""
import sys
def convert(item):
"""
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
"""
try:
return int(item)
except (ValueError, TypeError):
return -1
if __name__ == '__main__':
print(convert(sys.argv[1])) | <commit_before>"""A module to demonstrate exceptions."""
import sys
def convert(item):
'''
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
'''
try:
x = int(item)
print(str.format('Conversion succeeded! x= {}', x))
except ValueError:
print('Conversion Failed')
x = -1
return x
if __name__ == '__main__':
print(convert(sys.argv[1]))<commit_msg>Use two return statements and remove printing<commit_after>"""A module to demonstrate exceptions."""
import sys
def convert(item):
"""
Convert to an integer.
Args:
item: some object
Returns:
an integer representation of the object
Throws:
a ValueException
"""
try:
return int(item)
except (ValueError, TypeError):
return -1
if __name__ == '__main__':
print(convert(sys.argv[1])) |
972073f7d65fe4ea2910241a7f8ba42a78ab3a86 | fore/config.py | fore/config.py | # Basic config for stuff that can be easily changed, but which is git-managed.
# See also apikeys_sample.py for the configs which are _not_ git-managed.
app_name = "fore"
server_domain = "http://www.infiniteglitch.net"
lag_limit = 88200 # samples - how much we can lag by before dropping frames.
restart_timeout = 3 # seconds between polls to restart.txt
http_port = 8888
mini_http_port = 8193
uid = 1000 # User ID and group ID to drop privileges to
gid = 1000 # Set both to 0 to not drop privileges, eg if the server is started without privs
use_sudo_uid_gid = True # If set, uid/gid will be overridden with SUDO_UID/SUDO_GID if available
frontend_buffer = 20 # seconds of audio to buffer in frontend
past_played_buffer = 600 # seconds of audio to store track metadata for in the past
template_dir = "templates/"
drift_limit = 0.1 # seconds of audio after which drift should be corrected
max_track_length = 400
min_track_length = 90
# Default values when nothing exists
no_bpm_diff = 20
| # Basic config for stuff that can be easily changed, but which is git-managed.
# See also apikeys_sample.py for the configs which are _not_ git-managed.
app_name = "fore"
server_domain = "http://www.infiniteglitch.net"
lag_limit = 88200 # samples - how much we can lag by before dropping frames.
restart_timeout = 3 # seconds between polls to restart.txt
http_port = 8888
mini_http_port = 8193
uid = 0 # User ID and group ID to drop privileges to
gid = 0 # Set both to 0 to not drop privileges, eg if the server is started without privs
use_sudo_uid_gid = True # If set, uid/gid will be overridden with SUDO_UID/SUDO_GID if available
frontend_buffer = 20 # seconds of audio to buffer in frontend
past_played_buffer = 600 # seconds of audio to store track metadata for in the past
template_dir = "templates/"
drift_limit = 0.1 # seconds of audio after which drift should be corrected
max_track_length = 400
min_track_length = 90
# Default values when nothing exists
no_bpm_diff = 20
| Decrease user and group IDs from 1000 to 0. | Decrease user and group IDs from 1000 to 0.
| Python | artistic-2.0 | Rosuav/appension,MikeiLL/appension,Rosuav/appension,MikeiLL/appension,MikeiLL/appension,MikeiLL/appension,Rosuav/appension,Rosuav/appension | # Basic config for stuff that can be easily changed, but which is git-managed.
# See also apikeys_sample.py for the configs which are _not_ git-managed.
app_name = "fore"
server_domain = "http://www.infiniteglitch.net"
lag_limit = 88200 # samples - how much we can lag by before dropping frames.
restart_timeout = 3 # seconds between polls to restart.txt
http_port = 8888
mini_http_port = 8193
uid = 1000 # User ID and group ID to drop privileges to
gid = 1000 # Set both to 0 to not drop privileges, eg if the server is started without privs
use_sudo_uid_gid = True # If set, uid/gid will be overridden with SUDO_UID/SUDO_GID if available
frontend_buffer = 20 # seconds of audio to buffer in frontend
past_played_buffer = 600 # seconds of audio to store track metadata for in the past
template_dir = "templates/"
drift_limit = 0.1 # seconds of audio after which drift should be corrected
max_track_length = 400
min_track_length = 90
# Default values when nothing exists
no_bpm_diff = 20
Decrease user and group IDs from 1000 to 0. | # Basic config for stuff that can be easily changed, but which is git-managed.
# See also apikeys_sample.py for the configs which are _not_ git-managed.
app_name = "fore"
server_domain = "http://www.infiniteglitch.net"
lag_limit = 88200 # samples - how much we can lag by before dropping frames.
restart_timeout = 3 # seconds between polls to restart.txt
http_port = 8888
mini_http_port = 8193
uid = 0 # User ID and group ID to drop privileges to
gid = 0 # Set both to 0 to not drop privileges, eg if the server is started without privs
use_sudo_uid_gid = True # If set, uid/gid will be overridden with SUDO_UID/SUDO_GID if available
frontend_buffer = 20 # seconds of audio to buffer in frontend
past_played_buffer = 600 # seconds of audio to store track metadata for in the past
template_dir = "templates/"
drift_limit = 0.1 # seconds of audio after which drift should be corrected
max_track_length = 400
min_track_length = 90
# Default values when nothing exists
no_bpm_diff = 20
| <commit_before># Basic config for stuff that can be easily changed, but which is git-managed.
# See also apikeys_sample.py for the configs which are _not_ git-managed.
app_name = "fore"
server_domain = "http://www.infiniteglitch.net"
lag_limit = 88200 # samples - how much we can lag by before dropping frames.
restart_timeout = 3 # seconds between polls to restart.txt
http_port = 8888
mini_http_port = 8193
uid = 1000 # User ID and group ID to drop privileges to
gid = 1000 # Set both to 0 to not drop privileges, eg if the server is started without privs
use_sudo_uid_gid = True # If set, uid/gid will be overridden with SUDO_UID/SUDO_GID if available
frontend_buffer = 20 # seconds of audio to buffer in frontend
past_played_buffer = 600 # seconds of audio to store track metadata for in the past
template_dir = "templates/"
drift_limit = 0.1 # seconds of audio after which drift should be corrected
max_track_length = 400
min_track_length = 90
# Default values when nothing exists
no_bpm_diff = 20
<commit_msg>Decrease user and group IDs from 1000 to 0.<commit_after> | # Basic config for stuff that can be easily changed, but which is git-managed.
# See also apikeys_sample.py for the configs which are _not_ git-managed.
app_name = "fore"
server_domain = "http://www.infiniteglitch.net"
lag_limit = 88200 # samples - how much we can lag by before dropping frames.
restart_timeout = 3 # seconds between polls to restart.txt
http_port = 8888
mini_http_port = 8193
uid = 0 # User ID and group ID to drop privileges to
gid = 0 # Set both to 0 to not drop privileges, eg if the server is started without privs
use_sudo_uid_gid = True # If set, uid/gid will be overridden with SUDO_UID/SUDO_GID if available
frontend_buffer = 20 # seconds of audio to buffer in frontend
past_played_buffer = 600 # seconds of audio to store track metadata for in the past
template_dir = "templates/"
drift_limit = 0.1 # seconds of audio after which drift should be corrected
max_track_length = 400
min_track_length = 90
# Default values when nothing exists
no_bpm_diff = 20
| # Basic config for stuff that can be easily changed, but which is git-managed.
# See also apikeys_sample.py for the configs which are _not_ git-managed.
app_name = "fore"
server_domain = "http://www.infiniteglitch.net"
lag_limit = 88200 # samples - how much we can lag by before dropping frames.
restart_timeout = 3 # seconds between polls to restart.txt
http_port = 8888
mini_http_port = 8193
uid = 1000 # User ID and group ID to drop privileges to
gid = 1000 # Set both to 0 to not drop privileges, eg if the server is started without privs
use_sudo_uid_gid = True # If set, uid/gid will be overridden with SUDO_UID/SUDO_GID if available
frontend_buffer = 20 # seconds of audio to buffer in frontend
past_played_buffer = 600 # seconds of audio to store track metadata for in the past
template_dir = "templates/"
drift_limit = 0.1 # seconds of audio after which drift should be corrected
max_track_length = 400
min_track_length = 90
# Default values when nothing exists
no_bpm_diff = 20
Decrease user and group IDs from 1000 to 0.# Basic config for stuff that can be easily changed, but which is git-managed.
# See also apikeys_sample.py for the configs which are _not_ git-managed.
app_name = "fore"
server_domain = "http://www.infiniteglitch.net"
lag_limit = 88200 # samples - how much we can lag by before dropping frames.
restart_timeout = 3 # seconds between polls to restart.txt
http_port = 8888
mini_http_port = 8193
uid = 0 # User ID and group ID to drop privileges to
gid = 0 # Set both to 0 to not drop privileges, eg if the server is started without privs
use_sudo_uid_gid = True # If set, uid/gid will be overridden with SUDO_UID/SUDO_GID if available
frontend_buffer = 20 # seconds of audio to buffer in frontend
past_played_buffer = 600 # seconds of audio to store track metadata for in the past
template_dir = "templates/"
drift_limit = 0.1 # seconds of audio after which drift should be corrected
max_track_length = 400
min_track_length = 90
# Default values when nothing exists
no_bpm_diff = 20
| <commit_before># Basic config for stuff that can be easily changed, but which is git-managed.
# See also apikeys_sample.py for the configs which are _not_ git-managed.
app_name = "fore"
server_domain = "http://www.infiniteglitch.net"
lag_limit = 88200 # samples - how much we can lag by before dropping frames.
restart_timeout = 3 # seconds between polls to restart.txt
http_port = 8888
mini_http_port = 8193
uid = 1000 # User ID and group ID to drop privileges to
gid = 1000 # Set both to 0 to not drop privileges, eg if the server is started without privs
use_sudo_uid_gid = True # If set, uid/gid will be overridden with SUDO_UID/SUDO_GID if available
frontend_buffer = 20 # seconds of audio to buffer in frontend
past_played_buffer = 600 # seconds of audio to store track metadata for in the past
template_dir = "templates/"
drift_limit = 0.1 # seconds of audio after which drift should be corrected
max_track_length = 400
min_track_length = 90
# Default values when nothing exists
no_bpm_diff = 20
<commit_msg>Decrease user and group IDs from 1000 to 0.<commit_after># Basic config for stuff that can be easily changed, but which is git-managed.
# See also apikeys_sample.py for the configs which are _not_ git-managed.
app_name = "fore"
server_domain = "http://www.infiniteglitch.net"
lag_limit = 88200 # samples - how much we can lag by before dropping frames.
restart_timeout = 3 # seconds between polls to restart.txt
http_port = 8888
mini_http_port = 8193
uid = 0 # User ID and group ID to drop privileges to
gid = 0 # Set both to 0 to not drop privileges, eg if the server is started without privs
use_sudo_uid_gid = True # If set, uid/gid will be overridden with SUDO_UID/SUDO_GID if available
frontend_buffer = 20 # seconds of audio to buffer in frontend
past_played_buffer = 600 # seconds of audio to store track metadata for in the past
template_dir = "templates/"
drift_limit = 0.1 # seconds of audio after which drift should be corrected
max_track_length = 400
min_track_length = 90
# Default values when nothing exists
no_bpm_diff = 20
|
a0172116503f0b212a184fc4a1d2179115675e17 | fuzzyfinder/main.py | fuzzyfinder/main.py | # -*- coding: utf-8 -*-
import re
from . import export
@export
def fuzzyfinder(text, collection):
"""
Args:
text (str): A partial string which is typically entered by a user.
collection (iterable): A collection of strings which will be filtered
based on the input `text`.
Returns:
suggestions (generator): A generator object that produces a list of
suggestions narrowed down from `collections` using the `text`
input.
"""
suggestions = []
pat = '.*?'.join(map(re.escape, text))
regex = re.compile('%s' % pat)
for item in sorted(collection):
r = regex.search(item)
if r:
suggestions.append((len(r.group()), r.start(), item))
return (z for _, _, z in sorted(suggestions))
| # -*- coding: utf-8 -*-
import re
from . import export
@export
def fuzzyfinder(text, collection):
"""
Args:
text (str): A partial string which is typically entered by a user.
collection (iterable): A collection of strings which will be filtered
based on the input `text`.
Returns:
suggestions (generator): A generator object that produces a list of
suggestions narrowed down from `collections` using the `text`
input.
"""
suggestions = []
pat = '.*?'.join(map(re.escape, text))
regex = re.compile(pat)
for item in collection:
r = regex.search(item)
if r:
suggestions.append((len(r.group()), r.start(), item))
return (z for _, _, z in sorted(suggestions))
| Remove string interpolation and sorting the collection. | Remove string interpolation and sorting the collection. | Python | bsd-3-clause | adammenges/fuzzyfinder,amjith/fuzzyfinder,harrisonfeng/fuzzyfinder | # -*- coding: utf-8 -*-
import re
from . import export
@export
def fuzzyfinder(text, collection):
"""
Args:
text (str): A partial string which is typically entered by a user.
collection (iterable): A collection of strings which will be filtered
based on the input `text`.
Returns:
suggestions (generator): A generator object that produces a list of
suggestions narrowed down from `collections` using the `text`
input.
"""
suggestions = []
pat = '.*?'.join(map(re.escape, text))
regex = re.compile('%s' % pat)
for item in sorted(collection):
r = regex.search(item)
if r:
suggestions.append((len(r.group()), r.start(), item))
return (z for _, _, z in sorted(suggestions))
Remove string interpolation and sorting the collection. | # -*- coding: utf-8 -*-
import re
from . import export
@export
def fuzzyfinder(text, collection):
"""
Args:
text (str): A partial string which is typically entered by a user.
collection (iterable): A collection of strings which will be filtered
based on the input `text`.
Returns:
suggestions (generator): A generator object that produces a list of
suggestions narrowed down from `collections` using the `text`
input.
"""
suggestions = []
pat = '.*?'.join(map(re.escape, text))
regex = re.compile(pat)
for item in collection:
r = regex.search(item)
if r:
suggestions.append((len(r.group()), r.start(), item))
return (z for _, _, z in sorted(suggestions))
| <commit_before># -*- coding: utf-8 -*-
import re
from . import export
@export
def fuzzyfinder(text, collection):
"""
Args:
text (str): A partial string which is typically entered by a user.
collection (iterable): A collection of strings which will be filtered
based on the input `text`.
Returns:
suggestions (generator): A generator object that produces a list of
suggestions narrowed down from `collections` using the `text`
input.
"""
suggestions = []
pat = '.*?'.join(map(re.escape, text))
regex = re.compile('%s' % pat)
for item in sorted(collection):
r = regex.search(item)
if r:
suggestions.append((len(r.group()), r.start(), item))
return (z for _, _, z in sorted(suggestions))
<commit_msg>Remove string interpolation and sorting the collection.<commit_after> | # -*- coding: utf-8 -*-
import re
from . import export
@export
def fuzzyfinder(text, collection):
"""
Args:
text (str): A partial string which is typically entered by a user.
collection (iterable): A collection of strings which will be filtered
based on the input `text`.
Returns:
suggestions (generator): A generator object that produces a list of
suggestions narrowed down from `collections` using the `text`
input.
"""
suggestions = []
pat = '.*?'.join(map(re.escape, text))
regex = re.compile(pat)
for item in collection:
r = regex.search(item)
if r:
suggestions.append((len(r.group()), r.start(), item))
return (z for _, _, z in sorted(suggestions))
| # -*- coding: utf-8 -*-
import re
from . import export
@export
def fuzzyfinder(text, collection):
"""
Args:
text (str): A partial string which is typically entered by a user.
collection (iterable): A collection of strings which will be filtered
based on the input `text`.
Returns:
suggestions (generator): A generator object that produces a list of
suggestions narrowed down from `collections` using the `text`
input.
"""
suggestions = []
pat = '.*?'.join(map(re.escape, text))
regex = re.compile('%s' % pat)
for item in sorted(collection):
r = regex.search(item)
if r:
suggestions.append((len(r.group()), r.start(), item))
return (z for _, _, z in sorted(suggestions))
Remove string interpolation and sorting the collection.# -*- coding: utf-8 -*-
import re
from . import export
@export
def fuzzyfinder(text, collection):
"""
Args:
text (str): A partial string which is typically entered by a user.
collection (iterable): A collection of strings which will be filtered
based on the input `text`.
Returns:
suggestions (generator): A generator object that produces a list of
suggestions narrowed down from `collections` using the `text`
input.
"""
suggestions = []
pat = '.*?'.join(map(re.escape, text))
regex = re.compile(pat)
for item in collection:
r = regex.search(item)
if r:
suggestions.append((len(r.group()), r.start(), item))
return (z for _, _, z in sorted(suggestions))
| <commit_before># -*- coding: utf-8 -*-
import re
from . import export
@export
def fuzzyfinder(text, collection):
"""
Args:
text (str): A partial string which is typically entered by a user.
collection (iterable): A collection of strings which will be filtered
based on the input `text`.
Returns:
suggestions (generator): A generator object that produces a list of
suggestions narrowed down from `collections` using the `text`
input.
"""
suggestions = []
pat = '.*?'.join(map(re.escape, text))
regex = re.compile('%s' % pat)
for item in sorted(collection):
r = regex.search(item)
if r:
suggestions.append((len(r.group()), r.start(), item))
return (z for _, _, z in sorted(suggestions))
<commit_msg>Remove string interpolation and sorting the collection.<commit_after># -*- coding: utf-8 -*-
import re
from . import export
@export
def fuzzyfinder(text, collection):
"""
Args:
text (str): A partial string which is typically entered by a user.
collection (iterable): A collection of strings which will be filtered
based on the input `text`.
Returns:
suggestions (generator): A generator object that produces a list of
suggestions narrowed down from `collections` using the `text`
input.
"""
suggestions = []
pat = '.*?'.join(map(re.escape, text))
regex = re.compile(pat)
for item in collection:
r = regex.search(item)
if r:
suggestions.append((len(r.group()), r.start(), item))
return (z for _, _, z in sorted(suggestions))
|
caf0829191e9f3276fb144486ad602dcd482b60d | ignition/dsl/sfl/proteus_coefficient_printer.py | ignition/dsl/sfl/proteus_coefficient_printer.py | """Generator for Proteus coefficient evaluator"""
from .sfl_printer import SFLPrinter
from ...code_tools import comment_code, indent_code, PythonCodePrinter
coefficient_header = """\
Proteus Coefficient file generated from Ignition
"""
class_header = """\
class %{class_name}s(TC_base):
"""
class ProteusCoefficientPrinter(SFLPrinter):
"""Generator for Proteus Coefficient evaluator"""
language = 'Python'
comment_str = '//'
block_comment_tuple = ('"""', '"""')
def print_file(self, indent=0):
ret_code = ""
ret_code += self._print_header(indent)
ret_code += PythonCodePrinter(self._generator.class_dag)
return ret_code
| """Generator for Proteus coefficient evaluator"""
from .sfl_printer import SFLPrinter
from ...code_tools import comment_code, indent_code, PythonCodePrinter
coefficient_header = """\
Proteus Coefficient file generated from Ignition
"""
class ProteusCoefficientPrinter(SFLPrinter):
"""Generator for Proteus Coefficient evaluator"""
language = 'Python'
comment_str = '//'
block_comment_tuple = ('"""', '"""\n')
def _print_header(self, indent):
return comment_code(indent_code(coefficient_header, indent),
block_comment=self.block_comment_tuple)
def print_file(self, indent=0):
ret_code = ""
ret_code += self._print_header(indent)
ret_code += PythonCodePrinter(self._generator.class_dag).code_str()
return ret_code
| Print head, remove code for proteus python class head (use codeobj) | Print head, remove code for proteus python class head (use codeobj)
| Python | bsd-3-clause | IgnitionProject/ignition,IgnitionProject/ignition,IgnitionProject/ignition | """Generator for Proteus coefficient evaluator"""
from .sfl_printer import SFLPrinter
from ...code_tools import comment_code, indent_code, PythonCodePrinter
coefficient_header = """\
Proteus Coefficient file generated from Ignition
"""
class_header = """\
class %{class_name}s(TC_base):
"""
class ProteusCoefficientPrinter(SFLPrinter):
"""Generator for Proteus Coefficient evaluator"""
language = 'Python'
comment_str = '//'
block_comment_tuple = ('"""', '"""')
def print_file(self, indent=0):
ret_code = ""
ret_code += self._print_header(indent)
ret_code += PythonCodePrinter(self._generator.class_dag)
return ret_code
Print head, remove code for proteus python class head (use codeobj) | """Generator for Proteus coefficient evaluator"""
from .sfl_printer import SFLPrinter
from ...code_tools import comment_code, indent_code, PythonCodePrinter
coefficient_header = """\
Proteus Coefficient file generated from Ignition
"""
class ProteusCoefficientPrinter(SFLPrinter):
"""Generator for Proteus Coefficient evaluator"""
language = 'Python'
comment_str = '//'
block_comment_tuple = ('"""', '"""\n')
def _print_header(self, indent):
return comment_code(indent_code(coefficient_header, indent),
block_comment=self.block_comment_tuple)
def print_file(self, indent=0):
ret_code = ""
ret_code += self._print_header(indent)
ret_code += PythonCodePrinter(self._generator.class_dag).code_str()
return ret_code
| <commit_before>"""Generator for Proteus coefficient evaluator"""
from .sfl_printer import SFLPrinter
from ...code_tools import comment_code, indent_code, PythonCodePrinter
coefficient_header = """\
Proteus Coefficient file generated from Ignition
"""
class_header = """\
class %{class_name}s(TC_base):
"""
class ProteusCoefficientPrinter(SFLPrinter):
"""Generator for Proteus Coefficient evaluator"""
language = 'Python'
comment_str = '//'
block_comment_tuple = ('"""', '"""')
def print_file(self, indent=0):
ret_code = ""
ret_code += self._print_header(indent)
ret_code += PythonCodePrinter(self._generator.class_dag)
return ret_code
<commit_msg>Print head, remove code for proteus python class head (use codeobj)<commit_after> | """Generator for Proteus coefficient evaluator"""
from .sfl_printer import SFLPrinter
from ...code_tools import comment_code, indent_code, PythonCodePrinter
coefficient_header = """\
Proteus Coefficient file generated from Ignition
"""
class ProteusCoefficientPrinter(SFLPrinter):
"""Generator for Proteus Coefficient evaluator"""
language = 'Python'
comment_str = '//'
block_comment_tuple = ('"""', '"""\n')
def _print_header(self, indent):
return comment_code(indent_code(coefficient_header, indent),
block_comment=self.block_comment_tuple)
def print_file(self, indent=0):
ret_code = ""
ret_code += self._print_header(indent)
ret_code += PythonCodePrinter(self._generator.class_dag).code_str()
return ret_code
| """Generator for Proteus coefficient evaluator"""
from .sfl_printer import SFLPrinter
from ...code_tools import comment_code, indent_code, PythonCodePrinter
coefficient_header = """\
Proteus Coefficient file generated from Ignition
"""
class_header = """\
class %{class_name}s(TC_base):
"""
class ProteusCoefficientPrinter(SFLPrinter):
"""Generator for Proteus Coefficient evaluator"""
language = 'Python'
comment_str = '//'
block_comment_tuple = ('"""', '"""')
def print_file(self, indent=0):
ret_code = ""
ret_code += self._print_header(indent)
ret_code += PythonCodePrinter(self._generator.class_dag)
return ret_code
Print head, remove code for proteus python class head (use codeobj)"""Generator for Proteus coefficient evaluator"""
from .sfl_printer import SFLPrinter
from ...code_tools import comment_code, indent_code, PythonCodePrinter
coefficient_header = """\
Proteus Coefficient file generated from Ignition
"""
class ProteusCoefficientPrinter(SFLPrinter):
"""Generator for Proteus Coefficient evaluator"""
language = 'Python'
comment_str = '//'
block_comment_tuple = ('"""', '"""\n')
def _print_header(self, indent):
return comment_code(indent_code(coefficient_header, indent),
block_comment=self.block_comment_tuple)
def print_file(self, indent=0):
ret_code = ""
ret_code += self._print_header(indent)
ret_code += PythonCodePrinter(self._generator.class_dag).code_str()
return ret_code
| <commit_before>"""Generator for Proteus coefficient evaluator"""
from .sfl_printer import SFLPrinter
from ...code_tools import comment_code, indent_code, PythonCodePrinter
coefficient_header = """\
Proteus Coefficient file generated from Ignition
"""
class_header = """\
class %{class_name}s(TC_base):
"""
class ProteusCoefficientPrinter(SFLPrinter):
"""Generator for Proteus Coefficient evaluator"""
language = 'Python'
comment_str = '//'
block_comment_tuple = ('"""', '"""')
def print_file(self, indent=0):
ret_code = ""
ret_code += self._print_header(indent)
ret_code += PythonCodePrinter(self._generator.class_dag)
return ret_code
<commit_msg>Print head, remove code for proteus python class head (use codeobj)<commit_after>"""Generator for Proteus coefficient evaluator"""
from .sfl_printer import SFLPrinter
from ...code_tools import comment_code, indent_code, PythonCodePrinter
coefficient_header = """\
Proteus Coefficient file generated from Ignition
"""
class ProteusCoefficientPrinter(SFLPrinter):
"""Generator for Proteus Coefficient evaluator"""
language = 'Python'
comment_str = '//'
block_comment_tuple = ('"""', '"""\n')
def _print_header(self, indent):
return comment_code(indent_code(coefficient_header, indent),
block_comment=self.block_comment_tuple)
def print_file(self, indent=0):
ret_code = ""
ret_code += self._print_header(indent)
ret_code += PythonCodePrinter(self._generator.class_dag).code_str()
return ret_code
|
373e4e0e58cfec09b60983494f7b3bb4712e0ccd | 2/ConfNEP.py | 2/ConfNEP.py | """A custom ILAMB confrontation for net ecosystem productivity (nep)."""
import os
import numpy as np
from ILAMB.Confrontation import Confrontation
from ILAMB.Variable import Variable
import ilamblib as il
class ConfNEP(Confrontation):
"""Confront ``nep`` model outputs with ``nee`` observations.
Net ecosystem productivity (``nep``) is a CMIP5 standard output
provided by the MsTMIP models, and is the inverse of net ecosystem
exchange (``nee``), for which benchmark datasets are provided in
ILAMB.
"""
def __init__(self, **keywords):
super(ConfNEP, self).__init__(**keywords)
def stageData(self, m):
obs = Variable(filename=self.source,
variable_name=self.variable)
self._checkRegions(obs)
obs.data *= -1.0 # Reverse sign of benchmark data.
mod = m.extractTimeSeries(self.variable,
alt_vars=self.alternate_vars)
mod.data *= -1.0 # Reverse sign of modified model outputs.
obs, mod = il.MakeComparable(obs, mod, clip_ref=True,
logstring="[%s][%s]" %
(self.longname, m.name))
return obs, mod
| """A custom ILAMB confrontation for net ecosystem productivity (nep)."""
import os
import numpy as np
from ILAMB.Confrontation import Confrontation
from ILAMB.Variable import Variable
from ILAMB.ilamblib import MakeComparable
class ConfNEP(Confrontation):
"""Confront ``nep`` model outputs with ``nee`` observations.
Net ecosystem productivity (``nep``) is a CMIP5 standard output
provided by the MsTMIP models, and is the inverse of net ecosystem
exchange (``nee``), for which benchmark datasets are provided in
ILAMB.
"""
def __init__(self, **keywords):
super(ConfNEP, self).__init__(**keywords)
def stageData(self, m):
obs = Variable(filename=self.source,
variable_name=self.variable)
self._checkRegions(obs)
obs.data *= -1.0 # Reverse sign of benchmark data.
mod = m.extractTimeSeries(self.variable,
alt_vars=self.alternate_vars)
mod.data *= -1.0 # Reverse sign of modified model outputs.
obs, mod = MakeComparable(obs, mod, clip_ref=True,
logstring="[%s][%s]" %
(self.longname, m.name))
return obs, mod
| Change import so it works | Change import so it works
| Python | mit | permamodel/ILAMB-experiments | """A custom ILAMB confrontation for net ecosystem productivity (nep)."""
import os
import numpy as np
from ILAMB.Confrontation import Confrontation
from ILAMB.Variable import Variable
import ilamblib as il
class ConfNEP(Confrontation):
"""Confront ``nep`` model outputs with ``nee`` observations.
Net ecosystem productivity (``nep``) is a CMIP5 standard output
provided by the MsTMIP models, and is the inverse of net ecosystem
exchange (``nee``), for which benchmark datasets are provided in
ILAMB.
"""
def __init__(self, **keywords):
super(ConfNEP, self).__init__(**keywords)
def stageData(self, m):
obs = Variable(filename=self.source,
variable_name=self.variable)
self._checkRegions(obs)
obs.data *= -1.0 # Reverse sign of benchmark data.
mod = m.extractTimeSeries(self.variable,
alt_vars=self.alternate_vars)
mod.data *= -1.0 # Reverse sign of modified model outputs.
obs, mod = il.MakeComparable(obs, mod, clip_ref=True,
logstring="[%s][%s]" %
(self.longname, m.name))
return obs, mod
Change import so it works | """A custom ILAMB confrontation for net ecosystem productivity (nep)."""
import os
import numpy as np
from ILAMB.Confrontation import Confrontation
from ILAMB.Variable import Variable
from ILAMB.ilamblib import MakeComparable
class ConfNEP(Confrontation):
"""Confront ``nep`` model outputs with ``nee`` observations.
Net ecosystem productivity (``nep``) is a CMIP5 standard output
provided by the MsTMIP models, and is the inverse of net ecosystem
exchange (``nee``), for which benchmark datasets are provided in
ILAMB.
"""
def __init__(self, **keywords):
super(ConfNEP, self).__init__(**keywords)
def stageData(self, m):
obs = Variable(filename=self.source,
variable_name=self.variable)
self._checkRegions(obs)
obs.data *= -1.0 # Reverse sign of benchmark data.
mod = m.extractTimeSeries(self.variable,
alt_vars=self.alternate_vars)
mod.data *= -1.0 # Reverse sign of modified model outputs.
obs, mod = MakeComparable(obs, mod, clip_ref=True,
logstring="[%s][%s]" %
(self.longname, m.name))
return obs, mod
| <commit_before>"""A custom ILAMB confrontation for net ecosystem productivity (nep)."""
import os
import numpy as np
from ILAMB.Confrontation import Confrontation
from ILAMB.Variable import Variable
import ilamblib as il
class ConfNEP(Confrontation):
"""Confront ``nep`` model outputs with ``nee`` observations.
Net ecosystem productivity (``nep``) is a CMIP5 standard output
provided by the MsTMIP models, and is the inverse of net ecosystem
exchange (``nee``), for which benchmark datasets are provided in
ILAMB.
"""
def __init__(self, **keywords):
super(ConfNEP, self).__init__(**keywords)
def stageData(self, m):
obs = Variable(filename=self.source,
variable_name=self.variable)
self._checkRegions(obs)
obs.data *= -1.0 # Reverse sign of benchmark data.
mod = m.extractTimeSeries(self.variable,
alt_vars=self.alternate_vars)
mod.data *= -1.0 # Reverse sign of modified model outputs.
obs, mod = il.MakeComparable(obs, mod, clip_ref=True,
logstring="[%s][%s]" %
(self.longname, m.name))
return obs, mod
<commit_msg>Change import so it works<commit_after> | """A custom ILAMB confrontation for net ecosystem productivity (nep)."""
import os
import numpy as np
from ILAMB.Confrontation import Confrontation
from ILAMB.Variable import Variable
from ILAMB.ilamblib import MakeComparable
class ConfNEP(Confrontation):
"""Confront ``nep`` model outputs with ``nee`` observations.
Net ecosystem productivity (``nep``) is a CMIP5 standard output
provided by the MsTMIP models, and is the inverse of net ecosystem
exchange (``nee``), for which benchmark datasets are provided in
ILAMB.
"""
def __init__(self, **keywords):
super(ConfNEP, self).__init__(**keywords)
def stageData(self, m):
obs = Variable(filename=self.source,
variable_name=self.variable)
self._checkRegions(obs)
obs.data *= -1.0 # Reverse sign of benchmark data.
mod = m.extractTimeSeries(self.variable,
alt_vars=self.alternate_vars)
mod.data *= -1.0 # Reverse sign of modified model outputs.
obs, mod = MakeComparable(obs, mod, clip_ref=True,
logstring="[%s][%s]" %
(self.longname, m.name))
return obs, mod
| """A custom ILAMB confrontation for net ecosystem productivity (nep)."""
import os
import numpy as np
from ILAMB.Confrontation import Confrontation
from ILAMB.Variable import Variable
import ilamblib as il
class ConfNEP(Confrontation):
"""Confront ``nep`` model outputs with ``nee`` observations.
Net ecosystem productivity (``nep``) is a CMIP5 standard output
provided by the MsTMIP models, and is the inverse of net ecosystem
exchange (``nee``), for which benchmark datasets are provided in
ILAMB.
"""
def __init__(self, **keywords):
super(ConfNEP, self).__init__(**keywords)
def stageData(self, m):
obs = Variable(filename=self.source,
variable_name=self.variable)
self._checkRegions(obs)
obs.data *= -1.0 # Reverse sign of benchmark data.
mod = m.extractTimeSeries(self.variable,
alt_vars=self.alternate_vars)
mod.data *= -1.0 # Reverse sign of modified model outputs.
obs, mod = il.MakeComparable(obs, mod, clip_ref=True,
logstring="[%s][%s]" %
(self.longname, m.name))
return obs, mod
Change import so it works"""A custom ILAMB confrontation for net ecosystem productivity (nep)."""
import os
import numpy as np
from ILAMB.Confrontation import Confrontation
from ILAMB.Variable import Variable
from ILAMB.ilamblib import MakeComparable
class ConfNEP(Confrontation):
"""Confront ``nep`` model outputs with ``nee`` observations.
Net ecosystem productivity (``nep``) is a CMIP5 standard output
provided by the MsTMIP models, and is the inverse of net ecosystem
exchange (``nee``), for which benchmark datasets are provided in
ILAMB.
"""
def __init__(self, **keywords):
super(ConfNEP, self).__init__(**keywords)
def stageData(self, m):
obs = Variable(filename=self.source,
variable_name=self.variable)
self._checkRegions(obs)
obs.data *= -1.0 # Reverse sign of benchmark data.
mod = m.extractTimeSeries(self.variable,
alt_vars=self.alternate_vars)
mod.data *= -1.0 # Reverse sign of modified model outputs.
obs, mod = MakeComparable(obs, mod, clip_ref=True,
logstring="[%s][%s]" %
(self.longname, m.name))
return obs, mod
| <commit_before>"""A custom ILAMB confrontation for net ecosystem productivity (nep)."""
import os
import numpy as np
from ILAMB.Confrontation import Confrontation
from ILAMB.Variable import Variable
import ilamblib as il
class ConfNEP(Confrontation):
"""Confront ``nep`` model outputs with ``nee`` observations.
Net ecosystem productivity (``nep``) is a CMIP5 standard output
provided by the MsTMIP models, and is the inverse of net ecosystem
exchange (``nee``), for which benchmark datasets are provided in
ILAMB.
"""
def __init__(self, **keywords):
super(ConfNEP, self).__init__(**keywords)
def stageData(self, m):
obs = Variable(filename=self.source,
variable_name=self.variable)
self._checkRegions(obs)
obs.data *= -1.0 # Reverse sign of benchmark data.
mod = m.extractTimeSeries(self.variable,
alt_vars=self.alternate_vars)
mod.data *= -1.0 # Reverse sign of modified model outputs.
obs, mod = il.MakeComparable(obs, mod, clip_ref=True,
logstring="[%s][%s]" %
(self.longname, m.name))
return obs, mod
<commit_msg>Change import so it works<commit_after>"""A custom ILAMB confrontation for net ecosystem productivity (nep)."""
import os
import numpy as np
from ILAMB.Confrontation import Confrontation
from ILAMB.Variable import Variable
from ILAMB.ilamblib import MakeComparable
class ConfNEP(Confrontation):
"""Confront ``nep`` model outputs with ``nee`` observations.
Net ecosystem productivity (``nep``) is a CMIP5 standard output
provided by the MsTMIP models, and is the inverse of net ecosystem
exchange (``nee``), for which benchmark datasets are provided in
ILAMB.
"""
def __init__(self, **keywords):
super(ConfNEP, self).__init__(**keywords)
def stageData(self, m):
obs = Variable(filename=self.source,
variable_name=self.variable)
self._checkRegions(obs)
obs.data *= -1.0 # Reverse sign of benchmark data.
mod = m.extractTimeSeries(self.variable,
alt_vars=self.alternate_vars)
mod.data *= -1.0 # Reverse sign of modified model outputs.
obs, mod = MakeComparable(obs, mod, clip_ref=True,
logstring="[%s][%s]" %
(self.longname, m.name))
return obs, mod
|
23ec0899eaf60a9dc79f6671461a33eea7e7f464 | authtools/backends.py | authtools/backends.py | from django.contrib.auth.backends import ModelBackend
class CaseInsensitiveEmailBackend(ModelBackend):
"""
This authentication backend assumes that usernames are email addresses and simply lowercases
a username before an attempt is made to authenticate said username using Django's ModelBackend.
Example usage:
# In settings.py
AUTHENTICATION_BACKENDS = ('authtools.backends.CaseInsensitiveEmailBackend',)
NOTE:
A word of caution. Use of this backend presupposes a way to ensure that users cannot create
usernames that differ only in case (e.g., joe@test.org and JOE@test.org). Using this backend
in such a system is a huge security risk.
"""
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackend, self).authenticate(
username=username,
password=password,
**kwargs
)
| from django.contrib.auth.backends import ModelBackend
class CaseInsensitiveEmailBackendMixin(object):
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackendMixin, self).authenticate(
username=username,
password=password,
**kwargs
)
class CaseInsensitiveEmailBackend(ModelBackend):
"""
This authentication backend assumes that usernames are email addresses and simply lowercases
a username before an attempt is made to authenticate said username using Django's ModelBackend.
Example usage:
# In settings.py
AUTHENTICATION_BACKENDS = ('authtools.backends.CaseInsensitiveEmailBackend',)
NOTE:
A word of caution. Use of this backend presupposes a way to ensure that users cannot create
usernames that differ only in case (e.g., joe@test.org and JOE@test.org). Using this backend
in such a system is a huge security risk.
"""
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackend, self).authenticate(
username=username,
password=password,
**kwargs
)
| Add mixin to make the case-insensitive email auth backend more flexible | Add mixin to make the case-insensitive email auth backend more flexible
| Python | bsd-2-clause | fusionbox/django-authtools,vuchau/django-authtools,moreati/django-authtools,eevol/django-authtools,kivikakk/django-authtools | from django.contrib.auth.backends import ModelBackend
class CaseInsensitiveEmailBackend(ModelBackend):
"""
This authentication backend assumes that usernames are email addresses and simply lowercases
a username before an attempt is made to authenticate said username using Django's ModelBackend.
Example usage:
# In settings.py
AUTHENTICATION_BACKENDS = ('authtools.backends.CaseInsensitiveEmailBackend',)
NOTE:
A word of caution. Use of this backend presupposes a way to ensure that users cannot create
usernames that differ only in case (e.g., joe@test.org and JOE@test.org). Using this backend
in such a system is a huge security risk.
"""
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackend, self).authenticate(
username=username,
password=password,
**kwargs
)
Add mixin to make the case-insensitive email auth backend more flexible | from django.contrib.auth.backends import ModelBackend
class CaseInsensitiveEmailBackendMixin(object):
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackendMixin, self).authenticate(
username=username,
password=password,
**kwargs
)
class CaseInsensitiveEmailBackend(ModelBackend):
"""
This authentication backend assumes that usernames are email addresses and simply lowercases
a username before an attempt is made to authenticate said username using Django's ModelBackend.
Example usage:
# In settings.py
AUTHENTICATION_BACKENDS = ('authtools.backends.CaseInsensitiveEmailBackend',)
NOTE:
A word of caution. Use of this backend presupposes a way to ensure that users cannot create
usernames that differ only in case (e.g., joe@test.org and JOE@test.org). Using this backend
in such a system is a huge security risk.
"""
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackend, self).authenticate(
username=username,
password=password,
**kwargs
)
| <commit_before>from django.contrib.auth.backends import ModelBackend
class CaseInsensitiveEmailBackend(ModelBackend):
"""
This authentication backend assumes that usernames are email addresses and simply lowercases
a username before an attempt is made to authenticate said username using Django's ModelBackend.
Example usage:
# In settings.py
AUTHENTICATION_BACKENDS = ('authtools.backends.CaseInsensitiveEmailBackend',)
NOTE:
A word of caution. Use of this backend presupposes a way to ensure that users cannot create
usernames that differ only in case (e.g., joe@test.org and JOE@test.org). Using this backend
in such a system is a huge security risk.
"""
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackend, self).authenticate(
username=username,
password=password,
**kwargs
)
<commit_msg>Add mixin to make the case-insensitive email auth backend more flexible<commit_after> | from django.contrib.auth.backends import ModelBackend
class CaseInsensitiveEmailBackendMixin(object):
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackendMixin, self).authenticate(
username=username,
password=password,
**kwargs
)
class CaseInsensitiveEmailBackend(ModelBackend):
"""
This authentication backend assumes that usernames are email addresses and simply lowercases
a username before an attempt is made to authenticate said username using Django's ModelBackend.
Example usage:
# In settings.py
AUTHENTICATION_BACKENDS = ('authtools.backends.CaseInsensitiveEmailBackend',)
NOTE:
A word of caution. Use of this backend presupposes a way to ensure that users cannot create
usernames that differ only in case (e.g., joe@test.org and JOE@test.org). Using this backend
in such a system is a huge security risk.
"""
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackend, self).authenticate(
username=username,
password=password,
**kwargs
)
| from django.contrib.auth.backends import ModelBackend
class CaseInsensitiveEmailBackend(ModelBackend):
"""
This authentication backend assumes that usernames are email addresses and simply lowercases
a username before an attempt is made to authenticate said username using Django's ModelBackend.
Example usage:
# In settings.py
AUTHENTICATION_BACKENDS = ('authtools.backends.CaseInsensitiveEmailBackend',)
NOTE:
A word of caution. Use of this backend presupposes a way to ensure that users cannot create
usernames that differ only in case (e.g., joe@test.org and JOE@test.org). Using this backend
in such a system is a huge security risk.
"""
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackend, self).authenticate(
username=username,
password=password,
**kwargs
)
Add mixin to make the case-insensitive email auth backend more flexiblefrom django.contrib.auth.backends import ModelBackend
class CaseInsensitiveEmailBackendMixin(object):
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackendMixin, self).authenticate(
username=username,
password=password,
**kwargs
)
class CaseInsensitiveEmailBackend(ModelBackend):
"""
This authentication backend assumes that usernames are email addresses and simply lowercases
a username before an attempt is made to authenticate said username using Django's ModelBackend.
Example usage:
# In settings.py
AUTHENTICATION_BACKENDS = ('authtools.backends.CaseInsensitiveEmailBackend',)
NOTE:
A word of caution. Use of this backend presupposes a way to ensure that users cannot create
usernames that differ only in case (e.g., joe@test.org and JOE@test.org). Using this backend
in such a system is a huge security risk.
"""
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackend, self).authenticate(
username=username,
password=password,
**kwargs
)
| <commit_before>from django.contrib.auth.backends import ModelBackend
class CaseInsensitiveEmailBackend(ModelBackend):
"""
This authentication backend assumes that usernames are email addresses and simply lowercases
a username before an attempt is made to authenticate said username using Django's ModelBackend.
Example usage:
# In settings.py
AUTHENTICATION_BACKENDS = ('authtools.backends.CaseInsensitiveEmailBackend',)
NOTE:
A word of caution. Use of this backend presupposes a way to ensure that users cannot create
usernames that differ only in case (e.g., joe@test.org and JOE@test.org). Using this backend
in such a system is a huge security risk.
"""
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackend, self).authenticate(
username=username,
password=password,
**kwargs
)
<commit_msg>Add mixin to make the case-insensitive email auth backend more flexible<commit_after>from django.contrib.auth.backends import ModelBackend
class CaseInsensitiveEmailBackendMixin(object):
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackendMixin, self).authenticate(
username=username,
password=password,
**kwargs
)
class CaseInsensitiveEmailBackend(ModelBackend):
"""
This authentication backend assumes that usernames are email addresses and simply lowercases
a username before an attempt is made to authenticate said username using Django's ModelBackend.
Example usage:
# In settings.py
AUTHENTICATION_BACKENDS = ('authtools.backends.CaseInsensitiveEmailBackend',)
NOTE:
A word of caution. Use of this backend presupposes a way to ensure that users cannot create
usernames that differ only in case (e.g., joe@test.org and JOE@test.org). Using this backend
in such a system is a huge security risk.
"""
def authenticate(self, username=None, password=None, **kwargs):
if username is not None:
username = username.lower()
return super(CaseInsensitiveEmailBackend, self).authenticate(
username=username,
password=password,
**kwargs
)
|
e7d171b8b3721093c126560d1982e8eaebc4de6b | jay/urls.py | jay/urls.py | """jay URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from django.contrib.auth import views as auth_views
from django.views.generic import TemplateView
from . import demo_urls
from votes import urls as votes_urls
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', TemplateView.as_view(template_name="base/base.html")),
url(r'^demo/', include(demo_urls)),
url(r'^(?P<system_name>[\w-]+)/', include(votes_urls)),
url(r'^login/', auth_views.login, {'template_name': 'auth/login.html'}),
url(r'^logout/', auth_views.logout, {'template_name': 'auth/logout.html'}),
]
| """jay URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from django.contrib.auth import views as auth_views
from django.views.generic import TemplateView
from . import demo_urls
from votes import urls as votes_urls
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', TemplateView.as_view(template_name="base/base.html")),
url(r'^demo/', include(demo_urls)),
url(r'^(?P<system_name>[\w-]+)/', include(votes_urls, namespace='votes')),
url(r'^login/', auth_views.login, {'template_name': 'auth/login.html'}),
url(r'^logout/', auth_views.logout, {'template_name': 'auth/logout.html'}),
]
| Add namespace to votes app URLs | Add namespace to votes app URLs
| Python | mit | kuboschek/jay,OpenJUB/jay,OpenJUB/jay,kuboschek/jay,kuboschek/jay,OpenJUB/jay | """jay URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from django.contrib.auth import views as auth_views
from django.views.generic import TemplateView
from . import demo_urls
from votes import urls as votes_urls
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', TemplateView.as_view(template_name="base/base.html")),
url(r'^demo/', include(demo_urls)),
url(r'^(?P<system_name>[\w-]+)/', include(votes_urls)),
url(r'^login/', auth_views.login, {'template_name': 'auth/login.html'}),
url(r'^logout/', auth_views.logout, {'template_name': 'auth/logout.html'}),
]
Add namespace to votes app URLs | """jay URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from django.contrib.auth import views as auth_views
from django.views.generic import TemplateView
from . import demo_urls
from votes import urls as votes_urls
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', TemplateView.as_view(template_name="base/base.html")),
url(r'^demo/', include(demo_urls)),
url(r'^(?P<system_name>[\w-]+)/', include(votes_urls, namespace='votes')),
url(r'^login/', auth_views.login, {'template_name': 'auth/login.html'}),
url(r'^logout/', auth_views.logout, {'template_name': 'auth/logout.html'}),
]
| <commit_before>"""jay URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from django.contrib.auth import views as auth_views
from django.views.generic import TemplateView
from . import demo_urls
from votes import urls as votes_urls
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', TemplateView.as_view(template_name="base/base.html")),
url(r'^demo/', include(demo_urls)),
url(r'^(?P<system_name>[\w-]+)/', include(votes_urls)),
url(r'^login/', auth_views.login, {'template_name': 'auth/login.html'}),
url(r'^logout/', auth_views.logout, {'template_name': 'auth/logout.html'}),
]
<commit_msg>Add namespace to votes app URLs<commit_after> | """jay URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from django.contrib.auth import views as auth_views
from django.views.generic import TemplateView
from . import demo_urls
from votes import urls as votes_urls
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', TemplateView.as_view(template_name="base/base.html")),
url(r'^demo/', include(demo_urls)),
url(r'^(?P<system_name>[\w-]+)/', include(votes_urls, namespace='votes')),
url(r'^login/', auth_views.login, {'template_name': 'auth/login.html'}),
url(r'^logout/', auth_views.logout, {'template_name': 'auth/logout.html'}),
]
| """jay URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from django.contrib.auth import views as auth_views
from django.views.generic import TemplateView
from . import demo_urls
from votes import urls as votes_urls
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', TemplateView.as_view(template_name="base/base.html")),
url(r'^demo/', include(demo_urls)),
url(r'^(?P<system_name>[\w-]+)/', include(votes_urls)),
url(r'^login/', auth_views.login, {'template_name': 'auth/login.html'}),
url(r'^logout/', auth_views.logout, {'template_name': 'auth/logout.html'}),
]
Add namespace to votes app URLs"""jay URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from django.contrib.auth import views as auth_views
from django.views.generic import TemplateView
from . import demo_urls
from votes import urls as votes_urls
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', TemplateView.as_view(template_name="base/base.html")),
url(r'^demo/', include(demo_urls)),
url(r'^(?P<system_name>[\w-]+)/', include(votes_urls, namespace='votes')),
url(r'^login/', auth_views.login, {'template_name': 'auth/login.html'}),
url(r'^logout/', auth_views.logout, {'template_name': 'auth/logout.html'}),
]
| <commit_before>"""jay URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from django.contrib.auth import views as auth_views
from django.views.generic import TemplateView
from . import demo_urls
from votes import urls as votes_urls
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', TemplateView.as_view(template_name="base/base.html")),
url(r'^demo/', include(demo_urls)),
url(r'^(?P<system_name>[\w-]+)/', include(votes_urls)),
url(r'^login/', auth_views.login, {'template_name': 'auth/login.html'}),
url(r'^logout/', auth_views.logout, {'template_name': 'auth/logout.html'}),
]
<commit_msg>Add namespace to votes app URLs<commit_after>"""jay URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from django.contrib.auth import views as auth_views
from django.views.generic import TemplateView
from . import demo_urls
from votes import urls as votes_urls
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^$', TemplateView.as_view(template_name="base/base.html")),
url(r'^demo/', include(demo_urls)),
url(r'^(?P<system_name>[\w-]+)/', include(votes_urls, namespace='votes')),
url(r'^login/', auth_views.login, {'template_name': 'auth/login.html'}),
url(r'^logout/', auth_views.logout, {'template_name': 'auth/logout.html'}),
]
|
4874f1b1a7a3ff465493f601f5f056bf8f3f1921 | badgekit_webhooks/urls.py | badgekit_webhooks/urls.py | from __future__ import unicode_literals
from django.conf.urls import patterns, url
from . import views
from django.contrib.admin.views.decorators import staff_member_required
urlpatterns = patterns(
"",
url(r"^hello/$", "badgekit_webhooks.views.hello", name="badgekit_webhooks_hello"),
url(r"^issued/$", "badgekit_webhooks.views.badge_issued_hook",
name="badge_issued_hook"),
url(r"^instances/([-A-Za-z.0-9_]+)/$", staff_member_required(views.badge_instance_list),
name="badge_instance_list"),
url(r"^claim/([-A-Za-z0-9_]+)/$", 'badgekit_webhooks.views.claim_page'),
url(r"^claim/([-A-Za-z0-9_]+)/email/(html|text)$", 'badgekit_webhooks.views.show_claim_email',
name="show_claim_email"),
url(r"^issue/$", staff_member_required(views.SendClaimCodeView.as_view()),
name="badge_issue_form"),
url(r"^claimcode/([-A-Za-z.0-9_]+)/$",
views.ClaimCodeClaimView.as_view(), name='claimcode_claim'),
url(r"^badges/$", "badgekit_webhooks.views.list_badges_view", name="badges_list"),
)
| from __future__ import unicode_literals
from django.conf.urls import patterns, url
from . import views
from django.contrib.admin.views.decorators import staff_member_required
urlpatterns = patterns(
"",
url(r"^hello/$", "badgekit_webhooks.views.hello", name="badgekit_webhooks_hello"),
url(r"^issued/$", "badgekit_webhooks.views.badge_issued_hook",
name="badge_issued_hook"),
url(r"^claim/([-A-Za-z0-9_]+)/$", 'badgekit_webhooks.views.claim_page'),
url(r"^claim/([-A-Za-z0-9_]+)/email/(html|text)$", 'badgekit_webhooks.views.show_claim_email',
name="show_claim_email"),
url(r"^issue/$", staff_member_required(views.SendClaimCodeView.as_view()),
name="badge_issue_form"),
url(r"^claimcode/([-A-Za-z.0-9_]+)/$",
views.ClaimCodeClaimView.as_view(), name='claimcode_claim'),
url(r"^badges/$", "badgekit_webhooks.views.list_badges_view", name="badges_list"),
url(r"^badges/([-A-Za-z.0-9_]+)/instances/$", staff_member_required(views.badge_instance_list),
name="badge_instance_list"),
)
| Move badge instance list beneath 'badges' url | Move badge instance list beneath 'badges' url
| Python | mit | tgs/django-badgekit-webhooks | from __future__ import unicode_literals
from django.conf.urls import patterns, url
from . import views
from django.contrib.admin.views.decorators import staff_member_required
urlpatterns = patterns(
"",
url(r"^hello/$", "badgekit_webhooks.views.hello", name="badgekit_webhooks_hello"),
url(r"^issued/$", "badgekit_webhooks.views.badge_issued_hook",
name="badge_issued_hook"),
url(r"^instances/([-A-Za-z.0-9_]+)/$", staff_member_required(views.badge_instance_list),
name="badge_instance_list"),
url(r"^claim/([-A-Za-z0-9_]+)/$", 'badgekit_webhooks.views.claim_page'),
url(r"^claim/([-A-Za-z0-9_]+)/email/(html|text)$", 'badgekit_webhooks.views.show_claim_email',
name="show_claim_email"),
url(r"^issue/$", staff_member_required(views.SendClaimCodeView.as_view()),
name="badge_issue_form"),
url(r"^claimcode/([-A-Za-z.0-9_]+)/$",
views.ClaimCodeClaimView.as_view(), name='claimcode_claim'),
url(r"^badges/$", "badgekit_webhooks.views.list_badges_view", name="badges_list"),
)
Move badge instance list beneath 'badges' url | from __future__ import unicode_literals
from django.conf.urls import patterns, url
from . import views
from django.contrib.admin.views.decorators import staff_member_required
urlpatterns = patterns(
"",
url(r"^hello/$", "badgekit_webhooks.views.hello", name="badgekit_webhooks_hello"),
url(r"^issued/$", "badgekit_webhooks.views.badge_issued_hook",
name="badge_issued_hook"),
url(r"^claim/([-A-Za-z0-9_]+)/$", 'badgekit_webhooks.views.claim_page'),
url(r"^claim/([-A-Za-z0-9_]+)/email/(html|text)$", 'badgekit_webhooks.views.show_claim_email',
name="show_claim_email"),
url(r"^issue/$", staff_member_required(views.SendClaimCodeView.as_view()),
name="badge_issue_form"),
url(r"^claimcode/([-A-Za-z.0-9_]+)/$",
views.ClaimCodeClaimView.as_view(), name='claimcode_claim'),
url(r"^badges/$", "badgekit_webhooks.views.list_badges_view", name="badges_list"),
url(r"^badges/([-A-Za-z.0-9_]+)/instances/$", staff_member_required(views.badge_instance_list),
name="badge_instance_list"),
)
| <commit_before>from __future__ import unicode_literals
from django.conf.urls import patterns, url
from . import views
from django.contrib.admin.views.decorators import staff_member_required
urlpatterns = patterns(
"",
url(r"^hello/$", "badgekit_webhooks.views.hello", name="badgekit_webhooks_hello"),
url(r"^issued/$", "badgekit_webhooks.views.badge_issued_hook",
name="badge_issued_hook"),
url(r"^instances/([-A-Za-z.0-9_]+)/$", staff_member_required(views.badge_instance_list),
name="badge_instance_list"),
url(r"^claim/([-A-Za-z0-9_]+)/$", 'badgekit_webhooks.views.claim_page'),
url(r"^claim/([-A-Za-z0-9_]+)/email/(html|text)$", 'badgekit_webhooks.views.show_claim_email',
name="show_claim_email"),
url(r"^issue/$", staff_member_required(views.SendClaimCodeView.as_view()),
name="badge_issue_form"),
url(r"^claimcode/([-A-Za-z.0-9_]+)/$",
views.ClaimCodeClaimView.as_view(), name='claimcode_claim'),
url(r"^badges/$", "badgekit_webhooks.views.list_badges_view", name="badges_list"),
)
<commit_msg>Move badge instance list beneath 'badges' url<commit_after> | from __future__ import unicode_literals
from django.conf.urls import patterns, url
from . import views
from django.contrib.admin.views.decorators import staff_member_required
urlpatterns = patterns(
"",
url(r"^hello/$", "badgekit_webhooks.views.hello", name="badgekit_webhooks_hello"),
url(r"^issued/$", "badgekit_webhooks.views.badge_issued_hook",
name="badge_issued_hook"),
url(r"^claim/([-A-Za-z0-9_]+)/$", 'badgekit_webhooks.views.claim_page'),
url(r"^claim/([-A-Za-z0-9_]+)/email/(html|text)$", 'badgekit_webhooks.views.show_claim_email',
name="show_claim_email"),
url(r"^issue/$", staff_member_required(views.SendClaimCodeView.as_view()),
name="badge_issue_form"),
url(r"^claimcode/([-A-Za-z.0-9_]+)/$",
views.ClaimCodeClaimView.as_view(), name='claimcode_claim'),
url(r"^badges/$", "badgekit_webhooks.views.list_badges_view", name="badges_list"),
url(r"^badges/([-A-Za-z.0-9_]+)/instances/$", staff_member_required(views.badge_instance_list),
name="badge_instance_list"),
)
| from __future__ import unicode_literals
from django.conf.urls import patterns, url
from . import views
from django.contrib.admin.views.decorators import staff_member_required
urlpatterns = patterns(
"",
url(r"^hello/$", "badgekit_webhooks.views.hello", name="badgekit_webhooks_hello"),
url(r"^issued/$", "badgekit_webhooks.views.badge_issued_hook",
name="badge_issued_hook"),
url(r"^instances/([-A-Za-z.0-9_]+)/$", staff_member_required(views.badge_instance_list),
name="badge_instance_list"),
url(r"^claim/([-A-Za-z0-9_]+)/$", 'badgekit_webhooks.views.claim_page'),
url(r"^claim/([-A-Za-z0-9_]+)/email/(html|text)$", 'badgekit_webhooks.views.show_claim_email',
name="show_claim_email"),
url(r"^issue/$", staff_member_required(views.SendClaimCodeView.as_view()),
name="badge_issue_form"),
url(r"^claimcode/([-A-Za-z.0-9_]+)/$",
views.ClaimCodeClaimView.as_view(), name='claimcode_claim'),
url(r"^badges/$", "badgekit_webhooks.views.list_badges_view", name="badges_list"),
)
Move badge instance list beneath 'badges' urlfrom __future__ import unicode_literals
from django.conf.urls import patterns, url
from . import views
from django.contrib.admin.views.decorators import staff_member_required
urlpatterns = patterns(
"",
url(r"^hello/$", "badgekit_webhooks.views.hello", name="badgekit_webhooks_hello"),
url(r"^issued/$", "badgekit_webhooks.views.badge_issued_hook",
name="badge_issued_hook"),
url(r"^claim/([-A-Za-z0-9_]+)/$", 'badgekit_webhooks.views.claim_page'),
url(r"^claim/([-A-Za-z0-9_]+)/email/(html|text)$", 'badgekit_webhooks.views.show_claim_email',
name="show_claim_email"),
url(r"^issue/$", staff_member_required(views.SendClaimCodeView.as_view()),
name="badge_issue_form"),
url(r"^claimcode/([-A-Za-z.0-9_]+)/$",
views.ClaimCodeClaimView.as_view(), name='claimcode_claim'),
url(r"^badges/$", "badgekit_webhooks.views.list_badges_view", name="badges_list"),
url(r"^badges/([-A-Za-z.0-9_]+)/instances/$", staff_member_required(views.badge_instance_list),
name="badge_instance_list"),
)
| <commit_before>from __future__ import unicode_literals
from django.conf.urls import patterns, url
from . import views
from django.contrib.admin.views.decorators import staff_member_required
urlpatterns = patterns(
"",
url(r"^hello/$", "badgekit_webhooks.views.hello", name="badgekit_webhooks_hello"),
url(r"^issued/$", "badgekit_webhooks.views.badge_issued_hook",
name="badge_issued_hook"),
url(r"^instances/([-A-Za-z.0-9_]+)/$", staff_member_required(views.badge_instance_list),
name="badge_instance_list"),
url(r"^claim/([-A-Za-z0-9_]+)/$", 'badgekit_webhooks.views.claim_page'),
url(r"^claim/([-A-Za-z0-9_]+)/email/(html|text)$", 'badgekit_webhooks.views.show_claim_email',
name="show_claim_email"),
url(r"^issue/$", staff_member_required(views.SendClaimCodeView.as_view()),
name="badge_issue_form"),
url(r"^claimcode/([-A-Za-z.0-9_]+)/$",
views.ClaimCodeClaimView.as_view(), name='claimcode_claim'),
url(r"^badges/$", "badgekit_webhooks.views.list_badges_view", name="badges_list"),
)
<commit_msg>Move badge instance list beneath 'badges' url<commit_after>from __future__ import unicode_literals
from django.conf.urls import patterns, url
from . import views
from django.contrib.admin.views.decorators import staff_member_required
urlpatterns = patterns(
"",
url(r"^hello/$", "badgekit_webhooks.views.hello", name="badgekit_webhooks_hello"),
url(r"^issued/$", "badgekit_webhooks.views.badge_issued_hook",
name="badge_issued_hook"),
url(r"^claim/([-A-Za-z0-9_]+)/$", 'badgekit_webhooks.views.claim_page'),
url(r"^claim/([-A-Za-z0-9_]+)/email/(html|text)$", 'badgekit_webhooks.views.show_claim_email',
name="show_claim_email"),
url(r"^issue/$", staff_member_required(views.SendClaimCodeView.as_view()),
name="badge_issue_form"),
url(r"^claimcode/([-A-Za-z.0-9_]+)/$",
views.ClaimCodeClaimView.as_view(), name='claimcode_claim'),
url(r"^badges/$", "badgekit_webhooks.views.list_badges_view", name="badges_list"),
url(r"^badges/([-A-Za-z.0-9_]+)/instances/$", staff_member_required(views.badge_instance_list),
name="badge_instance_list"),
)
|
11312893661ac339212fad7d81a21c9ddc2533d3 | identities/tasks.py | identities/tasks.py | import json
import requests
from celery.task import Task
from django.conf import settings
class DeliverHook(Task):
def run(self, target, payload, instance=None, hook=None, **kwargs):
"""
target: the url to receive the payload.
payload: a python primitive data structure
instance: a possibly null "trigger" instance
hook: the defining Hook object
"""
requests.post(
url=target,
data=json.dumps(payload),
headers={
'Content-Type': 'application/json',
'Authorization': 'Token %s' % settings.HOOK_AUTH_TOKEN
}
)
def deliver_hook_wrapper(target, payload, instance, hook):
if instance is not None:
instance_id = instance.id
else:
instance_id = None
kwargs = dict(target=target, payload=payload,
instance_id=instance_id, hook_id=hook.id)
DeliverHook.apply_async(kwargs=kwargs)
| import json
import requests
from celery.task import Task
from django.conf import settings
class DeliverHook(Task):
def run(self, target, payload, instance_id=None, hook_id=None, **kwargs):
"""
target: the url to receive the payload.
payload: a python primitive data structure
instance_id: a possibly None "trigger" instance ID
hook_id: the ID of defining Hook object
"""
requests.post(
url=target,
data=json.dumps(payload),
headers={
'Content-Type': 'application/json',
'Authorization': 'Token %s' % settings.HOOK_AUTH_TOKEN
}
)
def deliver_hook_wrapper(target, payload, instance, hook):
if instance is not None:
instance_id = instance.id
else:
instance_id = None
kwargs = dict(target=target, payload=payload,
instance_id=instance_id, hook_id=hook.id)
DeliverHook.apply_async(kwargs=kwargs)
| Clean up params and docstrings | Clean up params and docstrings
| Python | bsd-3-clause | praekelt/seed-identity-store,praekelt/seed-identity-store | import json
import requests
from celery.task import Task
from django.conf import settings
class DeliverHook(Task):
def run(self, target, payload, instance=None, hook=None, **kwargs):
"""
target: the url to receive the payload.
payload: a python primitive data structure
instance: a possibly null "trigger" instance
hook: the defining Hook object
"""
requests.post(
url=target,
data=json.dumps(payload),
headers={
'Content-Type': 'application/json',
'Authorization': 'Token %s' % settings.HOOK_AUTH_TOKEN
}
)
def deliver_hook_wrapper(target, payload, instance, hook):
if instance is not None:
instance_id = instance.id
else:
instance_id = None
kwargs = dict(target=target, payload=payload,
instance_id=instance_id, hook_id=hook.id)
DeliverHook.apply_async(kwargs=kwargs)
Clean up params and docstrings | import json
import requests
from celery.task import Task
from django.conf import settings
class DeliverHook(Task):
def run(self, target, payload, instance_id=None, hook_id=None, **kwargs):
"""
target: the url to receive the payload.
payload: a python primitive data structure
instance_id: a possibly None "trigger" instance ID
hook_id: the ID of defining Hook object
"""
requests.post(
url=target,
data=json.dumps(payload),
headers={
'Content-Type': 'application/json',
'Authorization': 'Token %s' % settings.HOOK_AUTH_TOKEN
}
)
def deliver_hook_wrapper(target, payload, instance, hook):
if instance is not None:
instance_id = instance.id
else:
instance_id = None
kwargs = dict(target=target, payload=payload,
instance_id=instance_id, hook_id=hook.id)
DeliverHook.apply_async(kwargs=kwargs)
| <commit_before>import json
import requests
from celery.task import Task
from django.conf import settings
class DeliverHook(Task):
def run(self, target, payload, instance=None, hook=None, **kwargs):
"""
target: the url to receive the payload.
payload: a python primitive data structure
instance: a possibly null "trigger" instance
hook: the defining Hook object
"""
requests.post(
url=target,
data=json.dumps(payload),
headers={
'Content-Type': 'application/json',
'Authorization': 'Token %s' % settings.HOOK_AUTH_TOKEN
}
)
def deliver_hook_wrapper(target, payload, instance, hook):
if instance is not None:
instance_id = instance.id
else:
instance_id = None
kwargs = dict(target=target, payload=payload,
instance_id=instance_id, hook_id=hook.id)
DeliverHook.apply_async(kwargs=kwargs)
<commit_msg>Clean up params and docstrings<commit_after> | import json
import requests
from celery.task import Task
from django.conf import settings
class DeliverHook(Task):
def run(self, target, payload, instance_id=None, hook_id=None, **kwargs):
"""
target: the url to receive the payload.
payload: a python primitive data structure
instance_id: a possibly None "trigger" instance ID
hook_id: the ID of defining Hook object
"""
requests.post(
url=target,
data=json.dumps(payload),
headers={
'Content-Type': 'application/json',
'Authorization': 'Token %s' % settings.HOOK_AUTH_TOKEN
}
)
def deliver_hook_wrapper(target, payload, instance, hook):
if instance is not None:
instance_id = instance.id
else:
instance_id = None
kwargs = dict(target=target, payload=payload,
instance_id=instance_id, hook_id=hook.id)
DeliverHook.apply_async(kwargs=kwargs)
| import json
import requests
from celery.task import Task
from django.conf import settings
class DeliverHook(Task):
def run(self, target, payload, instance=None, hook=None, **kwargs):
"""
target: the url to receive the payload.
payload: a python primitive data structure
instance: a possibly null "trigger" instance
hook: the defining Hook object
"""
requests.post(
url=target,
data=json.dumps(payload),
headers={
'Content-Type': 'application/json',
'Authorization': 'Token %s' % settings.HOOK_AUTH_TOKEN
}
)
def deliver_hook_wrapper(target, payload, instance, hook):
if instance is not None:
instance_id = instance.id
else:
instance_id = None
kwargs = dict(target=target, payload=payload,
instance_id=instance_id, hook_id=hook.id)
DeliverHook.apply_async(kwargs=kwargs)
Clean up params and docstringsimport json
import requests
from celery.task import Task
from django.conf import settings
class DeliverHook(Task):
def run(self, target, payload, instance_id=None, hook_id=None, **kwargs):
"""
target: the url to receive the payload.
payload: a python primitive data structure
instance_id: a possibly None "trigger" instance ID
hook_id: the ID of defining Hook object
"""
requests.post(
url=target,
data=json.dumps(payload),
headers={
'Content-Type': 'application/json',
'Authorization': 'Token %s' % settings.HOOK_AUTH_TOKEN
}
)
def deliver_hook_wrapper(target, payload, instance, hook):
if instance is not None:
instance_id = instance.id
else:
instance_id = None
kwargs = dict(target=target, payload=payload,
instance_id=instance_id, hook_id=hook.id)
DeliverHook.apply_async(kwargs=kwargs)
| <commit_before>import json
import requests
from celery.task import Task
from django.conf import settings
class DeliverHook(Task):
def run(self, target, payload, instance=None, hook=None, **kwargs):
"""
target: the url to receive the payload.
payload: a python primitive data structure
instance: a possibly null "trigger" instance
hook: the defining Hook object
"""
requests.post(
url=target,
data=json.dumps(payload),
headers={
'Content-Type': 'application/json',
'Authorization': 'Token %s' % settings.HOOK_AUTH_TOKEN
}
)
def deliver_hook_wrapper(target, payload, instance, hook):
if instance is not None:
instance_id = instance.id
else:
instance_id = None
kwargs = dict(target=target, payload=payload,
instance_id=instance_id, hook_id=hook.id)
DeliverHook.apply_async(kwargs=kwargs)
<commit_msg>Clean up params and docstrings<commit_after>import json
import requests
from celery.task import Task
from django.conf import settings
class DeliverHook(Task):
def run(self, target, payload, instance_id=None, hook_id=None, **kwargs):
"""
target: the url to receive the payload.
payload: a python primitive data structure
instance_id: a possibly None "trigger" instance ID
hook_id: the ID of defining Hook object
"""
requests.post(
url=target,
data=json.dumps(payload),
headers={
'Content-Type': 'application/json',
'Authorization': 'Token %s' % settings.HOOK_AUTH_TOKEN
}
)
def deliver_hook_wrapper(target, payload, instance, hook):
if instance is not None:
instance_id = instance.id
else:
instance_id = None
kwargs = dict(target=target, payload=payload,
instance_id=instance_id, hook_id=hook.id)
DeliverHook.apply_async(kwargs=kwargs)
|
42a0fbf29168f12a4bc3afc53bbf7148b9d008f6 | spacy/tests/pipeline/test_textcat.py | spacy/tests/pipeline/test_textcat.py | from __future__ import unicode_literals
from ...language import Language
def test_simple_train():
nlp = Language()
nlp.add_pipe(nlp.create_pipe('textcat'))
nlp.get_pipe('textcat').add_label('is_good')
nlp.begin_training()
for i in range(5):
for text, answer in [('aaaa', 1.), ('bbbb', 0), ('aa', 1.),
('bbbbbbbbb', 0.), ('aaaaaa', 1)]:
nlp.update([text], [{'cats': {'answer': answer}}])
doc = nlp(u'aaa')
assert 'is_good' in doc.cats
assert doc.cats['is_good'] >= 0.5
| # coding: utf8
from __future__ import unicode_literals
from ...language import Language
def test_simple_train():
nlp = Language()
nlp.add_pipe(nlp.create_pipe('textcat'))
nlp.get_pipe('textcat').add_label('answer')
nlp.begin_training()
for i in range(5):
for text, answer in [('aaaa', 1.), ('bbbb', 0), ('aa', 1.),
('bbbbbbbbb', 0.), ('aaaaaa', 1)]:
nlp.update([text], [{'cats': {'answer': answer}}])
doc = nlp(u'aaa')
assert 'answer' in doc.cats
assert doc.cats['answer'] >= 0.5
| Fix textcat simple train example | Fix textcat simple train example
| Python | mit | aikramer2/spaCy,explosion/spaCy,explosion/spaCy,recognai/spaCy,explosion/spaCy,recognai/spaCy,aikramer2/spaCy,spacy-io/spaCy,aikramer2/spaCy,honnibal/spaCy,spacy-io/spaCy,spacy-io/spaCy,spacy-io/spaCy,honnibal/spaCy,spacy-io/spaCy,explosion/spaCy,explosion/spaCy,spacy-io/spaCy,recognai/spaCy,honnibal/spaCy,aikramer2/spaCy,recognai/spaCy,honnibal/spaCy,recognai/spaCy,recognai/spaCy,aikramer2/spaCy,explosion/spaCy,aikramer2/spaCy | from __future__ import unicode_literals
from ...language import Language
def test_simple_train():
nlp = Language()
nlp.add_pipe(nlp.create_pipe('textcat'))
nlp.get_pipe('textcat').add_label('is_good')
nlp.begin_training()
for i in range(5):
for text, answer in [('aaaa', 1.), ('bbbb', 0), ('aa', 1.),
('bbbbbbbbb', 0.), ('aaaaaa', 1)]:
nlp.update([text], [{'cats': {'answer': answer}}])
doc = nlp(u'aaa')
assert 'is_good' in doc.cats
assert doc.cats['is_good'] >= 0.5
Fix textcat simple train example | # coding: utf8
from __future__ import unicode_literals
from ...language import Language
def test_simple_train():
nlp = Language()
nlp.add_pipe(nlp.create_pipe('textcat'))
nlp.get_pipe('textcat').add_label('answer')
nlp.begin_training()
for i in range(5):
for text, answer in [('aaaa', 1.), ('bbbb', 0), ('aa', 1.),
('bbbbbbbbb', 0.), ('aaaaaa', 1)]:
nlp.update([text], [{'cats': {'answer': answer}}])
doc = nlp(u'aaa')
assert 'answer' in doc.cats
assert doc.cats['answer'] >= 0.5
| <commit_before>from __future__ import unicode_literals
from ...language import Language
def test_simple_train():
nlp = Language()
nlp.add_pipe(nlp.create_pipe('textcat'))
nlp.get_pipe('textcat').add_label('is_good')
nlp.begin_training()
for i in range(5):
for text, answer in [('aaaa', 1.), ('bbbb', 0), ('aa', 1.),
('bbbbbbbbb', 0.), ('aaaaaa', 1)]:
nlp.update([text], [{'cats': {'answer': answer}}])
doc = nlp(u'aaa')
assert 'is_good' in doc.cats
assert doc.cats['is_good'] >= 0.5
<commit_msg>Fix textcat simple train example<commit_after> | # coding: utf8
from __future__ import unicode_literals
from ...language import Language
def test_simple_train():
nlp = Language()
nlp.add_pipe(nlp.create_pipe('textcat'))
nlp.get_pipe('textcat').add_label('answer')
nlp.begin_training()
for i in range(5):
for text, answer in [('aaaa', 1.), ('bbbb', 0), ('aa', 1.),
('bbbbbbbbb', 0.), ('aaaaaa', 1)]:
nlp.update([text], [{'cats': {'answer': answer}}])
doc = nlp(u'aaa')
assert 'answer' in doc.cats
assert doc.cats['answer'] >= 0.5
| from __future__ import unicode_literals
from ...language import Language
def test_simple_train():
nlp = Language()
nlp.add_pipe(nlp.create_pipe('textcat'))
nlp.get_pipe('textcat').add_label('is_good')
nlp.begin_training()
for i in range(5):
for text, answer in [('aaaa', 1.), ('bbbb', 0), ('aa', 1.),
('bbbbbbbbb', 0.), ('aaaaaa', 1)]:
nlp.update([text], [{'cats': {'answer': answer}}])
doc = nlp(u'aaa')
assert 'is_good' in doc.cats
assert doc.cats['is_good'] >= 0.5
Fix textcat simple train example# coding: utf8
from __future__ import unicode_literals
from ...language import Language
def test_simple_train():
nlp = Language()
nlp.add_pipe(nlp.create_pipe('textcat'))
nlp.get_pipe('textcat').add_label('answer')
nlp.begin_training()
for i in range(5):
for text, answer in [('aaaa', 1.), ('bbbb', 0), ('aa', 1.),
('bbbbbbbbb', 0.), ('aaaaaa', 1)]:
nlp.update([text], [{'cats': {'answer': answer}}])
doc = nlp(u'aaa')
assert 'answer' in doc.cats
assert doc.cats['answer'] >= 0.5
| <commit_before>from __future__ import unicode_literals
from ...language import Language
def test_simple_train():
nlp = Language()
nlp.add_pipe(nlp.create_pipe('textcat'))
nlp.get_pipe('textcat').add_label('is_good')
nlp.begin_training()
for i in range(5):
for text, answer in [('aaaa', 1.), ('bbbb', 0), ('aa', 1.),
('bbbbbbbbb', 0.), ('aaaaaa', 1)]:
nlp.update([text], [{'cats': {'answer': answer}}])
doc = nlp(u'aaa')
assert 'is_good' in doc.cats
assert doc.cats['is_good'] >= 0.5
<commit_msg>Fix textcat simple train example<commit_after># coding: utf8
from __future__ import unicode_literals
from ...language import Language
def test_simple_train():
nlp = Language()
nlp.add_pipe(nlp.create_pipe('textcat'))
nlp.get_pipe('textcat').add_label('answer')
nlp.begin_training()
for i in range(5):
for text, answer in [('aaaa', 1.), ('bbbb', 0), ('aa', 1.),
('bbbbbbbbb', 0.), ('aaaaaa', 1)]:
nlp.update([text], [{'cats': {'answer': answer}}])
doc = nlp(u'aaa')
assert 'answer' in doc.cats
assert doc.cats['answer'] >= 0.5
|
0b5b25b5cc3b5fe59c0a263983dae05ebfdc8de9 | plenum/common/transactions.py | plenum/common/transactions.py | from enum import Enum
class Transactions(Enum):
def __str__(self):
return self.name
class PlenumTransactions(Transactions):
# These numeric constants CANNOT be changed once they have been used,
# because that would break backwards compatibility with the ledger
# Also the numeric constants CANNOT collide with transactions in dependent
# components.
NODE = "0"
NYM = "1"
AUDIT = "2"
GET_TXN = "3"
| from enum import Enum
class Transactions(Enum):
def __str__(self):
return self.name
class PlenumTransactions(Transactions):
# These numeric constants CANNOT be changed once they have been used,
# because that would break backwards compatibility with the ledger
# Also the numeric constants CANNOT collide with transactions in dependent
# components.
NODE = "0"
NYM = "1"
AUDIT = "2"
GET_TXN = "3"
TXN_AUTHOR_AGREEMENT = "4"
TXN_AUTHOR_AGREEMENT_AML = "5"
GET_TXN_AUTHOR_AGREEMENT = "6"
GET_TXN_AUTHOR_AGREEMENT_AML = "7"
| Add new TAA transaction codes | INDY-2066: Add new TAA transaction codes
Signed-off-by: Sergey Khoroshavin <b770466c7a06c5fe47531d5f0e31684f1131354d@dsr-corporation.com>
| Python | apache-2.0 | evernym/zeno,evernym/plenum | from enum import Enum
class Transactions(Enum):
def __str__(self):
return self.name
class PlenumTransactions(Transactions):
# These numeric constants CANNOT be changed once they have been used,
# because that would break backwards compatibility with the ledger
# Also the numeric constants CANNOT collide with transactions in dependent
# components.
NODE = "0"
NYM = "1"
AUDIT = "2"
GET_TXN = "3"
INDY-2066: Add new TAA transaction codes
Signed-off-by: Sergey Khoroshavin <b770466c7a06c5fe47531d5f0e31684f1131354d@dsr-corporation.com> | from enum import Enum
class Transactions(Enum):
def __str__(self):
return self.name
class PlenumTransactions(Transactions):
# These numeric constants CANNOT be changed once they have been used,
# because that would break backwards compatibility with the ledger
# Also the numeric constants CANNOT collide with transactions in dependent
# components.
NODE = "0"
NYM = "1"
AUDIT = "2"
GET_TXN = "3"
TXN_AUTHOR_AGREEMENT = "4"
TXN_AUTHOR_AGREEMENT_AML = "5"
GET_TXN_AUTHOR_AGREEMENT = "6"
GET_TXN_AUTHOR_AGREEMENT_AML = "7"
| <commit_before>from enum import Enum
class Transactions(Enum):
def __str__(self):
return self.name
class PlenumTransactions(Transactions):
# These numeric constants CANNOT be changed once they have been used,
# because that would break backwards compatibility with the ledger
# Also the numeric constants CANNOT collide with transactions in dependent
# components.
NODE = "0"
NYM = "1"
AUDIT = "2"
GET_TXN = "3"
<commit_msg>INDY-2066: Add new TAA transaction codes
Signed-off-by: Sergey Khoroshavin <b770466c7a06c5fe47531d5f0e31684f1131354d@dsr-corporation.com><commit_after> | from enum import Enum
class Transactions(Enum):
def __str__(self):
return self.name
class PlenumTransactions(Transactions):
# These numeric constants CANNOT be changed once they have been used,
# because that would break backwards compatibility with the ledger
# Also the numeric constants CANNOT collide with transactions in dependent
# components.
NODE = "0"
NYM = "1"
AUDIT = "2"
GET_TXN = "3"
TXN_AUTHOR_AGREEMENT = "4"
TXN_AUTHOR_AGREEMENT_AML = "5"
GET_TXN_AUTHOR_AGREEMENT = "6"
GET_TXN_AUTHOR_AGREEMENT_AML = "7"
| from enum import Enum
class Transactions(Enum):
def __str__(self):
return self.name
class PlenumTransactions(Transactions):
# These numeric constants CANNOT be changed once they have been used,
# because that would break backwards compatibility with the ledger
# Also the numeric constants CANNOT collide with transactions in dependent
# components.
NODE = "0"
NYM = "1"
AUDIT = "2"
GET_TXN = "3"
INDY-2066: Add new TAA transaction codes
Signed-off-by: Sergey Khoroshavin <b770466c7a06c5fe47531d5f0e31684f1131354d@dsr-corporation.com>from enum import Enum
class Transactions(Enum):
def __str__(self):
return self.name
class PlenumTransactions(Transactions):
# These numeric constants CANNOT be changed once they have been used,
# because that would break backwards compatibility with the ledger
# Also the numeric constants CANNOT collide with transactions in dependent
# components.
NODE = "0"
NYM = "1"
AUDIT = "2"
GET_TXN = "3"
TXN_AUTHOR_AGREEMENT = "4"
TXN_AUTHOR_AGREEMENT_AML = "5"
GET_TXN_AUTHOR_AGREEMENT = "6"
GET_TXN_AUTHOR_AGREEMENT_AML = "7"
| <commit_before>from enum import Enum
class Transactions(Enum):
def __str__(self):
return self.name
class PlenumTransactions(Transactions):
# These numeric constants CANNOT be changed once they have been used,
# because that would break backwards compatibility with the ledger
# Also the numeric constants CANNOT collide with transactions in dependent
# components.
NODE = "0"
NYM = "1"
AUDIT = "2"
GET_TXN = "3"
<commit_msg>INDY-2066: Add new TAA transaction codes
Signed-off-by: Sergey Khoroshavin <b770466c7a06c5fe47531d5f0e31684f1131354d@dsr-corporation.com><commit_after>from enum import Enum
class Transactions(Enum):
def __str__(self):
return self.name
class PlenumTransactions(Transactions):
# These numeric constants CANNOT be changed once they have been used,
# because that would break backwards compatibility with the ledger
# Also the numeric constants CANNOT collide with transactions in dependent
# components.
NODE = "0"
NYM = "1"
AUDIT = "2"
GET_TXN = "3"
TXN_AUTHOR_AGREEMENT = "4"
TXN_AUTHOR_AGREEMENT_AML = "5"
GET_TXN_AUTHOR_AGREEMENT = "6"
GET_TXN_AUTHOR_AGREEMENT_AML = "7"
|
6ad84b8cc930922296fbe8c583c8c69f5b94d9c9 | bots/humbug_git_config.py | bots/humbug_git_config.py | # Humbug Inc's internal git plugin configuration.
# The plugin and example config are under api/integrations/
# Leaving all the instructions out of this file to avoid having to
# sync them as we update the comments.
HUMBUG_USER = "humbug+commits@humbughq.com"
HUMBUG_API_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# commit_notice_destination() lets you customize where commit notices
# are sent to.
#
# It takes the following arguments:
# * repo = the name of the git repository
# * branch = the name of the branch that was pushed to
# * commit = the commit id
#
# Returns a dictionary encoding the stream and subject to send the
# notification to (or None to send no notification, e.g. for ).
#
# The default code below will send every commit pushed to "master" to
# * stream "commits"
# * subject "deploy => master" (using a pretty unicode right arrow)
# And similarly for branch "test-post-receive" (for use when testing).
def commit_notice_destination(repo, branch, commit):
if branch in ["master", "post-receive-test"]:
return dict(stream = "commits",
subject = u"deploy \u21D2 %s" % (branch,))
# Return None for cases where you don't want a notice sent
return None
HUMBUG_API_PATH = "/home/humbug/humbug/api"
HUMBUG_SITE = "https://staging.humbughq.com"
| # Humbug Inc's internal git plugin configuration.
# The plugin and example config are under api/integrations/
# Leaving all the instructions out of this file to avoid having to
# sync them as we update the comments.
HUMBUG_USER = "humbug+commits@humbughq.com"
HUMBUG_API_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# commit_notice_destination() lets you customize where commit notices
# are sent to.
#
# It takes the following arguments:
# * repo = the name of the git repository
# * branch = the name of the branch that was pushed to
# * commit = the commit id
#
# Returns a dictionary encoding the stream and subject to send the
# notification to (or None to send no notification, e.g. for ).
#
# The default code below will send every commit pushed to "master" to
# * stream "commits"
# * subject "deploy => master" (using a pretty unicode right arrow)
# And similarly for branch "test-post-receive" (for use when testing).
def commit_notice_destination(repo, branch, commit):
if branch in ["master", "prod", "post-receive-test"]:
return dict(stream = "commits",
subject = u"deploy \u21D2 %s" % (branch,))
# Return None for cases where you don't want a notice sent
return None
HUMBUG_API_PATH = "/home/humbug/humbug/api"
HUMBUG_SITE = "https://staging.humbughq.com"
| Put prod back on the list of branches to send notices about. | git: Put prod back on the list of branches to send notices about.
(imported from commit e608d7050b4e68045b03341dc41e8654e45a3af3)
| Python | apache-2.0 | MariaFaBella85/zulip,SmartPeople/zulip,blaze225/zulip,calvinleenyc/zulip,natanovia/zulip,moria/zulip,dnmfarrell/zulip,samatdav/zulip,tbutter/zulip,deer-hope/zulip,guiquanz/zulip,sharmaeklavya2/zulip,zulip/zulip,guiquanz/zulip,rishig/zulip,developerfm/zulip,schatt/zulip,AZtheAsian/zulip,praveenaki/zulip,glovebx/zulip,shaunstanislaus/zulip,sonali0901/zulip,KingxBanana/zulip,JanzTam/zulip,yocome/zulip,dawran6/zulip,hj3938/zulip,developerfm/zulip,shrikrishnaholla/zulip,adnanh/zulip,blaze225/zulip,bitemyapp/zulip,qq1012803704/zulip,bowlofstew/zulip,adnanh/zulip,wweiradio/zulip,LAndreas/zulip,KJin99/zulip,jainayush975/zulip,dwrpayne/zulip,akuseru/zulip,lfranchi/zulip,vikas-parashar/zulip,RobotCaleb/zulip,hayderimran7/zulip,Juanvulcano/zulip,Jianchun1/zulip,andersk/zulip,jessedhillon/zulip,eastlhu/zulip,pradiptad/zulip,aps-sids/zulip,ikasumiwt/zulip,atomic-labs/zulip,jerryge/zulip,RobotCaleb/zulip,bssrdf/zulip,zwily/zulip,jimmy54/zulip,technicalpickles/zulip,LeeRisk/zulip,zofuthan/zulip,Vallher/zulip,ryanbackman/zulip,natanovia/zulip,zachallaun/zulip,so0k/zulip,mansilladev/zulip,mohsenSy/zulip,jimmy54/zulip,johnnygaddarr/zulip,hustlzp/zulip,dhcrzf/zulip,hj3938/zulip,moria/zulip,udxxabp/zulip,avastu/zulip,lfranchi/zulip,jackrzhang/zulip,zhaoweigg/zulip,shubhamdhama/zulip,sonali0901/zulip,bowlofstew/zulip,m1ssou/zulip,MariaFaBella85/zulip,brockwhittaker/zulip,hengqujushi/zulip,MayB/zulip,sonali0901/zulip,aps-sids/zulip,grave-w-grave/zulip,wdaher/zulip,Drooids/zulip,joshisa/zulip,synicalsyntax/zulip,hafeez3000/zulip,jessedhillon/zulip,rht/zulip,Gabriel0402/zulip,punchagan/zulip,dattatreya303/zulip,udxxabp/zulip,zulip/zulip,schatt/zulip,vikas-parashar/zulip,kou/zulip,amallia/zulip,tommyip/zulip,isht3/zulip,thomasboyt/zulip,aps-sids/zulip,AZtheAsian/zulip,ryansnowboarder/zulip,bluesea/zulip,brainwane/zulip,huangkebo/zulip,hengqujushi/zulip,kaiyuanheshang/zulip,jackrzhang/zulip,itnihao/zulip,rishig/zulip,souravbadami/zulip,vabs22/zulip,ahmadassaf/zulip,bastianh/zulip,susansls/zulip,peguin40/zulip,Qgap/zulip,udxxabp/zulip,bluesea/zulip,MariaFaBella85/zulip,easyfmxu/zulip,gigawhitlocks/zulip,amanharitsh123/zulip,Juanvulcano/zulip,KingxBanana/zulip,hackerkid/zulip,lfranchi/zulip,voidException/zulip,ericzhou2008/zulip,atomic-labs/zulip,Qgap/zulip,MayB/zulip,armooo/zulip,krtkmj/zulip,sonali0901/zulip,sup95/zulip,bitemyapp/zulip,dawran6/zulip,Batterfii/zulip,dotcool/zulip,yocome/zulip,umkay/zulip,brainwane/zulip,akuseru/zulip,wdaher/zulip,arpitpanwar/zulip,gkotian/zulip,jrowan/zulip,brainwane/zulip,ashwinirudrappa/zulip,j831/zulip,littledogboy/zulip,aakash-cr7/zulip,fw1121/zulip,Diptanshu8/zulip,littledogboy/zulip,cosmicAsymmetry/zulip,vabs22/zulip,susansls/zulip,ipernet/zulip,jimmy54/zulip,nicholasbs/zulip,itnihao/zulip,KJin99/zulip,mdavid/zulip,isht3/zulip,ryansnowboarder/zulip,hustlzp/zulip,ikasumiwt/zulip,amanharitsh123/zulip,JPJPJPOPOP/zulip,brainwane/zulip,Juanvulcano/zulip,reyha/zulip,noroot/zulip,saitodisse/zulip,DazWorrall/zulip,kokoar/zulip,mdavid/zulip,LeeRisk/zulip,swinghu/zulip,tbutter/zulip,ashwinirudrappa/zulip,bastianh/zulip,ryanbackman/zulip,Qgap/zulip,MayB/zulip,jeffcao/zulip,developerfm/zulip,shubhamdhama/zulip,Cheppers/zulip,Batterfii/zulip,he15his/zulip,joshisa/zulip,deer-hope/zulip,jrowan/zulip,hackerkid/zulip,sup95/zulip,firstblade/zulip,zachallaun/zulip,timabbott/zulip,j831/zulip,vikas-parashar/zulip,mansilladev/zulip,vakila/zulip,atomic-labs/zulip,rht/zulip,brockwhittaker/zulip,mdavid/zulip,peguin40/zulip,eastlhu/zulip,samatdav/zulip,kou/zulip,easyfmxu/zulip,ufosky-server/zulip,joyhchen/zulip,hengqujushi/zulip,praveenaki/zulip,xuxiao/zulip,Frouk/zulip,bowlofstew/zulip,paxapy/zulip,zhaoweigg/zulip,amyliu345/zulip,themass/zulip,glovebx/zulip,samatdav/zulip,suxinde2009/zulip,kokoar/zulip,jimmy54/zulip,deer-hope/zulip,mahim97/zulip,grave-w-grave/zulip,amallia/zulip,mansilladev/zulip,so0k/zulip,JPJPJPOPOP/zulip,xuxiao/zulip,alliejones/zulip,PhilSk/zulip,LeeRisk/zulip,rishig/zulip,praveenaki/zulip,avastu/zulip,j831/zulip,suxinde2009/zulip,amanharitsh123/zulip,ipernet/zulip,ufosky-server/zulip,glovebx/zulip,aliceriot/zulip,LAndreas/zulip,timabbott/zulip,jeffcao/zulip,hackerkid/zulip,isht3/zulip,joyhchen/zulip,synicalsyntax/zulip,noroot/zulip,saitodisse/zulip,ikasumiwt/zulip,peguin40/zulip,Cheppers/zulip,bssrdf/zulip,eeshangarg/zulip,bitemyapp/zulip,amallia/zulip,tiansiyuan/zulip,xuanhan863/zulip,ahmadassaf/zulip,mansilladev/zulip,SmartPeople/zulip,natanovia/zulip,peguin40/zulip,Drooids/zulip,nicholasbs/zulip,PaulPetring/zulip,dawran6/zulip,zhaoweigg/zulip,punchagan/zulip,luyifan/zulip,codeKonami/zulip,yocome/zulip,qq1012803704/zulip,eastlhu/zulip,zorojean/zulip,Galexrt/zulip,brockwhittaker/zulip,KingxBanana/zulip,xuxiao/zulip,wweiradio/zulip,synicalsyntax/zulip,christi3k/zulip,mahim97/zulip,ryansnowboarder/zulip,PhilSk/zulip,tdr130/zulip,so0k/zulip,jerryge/zulip,dattatreya303/zulip,umkay/zulip,jonesgithub/zulip,Jianchun1/zulip,synicalsyntax/zulip,sonali0901/zulip,Frouk/zulip,kaiyuanheshang/zulip,ryanbackman/zulip,SmartPeople/zulip,mohsenSy/zulip,ryansnowboarder/zulip,souravbadami/zulip,themass/zulip,amanharitsh123/zulip,synicalsyntax/zulip,SmartPeople/zulip,hustlzp/zulip,praveenaki/zulip,bssrdf/zulip,armooo/zulip,zachallaun/zulip,synicalsyntax/zulip,dxq-git/zulip,bowlofstew/zulip,LAndreas/zulip,joshisa/zulip,xuanhan863/zulip,itnihao/zulip,ufosky-server/zulip,littledogboy/zulip,firstblade/zulip,hackerkid/zulip,armooo/zulip,amyliu345/zulip,calvinleenyc/zulip,andersk/zulip,atomic-labs/zulip,rht/zulip,alliejones/zulip,MariaFaBella85/zulip,praveenaki/zulip,sup95/zulip,AZtheAsian/zulip,MariaFaBella85/zulip,zacps/zulip,bluesea/zulip,Cheppers/zulip,verma-varsha/zulip,eastlhu/zulip,zacps/zulip,avastu/zulip,udxxabp/zulip,AZtheAsian/zulip,bitemyapp/zulip,ericzhou2008/zulip,zacps/zulip,technicalpickles/zulip,luyifan/zulip,zulip/zulip,Gabriel0402/zulip,themass/zulip,synicalsyntax/zulip,showell/zulip,avastu/zulip,PaulPetring/zulip,gigawhitlocks/zulip,Qgap/zulip,levixie/zulip,susansls/zulip,m1ssou/zulip,ApsOps/zulip,LeeRisk/zulip,umkay/zulip,easyfmxu/zulip,babbage/zulip,bssrdf/zulip,peguin40/zulip,yuvipanda/zulip,technicalpickles/zulip,technicalpickles/zulip,showell/zulip,blaze225/zulip,shrikrishnaholla/zulip,Gabriel0402/zulip,moria/zulip,paxapy/zulip,reyha/zulip,pradiptad/zulip,rht/zulip,ipernet/zulip,cosmicAsymmetry/zulip,jeffcao/zulip,shrikrishnaholla/zulip,firstblade/zulip,saitodisse/zulip,itnihao/zulip,fw1121/zulip,verma-varsha/zulip,guiquanz/zulip,timabbott/zulip,hj3938/zulip,hayderimran7/zulip,shrikrishnaholla/zulip,wavelets/zulip,bastianh/zulip,kou/zulip,Suninus/zulip,brockwhittaker/zulip,eeshangarg/zulip,reyha/zulip,wweiradio/zulip,zwily/zulip,huangkebo/zulip,Galexrt/zulip,Frouk/zulip,Gabriel0402/zulip,udxxabp/zulip,showell/zulip,qq1012803704/zulip,mahim97/zulip,hustlzp/zulip,bssrdf/zulip,ryansnowboarder/zulip,he15his/zulip,wangdeshui/zulip,schatt/zulip,jonesgithub/zulip,Galexrt/zulip,natanovia/zulip,levixie/zulip,LeeRisk/zulip,punchagan/zulip,DazWorrall/zulip,KJin99/zulip,andersk/zulip,willingc/zulip,samatdav/zulip,glovebx/zulip,AZtheAsian/zulip,jimmy54/zulip,pradiptad/zulip,blaze225/zulip,jrowan/zulip,tommyip/zulip,jonesgithub/zulip,wangdeshui/zulip,umkay/zulip,qq1012803704/zulip,Galexrt/zulip,zofuthan/zulip,dxq-git/zulip,DazWorrall/zulip,AZtheAsian/zulip,yocome/zulip,peiwei/zulip,zofuthan/zulip,proliming/zulip,johnnygaddarr/zulip,proliming/zulip,jessedhillon/zulip,timabbott/zulip,swinghu/zulip,MariaFaBella85/zulip,dattatreya303/zulip,shubhamdhama/zulip,PaulPetring/zulip,tbutter/zulip,so0k/zulip,jackrzhang/zulip,timabbott/zulip,Batterfii/zulip,jessedhillon/zulip,mahim97/zulip,seapasulli/zulip,calvinleenyc/zulip,joyhchen/zulip,firstblade/zulip,PaulPetring/zulip,grave-w-grave/zulip,he15his/zulip,zulip/zulip,grave-w-grave/zulip,mdavid/zulip,akuseru/zulip,johnny9/zulip,bastianh/zulip,schatt/zulip,stamhe/zulip,paxapy/zulip,deer-hope/zulip,alliejones/zulip,bssrdf/zulip,vaidap/zulip,ApsOps/zulip,krtkmj/zulip,PhilSk/zulip,codeKonami/zulip,calvinleenyc/zulip,isht3/zulip,developerfm/zulip,yocome/zulip,mohsenSy/zulip,dhcrzf/zulip,gigawhitlocks/zulip,jainayush975/zulip,hafeez3000/zulip,saitodisse/zulip,xuxiao/zulip,wdaher/zulip,themass/zulip,dattatreya303/zulip,tommyip/zulip,swinghu/zulip,bluesea/zulip,ipernet/zulip,arpith/zulip,JanzTam/zulip,zorojean/zulip,luyifan/zulip,LAndreas/zulip,ryanbackman/zulip,alliejones/zulip,avastu/zulip,xuanhan863/zulip,dawran6/zulip,j831/zulip,vikas-parashar/zulip,armooo/zulip,fw1121/zulip,dnmfarrell/zulip,gkotian/zulip,stamhe/zulip,JanzTam/zulip,sonali0901/zulip,fw1121/zulip,punchagan/zulip,swinghu/zulip,LAndreas/zulip,seapasulli/zulip,he15his/zulip,willingc/zulip,kou/zulip,RobotCaleb/zulip,jackrzhang/zulip,akuseru/zulip,stamhe/zulip,dxq-git/zulip,adnanh/zulip,avastu/zulip,wdaher/zulip,deer-hope/zulip,ApsOps/zulip,thomasboyt/zulip,bowlofstew/zulip,j831/zulip,DazWorrall/zulip,tdr130/zulip,vakila/zulip,sharmaeklavya2/zulip,littledogboy/zulip,wangdeshui/zulip,littledogboy/zulip,levixie/zulip,proliming/zulip,vakila/zulip,dxq-git/zulip,LeeRisk/zulip,pradiptad/zulip,cosmicAsymmetry/zulip,kokoar/zulip,nicholasbs/zulip,tiansiyuan/zulip,dawran6/zulip,itnihao/zulip,PaulPetring/zulip,JPJPJPOPOP/zulip,babbage/zulip,ApsOps/zulip,Juanvulcano/zulip,babbage/zulip,easyfmxu/zulip,zhaoweigg/zulip,souravbadami/zulip,johnnygaddarr/zulip,shaunstanislaus/zulip,proliming/zulip,dwrpayne/zulip,bitemyapp/zulip,dxq-git/zulip,aakash-cr7/zulip,jrowan/zulip,zorojean/zulip,krtkmj/zulip,yuvipanda/zulip,jphilipsen05/zulip,seapasulli/zulip,zwily/zulip,lfranchi/zulip,ahmadassaf/zulip,vikas-parashar/zulip,showell/zulip,dhcrzf/zulip,zwily/zulip,Qgap/zulip,KingxBanana/zulip,atomic-labs/zulip,gkotian/zulip,Batterfii/zulip,souravbadami/zulip,dnmfarrell/zulip,codeKonami/zulip,mansilladev/zulip,cosmicAsymmetry/zulip,wweiradio/zulip,Vallher/zulip,jrowan/zulip,ufosky-server/zulip,gigawhitlocks/zulip,zulip/zulip,dattatreya303/zulip,dnmfarrell/zulip,amanharitsh123/zulip,verma-varsha/zulip,DazWorrall/zulip,eeshangarg/zulip,JPJPJPOPOP/zulip,kou/zulip,dwrpayne/zulip,karamcnair/zulip,KingxBanana/zulip,dotcool/zulip,qq1012803704/zulip,TigorC/zulip,MayB/zulip,mohsenSy/zulip,stamhe/zulip,udxxabp/zulip,lfranchi/zulip,arpith/zulip,hengqujushi/zulip,dwrpayne/zulip,reyha/zulip,shubhamdhama/zulip,hj3938/zulip,mansilladev/zulip,ryansnowboarder/zulip,reyha/zulip,wangdeshui/zulip,bowlofstew/zulip,niftynei/zulip,shaunstanislaus/zulip,kokoar/zulip,vabs22/zulip,hayderimran7/zulip,umkay/zulip,saitodisse/zulip,ryansnowboarder/zulip,wavelets/zulip,tdr130/zulip,hustlzp/zulip,calvinleenyc/zulip,RobotCaleb/zulip,zwily/zulip,atomic-labs/zulip,itnihao/zulip,tiansiyuan/zulip,johnnygaddarr/zulip,xuanhan863/zulip,Vallher/zulip,amyliu345/zulip,dhcrzf/zulip,dnmfarrell/zulip,ipernet/zulip,tbutter/zulip,wangdeshui/zulip,thomasboyt/zulip,jerryge/zulip,jerryge/zulip,wavelets/zulip,Qgap/zulip,amallia/zulip,EasonYi/zulip,esander91/zulip,JPJPJPOPOP/zulip,deer-hope/zulip,ikasumiwt/zulip,Frouk/zulip,babbage/zulip,KingxBanana/zulip,stamhe/zulip,zachallaun/zulip,shaunstanislaus/zulip,christi3k/zulip,jphilipsen05/zulip,Jianchun1/zulip,qq1012803704/zulip,codeKonami/zulip,zhaoweigg/zulip,m1ssou/zulip,EasonYi/zulip,arpitpanwar/zulip,aliceriot/zulip,swinghu/zulip,kaiyuanheshang/zulip,sup95/zulip,suxinde2009/zulip,xuxiao/zulip,zorojean/zulip,moria/zulip,bssrdf/zulip,seapasulli/zulip,rishig/zulip,jphilipsen05/zulip,ashwinirudrappa/zulip,EasonYi/zulip,bluesea/zulip,dhcrzf/zulip,udxxabp/zulip,tiansiyuan/zulip,wweiradio/zulip,Suninus/zulip,tommyip/zulip,voidException/zulip,dawran6/zulip,hafeez3000/zulip,yuvipanda/zulip,blaze225/zulip,karamcnair/zulip,natanovia/zulip,vakila/zulip,willingc/zulip,dwrpayne/zulip,verma-varsha/zulip,noroot/zulip,ashwinirudrappa/zulip,themass/zulip,themass/zulip,jrowan/zulip,yuvipanda/zulip,jphilipsen05/zulip,jimmy54/zulip,jonesgithub/zulip,dotcool/zulip,andersk/zulip,Frouk/zulip,eastlhu/zulip,vikas-parashar/zulip,xuxiao/zulip,Qgap/zulip,xuxiao/zulip,punchagan/zulip,zachallaun/zulip,aps-sids/zulip,hengqujushi/zulip,PaulPetring/zulip,luyifan/zulip,nicholasbs/zulip,krtkmj/zulip,brainwane/zulip,vabs22/zulip,showell/zulip,niftynei/zulip,praveenaki/zulip,proliming/zulip,karamcnair/zulip,so0k/zulip,guiquanz/zulip,gigawhitlocks/zulip,Jianchun1/zulip,isht3/zulip,jimmy54/zulip,seapasulli/zulip,avastu/zulip,rht/zulip,zachallaun/zulip,suxinde2009/zulip,seapasulli/zulip,suxinde2009/zulip,arpitpanwar/zulip,arpitpanwar/zulip,aakash-cr7/zulip,schatt/zulip,joshisa/zulip,johnny9/zulip,Suninus/zulip,zwily/zulip,hafeez3000/zulip,ApsOps/zulip,ahmadassaf/zulip,vaidap/zulip,jerryge/zulip,technicalpickles/zulip,guiquanz/zulip,peiwei/zulip,joshisa/zulip,nicholasbs/zulip,punchagan/zulip,arpitpanwar/zulip,hackerkid/zulip,jeffcao/zulip,pradiptad/zulip,Drooids/zulip,ufosky-server/zulip,littledogboy/zulip,tdr130/zulip,thomasboyt/zulip,vakila/zulip,ikasumiwt/zulip,christi3k/zulip,ufosky-server/zulip,kokoar/zulip,ericzhou2008/zulip,johnny9/zulip,amanharitsh123/zulip,easyfmxu/zulip,amyliu345/zulip,zofuthan/zulip,m1ssou/zulip,codeKonami/zulip,sup95/zulip,ashwinirudrappa/zulip,kaiyuanheshang/zulip,TigorC/zulip,aps-sids/zulip,krtkmj/zulip,huangkebo/zulip,Drooids/zulip,voidException/zulip,jonesgithub/zulip,zacps/zulip,PaulPetring/zulip,he15his/zulip,pradiptad/zulip,fw1121/zulip,vaidap/zulip,Suninus/zulip,developerfm/zulip,ericzhou2008/zulip,armooo/zulip,huangkebo/zulip,developerfm/zulip,johnnygaddarr/zulip,gigawhitlocks/zulip,levixie/zulip,cosmicAsymmetry/zulip,tbutter/zulip,zulip/zulip,peiwei/zulip,aakash-cr7/zulip,jeffcao/zulip,wweiradio/zulip,JanzTam/zulip,zorojean/zulip,kou/zulip,vabs22/zulip,akuseru/zulip,shaunstanislaus/zulip,dxq-git/zulip,LeeRisk/zulip,amallia/zulip,zhaoweigg/zulip,xuanhan863/zulip,shubhamdhama/zulip,johnnygaddarr/zulip,niftynei/zulip,esander91/zulip,jackrzhang/zulip,amyliu345/zulip,yuvipanda/zulip,dhcrzf/zulip,niftynei/zulip,paxapy/zulip,eastlhu/zulip,peiwei/zulip,willingc/zulip,SmartPeople/zulip,hackerkid/zulip,aps-sids/zulip,huangkebo/zulip,moria/zulip,Gabriel0402/zulip,kou/zulip,JanzTam/zulip,gkotian/zulip,ericzhou2008/zulip,glovebx/zulip,jainayush975/zulip,Suninus/zulip,suxinde2009/zulip,grave-w-grave/zulip,christi3k/zulip,tommyip/zulip,seapasulli/zulip,LAndreas/zulip,tiansiyuan/zulip,stamhe/zulip,brockwhittaker/zulip,amallia/zulip,noroot/zulip,souravbadami/zulip,m1ssou/zulip,cosmicAsymmetry/zulip,jeffcao/zulip,babbage/zulip,christi3k/zulip,Galexrt/zulip,shaunstanislaus/zulip,vaidap/zulip,Vallher/zulip,karamcnair/zulip,samatdav/zulip,rishig/zulip,ikasumiwt/zulip,Vallher/zulip,jphilipsen05/zulip,joyhchen/zulip,hayderimran7/zulip,esander91/zulip,Jianchun1/zulip,saitodisse/zulip,schatt/zulip,karamcnair/zulip,bastianh/zulip,voidException/zulip,suxinde2009/zulip,Gabriel0402/zulip,xuanhan863/zulip,dhcrzf/zulip,mohsenSy/zulip,JanzTam/zulip,calvinleenyc/zulip,showell/zulip,technicalpickles/zulip,zulip/zulip,wavelets/zulip,firstblade/zulip,RobotCaleb/zulip,kaiyuanheshang/zulip,luyifan/zulip,Drooids/zulip,Juanvulcano/zulip,lfranchi/zulip,PhilSk/zulip,he15his/zulip,DazWorrall/zulip,huangkebo/zulip,peiwei/zulip,wavelets/zulip,jphilipsen05/zulip,Galexrt/zulip,aakash-cr7/zulip,aakash-cr7/zulip,amallia/zulip,armooo/zulip,alliejones/zulip,brockwhittaker/zulip,ericzhou2008/zulip,bluesea/zulip,umkay/zulip,schatt/zulip,hayderimran7/zulip,adnanh/zulip,arpitpanwar/zulip,saitodisse/zulip,andersk/zulip,ipernet/zulip,aps-sids/zulip,mohsenSy/zulip,blaze225/zulip,zhaoweigg/zulip,arpith/zulip,andersk/zulip,joshisa/zulip,arpith/zulip,dnmfarrell/zulip,paxapy/zulip,mahim97/zulip,wangdeshui/zulip,ahmadassaf/zulip,levixie/zulip,esander91/zulip,kaiyuanheshang/zulip,ashwinirudrappa/zulip,susansls/zulip,christi3k/zulip,nicholasbs/zulip,andersk/zulip,arpitpanwar/zulip,natanovia/zulip,jackrzhang/zulip,showell/zulip,jerryge/zulip,eeshangarg/zulip,aliceriot/zulip,j831/zulip,RobotCaleb/zulip,ryanbackman/zulip,thomasboyt/zulip,jessedhillon/zulip,shubhamdhama/zulip,ApsOps/zulip,RobotCaleb/zulip,PhilSk/zulip,eastlhu/zulip,amyliu345/zulip,lfranchi/zulip,zofuthan/zulip,yuvipanda/zulip,niftynei/zulip,firstblade/zulip,xuanhan863/zulip,kokoar/zulip,gigawhitlocks/zulip,shrikrishnaholla/zulip,willingc/zulip,guiquanz/zulip,hafeez3000/zulip,gkotian/zulip,adnanh/zulip,huangkebo/zulip,umkay/zulip,EasonYi/zulip,sharmaeklavya2/zulip,kaiyuanheshang/zulip,hj3938/zulip,hafeez3000/zulip,wdaher/zulip,moria/zulip,MayB/zulip,KJin99/zulip,SmartPeople/zulip,zwily/zulip,praveenaki/zulip,tiansiyuan/zulip,bitemyapp/zulip,hengqujushi/zulip,technicalpickles/zulip,hustlzp/zulip,jainayush975/zulip,Batterfii/zulip,tdr130/zulip,sup95/zulip,Diptanshu8/zulip,babbage/zulip,akuseru/zulip,zorojean/zulip,proliming/zulip,bastianh/zulip,johnny9/zulip,Cheppers/zulip,Vallher/zulip,voidException/zulip,ApsOps/zulip,shrikrishnaholla/zulip,voidException/zulip,mdavid/zulip,verma-varsha/zulip,brainwane/zulip,krtkmj/zulip,vabs22/zulip,Batterfii/zulip,shrikrishnaholla/zulip,wavelets/zulip,susansls/zulip,natanovia/zulip,eeshangarg/zulip,willingc/zulip,esander91/zulip,wweiradio/zulip,shubhamdhama/zulip,zorojean/zulip,Jianchun1/zulip,tiansiyuan/zulip,littledogboy/zulip,dnmfarrell/zulip,tbutter/zulip,fw1121/zulip,TigorC/zulip,reyha/zulip,m1ssou/zulip,qq1012803704/zulip,stamhe/zulip,armooo/zulip,sharmaeklavya2/zulip,KJin99/zulip,joshisa/zulip,souravbadami/zulip,m1ssou/zulip,jessedhillon/zulip,KJin99/zulip,mdavid/zulip,Batterfii/zulip,babbage/zulip,Gabriel0402/zulip,bluesea/zulip,tdr130/zulip,luyifan/zulip,alliejones/zulip,DazWorrall/zulip,sharmaeklavya2/zulip,nicholasbs/zulip,kokoar/zulip,Vallher/zulip,TigorC/zulip,punchagan/zulip,dotcool/zulip,MariaFaBella85/zulip,swinghu/zulip,guiquanz/zulip,yuvipanda/zulip,krtkmj/zulip,esander91/zulip,karamcnair/zulip,zofuthan/zulip,ufosky-server/zulip,dotcool/zulip,JPJPJPOPOP/zulip,hengqujushi/zulip,tommyip/zulip,johnny9/zulip,ryanbackman/zulip,vaidap/zulip,thomasboyt/zulip,jerryge/zulip,hj3938/zulip,Suninus/zulip,yocome/zulip,noroot/zulip,fw1121/zulip,JanzTam/zulip,verma-varsha/zulip,thomasboyt/zulip,itnihao/zulip,atomic-labs/zulip,adnanh/zulip,dxq-git/zulip,johnnygaddarr/zulip,proliming/zulip,aliceriot/zulip,brainwane/zulip,paxapy/zulip,EasonYi/zulip,esander91/zulip,so0k/zulip,peiwei/zulip,dwrpayne/zulip,willingc/zulip,Cheppers/zulip,easyfmxu/zulip,bastianh/zulip,joyhchen/zulip,so0k/zulip,adnanh/zulip,codeKonami/zulip,codeKonami/zulip,tommyip/zulip,he15his/zulip,rht/zulip,deer-hope/zulip,wangdeshui/zulip,timabbott/zulip,jeffcao/zulip,ericzhou2008/zulip,ikasumiwt/zulip,ipernet/zulip,ahmadassaf/zulip,Diptanshu8/zulip,moria/zulip,akuseru/zulip,Juanvulcano/zulip,hackerkid/zulip,themass/zulip,dattatreya303/zulip,noroot/zulip,jessedhillon/zulip,hustlzp/zulip,Drooids/zulip,arpith/zulip,luyifan/zulip,alliejones/zulip,wdaher/zulip,hafeez3000/zulip,Drooids/zulip,glovebx/zulip,susansls/zulip,yocome/zulip,wdaher/zulip,PhilSk/zulip,Frouk/zulip,vakila/zulip,mansilladev/zulip,bowlofstew/zulip,MayB/zulip,pradiptad/zulip,eeshangarg/zulip,rishig/zulip,voidException/zulip,gkotian/zulip,rishig/zulip,shaunstanislaus/zulip,glovebx/zulip,EasonYi/zulip,jackrzhang/zulip,tbutter/zulip,Diptanshu8/zulip,swinghu/zulip,dotcool/zulip,sharmaeklavya2/zulip,ashwinirudrappa/zulip,johnny9/zulip,KJin99/zulip,zacps/zulip,samatdav/zulip,dwrpayne/zulip,Galexrt/zulip,eeshangarg/zulip,peguin40/zulip,Diptanshu8/zulip,jonesgithub/zulip,timabbott/zulip,joyhchen/zulip,LAndreas/zulip,jonesgithub/zulip,Cheppers/zulip,developerfm/zulip,karamcnair/zulip,mahim97/zulip,EasonYi/zulip,easyfmxu/zulip,firstblade/zulip,hayderimran7/zulip,hj3938/zulip,aliceriot/zulip,ahmadassaf/zulip,arpith/zulip,peiwei/zulip,zachallaun/zulip,mdavid/zulip,vakila/zulip,niftynei/zulip,vaidap/zulip,aliceriot/zulip,Diptanshu8/zulip,dotcool/zulip,Cheppers/zulip,levixie/zulip,zofuthan/zulip,TigorC/zulip,grave-w-grave/zulip,isht3/zulip,wavelets/zulip,johnny9/zulip,zacps/zulip,tdr130/zulip,aliceriot/zulip,noroot/zulip,MayB/zulip,levixie/zulip,jainayush975/zulip,rht/zulip,hayderimran7/zulip,Frouk/zulip,gkotian/zulip,Suninus/zulip,TigorC/zulip,bitemyapp/zulip,jainayush975/zulip | # Humbug Inc's internal git plugin configuration.
# The plugin and example config are under api/integrations/
# Leaving all the instructions out of this file to avoid having to
# sync them as we update the comments.
HUMBUG_USER = "humbug+commits@humbughq.com"
HUMBUG_API_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# commit_notice_destination() lets you customize where commit notices
# are sent to.
#
# It takes the following arguments:
# * repo = the name of the git repository
# * branch = the name of the branch that was pushed to
# * commit = the commit id
#
# Returns a dictionary encoding the stream and subject to send the
# notification to (or None to send no notification, e.g. for ).
#
# The default code below will send every commit pushed to "master" to
# * stream "commits"
# * subject "deploy => master" (using a pretty unicode right arrow)
# And similarly for branch "test-post-receive" (for use when testing).
def commit_notice_destination(repo, branch, commit):
if branch in ["master", "post-receive-test"]:
return dict(stream = "commits",
subject = u"deploy \u21D2 %s" % (branch,))
# Return None for cases where you don't want a notice sent
return None
HUMBUG_API_PATH = "/home/humbug/humbug/api"
HUMBUG_SITE = "https://staging.humbughq.com"
git: Put prod back on the list of branches to send notices about.
(imported from commit e608d7050b4e68045b03341dc41e8654e45a3af3) | # Humbug Inc's internal git plugin configuration.
# The plugin and example config are under api/integrations/
# Leaving all the instructions out of this file to avoid having to
# sync them as we update the comments.
HUMBUG_USER = "humbug+commits@humbughq.com"
HUMBUG_API_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# commit_notice_destination() lets you customize where commit notices
# are sent to.
#
# It takes the following arguments:
# * repo = the name of the git repository
# * branch = the name of the branch that was pushed to
# * commit = the commit id
#
# Returns a dictionary encoding the stream and subject to send the
# notification to (or None to send no notification, e.g. for ).
#
# The default code below will send every commit pushed to "master" to
# * stream "commits"
# * subject "deploy => master" (using a pretty unicode right arrow)
# And similarly for branch "test-post-receive" (for use when testing).
def commit_notice_destination(repo, branch, commit):
if branch in ["master", "prod", "post-receive-test"]:
return dict(stream = "commits",
subject = u"deploy \u21D2 %s" % (branch,))
# Return None for cases where you don't want a notice sent
return None
HUMBUG_API_PATH = "/home/humbug/humbug/api"
HUMBUG_SITE = "https://staging.humbughq.com"
| <commit_before># Humbug Inc's internal git plugin configuration.
# The plugin and example config are under api/integrations/
# Leaving all the instructions out of this file to avoid having to
# sync them as we update the comments.
HUMBUG_USER = "humbug+commits@humbughq.com"
HUMBUG_API_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# commit_notice_destination() lets you customize where commit notices
# are sent to.
#
# It takes the following arguments:
# * repo = the name of the git repository
# * branch = the name of the branch that was pushed to
# * commit = the commit id
#
# Returns a dictionary encoding the stream and subject to send the
# notification to (or None to send no notification, e.g. for ).
#
# The default code below will send every commit pushed to "master" to
# * stream "commits"
# * subject "deploy => master" (using a pretty unicode right arrow)
# And similarly for branch "test-post-receive" (for use when testing).
def commit_notice_destination(repo, branch, commit):
if branch in ["master", "post-receive-test"]:
return dict(stream = "commits",
subject = u"deploy \u21D2 %s" % (branch,))
# Return None for cases where you don't want a notice sent
return None
HUMBUG_API_PATH = "/home/humbug/humbug/api"
HUMBUG_SITE = "https://staging.humbughq.com"
<commit_msg>git: Put prod back on the list of branches to send notices about.
(imported from commit e608d7050b4e68045b03341dc41e8654e45a3af3)<commit_after> | # Humbug Inc's internal git plugin configuration.
# The plugin and example config are under api/integrations/
# Leaving all the instructions out of this file to avoid having to
# sync them as we update the comments.
HUMBUG_USER = "humbug+commits@humbughq.com"
HUMBUG_API_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# commit_notice_destination() lets you customize where commit notices
# are sent to.
#
# It takes the following arguments:
# * repo = the name of the git repository
# * branch = the name of the branch that was pushed to
# * commit = the commit id
#
# Returns a dictionary encoding the stream and subject to send the
# notification to (or None to send no notification, e.g. for ).
#
# The default code below will send every commit pushed to "master" to
# * stream "commits"
# * subject "deploy => master" (using a pretty unicode right arrow)
# And similarly for branch "test-post-receive" (for use when testing).
def commit_notice_destination(repo, branch, commit):
if branch in ["master", "prod", "post-receive-test"]:
return dict(stream = "commits",
subject = u"deploy \u21D2 %s" % (branch,))
# Return None for cases where you don't want a notice sent
return None
HUMBUG_API_PATH = "/home/humbug/humbug/api"
HUMBUG_SITE = "https://staging.humbughq.com"
| # Humbug Inc's internal git plugin configuration.
# The plugin and example config are under api/integrations/
# Leaving all the instructions out of this file to avoid having to
# sync them as we update the comments.
HUMBUG_USER = "humbug+commits@humbughq.com"
HUMBUG_API_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# commit_notice_destination() lets you customize where commit notices
# are sent to.
#
# It takes the following arguments:
# * repo = the name of the git repository
# * branch = the name of the branch that was pushed to
# * commit = the commit id
#
# Returns a dictionary encoding the stream and subject to send the
# notification to (or None to send no notification, e.g. for ).
#
# The default code below will send every commit pushed to "master" to
# * stream "commits"
# * subject "deploy => master" (using a pretty unicode right arrow)
# And similarly for branch "test-post-receive" (for use when testing).
def commit_notice_destination(repo, branch, commit):
if branch in ["master", "post-receive-test"]:
return dict(stream = "commits",
subject = u"deploy \u21D2 %s" % (branch,))
# Return None for cases where you don't want a notice sent
return None
HUMBUG_API_PATH = "/home/humbug/humbug/api"
HUMBUG_SITE = "https://staging.humbughq.com"
git: Put prod back on the list of branches to send notices about.
(imported from commit e608d7050b4e68045b03341dc41e8654e45a3af3)# Humbug Inc's internal git plugin configuration.
# The plugin and example config are under api/integrations/
# Leaving all the instructions out of this file to avoid having to
# sync them as we update the comments.
HUMBUG_USER = "humbug+commits@humbughq.com"
HUMBUG_API_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# commit_notice_destination() lets you customize where commit notices
# are sent to.
#
# It takes the following arguments:
# * repo = the name of the git repository
# * branch = the name of the branch that was pushed to
# * commit = the commit id
#
# Returns a dictionary encoding the stream and subject to send the
# notification to (or None to send no notification, e.g. for ).
#
# The default code below will send every commit pushed to "master" to
# * stream "commits"
# * subject "deploy => master" (using a pretty unicode right arrow)
# And similarly for branch "test-post-receive" (for use when testing).
def commit_notice_destination(repo, branch, commit):
if branch in ["master", "prod", "post-receive-test"]:
return dict(stream = "commits",
subject = u"deploy \u21D2 %s" % (branch,))
# Return None for cases where you don't want a notice sent
return None
HUMBUG_API_PATH = "/home/humbug/humbug/api"
HUMBUG_SITE = "https://staging.humbughq.com"
| <commit_before># Humbug Inc's internal git plugin configuration.
# The plugin and example config are under api/integrations/
# Leaving all the instructions out of this file to avoid having to
# sync them as we update the comments.
HUMBUG_USER = "humbug+commits@humbughq.com"
HUMBUG_API_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# commit_notice_destination() lets you customize where commit notices
# are sent to.
#
# It takes the following arguments:
# * repo = the name of the git repository
# * branch = the name of the branch that was pushed to
# * commit = the commit id
#
# Returns a dictionary encoding the stream and subject to send the
# notification to (or None to send no notification, e.g. for ).
#
# The default code below will send every commit pushed to "master" to
# * stream "commits"
# * subject "deploy => master" (using a pretty unicode right arrow)
# And similarly for branch "test-post-receive" (for use when testing).
def commit_notice_destination(repo, branch, commit):
if branch in ["master", "post-receive-test"]:
return dict(stream = "commits",
subject = u"deploy \u21D2 %s" % (branch,))
# Return None for cases where you don't want a notice sent
return None
HUMBUG_API_PATH = "/home/humbug/humbug/api"
HUMBUG_SITE = "https://staging.humbughq.com"
<commit_msg>git: Put prod back on the list of branches to send notices about.
(imported from commit e608d7050b4e68045b03341dc41e8654e45a3af3)<commit_after># Humbug Inc's internal git plugin configuration.
# The plugin and example config are under api/integrations/
# Leaving all the instructions out of this file to avoid having to
# sync them as we update the comments.
HUMBUG_USER = "humbug+commits@humbughq.com"
HUMBUG_API_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
# commit_notice_destination() lets you customize where commit notices
# are sent to.
#
# It takes the following arguments:
# * repo = the name of the git repository
# * branch = the name of the branch that was pushed to
# * commit = the commit id
#
# Returns a dictionary encoding the stream and subject to send the
# notification to (or None to send no notification, e.g. for ).
#
# The default code below will send every commit pushed to "master" to
# * stream "commits"
# * subject "deploy => master" (using a pretty unicode right arrow)
# And similarly for branch "test-post-receive" (for use when testing).
def commit_notice_destination(repo, branch, commit):
if branch in ["master", "prod", "post-receive-test"]:
return dict(stream = "commits",
subject = u"deploy \u21D2 %s" % (branch,))
# Return None for cases where you don't want a notice sent
return None
HUMBUG_API_PATH = "/home/humbug/humbug/api"
HUMBUG_SITE = "https://staging.humbughq.com"
|
e1c58062db9c107c358e3617793aaed7cdb3a133 | jobmon/util.py | jobmon/util.py | import os
import threading
class TerminableThreadMixin:
"""
TerminableThreadMixin is useful for threads that need to be terminated
from the outside. It provides a method called 'terminate', which communicates
to the thread that it needs to die, and then waits for the death to occur.
It imposes the following rules:
1. Call it's .__init__ inside of your __init__
2. Use the .reader inside of the thread - when it has data written on it,
that is an exit request
3. Call it's .cleanup method before exiting.
"""
def __init__(self):
reader, writer = os.pipe()
self.exit_reader = os.fdopen(reader, 'rb')
self.exit_writer = os.fdopen(writer, 'wb')
def cleanup(self):
self.exit_reader.close()
self.exit_writer.close()
def terminate(self):
"""
Asynchronously terminates the thread, without waiting for it to exit.
"""
try:
self.exit_writer.write(b' ')
self.exit_writer.flush()
except ValueError:
pass
def wait_for_exit(self):
"""
Waits for the thread to exit - should be run only after terminating
the thread.
"""
self.join()
| import logging
import os
import threading
def reset_loggers():
"""
Removes all handlers from the current loggers to allow for a new basicConfig.
"""
root = logging.getLogger()
for handler in root.handlers[:]:
root.removeHandler(handler)
class TerminableThreadMixin:
"""
TerminableThreadMixin is useful for threads that need to be terminated
from the outside. It provides a method called 'terminate', which communicates
to the thread that it needs to die, and then waits for the death to occur.
It imposes the following rules:
1. Call it's .__init__ inside of your __init__
2. Use the .reader inside of the thread - when it has data written on it,
that is an exit request
3. Call it's .cleanup method before exiting.
"""
def __init__(self):
reader, writer = os.pipe()
self.exit_reader = os.fdopen(reader, 'rb')
self.exit_writer = os.fdopen(writer, 'wb')
def cleanup(self):
self.exit_reader.close()
self.exit_writer.close()
def terminate(self):
"""
Asynchronously terminates the thread, without waiting for it to exit.
"""
try:
self.exit_writer.write(b' ')
self.exit_writer.flush()
except ValueError:
pass
def wait_for_exit(self):
"""
Waits for the thread to exit - should be run only after terminating
the thread.
"""
self.join()
| Add way to reset loggers | Add way to reset loggers
logging.basicConfig doesn't reset the current logging infrastructure, which
was messing up some tests that expected logs to be one place even though
they ended up in another
| Python | bsd-2-clause | adamnew123456/jobmon | import os
import threading
class TerminableThreadMixin:
"""
TerminableThreadMixin is useful for threads that need to be terminated
from the outside. It provides a method called 'terminate', which communicates
to the thread that it needs to die, and then waits for the death to occur.
It imposes the following rules:
1. Call it's .__init__ inside of your __init__
2. Use the .reader inside of the thread - when it has data written on it,
that is an exit request
3. Call it's .cleanup method before exiting.
"""
def __init__(self):
reader, writer = os.pipe()
self.exit_reader = os.fdopen(reader, 'rb')
self.exit_writer = os.fdopen(writer, 'wb')
def cleanup(self):
self.exit_reader.close()
self.exit_writer.close()
def terminate(self):
"""
Asynchronously terminates the thread, without waiting for it to exit.
"""
try:
self.exit_writer.write(b' ')
self.exit_writer.flush()
except ValueError:
pass
def wait_for_exit(self):
"""
Waits for the thread to exit - should be run only after terminating
the thread.
"""
self.join()
Add way to reset loggers
logging.basicConfig doesn't reset the current logging infrastructure, which
was messing up some tests that expected logs to be one place even though
they ended up in another | import logging
import os
import threading
def reset_loggers():
"""
Removes all handlers from the current loggers to allow for a new basicConfig.
"""
root = logging.getLogger()
for handler in root.handlers[:]:
root.removeHandler(handler)
class TerminableThreadMixin:
"""
TerminableThreadMixin is useful for threads that need to be terminated
from the outside. It provides a method called 'terminate', which communicates
to the thread that it needs to die, and then waits for the death to occur.
It imposes the following rules:
1. Call it's .__init__ inside of your __init__
2. Use the .reader inside of the thread - when it has data written on it,
that is an exit request
3. Call it's .cleanup method before exiting.
"""
def __init__(self):
reader, writer = os.pipe()
self.exit_reader = os.fdopen(reader, 'rb')
self.exit_writer = os.fdopen(writer, 'wb')
def cleanup(self):
self.exit_reader.close()
self.exit_writer.close()
def terminate(self):
"""
Asynchronously terminates the thread, without waiting for it to exit.
"""
try:
self.exit_writer.write(b' ')
self.exit_writer.flush()
except ValueError:
pass
def wait_for_exit(self):
"""
Waits for the thread to exit - should be run only after terminating
the thread.
"""
self.join()
| <commit_before>import os
import threading
class TerminableThreadMixin:
"""
TerminableThreadMixin is useful for threads that need to be terminated
from the outside. It provides a method called 'terminate', which communicates
to the thread that it needs to die, and then waits for the death to occur.
It imposes the following rules:
1. Call it's .__init__ inside of your __init__
2. Use the .reader inside of the thread - when it has data written on it,
that is an exit request
3. Call it's .cleanup method before exiting.
"""
def __init__(self):
reader, writer = os.pipe()
self.exit_reader = os.fdopen(reader, 'rb')
self.exit_writer = os.fdopen(writer, 'wb')
def cleanup(self):
self.exit_reader.close()
self.exit_writer.close()
def terminate(self):
"""
Asynchronously terminates the thread, without waiting for it to exit.
"""
try:
self.exit_writer.write(b' ')
self.exit_writer.flush()
except ValueError:
pass
def wait_for_exit(self):
"""
Waits for the thread to exit - should be run only after terminating
the thread.
"""
self.join()
<commit_msg>Add way to reset loggers
logging.basicConfig doesn't reset the current logging infrastructure, which
was messing up some tests that expected logs to be one place even though
they ended up in another<commit_after> | import logging
import os
import threading
def reset_loggers():
"""
Removes all handlers from the current loggers to allow for a new basicConfig.
"""
root = logging.getLogger()
for handler in root.handlers[:]:
root.removeHandler(handler)
class TerminableThreadMixin:
"""
TerminableThreadMixin is useful for threads that need to be terminated
from the outside. It provides a method called 'terminate', which communicates
to the thread that it needs to die, and then waits for the death to occur.
It imposes the following rules:
1. Call it's .__init__ inside of your __init__
2. Use the .reader inside of the thread - when it has data written on it,
that is an exit request
3. Call it's .cleanup method before exiting.
"""
def __init__(self):
reader, writer = os.pipe()
self.exit_reader = os.fdopen(reader, 'rb')
self.exit_writer = os.fdopen(writer, 'wb')
def cleanup(self):
self.exit_reader.close()
self.exit_writer.close()
def terminate(self):
"""
Asynchronously terminates the thread, without waiting for it to exit.
"""
try:
self.exit_writer.write(b' ')
self.exit_writer.flush()
except ValueError:
pass
def wait_for_exit(self):
"""
Waits for the thread to exit - should be run only after terminating
the thread.
"""
self.join()
| import os
import threading
class TerminableThreadMixin:
"""
TerminableThreadMixin is useful for threads that need to be terminated
from the outside. It provides a method called 'terminate', which communicates
to the thread that it needs to die, and then waits for the death to occur.
It imposes the following rules:
1. Call it's .__init__ inside of your __init__
2. Use the .reader inside of the thread - when it has data written on it,
that is an exit request
3. Call it's .cleanup method before exiting.
"""
def __init__(self):
reader, writer = os.pipe()
self.exit_reader = os.fdopen(reader, 'rb')
self.exit_writer = os.fdopen(writer, 'wb')
def cleanup(self):
self.exit_reader.close()
self.exit_writer.close()
def terminate(self):
"""
Asynchronously terminates the thread, without waiting for it to exit.
"""
try:
self.exit_writer.write(b' ')
self.exit_writer.flush()
except ValueError:
pass
def wait_for_exit(self):
"""
Waits for the thread to exit - should be run only after terminating
the thread.
"""
self.join()
Add way to reset loggers
logging.basicConfig doesn't reset the current logging infrastructure, which
was messing up some tests that expected logs to be one place even though
they ended up in anotherimport logging
import os
import threading
def reset_loggers():
"""
Removes all handlers from the current loggers to allow for a new basicConfig.
"""
root = logging.getLogger()
for handler in root.handlers[:]:
root.removeHandler(handler)
class TerminableThreadMixin:
"""
TerminableThreadMixin is useful for threads that need to be terminated
from the outside. It provides a method called 'terminate', which communicates
to the thread that it needs to die, and then waits for the death to occur.
It imposes the following rules:
1. Call it's .__init__ inside of your __init__
2. Use the .reader inside of the thread - when it has data written on it,
that is an exit request
3. Call it's .cleanup method before exiting.
"""
def __init__(self):
reader, writer = os.pipe()
self.exit_reader = os.fdopen(reader, 'rb')
self.exit_writer = os.fdopen(writer, 'wb')
def cleanup(self):
self.exit_reader.close()
self.exit_writer.close()
def terminate(self):
"""
Asynchronously terminates the thread, without waiting for it to exit.
"""
try:
self.exit_writer.write(b' ')
self.exit_writer.flush()
except ValueError:
pass
def wait_for_exit(self):
"""
Waits for the thread to exit - should be run only after terminating
the thread.
"""
self.join()
| <commit_before>import os
import threading
class TerminableThreadMixin:
"""
TerminableThreadMixin is useful for threads that need to be terminated
from the outside. It provides a method called 'terminate', which communicates
to the thread that it needs to die, and then waits for the death to occur.
It imposes the following rules:
1. Call it's .__init__ inside of your __init__
2. Use the .reader inside of the thread - when it has data written on it,
that is an exit request
3. Call it's .cleanup method before exiting.
"""
def __init__(self):
reader, writer = os.pipe()
self.exit_reader = os.fdopen(reader, 'rb')
self.exit_writer = os.fdopen(writer, 'wb')
def cleanup(self):
self.exit_reader.close()
self.exit_writer.close()
def terminate(self):
"""
Asynchronously terminates the thread, without waiting for it to exit.
"""
try:
self.exit_writer.write(b' ')
self.exit_writer.flush()
except ValueError:
pass
def wait_for_exit(self):
"""
Waits for the thread to exit - should be run only after terminating
the thread.
"""
self.join()
<commit_msg>Add way to reset loggers
logging.basicConfig doesn't reset the current logging infrastructure, which
was messing up some tests that expected logs to be one place even though
they ended up in another<commit_after>import logging
import os
import threading
def reset_loggers():
"""
Removes all handlers from the current loggers to allow for a new basicConfig.
"""
root = logging.getLogger()
for handler in root.handlers[:]:
root.removeHandler(handler)
class TerminableThreadMixin:
"""
TerminableThreadMixin is useful for threads that need to be terminated
from the outside. It provides a method called 'terminate', which communicates
to the thread that it needs to die, and then waits for the death to occur.
It imposes the following rules:
1. Call it's .__init__ inside of your __init__
2. Use the .reader inside of the thread - when it has data written on it,
that is an exit request
3. Call it's .cleanup method before exiting.
"""
def __init__(self):
reader, writer = os.pipe()
self.exit_reader = os.fdopen(reader, 'rb')
self.exit_writer = os.fdopen(writer, 'wb')
def cleanup(self):
self.exit_reader.close()
self.exit_writer.close()
def terminate(self):
"""
Asynchronously terminates the thread, without waiting for it to exit.
"""
try:
self.exit_writer.write(b' ')
self.exit_writer.flush()
except ValueError:
pass
def wait_for_exit(self):
"""
Waits for the thread to exit - should be run only after terminating
the thread.
"""
self.join()
|
232a9fd87f15a8b118c835d6f888d6bb9e236d19 | cat/search_indexes.py | cat/search_indexes.py | from haystack.indexes import SearchIndex, CharField, MultiValueField, BooleanField
from haystack import site
from .models import MuseumObject
class MuseumObjectIndex(SearchIndex):
text = CharField(document=True, use_template=True)
categories = MultiValueField(faceted=True)
item_name = CharField(model_attr='artefact_type', faceted=True)
global_region = CharField(model_attr='global_region', faceted=True)
country = CharField(model_attr='country', faceted=True)
people = MultiValueField(faceted=True)
has_images = BooleanField(faceted=True)
def prepare_categories(self, object):
return [unicode(cat) for cat in object.category.all()]
def prepare_people(self, object):
people = set()
if object.maker:
people.add(unicode(object.maker))
people.add(unicode(object.donor))
people.add(unicode(object.collector))
return list(people)
def prepare_has_images(self, object):
return object.artefactrepresentation_set.exists()
def get_model(self):
return MuseumObject
def index_queryset(self):
"""
Used when the entire index for model is updated.
"""
### TODO ###
# Ignore private/reserved etc objects
return self.get_model().objects.all()
site.register(MuseumObject, MuseumObjectIndex)
| from haystack.indexes import SearchIndex, CharField, MultiValueField, BooleanField
from haystack import site
from .models import MuseumObject
class MuseumObjectIndex(SearchIndex):
text = CharField(document=True, use_template=True)
categories = MultiValueField(faceted=True)
item_name = CharField(model_attr='artefact_type', faceted=True)
global_region = CharField(model_attr='global_region', faceted=True)
country = CharField(model_attr='country', faceted=True, default='')
people = MultiValueField(faceted=True)
has_images = BooleanField(faceted=True)
def prepare_categories(self, object):
return [unicode(cat) for cat in object.category.all()]
def prepare_people(self, object):
people = set()
if object.maker:
people.add(unicode(object.maker))
people.add(unicode(object.donor))
people.add(unicode(object.collector))
return list(people)
def prepare_has_images(self, object):
return object.artefactrepresentation_set.exists()
def get_model(self):
return MuseumObject
def index_queryset(self):
"""
Used when the entire index for model is updated.
"""
### TODO ###
# Ignore private/reserved etc objects
return self.get_model().objects.all()
site.register(MuseumObject, MuseumObjectIndex)
| Allow indexing blank country field | Allow indexing blank country field
| Python | bsd-3-clause | uq-eresearch/uqam,uq-eresearch/uqam,uq-eresearch/uqam,uq-eresearch/uqam | from haystack.indexes import SearchIndex, CharField, MultiValueField, BooleanField
from haystack import site
from .models import MuseumObject
class MuseumObjectIndex(SearchIndex):
text = CharField(document=True, use_template=True)
categories = MultiValueField(faceted=True)
item_name = CharField(model_attr='artefact_type', faceted=True)
global_region = CharField(model_attr='global_region', faceted=True)
country = CharField(model_attr='country', faceted=True)
people = MultiValueField(faceted=True)
has_images = BooleanField(faceted=True)
def prepare_categories(self, object):
return [unicode(cat) for cat in object.category.all()]
def prepare_people(self, object):
people = set()
if object.maker:
people.add(unicode(object.maker))
people.add(unicode(object.donor))
people.add(unicode(object.collector))
return list(people)
def prepare_has_images(self, object):
return object.artefactrepresentation_set.exists()
def get_model(self):
return MuseumObject
def index_queryset(self):
"""
Used when the entire index for model is updated.
"""
### TODO ###
# Ignore private/reserved etc objects
return self.get_model().objects.all()
site.register(MuseumObject, MuseumObjectIndex)
Allow indexing blank country field | from haystack.indexes import SearchIndex, CharField, MultiValueField, BooleanField
from haystack import site
from .models import MuseumObject
class MuseumObjectIndex(SearchIndex):
text = CharField(document=True, use_template=True)
categories = MultiValueField(faceted=True)
item_name = CharField(model_attr='artefact_type', faceted=True)
global_region = CharField(model_attr='global_region', faceted=True)
country = CharField(model_attr='country', faceted=True, default='')
people = MultiValueField(faceted=True)
has_images = BooleanField(faceted=True)
def prepare_categories(self, object):
return [unicode(cat) for cat in object.category.all()]
def prepare_people(self, object):
people = set()
if object.maker:
people.add(unicode(object.maker))
people.add(unicode(object.donor))
people.add(unicode(object.collector))
return list(people)
def prepare_has_images(self, object):
return object.artefactrepresentation_set.exists()
def get_model(self):
return MuseumObject
def index_queryset(self):
"""
Used when the entire index for model is updated.
"""
### TODO ###
# Ignore private/reserved etc objects
return self.get_model().objects.all()
site.register(MuseumObject, MuseumObjectIndex)
| <commit_before>from haystack.indexes import SearchIndex, CharField, MultiValueField, BooleanField
from haystack import site
from .models import MuseumObject
class MuseumObjectIndex(SearchIndex):
text = CharField(document=True, use_template=True)
categories = MultiValueField(faceted=True)
item_name = CharField(model_attr='artefact_type', faceted=True)
global_region = CharField(model_attr='global_region', faceted=True)
country = CharField(model_attr='country', faceted=True)
people = MultiValueField(faceted=True)
has_images = BooleanField(faceted=True)
def prepare_categories(self, object):
return [unicode(cat) for cat in object.category.all()]
def prepare_people(self, object):
people = set()
if object.maker:
people.add(unicode(object.maker))
people.add(unicode(object.donor))
people.add(unicode(object.collector))
return list(people)
def prepare_has_images(self, object):
return object.artefactrepresentation_set.exists()
def get_model(self):
return MuseumObject
def index_queryset(self):
"""
Used when the entire index for model is updated.
"""
### TODO ###
# Ignore private/reserved etc objects
return self.get_model().objects.all()
site.register(MuseumObject, MuseumObjectIndex)
<commit_msg>Allow indexing blank country field<commit_after> | from haystack.indexes import SearchIndex, CharField, MultiValueField, BooleanField
from haystack import site
from .models import MuseumObject
class MuseumObjectIndex(SearchIndex):
text = CharField(document=True, use_template=True)
categories = MultiValueField(faceted=True)
item_name = CharField(model_attr='artefact_type', faceted=True)
global_region = CharField(model_attr='global_region', faceted=True)
country = CharField(model_attr='country', faceted=True, default='')
people = MultiValueField(faceted=True)
has_images = BooleanField(faceted=True)
def prepare_categories(self, object):
return [unicode(cat) for cat in object.category.all()]
def prepare_people(self, object):
people = set()
if object.maker:
people.add(unicode(object.maker))
people.add(unicode(object.donor))
people.add(unicode(object.collector))
return list(people)
def prepare_has_images(self, object):
return object.artefactrepresentation_set.exists()
def get_model(self):
return MuseumObject
def index_queryset(self):
"""
Used when the entire index for model is updated.
"""
### TODO ###
# Ignore private/reserved etc objects
return self.get_model().objects.all()
site.register(MuseumObject, MuseumObjectIndex)
| from haystack.indexes import SearchIndex, CharField, MultiValueField, BooleanField
from haystack import site
from .models import MuseumObject
class MuseumObjectIndex(SearchIndex):
text = CharField(document=True, use_template=True)
categories = MultiValueField(faceted=True)
item_name = CharField(model_attr='artefact_type', faceted=True)
global_region = CharField(model_attr='global_region', faceted=True)
country = CharField(model_attr='country', faceted=True)
people = MultiValueField(faceted=True)
has_images = BooleanField(faceted=True)
def prepare_categories(self, object):
return [unicode(cat) for cat in object.category.all()]
def prepare_people(self, object):
people = set()
if object.maker:
people.add(unicode(object.maker))
people.add(unicode(object.donor))
people.add(unicode(object.collector))
return list(people)
def prepare_has_images(self, object):
return object.artefactrepresentation_set.exists()
def get_model(self):
return MuseumObject
def index_queryset(self):
"""
Used when the entire index for model is updated.
"""
### TODO ###
# Ignore private/reserved etc objects
return self.get_model().objects.all()
site.register(MuseumObject, MuseumObjectIndex)
Allow indexing blank country fieldfrom haystack.indexes import SearchIndex, CharField, MultiValueField, BooleanField
from haystack import site
from .models import MuseumObject
class MuseumObjectIndex(SearchIndex):
text = CharField(document=True, use_template=True)
categories = MultiValueField(faceted=True)
item_name = CharField(model_attr='artefact_type', faceted=True)
global_region = CharField(model_attr='global_region', faceted=True)
country = CharField(model_attr='country', faceted=True, default='')
people = MultiValueField(faceted=True)
has_images = BooleanField(faceted=True)
def prepare_categories(self, object):
return [unicode(cat) for cat in object.category.all()]
def prepare_people(self, object):
people = set()
if object.maker:
people.add(unicode(object.maker))
people.add(unicode(object.donor))
people.add(unicode(object.collector))
return list(people)
def prepare_has_images(self, object):
return object.artefactrepresentation_set.exists()
def get_model(self):
return MuseumObject
def index_queryset(self):
"""
Used when the entire index for model is updated.
"""
### TODO ###
# Ignore private/reserved etc objects
return self.get_model().objects.all()
site.register(MuseumObject, MuseumObjectIndex)
| <commit_before>from haystack.indexes import SearchIndex, CharField, MultiValueField, BooleanField
from haystack import site
from .models import MuseumObject
class MuseumObjectIndex(SearchIndex):
text = CharField(document=True, use_template=True)
categories = MultiValueField(faceted=True)
item_name = CharField(model_attr='artefact_type', faceted=True)
global_region = CharField(model_attr='global_region', faceted=True)
country = CharField(model_attr='country', faceted=True)
people = MultiValueField(faceted=True)
has_images = BooleanField(faceted=True)
def prepare_categories(self, object):
return [unicode(cat) for cat in object.category.all()]
def prepare_people(self, object):
people = set()
if object.maker:
people.add(unicode(object.maker))
people.add(unicode(object.donor))
people.add(unicode(object.collector))
return list(people)
def prepare_has_images(self, object):
return object.artefactrepresentation_set.exists()
def get_model(self):
return MuseumObject
def index_queryset(self):
"""
Used when the entire index for model is updated.
"""
### TODO ###
# Ignore private/reserved etc objects
return self.get_model().objects.all()
site.register(MuseumObject, MuseumObjectIndex)
<commit_msg>Allow indexing blank country field<commit_after>from haystack.indexes import SearchIndex, CharField, MultiValueField, BooleanField
from haystack import site
from .models import MuseumObject
class MuseumObjectIndex(SearchIndex):
text = CharField(document=True, use_template=True)
categories = MultiValueField(faceted=True)
item_name = CharField(model_attr='artefact_type', faceted=True)
global_region = CharField(model_attr='global_region', faceted=True)
country = CharField(model_attr='country', faceted=True, default='')
people = MultiValueField(faceted=True)
has_images = BooleanField(faceted=True)
def prepare_categories(self, object):
return [unicode(cat) for cat in object.category.all()]
def prepare_people(self, object):
people = set()
if object.maker:
people.add(unicode(object.maker))
people.add(unicode(object.donor))
people.add(unicode(object.collector))
return list(people)
def prepare_has_images(self, object):
return object.artefactrepresentation_set.exists()
def get_model(self):
return MuseumObject
def index_queryset(self):
"""
Used when the entire index for model is updated.
"""
### TODO ###
# Ignore private/reserved etc objects
return self.get_model().objects.all()
site.register(MuseumObject, MuseumObjectIndex)
|
76ed5eb9d9d2f3a453de6976f52221e6970b6b71 | tests/unit/test_offline_compression.py | tests/unit/test_offline_compression.py | import os
import shutil
import tempfile
from django.test import TestCase
from django.core.management import call_command
from django.test.utils import override_settings
TMP_STATIC_DIR = tempfile.mkdtemp()
@override_settings(
COMPRESS_ENABLED=True,
COMPRESS_OFFLINE=True,
COMPRESS_ROOT=TMP_STATIC_DIR
)
class TestOfflineCompression(TestCase):
def tearDown(self):
super(TestOfflineCompression, self).tearDown()
if os.path.exists(TMP_STATIC_DIR):
shutil.rmtree(TMP_STATIC_DIR)
def test_(self):
call_command('compress')
| import os
import shutil
import tempfile
from django.test import TestCase
from django.core.management import call_command
from django.test.utils import override_settings
TMP_STATIC_DIR = tempfile.mkdtemp()
@override_settings(
COMPRESS_ENABLED=True,
COMPRESS_OFFLINE=True,
COMPRESS_ROOT=TMP_STATIC_DIR
)
class TestOfflineCompression(TestCase):
def tearDown(self):
super(TestOfflineCompression, self).tearDown()
if os.path.exists(TMP_STATIC_DIR):
shutil.rmtree(TMP_STATIC_DIR)
def test_(self):
call_command('compress', verbosity=0)
| Add test for offline compression using django_compressor | Add test for offline compression using django_compressor
| Python | bsd-3-clause | tangentlabs/django-fancypages,socradev/django-fancypages,tangentlabs/django-fancypages,socradev/django-fancypages,socradev/django-fancypages,tangentlabs/django-fancypages | import os
import shutil
import tempfile
from django.test import TestCase
from django.core.management import call_command
from django.test.utils import override_settings
TMP_STATIC_DIR = tempfile.mkdtemp()
@override_settings(
COMPRESS_ENABLED=True,
COMPRESS_OFFLINE=True,
COMPRESS_ROOT=TMP_STATIC_DIR
)
class TestOfflineCompression(TestCase):
def tearDown(self):
super(TestOfflineCompression, self).tearDown()
if os.path.exists(TMP_STATIC_DIR):
shutil.rmtree(TMP_STATIC_DIR)
def test_(self):
call_command('compress')
Add test for offline compression using django_compressor | import os
import shutil
import tempfile
from django.test import TestCase
from django.core.management import call_command
from django.test.utils import override_settings
TMP_STATIC_DIR = tempfile.mkdtemp()
@override_settings(
COMPRESS_ENABLED=True,
COMPRESS_OFFLINE=True,
COMPRESS_ROOT=TMP_STATIC_DIR
)
class TestOfflineCompression(TestCase):
def tearDown(self):
super(TestOfflineCompression, self).tearDown()
if os.path.exists(TMP_STATIC_DIR):
shutil.rmtree(TMP_STATIC_DIR)
def test_(self):
call_command('compress', verbosity=0)
| <commit_before>import os
import shutil
import tempfile
from django.test import TestCase
from django.core.management import call_command
from django.test.utils import override_settings
TMP_STATIC_DIR = tempfile.mkdtemp()
@override_settings(
COMPRESS_ENABLED=True,
COMPRESS_OFFLINE=True,
COMPRESS_ROOT=TMP_STATIC_DIR
)
class TestOfflineCompression(TestCase):
def tearDown(self):
super(TestOfflineCompression, self).tearDown()
if os.path.exists(TMP_STATIC_DIR):
shutil.rmtree(TMP_STATIC_DIR)
def test_(self):
call_command('compress')
<commit_msg>Add test for offline compression using django_compressor<commit_after> | import os
import shutil
import tempfile
from django.test import TestCase
from django.core.management import call_command
from django.test.utils import override_settings
TMP_STATIC_DIR = tempfile.mkdtemp()
@override_settings(
COMPRESS_ENABLED=True,
COMPRESS_OFFLINE=True,
COMPRESS_ROOT=TMP_STATIC_DIR
)
class TestOfflineCompression(TestCase):
def tearDown(self):
super(TestOfflineCompression, self).tearDown()
if os.path.exists(TMP_STATIC_DIR):
shutil.rmtree(TMP_STATIC_DIR)
def test_(self):
call_command('compress', verbosity=0)
| import os
import shutil
import tempfile
from django.test import TestCase
from django.core.management import call_command
from django.test.utils import override_settings
TMP_STATIC_DIR = tempfile.mkdtemp()
@override_settings(
COMPRESS_ENABLED=True,
COMPRESS_OFFLINE=True,
COMPRESS_ROOT=TMP_STATIC_DIR
)
class TestOfflineCompression(TestCase):
def tearDown(self):
super(TestOfflineCompression, self).tearDown()
if os.path.exists(TMP_STATIC_DIR):
shutil.rmtree(TMP_STATIC_DIR)
def test_(self):
call_command('compress')
Add test for offline compression using django_compressorimport os
import shutil
import tempfile
from django.test import TestCase
from django.core.management import call_command
from django.test.utils import override_settings
TMP_STATIC_DIR = tempfile.mkdtemp()
@override_settings(
COMPRESS_ENABLED=True,
COMPRESS_OFFLINE=True,
COMPRESS_ROOT=TMP_STATIC_DIR
)
class TestOfflineCompression(TestCase):
def tearDown(self):
super(TestOfflineCompression, self).tearDown()
if os.path.exists(TMP_STATIC_DIR):
shutil.rmtree(TMP_STATIC_DIR)
def test_(self):
call_command('compress', verbosity=0)
| <commit_before>import os
import shutil
import tempfile
from django.test import TestCase
from django.core.management import call_command
from django.test.utils import override_settings
TMP_STATIC_DIR = tempfile.mkdtemp()
@override_settings(
COMPRESS_ENABLED=True,
COMPRESS_OFFLINE=True,
COMPRESS_ROOT=TMP_STATIC_DIR
)
class TestOfflineCompression(TestCase):
def tearDown(self):
super(TestOfflineCompression, self).tearDown()
if os.path.exists(TMP_STATIC_DIR):
shutil.rmtree(TMP_STATIC_DIR)
def test_(self):
call_command('compress')
<commit_msg>Add test for offline compression using django_compressor<commit_after>import os
import shutil
import tempfile
from django.test import TestCase
from django.core.management import call_command
from django.test.utils import override_settings
TMP_STATIC_DIR = tempfile.mkdtemp()
@override_settings(
COMPRESS_ENABLED=True,
COMPRESS_OFFLINE=True,
COMPRESS_ROOT=TMP_STATIC_DIR
)
class TestOfflineCompression(TestCase):
def tearDown(self):
super(TestOfflineCompression, self).tearDown()
if os.path.exists(TMP_STATIC_DIR):
shutil.rmtree(TMP_STATIC_DIR)
def test_(self):
call_command('compress', verbosity=0)
|
d7e9eba6fb3628f0736bd468ae76e05099b9d651 | space/decorators.py | space/decorators.py | from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from incubator.settings import STATUS_SECRETS
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
| from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from django.conf import settings
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in settings.STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
| Use from django.conf import settings | Use from django.conf import settings
| Python | agpl-3.0 | UrLab/incubator,UrLab/incubator,UrLab/incubator,UrLab/incubator | from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from incubator.settings import STATUS_SECRETS
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
Use from django.conf import settings | from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from django.conf import settings
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in settings.STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
| <commit_before>from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from incubator.settings import STATUS_SECRETS
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
<commit_msg>Use from django.conf import settings<commit_after> | from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from django.conf import settings
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in settings.STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
| from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from incubator.settings import STATUS_SECRETS
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
Use from django.conf import settingsfrom django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from django.conf import settings
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in settings.STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
| <commit_before>from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from incubator.settings import STATUS_SECRETS
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
<commit_msg>Use from django.conf import settings<commit_after>from django.http import HttpResponseBadRequest, HttpResponseForbidden
from django.views.decorators.csrf import csrf_exempt
from django.conf import settings
def one_or_zero(arg):
"""Typecast to 1 or 0"""
if arg == '1':
return 1
elif arg == '0':
return 0
raise ValueError("not one or zero")
def private_api(**required_params):
"""
Filter incoming private API requests, and perform parameter validation and
extraction
"""
def outer(some_view):
@csrf_exempt
def inner(request, *args, **kwargs):
if request.method != 'POST':
return HttpResponseBadRequest("Only POST is allowed")
if 'secret' not in request.POST.keys():
return HttpResponseBadRequest(
"You must query this endpoint with a secret.")
if request.POST['secret'] not in settings.STATUS_SECRETS:
message = 'Bad secret {} is not in the allowed list'.format(
request.POST['secret'])
return HttpResponseForbidden(message)
params = {}
for name, typecast in required_params.items():
if name not in request.POST.keys():
return HttpResponseBadRequest(
"Parameter %s is required" % name)
try:
params[name] = typecast(request.POST[name])
except ValueError:
return HttpResponseBadRequest(
"Did not understood %s=%s" % (name, request.POST[name]))
return some_view(request, **params)
return inner
return outer
|
ce3c7daff5eaaf8eefecf3f4e5bd9fbca40a7a2a | cob/subsystems/tasks_subsystem.py | cob/subsystems/tasks_subsystem.py | import os
import logbook
from .base import SubsystemBase
_logger = logbook.Logger(__name__)
class TasksSubsystem(SubsystemBase):
NAME = 'tasks'
def activate(self, flask_app):
from ..celery.app import celery_app
self._config = self.project.config.get('celery', {})
# ensure critical celery config exists
self._config.setdefault('broker_url', 'amqp://guest:guest@localhost/')
override_broker_url = os.environ.get('COB_CELERY_BROKER_URL')
if override_broker_url is not None:
self._config['broker_url'] = override_broker_url
celery_app.conf.broker_url = self._config['broker_url']
self.queues = set()
def get_queue_names(self):
names = {queue_name for grain in self.grains for queue_name in grain.config.get('queue_names', [])}
names.add('celery')
return sorted(names)
def configure_grain(self, grain, flask_app):
_ = grain.load()
def configure_app(self, flask_app):
super().configure_app(flask_app)
from ..celery.app import celery_app
for task in celery_app.tasks.values():
queue_name = getattr(task, 'queue', None)
if queue_name is not None:
self.queues.add(queue_name)
def iter_locations(self):
return None
| import os
import logbook
from .base import SubsystemBase
_logger = logbook.Logger(__name__)
class TasksSubsystem(SubsystemBase):
NAME = 'tasks'
def activate(self, flask_app):
from ..celery.app import celery_app
self._config = self.project.config.get('celery', {})
# ensure critical celery config exists
self._config.setdefault('broker_url', 'amqp://guest:guest@localhost/')
override_broker_url = os.environ.get('COB_CELERY_BROKER_URL')
if override_broker_url is not None:
self._config['broker_url'] = override_broker_url
celery_app.conf.update(self._config)
self.queues = set()
def get_queue_names(self):
names = {queue_name for grain in self.grains for queue_name in grain.config.get('queue_names', [])}
names.add('celery')
return sorted(names)
def configure_grain(self, grain, flask_app):
_ = grain.load()
def configure_app(self, flask_app):
super().configure_app(flask_app)
from ..celery.app import celery_app
for task in celery_app.tasks.values():
queue_name = getattr(task, 'queue', None)
if queue_name is not None:
self.queues.add(queue_name)
def iter_locations(self):
return None
| Allow passing Celery configuration under the project's config | Allow passing Celery configuration under the project's config
| Python | bsd-3-clause | getweber/weber-cli | import os
import logbook
from .base import SubsystemBase
_logger = logbook.Logger(__name__)
class TasksSubsystem(SubsystemBase):
NAME = 'tasks'
def activate(self, flask_app):
from ..celery.app import celery_app
self._config = self.project.config.get('celery', {})
# ensure critical celery config exists
self._config.setdefault('broker_url', 'amqp://guest:guest@localhost/')
override_broker_url = os.environ.get('COB_CELERY_BROKER_URL')
if override_broker_url is not None:
self._config['broker_url'] = override_broker_url
celery_app.conf.broker_url = self._config['broker_url']
self.queues = set()
def get_queue_names(self):
names = {queue_name for grain in self.grains for queue_name in grain.config.get('queue_names', [])}
names.add('celery')
return sorted(names)
def configure_grain(self, grain, flask_app):
_ = grain.load()
def configure_app(self, flask_app):
super().configure_app(flask_app)
from ..celery.app import celery_app
for task in celery_app.tasks.values():
queue_name = getattr(task, 'queue', None)
if queue_name is not None:
self.queues.add(queue_name)
def iter_locations(self):
return None
Allow passing Celery configuration under the project's config | import os
import logbook
from .base import SubsystemBase
_logger = logbook.Logger(__name__)
class TasksSubsystem(SubsystemBase):
NAME = 'tasks'
def activate(self, flask_app):
from ..celery.app import celery_app
self._config = self.project.config.get('celery', {})
# ensure critical celery config exists
self._config.setdefault('broker_url', 'amqp://guest:guest@localhost/')
override_broker_url = os.environ.get('COB_CELERY_BROKER_URL')
if override_broker_url is not None:
self._config['broker_url'] = override_broker_url
celery_app.conf.update(self._config)
self.queues = set()
def get_queue_names(self):
names = {queue_name for grain in self.grains for queue_name in grain.config.get('queue_names', [])}
names.add('celery')
return sorted(names)
def configure_grain(self, grain, flask_app):
_ = grain.load()
def configure_app(self, flask_app):
super().configure_app(flask_app)
from ..celery.app import celery_app
for task in celery_app.tasks.values():
queue_name = getattr(task, 'queue', None)
if queue_name is not None:
self.queues.add(queue_name)
def iter_locations(self):
return None
| <commit_before>import os
import logbook
from .base import SubsystemBase
_logger = logbook.Logger(__name__)
class TasksSubsystem(SubsystemBase):
NAME = 'tasks'
def activate(self, flask_app):
from ..celery.app import celery_app
self._config = self.project.config.get('celery', {})
# ensure critical celery config exists
self._config.setdefault('broker_url', 'amqp://guest:guest@localhost/')
override_broker_url = os.environ.get('COB_CELERY_BROKER_URL')
if override_broker_url is not None:
self._config['broker_url'] = override_broker_url
celery_app.conf.broker_url = self._config['broker_url']
self.queues = set()
def get_queue_names(self):
names = {queue_name for grain in self.grains for queue_name in grain.config.get('queue_names', [])}
names.add('celery')
return sorted(names)
def configure_grain(self, grain, flask_app):
_ = grain.load()
def configure_app(self, flask_app):
super().configure_app(flask_app)
from ..celery.app import celery_app
for task in celery_app.tasks.values():
queue_name = getattr(task, 'queue', None)
if queue_name is not None:
self.queues.add(queue_name)
def iter_locations(self):
return None
<commit_msg>Allow passing Celery configuration under the project's config<commit_after> | import os
import logbook
from .base import SubsystemBase
_logger = logbook.Logger(__name__)
class TasksSubsystem(SubsystemBase):
NAME = 'tasks'
def activate(self, flask_app):
from ..celery.app import celery_app
self._config = self.project.config.get('celery', {})
# ensure critical celery config exists
self._config.setdefault('broker_url', 'amqp://guest:guest@localhost/')
override_broker_url = os.environ.get('COB_CELERY_BROKER_URL')
if override_broker_url is not None:
self._config['broker_url'] = override_broker_url
celery_app.conf.update(self._config)
self.queues = set()
def get_queue_names(self):
names = {queue_name for grain in self.grains for queue_name in grain.config.get('queue_names', [])}
names.add('celery')
return sorted(names)
def configure_grain(self, grain, flask_app):
_ = grain.load()
def configure_app(self, flask_app):
super().configure_app(flask_app)
from ..celery.app import celery_app
for task in celery_app.tasks.values():
queue_name = getattr(task, 'queue', None)
if queue_name is not None:
self.queues.add(queue_name)
def iter_locations(self):
return None
| import os
import logbook
from .base import SubsystemBase
_logger = logbook.Logger(__name__)
class TasksSubsystem(SubsystemBase):
NAME = 'tasks'
def activate(self, flask_app):
from ..celery.app import celery_app
self._config = self.project.config.get('celery', {})
# ensure critical celery config exists
self._config.setdefault('broker_url', 'amqp://guest:guest@localhost/')
override_broker_url = os.environ.get('COB_CELERY_BROKER_URL')
if override_broker_url is not None:
self._config['broker_url'] = override_broker_url
celery_app.conf.broker_url = self._config['broker_url']
self.queues = set()
def get_queue_names(self):
names = {queue_name for grain in self.grains for queue_name in grain.config.get('queue_names', [])}
names.add('celery')
return sorted(names)
def configure_grain(self, grain, flask_app):
_ = grain.load()
def configure_app(self, flask_app):
super().configure_app(flask_app)
from ..celery.app import celery_app
for task in celery_app.tasks.values():
queue_name = getattr(task, 'queue', None)
if queue_name is not None:
self.queues.add(queue_name)
def iter_locations(self):
return None
Allow passing Celery configuration under the project's configimport os
import logbook
from .base import SubsystemBase
_logger = logbook.Logger(__name__)
class TasksSubsystem(SubsystemBase):
NAME = 'tasks'
def activate(self, flask_app):
from ..celery.app import celery_app
self._config = self.project.config.get('celery', {})
# ensure critical celery config exists
self._config.setdefault('broker_url', 'amqp://guest:guest@localhost/')
override_broker_url = os.environ.get('COB_CELERY_BROKER_URL')
if override_broker_url is not None:
self._config['broker_url'] = override_broker_url
celery_app.conf.update(self._config)
self.queues = set()
def get_queue_names(self):
names = {queue_name for grain in self.grains for queue_name in grain.config.get('queue_names', [])}
names.add('celery')
return sorted(names)
def configure_grain(self, grain, flask_app):
_ = grain.load()
def configure_app(self, flask_app):
super().configure_app(flask_app)
from ..celery.app import celery_app
for task in celery_app.tasks.values():
queue_name = getattr(task, 'queue', None)
if queue_name is not None:
self.queues.add(queue_name)
def iter_locations(self):
return None
| <commit_before>import os
import logbook
from .base import SubsystemBase
_logger = logbook.Logger(__name__)
class TasksSubsystem(SubsystemBase):
NAME = 'tasks'
def activate(self, flask_app):
from ..celery.app import celery_app
self._config = self.project.config.get('celery', {})
# ensure critical celery config exists
self._config.setdefault('broker_url', 'amqp://guest:guest@localhost/')
override_broker_url = os.environ.get('COB_CELERY_BROKER_URL')
if override_broker_url is not None:
self._config['broker_url'] = override_broker_url
celery_app.conf.broker_url = self._config['broker_url']
self.queues = set()
def get_queue_names(self):
names = {queue_name for grain in self.grains for queue_name in grain.config.get('queue_names', [])}
names.add('celery')
return sorted(names)
def configure_grain(self, grain, flask_app):
_ = grain.load()
def configure_app(self, flask_app):
super().configure_app(flask_app)
from ..celery.app import celery_app
for task in celery_app.tasks.values():
queue_name = getattr(task, 'queue', None)
if queue_name is not None:
self.queues.add(queue_name)
def iter_locations(self):
return None
<commit_msg>Allow passing Celery configuration under the project's config<commit_after>import os
import logbook
from .base import SubsystemBase
_logger = logbook.Logger(__name__)
class TasksSubsystem(SubsystemBase):
NAME = 'tasks'
def activate(self, flask_app):
from ..celery.app import celery_app
self._config = self.project.config.get('celery', {})
# ensure critical celery config exists
self._config.setdefault('broker_url', 'amqp://guest:guest@localhost/')
override_broker_url = os.environ.get('COB_CELERY_BROKER_URL')
if override_broker_url is not None:
self._config['broker_url'] = override_broker_url
celery_app.conf.update(self._config)
self.queues = set()
def get_queue_names(self):
names = {queue_name for grain in self.grains for queue_name in grain.config.get('queue_names', [])}
names.add('celery')
return sorted(names)
def configure_grain(self, grain, flask_app):
_ = grain.load()
def configure_app(self, flask_app):
super().configure_app(flask_app)
from ..celery.app import celery_app
for task in celery_app.tasks.values():
queue_name = getattr(task, 'queue', None)
if queue_name is not None:
self.queues.add(queue_name)
def iter_locations(self):
return None
|
68d1943b591afe55f75fa31dfb3c4c61b8f4297f | daybed/tests/support.py | daybed/tests/support.py | import os
from uuid import uuid4
from unittest import TestCase
import webtest
from daybed.db import DatabaseConnection
HERE = os.path.dirname(os.path.abspath(__file__))
class BaseWebTest(TestCase):
"""Base Web Test to test your cornice service.
It setups the database before each test and delete it after.
"""
def setUp(self):
self.db_name = os.environ['DB_NAME'] = 'daybed-tests-%s' % uuid4()
self.db_server = self.app.app.registry.settings['db_server']
self.db_base = self.db_server[self.db_name]
self.db = DatabaseConnection(self.db_base)
self.app = webtest.TestApp("config:tests.ini", relative_to=HERE)
def tearDown(self):
# Delete Test DB
self.db_server.delete(self.db_name)
def put_valid_definition(self):
"""Create a valid definition named "todo".
"""
# Put a valid definition
self.app.put_json('/definitions/todo',
self.valid_definition,
headers=self.headers)
| import os
from uuid import uuid4
from unittest import TestCase
import webtest
from daybed.db import DatabaseConnection
HERE = os.path.dirname(os.path.abspath(__file__))
class BaseWebTest(TestCase):
"""Base Web Test to test your cornice service.
It setups the database before each test and delete it after.
"""
def setUp(self):
self.db_name = os.environ['DB_NAME'] = 'daybed-tests-%s' % uuid4()
self.app = webtest.TestApp("config:tests.ini", relative_to=HERE)
self.db_server = self.app.app.registry.settings['db_server']
self.db_base = self.db_server[self.db_name]
self.db = DatabaseConnection(self.db_base)
def tearDown(self):
# Delete Test DB
self.db_server.delete(self.db_name)
def put_valid_definition(self):
"""Create a valid definition named "todo".
"""
# Put a valid definition
self.app.put_json('/definitions/todo',
self.valid_definition,
headers=self.headers)
| Fix db deletion, take 2 | Fix db deletion, take 2
| Python | bsd-3-clause | spiral-project/daybed,spiral-project/daybed | import os
from uuid import uuid4
from unittest import TestCase
import webtest
from daybed.db import DatabaseConnection
HERE = os.path.dirname(os.path.abspath(__file__))
class BaseWebTest(TestCase):
"""Base Web Test to test your cornice service.
It setups the database before each test and delete it after.
"""
def setUp(self):
self.db_name = os.environ['DB_NAME'] = 'daybed-tests-%s' % uuid4()
self.db_server = self.app.app.registry.settings['db_server']
self.db_base = self.db_server[self.db_name]
self.db = DatabaseConnection(self.db_base)
self.app = webtest.TestApp("config:tests.ini", relative_to=HERE)
def tearDown(self):
# Delete Test DB
self.db_server.delete(self.db_name)
def put_valid_definition(self):
"""Create a valid definition named "todo".
"""
# Put a valid definition
self.app.put_json('/definitions/todo',
self.valid_definition,
headers=self.headers)
Fix db deletion, take 2 | import os
from uuid import uuid4
from unittest import TestCase
import webtest
from daybed.db import DatabaseConnection
HERE = os.path.dirname(os.path.abspath(__file__))
class BaseWebTest(TestCase):
"""Base Web Test to test your cornice service.
It setups the database before each test and delete it after.
"""
def setUp(self):
self.db_name = os.environ['DB_NAME'] = 'daybed-tests-%s' % uuid4()
self.app = webtest.TestApp("config:tests.ini", relative_to=HERE)
self.db_server = self.app.app.registry.settings['db_server']
self.db_base = self.db_server[self.db_name]
self.db = DatabaseConnection(self.db_base)
def tearDown(self):
# Delete Test DB
self.db_server.delete(self.db_name)
def put_valid_definition(self):
"""Create a valid definition named "todo".
"""
# Put a valid definition
self.app.put_json('/definitions/todo',
self.valid_definition,
headers=self.headers)
| <commit_before>import os
from uuid import uuid4
from unittest import TestCase
import webtest
from daybed.db import DatabaseConnection
HERE = os.path.dirname(os.path.abspath(__file__))
class BaseWebTest(TestCase):
"""Base Web Test to test your cornice service.
It setups the database before each test and delete it after.
"""
def setUp(self):
self.db_name = os.environ['DB_NAME'] = 'daybed-tests-%s' % uuid4()
self.db_server = self.app.app.registry.settings['db_server']
self.db_base = self.db_server[self.db_name]
self.db = DatabaseConnection(self.db_base)
self.app = webtest.TestApp("config:tests.ini", relative_to=HERE)
def tearDown(self):
# Delete Test DB
self.db_server.delete(self.db_name)
def put_valid_definition(self):
"""Create a valid definition named "todo".
"""
# Put a valid definition
self.app.put_json('/definitions/todo',
self.valid_definition,
headers=self.headers)
<commit_msg>Fix db deletion, take 2<commit_after> | import os
from uuid import uuid4
from unittest import TestCase
import webtest
from daybed.db import DatabaseConnection
HERE = os.path.dirname(os.path.abspath(__file__))
class BaseWebTest(TestCase):
"""Base Web Test to test your cornice service.
It setups the database before each test and delete it after.
"""
def setUp(self):
self.db_name = os.environ['DB_NAME'] = 'daybed-tests-%s' % uuid4()
self.app = webtest.TestApp("config:tests.ini", relative_to=HERE)
self.db_server = self.app.app.registry.settings['db_server']
self.db_base = self.db_server[self.db_name]
self.db = DatabaseConnection(self.db_base)
def tearDown(self):
# Delete Test DB
self.db_server.delete(self.db_name)
def put_valid_definition(self):
"""Create a valid definition named "todo".
"""
# Put a valid definition
self.app.put_json('/definitions/todo',
self.valid_definition,
headers=self.headers)
| import os
from uuid import uuid4
from unittest import TestCase
import webtest
from daybed.db import DatabaseConnection
HERE = os.path.dirname(os.path.abspath(__file__))
class BaseWebTest(TestCase):
"""Base Web Test to test your cornice service.
It setups the database before each test and delete it after.
"""
def setUp(self):
self.db_name = os.environ['DB_NAME'] = 'daybed-tests-%s' % uuid4()
self.db_server = self.app.app.registry.settings['db_server']
self.db_base = self.db_server[self.db_name]
self.db = DatabaseConnection(self.db_base)
self.app = webtest.TestApp("config:tests.ini", relative_to=HERE)
def tearDown(self):
# Delete Test DB
self.db_server.delete(self.db_name)
def put_valid_definition(self):
"""Create a valid definition named "todo".
"""
# Put a valid definition
self.app.put_json('/definitions/todo',
self.valid_definition,
headers=self.headers)
Fix db deletion, take 2import os
from uuid import uuid4
from unittest import TestCase
import webtest
from daybed.db import DatabaseConnection
HERE = os.path.dirname(os.path.abspath(__file__))
class BaseWebTest(TestCase):
"""Base Web Test to test your cornice service.
It setups the database before each test and delete it after.
"""
def setUp(self):
self.db_name = os.environ['DB_NAME'] = 'daybed-tests-%s' % uuid4()
self.app = webtest.TestApp("config:tests.ini", relative_to=HERE)
self.db_server = self.app.app.registry.settings['db_server']
self.db_base = self.db_server[self.db_name]
self.db = DatabaseConnection(self.db_base)
def tearDown(self):
# Delete Test DB
self.db_server.delete(self.db_name)
def put_valid_definition(self):
"""Create a valid definition named "todo".
"""
# Put a valid definition
self.app.put_json('/definitions/todo',
self.valid_definition,
headers=self.headers)
| <commit_before>import os
from uuid import uuid4
from unittest import TestCase
import webtest
from daybed.db import DatabaseConnection
HERE = os.path.dirname(os.path.abspath(__file__))
class BaseWebTest(TestCase):
"""Base Web Test to test your cornice service.
It setups the database before each test and delete it after.
"""
def setUp(self):
self.db_name = os.environ['DB_NAME'] = 'daybed-tests-%s' % uuid4()
self.db_server = self.app.app.registry.settings['db_server']
self.db_base = self.db_server[self.db_name]
self.db = DatabaseConnection(self.db_base)
self.app = webtest.TestApp("config:tests.ini", relative_to=HERE)
def tearDown(self):
# Delete Test DB
self.db_server.delete(self.db_name)
def put_valid_definition(self):
"""Create a valid definition named "todo".
"""
# Put a valid definition
self.app.put_json('/definitions/todo',
self.valid_definition,
headers=self.headers)
<commit_msg>Fix db deletion, take 2<commit_after>import os
from uuid import uuid4
from unittest import TestCase
import webtest
from daybed.db import DatabaseConnection
HERE = os.path.dirname(os.path.abspath(__file__))
class BaseWebTest(TestCase):
"""Base Web Test to test your cornice service.
It setups the database before each test and delete it after.
"""
def setUp(self):
self.db_name = os.environ['DB_NAME'] = 'daybed-tests-%s' % uuid4()
self.app = webtest.TestApp("config:tests.ini", relative_to=HERE)
self.db_server = self.app.app.registry.settings['db_server']
self.db_base = self.db_server[self.db_name]
self.db = DatabaseConnection(self.db_base)
def tearDown(self):
# Delete Test DB
self.db_server.delete(self.db_name)
def put_valid_definition(self):
"""Create a valid definition named "todo".
"""
# Put a valid definition
self.app.put_json('/definitions/todo',
self.valid_definition,
headers=self.headers)
|
455c8ed93dcac20b6393cf781e19971fa3b92cdb | tests/test_cookies.py | tests/test_cookies.py | # -*- coding: utf-8 -*-
def test_bar_fixture(testdir):
"""Make sure that pytest accepts our fixture."""
# create a temporary pytest test module
testdir.makepyfile("""
def test_sth(bar):
assert bar == "europython2015"
""")
# run pytest with the following cmd args
result = testdir.runpytest(
'--foo=europython2015',
'-v'
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_sth PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
def test_help_message(testdir):
result = testdir.runpytest(
'--help',
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'cookies:',
'*--foo=DEST_FOO*Set the value for the fixture "bar".',
])
def test_hello_ini_setting(testdir):
testdir.makeini("""
[pytest]
HELLO = world
""")
testdir.makepyfile("""
import pytest
@pytest.fixture
def hello(request):
return request.config.getini('HELLO')
def test_hello_world(hello):
assert hello == 'world'
""")
result = testdir.runpytest('-v')
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_hello_world PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
| # -*- coding: utf-8 -*-
def test_cookies_fixture(testdir):
"""Make sure that pytest accepts the `cookies` fixture."""
# create a temporary pytest test module
testdir.makepyfile("""
def test_valid_fixture(cookies):
assert hasattr(cookies, 'bake')
assert callable(cookies.bake)
assert callable(cookies.bake)
""")
# run pytest with the following cmd args
result = testdir.runpytest('-v')
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_valid_fixture PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
def test_help_message(testdir):
result = testdir.runpytest(
'--help',
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'cookies:',
])
| Implement simple tests for the 'cookies' fixture | Implement simple tests for the 'cookies' fixture
| Python | mit | hackebrot/pytest-cookies | # -*- coding: utf-8 -*-
def test_bar_fixture(testdir):
"""Make sure that pytest accepts our fixture."""
# create a temporary pytest test module
testdir.makepyfile("""
def test_sth(bar):
assert bar == "europython2015"
""")
# run pytest with the following cmd args
result = testdir.runpytest(
'--foo=europython2015',
'-v'
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_sth PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
def test_help_message(testdir):
result = testdir.runpytest(
'--help',
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'cookies:',
'*--foo=DEST_FOO*Set the value for the fixture "bar".',
])
def test_hello_ini_setting(testdir):
testdir.makeini("""
[pytest]
HELLO = world
""")
testdir.makepyfile("""
import pytest
@pytest.fixture
def hello(request):
return request.config.getini('HELLO')
def test_hello_world(hello):
assert hello == 'world'
""")
result = testdir.runpytest('-v')
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_hello_world PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
Implement simple tests for the 'cookies' fixture | # -*- coding: utf-8 -*-
def test_cookies_fixture(testdir):
"""Make sure that pytest accepts the `cookies` fixture."""
# create a temporary pytest test module
testdir.makepyfile("""
def test_valid_fixture(cookies):
assert hasattr(cookies, 'bake')
assert callable(cookies.bake)
assert callable(cookies.bake)
""")
# run pytest with the following cmd args
result = testdir.runpytest('-v')
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_valid_fixture PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
def test_help_message(testdir):
result = testdir.runpytest(
'--help',
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'cookies:',
])
| <commit_before># -*- coding: utf-8 -*-
def test_bar_fixture(testdir):
"""Make sure that pytest accepts our fixture."""
# create a temporary pytest test module
testdir.makepyfile("""
def test_sth(bar):
assert bar == "europython2015"
""")
# run pytest with the following cmd args
result = testdir.runpytest(
'--foo=europython2015',
'-v'
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_sth PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
def test_help_message(testdir):
result = testdir.runpytest(
'--help',
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'cookies:',
'*--foo=DEST_FOO*Set the value for the fixture "bar".',
])
def test_hello_ini_setting(testdir):
testdir.makeini("""
[pytest]
HELLO = world
""")
testdir.makepyfile("""
import pytest
@pytest.fixture
def hello(request):
return request.config.getini('HELLO')
def test_hello_world(hello):
assert hello == 'world'
""")
result = testdir.runpytest('-v')
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_hello_world PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
<commit_msg>Implement simple tests for the 'cookies' fixture<commit_after> | # -*- coding: utf-8 -*-
def test_cookies_fixture(testdir):
"""Make sure that pytest accepts the `cookies` fixture."""
# create a temporary pytest test module
testdir.makepyfile("""
def test_valid_fixture(cookies):
assert hasattr(cookies, 'bake')
assert callable(cookies.bake)
assert callable(cookies.bake)
""")
# run pytest with the following cmd args
result = testdir.runpytest('-v')
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_valid_fixture PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
def test_help_message(testdir):
result = testdir.runpytest(
'--help',
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'cookies:',
])
| # -*- coding: utf-8 -*-
def test_bar_fixture(testdir):
"""Make sure that pytest accepts our fixture."""
# create a temporary pytest test module
testdir.makepyfile("""
def test_sth(bar):
assert bar == "europython2015"
""")
# run pytest with the following cmd args
result = testdir.runpytest(
'--foo=europython2015',
'-v'
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_sth PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
def test_help_message(testdir):
result = testdir.runpytest(
'--help',
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'cookies:',
'*--foo=DEST_FOO*Set the value for the fixture "bar".',
])
def test_hello_ini_setting(testdir):
testdir.makeini("""
[pytest]
HELLO = world
""")
testdir.makepyfile("""
import pytest
@pytest.fixture
def hello(request):
return request.config.getini('HELLO')
def test_hello_world(hello):
assert hello == 'world'
""")
result = testdir.runpytest('-v')
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_hello_world PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
Implement simple tests for the 'cookies' fixture# -*- coding: utf-8 -*-
def test_cookies_fixture(testdir):
"""Make sure that pytest accepts the `cookies` fixture."""
# create a temporary pytest test module
testdir.makepyfile("""
def test_valid_fixture(cookies):
assert hasattr(cookies, 'bake')
assert callable(cookies.bake)
assert callable(cookies.bake)
""")
# run pytest with the following cmd args
result = testdir.runpytest('-v')
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_valid_fixture PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
def test_help_message(testdir):
result = testdir.runpytest(
'--help',
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'cookies:',
])
| <commit_before># -*- coding: utf-8 -*-
def test_bar_fixture(testdir):
"""Make sure that pytest accepts our fixture."""
# create a temporary pytest test module
testdir.makepyfile("""
def test_sth(bar):
assert bar == "europython2015"
""")
# run pytest with the following cmd args
result = testdir.runpytest(
'--foo=europython2015',
'-v'
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_sth PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
def test_help_message(testdir):
result = testdir.runpytest(
'--help',
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'cookies:',
'*--foo=DEST_FOO*Set the value for the fixture "bar".',
])
def test_hello_ini_setting(testdir):
testdir.makeini("""
[pytest]
HELLO = world
""")
testdir.makepyfile("""
import pytest
@pytest.fixture
def hello(request):
return request.config.getini('HELLO')
def test_hello_world(hello):
assert hello == 'world'
""")
result = testdir.runpytest('-v')
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_hello_world PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
<commit_msg>Implement simple tests for the 'cookies' fixture<commit_after># -*- coding: utf-8 -*-
def test_cookies_fixture(testdir):
"""Make sure that pytest accepts the `cookies` fixture."""
# create a temporary pytest test module
testdir.makepyfile("""
def test_valid_fixture(cookies):
assert hasattr(cookies, 'bake')
assert callable(cookies.bake)
assert callable(cookies.bake)
""")
# run pytest with the following cmd args
result = testdir.runpytest('-v')
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'*::test_valid_fixture PASSED',
])
# make sure that that we get a '0' exit code for the testsuite
assert result.ret == 0
def test_help_message(testdir):
result = testdir.runpytest(
'--help',
)
# fnmatch_lines does an assertion internally
result.stdout.fnmatch_lines([
'cookies:',
])
|
134fcbd6e82957ac3abd2eebdc296fd4ccb457e9 | alexandria/api/books.py | alexandria/api/books.py | from . import app, mongo
from alexandria.decorators import *
from flask import request, jsonify, url_for, session
from flask.ext.classy import FlaskView, route
import json
from bson import json_util
class BooksView(FlaskView):
route_prefix = '/api/'
@authenticated
def index(self):
query = mongo.Books.find()
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
@authenticated
def genre(self, id):
query = mongo.Books.find({'genres':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
@authenticated
def author(self, id):
query = mongo.Books.find({'authors':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
BooksView.register(app)
| from . import app, mongo
from alexandria.decorators import *
from flask import request, jsonify, url_for, session
from flask.ext.classy import FlaskView, route
import json
from bson import json_util
class BooksView(FlaskView):
route_prefix = '/api/'
@authenticated
def index(self):
query = mongo.Books.find()
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
@authenticated
def genre(self, id):
query = mongo.Books.find({'genres':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
@authenticated
def author(self, id):
query = mongo.Books.find({'authors':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
BooksView.register(app)
| Set value of 'owner' to the value of the ObjectId | Set value of 'owner' to the value of the ObjectId
| Python | mit | citruspi/Alexandria,citruspi/Alexandria | from . import app, mongo
from alexandria.decorators import *
from flask import request, jsonify, url_for, session
from flask.ext.classy import FlaskView, route
import json
from bson import json_util
class BooksView(FlaskView):
route_prefix = '/api/'
@authenticated
def index(self):
query = mongo.Books.find()
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
@authenticated
def genre(self, id):
query = mongo.Books.find({'genres':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
@authenticated
def author(self, id):
query = mongo.Books.find({'authors':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
BooksView.register(app)
Set value of 'owner' to the value of the ObjectId | from . import app, mongo
from alexandria.decorators import *
from flask import request, jsonify, url_for, session
from flask.ext.classy import FlaskView, route
import json
from bson import json_util
class BooksView(FlaskView):
route_prefix = '/api/'
@authenticated
def index(self):
query = mongo.Books.find()
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
@authenticated
def genre(self, id):
query = mongo.Books.find({'genres':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
@authenticated
def author(self, id):
query = mongo.Books.find({'authors':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
BooksView.register(app)
| <commit_before>from . import app, mongo
from alexandria.decorators import *
from flask import request, jsonify, url_for, session
from flask.ext.classy import FlaskView, route
import json
from bson import json_util
class BooksView(FlaskView):
route_prefix = '/api/'
@authenticated
def index(self):
query = mongo.Books.find()
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
@authenticated
def genre(self, id):
query = mongo.Books.find({'genres':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
@authenticated
def author(self, id):
query = mongo.Books.find({'authors':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
BooksView.register(app)
<commit_msg>Set value of 'owner' to the value of the ObjectId<commit_after> | from . import app, mongo
from alexandria.decorators import *
from flask import request, jsonify, url_for, session
from flask.ext.classy import FlaskView, route
import json
from bson import json_util
class BooksView(FlaskView):
route_prefix = '/api/'
@authenticated
def index(self):
query = mongo.Books.find()
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
@authenticated
def genre(self, id):
query = mongo.Books.find({'genres':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
@authenticated
def author(self, id):
query = mongo.Books.find({'authors':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
BooksView.register(app)
| from . import app, mongo
from alexandria.decorators import *
from flask import request, jsonify, url_for, session
from flask.ext.classy import FlaskView, route
import json
from bson import json_util
class BooksView(FlaskView):
route_prefix = '/api/'
@authenticated
def index(self):
query = mongo.Books.find()
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
@authenticated
def genre(self, id):
query = mongo.Books.find({'genres':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
@authenticated
def author(self, id):
query = mongo.Books.find({'authors':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
BooksView.register(app)
Set value of 'owner' to the value of the ObjectIdfrom . import app, mongo
from alexandria.decorators import *
from flask import request, jsonify, url_for, session
from flask.ext.classy import FlaskView, route
import json
from bson import json_util
class BooksView(FlaskView):
route_prefix = '/api/'
@authenticated
def index(self):
query = mongo.Books.find()
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
@authenticated
def genre(self, id):
query = mongo.Books.find({'genres':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
@authenticated
def author(self, id):
query = mongo.Books.find({'authors':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
BooksView.register(app)
| <commit_before>from . import app, mongo
from alexandria.decorators import *
from flask import request, jsonify, url_for, session
from flask.ext.classy import FlaskView, route
import json
from bson import json_util
class BooksView(FlaskView):
route_prefix = '/api/'
@authenticated
def index(self):
query = mongo.Books.find()
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
@authenticated
def genre(self, id):
query = mongo.Books.find({'genres':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
@authenticated
def author(self, id):
query = mongo.Books.find({'authors':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
return jsonify(books=books)
BooksView.register(app)
<commit_msg>Set value of 'owner' to the value of the ObjectId<commit_after>from . import app, mongo
from alexandria.decorators import *
from flask import request, jsonify, url_for, session
from flask.ext.classy import FlaskView, route
import json
from bson import json_util
class BooksView(FlaskView):
route_prefix = '/api/'
@authenticated
def index(self):
query = mongo.Books.find()
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
@authenticated
def genre(self, id):
query = mongo.Books.find({'genres':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
@authenticated
def author(self, id):
query = mongo.Books.find({'authors':id})
books = json.loads(json_util.dumps(query, default=json_util.default))
for book in books:
book['id'] = book['_id']['$oid']
book.pop('_id')
book['owner'] = book['owner']['$oid']
return jsonify(books=books)
BooksView.register(app)
|
233e74abcc4a70f573e199074f5184b30bdfe1d2 | seam/__init__.py | seam/__init__.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
""" __init__.py
Seam
====
Seam is a simple layer between existing neuroimaging tools and your data.
While it is opinionated in how to execute tools, it makes no decisions
as to how data is organized or how the scripts are ultimately run. These
decisions are are up to you.
This is the main API.
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
__version__ = '0.0'
import freesurfer | #!/usr/bin/env python
# -*- coding: utf-8 -*-
""" __init__.py
Seam
====
Seam is a simple layer between existing neuroimaging tools and your data.
While it is opinionated in how to execute tools, it makes no decisions
as to how data is organized or how the scripts are ultimately run. These
decisions are are up to you.
This is the main API.
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
__version__ = '0.0'
from . import freesurfer
__all__ = ['freesurfer', ]
| Fix py3k relative import error | Fix py3k relative import error
| Python | mit | VUIIS/seam,VUIIS/seam | #!/usr/bin/env python
# -*- coding: utf-8 -*-
""" __init__.py
Seam
====
Seam is a simple layer between existing neuroimaging tools and your data.
While it is opinionated in how to execute tools, it makes no decisions
as to how data is organized or how the scripts are ultimately run. These
decisions are are up to you.
This is the main API.
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
__version__ = '0.0'
import freesurferFix py3k relative import error | #!/usr/bin/env python
# -*- coding: utf-8 -*-
""" __init__.py
Seam
====
Seam is a simple layer between existing neuroimaging tools and your data.
While it is opinionated in how to execute tools, it makes no decisions
as to how data is organized or how the scripts are ultimately run. These
decisions are are up to you.
This is the main API.
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
__version__ = '0.0'
from . import freesurfer
__all__ = ['freesurfer', ]
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
""" __init__.py
Seam
====
Seam is a simple layer between existing neuroimaging tools and your data.
While it is opinionated in how to execute tools, it makes no decisions
as to how data is organized or how the scripts are ultimately run. These
decisions are are up to you.
This is the main API.
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
__version__ = '0.0'
import freesurfer<commit_msg>Fix py3k relative import error<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
""" __init__.py
Seam
====
Seam is a simple layer between existing neuroimaging tools and your data.
While it is opinionated in how to execute tools, it makes no decisions
as to how data is organized or how the scripts are ultimately run. These
decisions are are up to you.
This is the main API.
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
__version__ = '0.0'
from . import freesurfer
__all__ = ['freesurfer', ]
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
""" __init__.py
Seam
====
Seam is a simple layer between existing neuroimaging tools and your data.
While it is opinionated in how to execute tools, it makes no decisions
as to how data is organized or how the scripts are ultimately run. These
decisions are are up to you.
This is the main API.
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
__version__ = '0.0'
import freesurferFix py3k relative import error#!/usr/bin/env python
# -*- coding: utf-8 -*-
""" __init__.py
Seam
====
Seam is a simple layer between existing neuroimaging tools and your data.
While it is opinionated in how to execute tools, it makes no decisions
as to how data is organized or how the scripts are ultimately run. These
decisions are are up to you.
This is the main API.
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
__version__ = '0.0'
from . import freesurfer
__all__ = ['freesurfer', ]
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
""" __init__.py
Seam
====
Seam is a simple layer between existing neuroimaging tools and your data.
While it is opinionated in how to execute tools, it makes no decisions
as to how data is organized or how the scripts are ultimately run. These
decisions are are up to you.
This is the main API.
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
__version__ = '0.0'
import freesurfer<commit_msg>Fix py3k relative import error<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
""" __init__.py
Seam
====
Seam is a simple layer between existing neuroimaging tools and your data.
While it is opinionated in how to execute tools, it makes no decisions
as to how data is organized or how the scripts are ultimately run. These
decisions are are up to you.
This is the main API.
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
__version__ = '0.0'
from . import freesurfer
__all__ = ['freesurfer', ]
|
b193df5080cc8076739509523cf391f5b7132d56 | kerastuner/utils.py | kerastuner/utils.py | # Copyright 2019 The Keras Tuner Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import gc
import numpy as np
from tensorflow.python import Session, ConfigProto
from tensorflow.python.keras import backend as K
def compute_model_size(model):
"comput the size of a given model"
params = [K.count_params(p) for p in set(model.trainable_weights)]
return int(np.sum(params))
def clear_tf_session():
"Clear tensorflow graph to avoid OOM issues"
K.clear_session()
# K.get_session().close() # unsure if it is needed
gc.collect()
if hasattr(K, 'set_session'):
cfg = ConfigProto()
cfg.gpu_options.allow_growth = True # pylint: disable=no-member
K.set_session(Session(config=cfg))
| # Copyright 2019 The Keras Tuner Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import gc
import numpy as np
from tensorflow.python import Session, ConfigProto
from tensorflow.python.keras import backend as K
def compute_model_size(model):
"comput the size of a given model"
params = [K.count_params(p) for p in model.trainable_weights]
return int(np.sum(params))
def clear_tf_session():
"Clear tensorflow graph to avoid OOM issues"
K.clear_session()
# K.get_session().close() # unsure if it is needed
gc.collect()
if hasattr(K, 'set_session'):
cfg = ConfigProto()
cfg.gpu_options.allow_growth = True # pylint: disable=no-member
K.set_session(Session(config=cfg))
| Fix bugs for updating from tf 2.0b0 to rc0 | Fix bugs for updating from tf 2.0b0 to rc0
Updating the dependency of tensorflow from 2.0.0b1 to 2.0.0rc0 is causing crashes in keras-tuner.
Not sure the "set" operation is really useful, but it is causing the crash because it tries to hash a tensor with equality enabled.
```
../../.virtualenvs/ak/lib/python3.6/site-packages/kerastuner/utils.py:23: in compute_model_size
params = [K.count_params(p) for p in set(model.trainable_weights)]
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <tf.Variable 'dense/kernel:0' shape=(32, 32) dtype=float32, numpy=
array([[ 0.03147817, -0.30117437, -0.08552131, ...,...],
[ 0.23273078, -0.10847142, 0.0151484 , ..., 0.16468436,
-0.21210298, -0.1087181 ]], dtype=float32)>
def __hash__(self):
if ops.Tensor._USE_EQUALITY and ops.executing_eagerly_outside_functions(): # pylint: disable=protected-access
> raise TypeError("Variable is unhashable if Tensor equality is enabled. "
"Instead, use tensor.experimental_ref() as the key.")
E TypeError: Variable is unhashable if Tensor equality is enabled. Instead, use tensor.experimental_ref() as the key.
../../.virtualenvs/ak/lib/python3.6/site-packages/tensorflow_core/python/ops/variables.py:1085: TypeError
``` | Python | apache-2.0 | keras-team/keras-tuner,keras-team/keras-tuner | # Copyright 2019 The Keras Tuner Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import gc
import numpy as np
from tensorflow.python import Session, ConfigProto
from tensorflow.python.keras import backend as K
def compute_model_size(model):
"comput the size of a given model"
params = [K.count_params(p) for p in set(model.trainable_weights)]
return int(np.sum(params))
def clear_tf_session():
"Clear tensorflow graph to avoid OOM issues"
K.clear_session()
# K.get_session().close() # unsure if it is needed
gc.collect()
if hasattr(K, 'set_session'):
cfg = ConfigProto()
cfg.gpu_options.allow_growth = True # pylint: disable=no-member
K.set_session(Session(config=cfg))
Fix bugs for updating from tf 2.0b0 to rc0
Updating the dependency of tensorflow from 2.0.0b1 to 2.0.0rc0 is causing crashes in keras-tuner.
Not sure the "set" operation is really useful, but it is causing the crash because it tries to hash a tensor with equality enabled.
```
../../.virtualenvs/ak/lib/python3.6/site-packages/kerastuner/utils.py:23: in compute_model_size
params = [K.count_params(p) for p in set(model.trainable_weights)]
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <tf.Variable 'dense/kernel:0' shape=(32, 32) dtype=float32, numpy=
array([[ 0.03147817, -0.30117437, -0.08552131, ...,...],
[ 0.23273078, -0.10847142, 0.0151484 , ..., 0.16468436,
-0.21210298, -0.1087181 ]], dtype=float32)>
def __hash__(self):
if ops.Tensor._USE_EQUALITY and ops.executing_eagerly_outside_functions(): # pylint: disable=protected-access
> raise TypeError("Variable is unhashable if Tensor equality is enabled. "
"Instead, use tensor.experimental_ref() as the key.")
E TypeError: Variable is unhashable if Tensor equality is enabled. Instead, use tensor.experimental_ref() as the key.
../../.virtualenvs/ak/lib/python3.6/site-packages/tensorflow_core/python/ops/variables.py:1085: TypeError
``` | # Copyright 2019 The Keras Tuner Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import gc
import numpy as np
from tensorflow.python import Session, ConfigProto
from tensorflow.python.keras import backend as K
def compute_model_size(model):
"comput the size of a given model"
params = [K.count_params(p) for p in model.trainable_weights]
return int(np.sum(params))
def clear_tf_session():
"Clear tensorflow graph to avoid OOM issues"
K.clear_session()
# K.get_session().close() # unsure if it is needed
gc.collect()
if hasattr(K, 'set_session'):
cfg = ConfigProto()
cfg.gpu_options.allow_growth = True # pylint: disable=no-member
K.set_session(Session(config=cfg))
| <commit_before># Copyright 2019 The Keras Tuner Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import gc
import numpy as np
from tensorflow.python import Session, ConfigProto
from tensorflow.python.keras import backend as K
def compute_model_size(model):
"comput the size of a given model"
params = [K.count_params(p) for p in set(model.trainable_weights)]
return int(np.sum(params))
def clear_tf_session():
"Clear tensorflow graph to avoid OOM issues"
K.clear_session()
# K.get_session().close() # unsure if it is needed
gc.collect()
if hasattr(K, 'set_session'):
cfg = ConfigProto()
cfg.gpu_options.allow_growth = True # pylint: disable=no-member
K.set_session(Session(config=cfg))
<commit_msg>Fix bugs for updating from tf 2.0b0 to rc0
Updating the dependency of tensorflow from 2.0.0b1 to 2.0.0rc0 is causing crashes in keras-tuner.
Not sure the "set" operation is really useful, but it is causing the crash because it tries to hash a tensor with equality enabled.
```
../../.virtualenvs/ak/lib/python3.6/site-packages/kerastuner/utils.py:23: in compute_model_size
params = [K.count_params(p) for p in set(model.trainable_weights)]
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <tf.Variable 'dense/kernel:0' shape=(32, 32) dtype=float32, numpy=
array([[ 0.03147817, -0.30117437, -0.08552131, ...,...],
[ 0.23273078, -0.10847142, 0.0151484 , ..., 0.16468436,
-0.21210298, -0.1087181 ]], dtype=float32)>
def __hash__(self):
if ops.Tensor._USE_EQUALITY and ops.executing_eagerly_outside_functions(): # pylint: disable=protected-access
> raise TypeError("Variable is unhashable if Tensor equality is enabled. "
"Instead, use tensor.experimental_ref() as the key.")
E TypeError: Variable is unhashable if Tensor equality is enabled. Instead, use tensor.experimental_ref() as the key.
../../.virtualenvs/ak/lib/python3.6/site-packages/tensorflow_core/python/ops/variables.py:1085: TypeError
```<commit_after> | # Copyright 2019 The Keras Tuner Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import gc
import numpy as np
from tensorflow.python import Session, ConfigProto
from tensorflow.python.keras import backend as K
def compute_model_size(model):
"comput the size of a given model"
params = [K.count_params(p) for p in model.trainable_weights]
return int(np.sum(params))
def clear_tf_session():
"Clear tensorflow graph to avoid OOM issues"
K.clear_session()
# K.get_session().close() # unsure if it is needed
gc.collect()
if hasattr(K, 'set_session'):
cfg = ConfigProto()
cfg.gpu_options.allow_growth = True # pylint: disable=no-member
K.set_session(Session(config=cfg))
| # Copyright 2019 The Keras Tuner Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import gc
import numpy as np
from tensorflow.python import Session, ConfigProto
from tensorflow.python.keras import backend as K
def compute_model_size(model):
"comput the size of a given model"
params = [K.count_params(p) for p in set(model.trainable_weights)]
return int(np.sum(params))
def clear_tf_session():
"Clear tensorflow graph to avoid OOM issues"
K.clear_session()
# K.get_session().close() # unsure if it is needed
gc.collect()
if hasattr(K, 'set_session'):
cfg = ConfigProto()
cfg.gpu_options.allow_growth = True # pylint: disable=no-member
K.set_session(Session(config=cfg))
Fix bugs for updating from tf 2.0b0 to rc0
Updating the dependency of tensorflow from 2.0.0b1 to 2.0.0rc0 is causing crashes in keras-tuner.
Not sure the "set" operation is really useful, but it is causing the crash because it tries to hash a tensor with equality enabled.
```
../../.virtualenvs/ak/lib/python3.6/site-packages/kerastuner/utils.py:23: in compute_model_size
params = [K.count_params(p) for p in set(model.trainable_weights)]
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <tf.Variable 'dense/kernel:0' shape=(32, 32) dtype=float32, numpy=
array([[ 0.03147817, -0.30117437, -0.08552131, ...,...],
[ 0.23273078, -0.10847142, 0.0151484 , ..., 0.16468436,
-0.21210298, -0.1087181 ]], dtype=float32)>
def __hash__(self):
if ops.Tensor._USE_EQUALITY and ops.executing_eagerly_outside_functions(): # pylint: disable=protected-access
> raise TypeError("Variable is unhashable if Tensor equality is enabled. "
"Instead, use tensor.experimental_ref() as the key.")
E TypeError: Variable is unhashable if Tensor equality is enabled. Instead, use tensor.experimental_ref() as the key.
../../.virtualenvs/ak/lib/python3.6/site-packages/tensorflow_core/python/ops/variables.py:1085: TypeError
```# Copyright 2019 The Keras Tuner Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import gc
import numpy as np
from tensorflow.python import Session, ConfigProto
from tensorflow.python.keras import backend as K
def compute_model_size(model):
"comput the size of a given model"
params = [K.count_params(p) for p in model.trainable_weights]
return int(np.sum(params))
def clear_tf_session():
"Clear tensorflow graph to avoid OOM issues"
K.clear_session()
# K.get_session().close() # unsure if it is needed
gc.collect()
if hasattr(K, 'set_session'):
cfg = ConfigProto()
cfg.gpu_options.allow_growth = True # pylint: disable=no-member
K.set_session(Session(config=cfg))
| <commit_before># Copyright 2019 The Keras Tuner Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import gc
import numpy as np
from tensorflow.python import Session, ConfigProto
from tensorflow.python.keras import backend as K
def compute_model_size(model):
"comput the size of a given model"
params = [K.count_params(p) for p in set(model.trainable_weights)]
return int(np.sum(params))
def clear_tf_session():
"Clear tensorflow graph to avoid OOM issues"
K.clear_session()
# K.get_session().close() # unsure if it is needed
gc.collect()
if hasattr(K, 'set_session'):
cfg = ConfigProto()
cfg.gpu_options.allow_growth = True # pylint: disable=no-member
K.set_session(Session(config=cfg))
<commit_msg>Fix bugs for updating from tf 2.0b0 to rc0
Updating the dependency of tensorflow from 2.0.0b1 to 2.0.0rc0 is causing crashes in keras-tuner.
Not sure the "set" operation is really useful, but it is causing the crash because it tries to hash a tensor with equality enabled.
```
../../.virtualenvs/ak/lib/python3.6/site-packages/kerastuner/utils.py:23: in compute_model_size
params = [K.count_params(p) for p in set(model.trainable_weights)]
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <tf.Variable 'dense/kernel:0' shape=(32, 32) dtype=float32, numpy=
array([[ 0.03147817, -0.30117437, -0.08552131, ...,...],
[ 0.23273078, -0.10847142, 0.0151484 , ..., 0.16468436,
-0.21210298, -0.1087181 ]], dtype=float32)>
def __hash__(self):
if ops.Tensor._USE_EQUALITY and ops.executing_eagerly_outside_functions(): # pylint: disable=protected-access
> raise TypeError("Variable is unhashable if Tensor equality is enabled. "
"Instead, use tensor.experimental_ref() as the key.")
E TypeError: Variable is unhashable if Tensor equality is enabled. Instead, use tensor.experimental_ref() as the key.
../../.virtualenvs/ak/lib/python3.6/site-packages/tensorflow_core/python/ops/variables.py:1085: TypeError
```<commit_after># Copyright 2019 The Keras Tuner Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import gc
import numpy as np
from tensorflow.python import Session, ConfigProto
from tensorflow.python.keras import backend as K
def compute_model_size(model):
"comput the size of a given model"
params = [K.count_params(p) for p in model.trainable_weights]
return int(np.sum(params))
def clear_tf_session():
"Clear tensorflow graph to avoid OOM issues"
K.clear_session()
# K.get_session().close() # unsure if it is needed
gc.collect()
if hasattr(K, 'set_session'):
cfg = ConfigProto()
cfg.gpu_options.allow_growth = True # pylint: disable=no-member
K.set_session(Session(config=cfg))
|
3a21137a09c58612044580f5835be1bbe2765500 | markdown-pp.py | markdown-pp.py | #!/usr/bin/env python
# Copyright (C) 2010 John Reese
# Licensed under the MIT license
import sys
import MarkdownPP
if len(sys.argv) > 2:
mdpp = open(sys.argv[1], "r")
md = open(sys.argv[2], "w")
elif len(sys.argv) > 1:
mdpp = open(sys.argv[1], "r")
md = sys.stdout
else:
sys.exit(1)
MarkdownPP.MarkdownPP(input=mdpp, output=md, modules=MarkdownPP.modules.keys())
mdpp.close()
md.close()
| #!/usr/bin/env python2
# Copyright (C) 2010 John Reese
# Licensed under the MIT license
import sys
import MarkdownPP
if len(sys.argv) > 2:
mdpp = open(sys.argv[1], "r")
md = open(sys.argv[2], "w")
elif len(sys.argv) > 1:
mdpp = open(sys.argv[1], "r")
md = sys.stdout
else:
sys.exit(1)
MarkdownPP.MarkdownPP(input=mdpp, output=md, modules=MarkdownPP.modules.keys())
mdpp.close()
md.close()
| Allow specifying depth of TOC headers and rewrites | Allow specifying depth of TOC headers and rewrites
| Python | mit | triAGENS/markdown-pp,jreese/markdown-pp | #!/usr/bin/env python
# Copyright (C) 2010 John Reese
# Licensed under the MIT license
import sys
import MarkdownPP
if len(sys.argv) > 2:
mdpp = open(sys.argv[1], "r")
md = open(sys.argv[2], "w")
elif len(sys.argv) > 1:
mdpp = open(sys.argv[1], "r")
md = sys.stdout
else:
sys.exit(1)
MarkdownPP.MarkdownPP(input=mdpp, output=md, modules=MarkdownPP.modules.keys())
mdpp.close()
md.close()
Allow specifying depth of TOC headers and rewrites | #!/usr/bin/env python2
# Copyright (C) 2010 John Reese
# Licensed under the MIT license
import sys
import MarkdownPP
if len(sys.argv) > 2:
mdpp = open(sys.argv[1], "r")
md = open(sys.argv[2], "w")
elif len(sys.argv) > 1:
mdpp = open(sys.argv[1], "r")
md = sys.stdout
else:
sys.exit(1)
MarkdownPP.MarkdownPP(input=mdpp, output=md, modules=MarkdownPP.modules.keys())
mdpp.close()
md.close()
| <commit_before>#!/usr/bin/env python
# Copyright (C) 2010 John Reese
# Licensed under the MIT license
import sys
import MarkdownPP
if len(sys.argv) > 2:
mdpp = open(sys.argv[1], "r")
md = open(sys.argv[2], "w")
elif len(sys.argv) > 1:
mdpp = open(sys.argv[1], "r")
md = sys.stdout
else:
sys.exit(1)
MarkdownPP.MarkdownPP(input=mdpp, output=md, modules=MarkdownPP.modules.keys())
mdpp.close()
md.close()
<commit_msg>Allow specifying depth of TOC headers and rewrites<commit_after> | #!/usr/bin/env python2
# Copyright (C) 2010 John Reese
# Licensed under the MIT license
import sys
import MarkdownPP
if len(sys.argv) > 2:
mdpp = open(sys.argv[1], "r")
md = open(sys.argv[2], "w")
elif len(sys.argv) > 1:
mdpp = open(sys.argv[1], "r")
md = sys.stdout
else:
sys.exit(1)
MarkdownPP.MarkdownPP(input=mdpp, output=md, modules=MarkdownPP.modules.keys())
mdpp.close()
md.close()
| #!/usr/bin/env python
# Copyright (C) 2010 John Reese
# Licensed under the MIT license
import sys
import MarkdownPP
if len(sys.argv) > 2:
mdpp = open(sys.argv[1], "r")
md = open(sys.argv[2], "w")
elif len(sys.argv) > 1:
mdpp = open(sys.argv[1], "r")
md = sys.stdout
else:
sys.exit(1)
MarkdownPP.MarkdownPP(input=mdpp, output=md, modules=MarkdownPP.modules.keys())
mdpp.close()
md.close()
Allow specifying depth of TOC headers and rewrites#!/usr/bin/env python2
# Copyright (C) 2010 John Reese
# Licensed under the MIT license
import sys
import MarkdownPP
if len(sys.argv) > 2:
mdpp = open(sys.argv[1], "r")
md = open(sys.argv[2], "w")
elif len(sys.argv) > 1:
mdpp = open(sys.argv[1], "r")
md = sys.stdout
else:
sys.exit(1)
MarkdownPP.MarkdownPP(input=mdpp, output=md, modules=MarkdownPP.modules.keys())
mdpp.close()
md.close()
| <commit_before>#!/usr/bin/env python
# Copyright (C) 2010 John Reese
# Licensed under the MIT license
import sys
import MarkdownPP
if len(sys.argv) > 2:
mdpp = open(sys.argv[1], "r")
md = open(sys.argv[2], "w")
elif len(sys.argv) > 1:
mdpp = open(sys.argv[1], "r")
md = sys.stdout
else:
sys.exit(1)
MarkdownPP.MarkdownPP(input=mdpp, output=md, modules=MarkdownPP.modules.keys())
mdpp.close()
md.close()
<commit_msg>Allow specifying depth of TOC headers and rewrites<commit_after>#!/usr/bin/env python2
# Copyright (C) 2010 John Reese
# Licensed under the MIT license
import sys
import MarkdownPP
if len(sys.argv) > 2:
mdpp = open(sys.argv[1], "r")
md = open(sys.argv[2], "w")
elif len(sys.argv) > 1:
mdpp = open(sys.argv[1], "r")
md = sys.stdout
else:
sys.exit(1)
MarkdownPP.MarkdownPP(input=mdpp, output=md, modules=MarkdownPP.modules.keys())
mdpp.close()
md.close()
|
418e7a7d8c8261578df046d251041ab0794d1580 | decorators.py | decorators.py | #!/usr/bin/env python
class RequiresType(object):
"""
Checks that the first (or position given by the keyword argument 'position'
argument to the function is an instance of one of the types given in the
positional decorator arguments
"""
def __init__(self, *types, **kwargs):
self.types = types
self.position = 0
self.returnvalue = False
if 'position' in kwargs:
self.position = int(kwargs['position']) - 1
if 'returnvalue' in kwargs:
self.returnvalue = kwargs['returnvalue']
def __call__(self, f):
def wrapped_f(*args, **kwargs):
if type(args[self.position]) not in self.types:
return self.returnvalue
return f(*args, **kwargs)
return wrapped_f
| #!/usr/bin/env python
class RequiresType(object):
"""
Checks that the first (or position given by the keyword argument 'position'
argument to the function is an instance of one of the types given in the
positional decorator arguments
"""
def __init__(self, *types, **kwargs):
self.types = types
self.position = 0
if 'position' in kwargs:
self.position = int(kwargs['position']) - 1
def __call__(self, f):
def wrapped_f(*args, **kwargs):
if type(args[self.position]) not in self.types:
raise TypeError("Invalid argument type '%s' at position %d. " +
"Expected one of (%s)" % (
type(args[self.position]).__name__, self.position,
", ".join([t.__name__ for t in self.types])))
return f(*args, **kwargs)
return wrapped_f
| Raise TypeError instead of returning | Raise TypeError instead of returning
| Python | bsd-3-clause | rasher/reddit-modbot | #!/usr/bin/env python
class RequiresType(object):
"""
Checks that the first (or position given by the keyword argument 'position'
argument to the function is an instance of one of the types given in the
positional decorator arguments
"""
def __init__(self, *types, **kwargs):
self.types = types
self.position = 0
self.returnvalue = False
if 'position' in kwargs:
self.position = int(kwargs['position']) - 1
if 'returnvalue' in kwargs:
self.returnvalue = kwargs['returnvalue']
def __call__(self, f):
def wrapped_f(*args, **kwargs):
if type(args[self.position]) not in self.types:
return self.returnvalue
return f(*args, **kwargs)
return wrapped_f
Raise TypeError instead of returning | #!/usr/bin/env python
class RequiresType(object):
"""
Checks that the first (or position given by the keyword argument 'position'
argument to the function is an instance of one of the types given in the
positional decorator arguments
"""
def __init__(self, *types, **kwargs):
self.types = types
self.position = 0
if 'position' in kwargs:
self.position = int(kwargs['position']) - 1
def __call__(self, f):
def wrapped_f(*args, **kwargs):
if type(args[self.position]) not in self.types:
raise TypeError("Invalid argument type '%s' at position %d. " +
"Expected one of (%s)" % (
type(args[self.position]).__name__, self.position,
", ".join([t.__name__ for t in self.types])))
return f(*args, **kwargs)
return wrapped_f
| <commit_before>#!/usr/bin/env python
class RequiresType(object):
"""
Checks that the first (or position given by the keyword argument 'position'
argument to the function is an instance of one of the types given in the
positional decorator arguments
"""
def __init__(self, *types, **kwargs):
self.types = types
self.position = 0
self.returnvalue = False
if 'position' in kwargs:
self.position = int(kwargs['position']) - 1
if 'returnvalue' in kwargs:
self.returnvalue = kwargs['returnvalue']
def __call__(self, f):
def wrapped_f(*args, **kwargs):
if type(args[self.position]) not in self.types:
return self.returnvalue
return f(*args, **kwargs)
return wrapped_f
<commit_msg>Raise TypeError instead of returning<commit_after> | #!/usr/bin/env python
class RequiresType(object):
"""
Checks that the first (or position given by the keyword argument 'position'
argument to the function is an instance of one of the types given in the
positional decorator arguments
"""
def __init__(self, *types, **kwargs):
self.types = types
self.position = 0
if 'position' in kwargs:
self.position = int(kwargs['position']) - 1
def __call__(self, f):
def wrapped_f(*args, **kwargs):
if type(args[self.position]) not in self.types:
raise TypeError("Invalid argument type '%s' at position %d. " +
"Expected one of (%s)" % (
type(args[self.position]).__name__, self.position,
", ".join([t.__name__ for t in self.types])))
return f(*args, **kwargs)
return wrapped_f
| #!/usr/bin/env python
class RequiresType(object):
"""
Checks that the first (or position given by the keyword argument 'position'
argument to the function is an instance of one of the types given in the
positional decorator arguments
"""
def __init__(self, *types, **kwargs):
self.types = types
self.position = 0
self.returnvalue = False
if 'position' in kwargs:
self.position = int(kwargs['position']) - 1
if 'returnvalue' in kwargs:
self.returnvalue = kwargs['returnvalue']
def __call__(self, f):
def wrapped_f(*args, **kwargs):
if type(args[self.position]) not in self.types:
return self.returnvalue
return f(*args, **kwargs)
return wrapped_f
Raise TypeError instead of returning#!/usr/bin/env python
class RequiresType(object):
"""
Checks that the first (or position given by the keyword argument 'position'
argument to the function is an instance of one of the types given in the
positional decorator arguments
"""
def __init__(self, *types, **kwargs):
self.types = types
self.position = 0
if 'position' in kwargs:
self.position = int(kwargs['position']) - 1
def __call__(self, f):
def wrapped_f(*args, **kwargs):
if type(args[self.position]) not in self.types:
raise TypeError("Invalid argument type '%s' at position %d. " +
"Expected one of (%s)" % (
type(args[self.position]).__name__, self.position,
", ".join([t.__name__ for t in self.types])))
return f(*args, **kwargs)
return wrapped_f
| <commit_before>#!/usr/bin/env python
class RequiresType(object):
"""
Checks that the first (or position given by the keyword argument 'position'
argument to the function is an instance of one of the types given in the
positional decorator arguments
"""
def __init__(self, *types, **kwargs):
self.types = types
self.position = 0
self.returnvalue = False
if 'position' in kwargs:
self.position = int(kwargs['position']) - 1
if 'returnvalue' in kwargs:
self.returnvalue = kwargs['returnvalue']
def __call__(self, f):
def wrapped_f(*args, **kwargs):
if type(args[self.position]) not in self.types:
return self.returnvalue
return f(*args, **kwargs)
return wrapped_f
<commit_msg>Raise TypeError instead of returning<commit_after>#!/usr/bin/env python
class RequiresType(object):
"""
Checks that the first (or position given by the keyword argument 'position'
argument to the function is an instance of one of the types given in the
positional decorator arguments
"""
def __init__(self, *types, **kwargs):
self.types = types
self.position = 0
if 'position' in kwargs:
self.position = int(kwargs['position']) - 1
def __call__(self, f):
def wrapped_f(*args, **kwargs):
if type(args[self.position]) not in self.types:
raise TypeError("Invalid argument type '%s' at position %d. " +
"Expected one of (%s)" % (
type(args[self.position]).__name__, self.position,
", ".join([t.__name__ for t in self.types])))
return f(*args, **kwargs)
return wrapped_f
|
94d54e20fe5590fad0449bef79366654b3c7f23d | swingtime/urls.py | swingtime/urls.py | from django.conf.urls import patterns, url
from swingtime import views
urlpatterns = patterns('',
url(
r'^(?:calendar/)?$',
views.CalendarView.as_view(),
name='swingtime-calendar'
),
url(
r'^calendar/json/$',
views.CalendarJSONView.as_view(),
name='swingtime-calendar-json'
),
url(
r'^calendar/(?P<year>\d{4})/$',
views.CalendarView.as_view(),
name='swingtime-yearly-view'
),
url(
r'^calendar/(\d{4})/(0?[1-9]|1[012])/$',
views.CalendarView.as_view(),
name='swingtime-monthly-view'
),
url(
r'^agenda/$',
views.AgendaView.as_view(),
name='swingtime-events'
),
url(
r'^event/(\d+)/$',
views.event_view,
name='swingtime-event'
),
url(
r'^event/(\d+)/occurrence/(\d+)/$',
views.occurrence_view,
name='swingtime-occurrence'
),
)
| from django.conf.urls import patterns, url
from swingtime import views
urlpatterns = patterns('',
url(
r'^(?:calendar/)?$',
views.CalendarView.as_view(),
name='swingtime-calendar'
),
url(
r'^calendar.json$',
views.CalendarJSONView.as_view(),
name='swingtime-calendar-json'
),
url(
r'^calendar/(?P<year>\d{4})/$',
views.CalendarView.as_view(),
name='swingtime-yearly-view'
),
url(
r'^calendar/(\d{4})/(0?[1-9]|1[012])/$',
views.CalendarView.as_view(),
name='swingtime-monthly-view'
),
url(
r'^agenda/$',
views.AgendaView.as_view(),
name='swingtime-events'
),
url(
r'^event/(\d+)/$',
views.event_view,
name='swingtime-event'
),
url(
r'^event/(\d+)/occurrence/(\d+)/$',
views.occurrence_view,
name='swingtime-occurrence'
),
)
| Change calendar JSON view url | Change calendar JSON view url
| Python | mit | jonge-democraten/mezzanine-fullcalendar | from django.conf.urls import patterns, url
from swingtime import views
urlpatterns = patterns('',
url(
r'^(?:calendar/)?$',
views.CalendarView.as_view(),
name='swingtime-calendar'
),
url(
r'^calendar/json/$',
views.CalendarJSONView.as_view(),
name='swingtime-calendar-json'
),
url(
r'^calendar/(?P<year>\d{4})/$',
views.CalendarView.as_view(),
name='swingtime-yearly-view'
),
url(
r'^calendar/(\d{4})/(0?[1-9]|1[012])/$',
views.CalendarView.as_view(),
name='swingtime-monthly-view'
),
url(
r'^agenda/$',
views.AgendaView.as_view(),
name='swingtime-events'
),
url(
r'^event/(\d+)/$',
views.event_view,
name='swingtime-event'
),
url(
r'^event/(\d+)/occurrence/(\d+)/$',
views.occurrence_view,
name='swingtime-occurrence'
),
)
Change calendar JSON view url | from django.conf.urls import patterns, url
from swingtime import views
urlpatterns = patterns('',
url(
r'^(?:calendar/)?$',
views.CalendarView.as_view(),
name='swingtime-calendar'
),
url(
r'^calendar.json$',
views.CalendarJSONView.as_view(),
name='swingtime-calendar-json'
),
url(
r'^calendar/(?P<year>\d{4})/$',
views.CalendarView.as_view(),
name='swingtime-yearly-view'
),
url(
r'^calendar/(\d{4})/(0?[1-9]|1[012])/$',
views.CalendarView.as_view(),
name='swingtime-monthly-view'
),
url(
r'^agenda/$',
views.AgendaView.as_view(),
name='swingtime-events'
),
url(
r'^event/(\d+)/$',
views.event_view,
name='swingtime-event'
),
url(
r'^event/(\d+)/occurrence/(\d+)/$',
views.occurrence_view,
name='swingtime-occurrence'
),
)
| <commit_before>from django.conf.urls import patterns, url
from swingtime import views
urlpatterns = patterns('',
url(
r'^(?:calendar/)?$',
views.CalendarView.as_view(),
name='swingtime-calendar'
),
url(
r'^calendar/json/$',
views.CalendarJSONView.as_view(),
name='swingtime-calendar-json'
),
url(
r'^calendar/(?P<year>\d{4})/$',
views.CalendarView.as_view(),
name='swingtime-yearly-view'
),
url(
r'^calendar/(\d{4})/(0?[1-9]|1[012])/$',
views.CalendarView.as_view(),
name='swingtime-monthly-view'
),
url(
r'^agenda/$',
views.AgendaView.as_view(),
name='swingtime-events'
),
url(
r'^event/(\d+)/$',
views.event_view,
name='swingtime-event'
),
url(
r'^event/(\d+)/occurrence/(\d+)/$',
views.occurrence_view,
name='swingtime-occurrence'
),
)
<commit_msg>Change calendar JSON view url<commit_after> | from django.conf.urls import patterns, url
from swingtime import views
urlpatterns = patterns('',
url(
r'^(?:calendar/)?$',
views.CalendarView.as_view(),
name='swingtime-calendar'
),
url(
r'^calendar.json$',
views.CalendarJSONView.as_view(),
name='swingtime-calendar-json'
),
url(
r'^calendar/(?P<year>\d{4})/$',
views.CalendarView.as_view(),
name='swingtime-yearly-view'
),
url(
r'^calendar/(\d{4})/(0?[1-9]|1[012])/$',
views.CalendarView.as_view(),
name='swingtime-monthly-view'
),
url(
r'^agenda/$',
views.AgendaView.as_view(),
name='swingtime-events'
),
url(
r'^event/(\d+)/$',
views.event_view,
name='swingtime-event'
),
url(
r'^event/(\d+)/occurrence/(\d+)/$',
views.occurrence_view,
name='swingtime-occurrence'
),
)
| from django.conf.urls import patterns, url
from swingtime import views
urlpatterns = patterns('',
url(
r'^(?:calendar/)?$',
views.CalendarView.as_view(),
name='swingtime-calendar'
),
url(
r'^calendar/json/$',
views.CalendarJSONView.as_view(),
name='swingtime-calendar-json'
),
url(
r'^calendar/(?P<year>\d{4})/$',
views.CalendarView.as_view(),
name='swingtime-yearly-view'
),
url(
r'^calendar/(\d{4})/(0?[1-9]|1[012])/$',
views.CalendarView.as_view(),
name='swingtime-monthly-view'
),
url(
r'^agenda/$',
views.AgendaView.as_view(),
name='swingtime-events'
),
url(
r'^event/(\d+)/$',
views.event_view,
name='swingtime-event'
),
url(
r'^event/(\d+)/occurrence/(\d+)/$',
views.occurrence_view,
name='swingtime-occurrence'
),
)
Change calendar JSON view urlfrom django.conf.urls import patterns, url
from swingtime import views
urlpatterns = patterns('',
url(
r'^(?:calendar/)?$',
views.CalendarView.as_view(),
name='swingtime-calendar'
),
url(
r'^calendar.json$',
views.CalendarJSONView.as_view(),
name='swingtime-calendar-json'
),
url(
r'^calendar/(?P<year>\d{4})/$',
views.CalendarView.as_view(),
name='swingtime-yearly-view'
),
url(
r'^calendar/(\d{4})/(0?[1-9]|1[012])/$',
views.CalendarView.as_view(),
name='swingtime-monthly-view'
),
url(
r'^agenda/$',
views.AgendaView.as_view(),
name='swingtime-events'
),
url(
r'^event/(\d+)/$',
views.event_view,
name='swingtime-event'
),
url(
r'^event/(\d+)/occurrence/(\d+)/$',
views.occurrence_view,
name='swingtime-occurrence'
),
)
| <commit_before>from django.conf.urls import patterns, url
from swingtime import views
urlpatterns = patterns('',
url(
r'^(?:calendar/)?$',
views.CalendarView.as_view(),
name='swingtime-calendar'
),
url(
r'^calendar/json/$',
views.CalendarJSONView.as_view(),
name='swingtime-calendar-json'
),
url(
r'^calendar/(?P<year>\d{4})/$',
views.CalendarView.as_view(),
name='swingtime-yearly-view'
),
url(
r'^calendar/(\d{4})/(0?[1-9]|1[012])/$',
views.CalendarView.as_view(),
name='swingtime-monthly-view'
),
url(
r'^agenda/$',
views.AgendaView.as_view(),
name='swingtime-events'
),
url(
r'^event/(\d+)/$',
views.event_view,
name='swingtime-event'
),
url(
r'^event/(\d+)/occurrence/(\d+)/$',
views.occurrence_view,
name='swingtime-occurrence'
),
)
<commit_msg>Change calendar JSON view url<commit_after>from django.conf.urls import patterns, url
from swingtime import views
urlpatterns = patterns('',
url(
r'^(?:calendar/)?$',
views.CalendarView.as_view(),
name='swingtime-calendar'
),
url(
r'^calendar.json$',
views.CalendarJSONView.as_view(),
name='swingtime-calendar-json'
),
url(
r'^calendar/(?P<year>\d{4})/$',
views.CalendarView.as_view(),
name='swingtime-yearly-view'
),
url(
r'^calendar/(\d{4})/(0?[1-9]|1[012])/$',
views.CalendarView.as_view(),
name='swingtime-monthly-view'
),
url(
r'^agenda/$',
views.AgendaView.as_view(),
name='swingtime-events'
),
url(
r'^event/(\d+)/$',
views.event_view,
name='swingtime-event'
),
url(
r'^event/(\d+)/occurrence/(\d+)/$',
views.occurrence_view,
name='swingtime-occurrence'
),
)
|
0d0d43f957cb79a99eaacef0623cd57351ca40f6 | test/factories.py | test/factories.py | # coding: utf-8
import factory
from django.contrib.auth.models import User
class UserFactory(factory.Factory):
FACTORY_FOR = User
first_name = "Boy"
last_name = "Factory"
email = factory.LazyAttribute(
lambda a:
"{0}_{1}@example.com".format(a.first_name, a.last_name).lower())
username = factory.Sequence(lambda n: "username_%s" % n)
| # coding: utf-8
import factory
from django.contrib.auth.models import User
class UserFactory(factory.Factory):
FACTORY_FOR = User
first_name = "Boy"
last_name = "Factory"
email = factory.LazyAttribute(
lambda a:
"{0}_{1}@example.com".format(a.first_name, a.last_name).lower())
username = factory.Sequence(lambda n: "username_%s" % n)
is_active = False
is_staff = False
is_superuser = False
| Fix userfactory - set user flags | Fix userfactory - set user flags
| Python | mit | sarutobi/Rynda,sarutobi/ritmserdtsa,sarutobi/flowofkindness,sarutobi/Rynda,sarutobi/flowofkindness,sarutobi/ritmserdtsa,sarutobi/flowofkindness,sarutobi/ritmserdtsa,sarutobi/Rynda,sarutobi/Rynda,sarutobi/flowofkindness,sarutobi/ritmserdtsa | # coding: utf-8
import factory
from django.contrib.auth.models import User
class UserFactory(factory.Factory):
FACTORY_FOR = User
first_name = "Boy"
last_name = "Factory"
email = factory.LazyAttribute(
lambda a:
"{0}_{1}@example.com".format(a.first_name, a.last_name).lower())
username = factory.Sequence(lambda n: "username_%s" % n)
Fix userfactory - set user flags | # coding: utf-8
import factory
from django.contrib.auth.models import User
class UserFactory(factory.Factory):
FACTORY_FOR = User
first_name = "Boy"
last_name = "Factory"
email = factory.LazyAttribute(
lambda a:
"{0}_{1}@example.com".format(a.first_name, a.last_name).lower())
username = factory.Sequence(lambda n: "username_%s" % n)
is_active = False
is_staff = False
is_superuser = False
| <commit_before># coding: utf-8
import factory
from django.contrib.auth.models import User
class UserFactory(factory.Factory):
FACTORY_FOR = User
first_name = "Boy"
last_name = "Factory"
email = factory.LazyAttribute(
lambda a:
"{0}_{1}@example.com".format(a.first_name, a.last_name).lower())
username = factory.Sequence(lambda n: "username_%s" % n)
<commit_msg>Fix userfactory - set user flags<commit_after> | # coding: utf-8
import factory
from django.contrib.auth.models import User
class UserFactory(factory.Factory):
FACTORY_FOR = User
first_name = "Boy"
last_name = "Factory"
email = factory.LazyAttribute(
lambda a:
"{0}_{1}@example.com".format(a.first_name, a.last_name).lower())
username = factory.Sequence(lambda n: "username_%s" % n)
is_active = False
is_staff = False
is_superuser = False
| # coding: utf-8
import factory
from django.contrib.auth.models import User
class UserFactory(factory.Factory):
FACTORY_FOR = User
first_name = "Boy"
last_name = "Factory"
email = factory.LazyAttribute(
lambda a:
"{0}_{1}@example.com".format(a.first_name, a.last_name).lower())
username = factory.Sequence(lambda n: "username_%s" % n)
Fix userfactory - set user flags# coding: utf-8
import factory
from django.contrib.auth.models import User
class UserFactory(factory.Factory):
FACTORY_FOR = User
first_name = "Boy"
last_name = "Factory"
email = factory.LazyAttribute(
lambda a:
"{0}_{1}@example.com".format(a.first_name, a.last_name).lower())
username = factory.Sequence(lambda n: "username_%s" % n)
is_active = False
is_staff = False
is_superuser = False
| <commit_before># coding: utf-8
import factory
from django.contrib.auth.models import User
class UserFactory(factory.Factory):
FACTORY_FOR = User
first_name = "Boy"
last_name = "Factory"
email = factory.LazyAttribute(
lambda a:
"{0}_{1}@example.com".format(a.first_name, a.last_name).lower())
username = factory.Sequence(lambda n: "username_%s" % n)
<commit_msg>Fix userfactory - set user flags<commit_after># coding: utf-8
import factory
from django.contrib.auth.models import User
class UserFactory(factory.Factory):
FACTORY_FOR = User
first_name = "Boy"
last_name = "Factory"
email = factory.LazyAttribute(
lambda a:
"{0}_{1}@example.com".format(a.first_name, a.last_name).lower())
username = factory.Sequence(lambda n: "username_%s" % n)
is_active = False
is_staff = False
is_superuser = False
|
f4fdba652a1822698778c65df66a2639ec0fc5ad | tests/conftest.py | tests/conftest.py | import pytest
from .app import setup, teardown
@pytest.fixture(autouse=True, scope='session')
def db_migration(request):
setup()
request.addfinalizer(teardown)
| from __future__ import absolute_import
import pytest
from .app import setup, teardown
from app import create_app
from app.models import db, Framework
@pytest.fixture(autouse=True, scope='session')
def db_migration(request):
setup()
request.addfinalizer(teardown)
@pytest.fixture(scope='session')
def app(request):
return create_app('test')
@pytest.fixture()
def live_framework(request, app, status='live'):
with app.app_context():
framework = Framework.query.filter(
Framework.slug == 'digital-outcomes-and-specialists'
).first()
original_framework_status = framework.status
framework.status = status
db.session.add(framework)
db.session.commit()
def teardown():
with app.app_context():
framework = Framework.query.filter(
Framework.slug == 'digital-outcomes-and-specialists'
).first()
framework.status = original_framework_status
db.session.add(framework)
db.session.commit()
request.addfinalizer(teardown)
@pytest.fixture()
def expired_framework(request, app):
return live_framework(request, app, status='expired')
| Add live and expired framework pytest fixtures | Add live and expired framework pytest fixtures
This is a tiny attempt to move away from relying on database migrations
to set up framework records for tests. Using migration framework records
means we need to use actual (sometimes expired) frameworks to write tests
ties us to existing frameworks and require manual rollbacks for any record
changes since frameworks/lots tables are not reset between tests.
Using fixtures for frameworks makes it possible to depend explicitly
on database state required to run the test. This works better than
creating objects in setup/teardown since we can create more complicated
objects by using fixture dependencies (eg a service fixture relying on
framework and supplier fixtures etc.).
The fixture is still using DOS for now, but we could change it to create
a test framework with test lots as long as any additional test objects
are created using the fixture dependencies.
| Python | mit | alphagov/digitalmarketplace-api,alphagov/digitalmarketplace-api,alphagov/digitalmarketplace-api | import pytest
from .app import setup, teardown
@pytest.fixture(autouse=True, scope='session')
def db_migration(request):
setup()
request.addfinalizer(teardown)
Add live and expired framework pytest fixtures
This is a tiny attempt to move away from relying on database migrations
to set up framework records for tests. Using migration framework records
means we need to use actual (sometimes expired) frameworks to write tests
ties us to existing frameworks and require manual rollbacks for any record
changes since frameworks/lots tables are not reset between tests.
Using fixtures for frameworks makes it possible to depend explicitly
on database state required to run the test. This works better than
creating objects in setup/teardown since we can create more complicated
objects by using fixture dependencies (eg a service fixture relying on
framework and supplier fixtures etc.).
The fixture is still using DOS for now, but we could change it to create
a test framework with test lots as long as any additional test objects
are created using the fixture dependencies. | from __future__ import absolute_import
import pytest
from .app import setup, teardown
from app import create_app
from app.models import db, Framework
@pytest.fixture(autouse=True, scope='session')
def db_migration(request):
setup()
request.addfinalizer(teardown)
@pytest.fixture(scope='session')
def app(request):
return create_app('test')
@pytest.fixture()
def live_framework(request, app, status='live'):
with app.app_context():
framework = Framework.query.filter(
Framework.slug == 'digital-outcomes-and-specialists'
).first()
original_framework_status = framework.status
framework.status = status
db.session.add(framework)
db.session.commit()
def teardown():
with app.app_context():
framework = Framework.query.filter(
Framework.slug == 'digital-outcomes-and-specialists'
).first()
framework.status = original_framework_status
db.session.add(framework)
db.session.commit()
request.addfinalizer(teardown)
@pytest.fixture()
def expired_framework(request, app):
return live_framework(request, app, status='expired')
| <commit_before>import pytest
from .app import setup, teardown
@pytest.fixture(autouse=True, scope='session')
def db_migration(request):
setup()
request.addfinalizer(teardown)
<commit_msg>Add live and expired framework pytest fixtures
This is a tiny attempt to move away from relying on database migrations
to set up framework records for tests. Using migration framework records
means we need to use actual (sometimes expired) frameworks to write tests
ties us to existing frameworks and require manual rollbacks for any record
changes since frameworks/lots tables are not reset between tests.
Using fixtures for frameworks makes it possible to depend explicitly
on database state required to run the test. This works better than
creating objects in setup/teardown since we can create more complicated
objects by using fixture dependencies (eg a service fixture relying on
framework and supplier fixtures etc.).
The fixture is still using DOS for now, but we could change it to create
a test framework with test lots as long as any additional test objects
are created using the fixture dependencies.<commit_after> | from __future__ import absolute_import
import pytest
from .app import setup, teardown
from app import create_app
from app.models import db, Framework
@pytest.fixture(autouse=True, scope='session')
def db_migration(request):
setup()
request.addfinalizer(teardown)
@pytest.fixture(scope='session')
def app(request):
return create_app('test')
@pytest.fixture()
def live_framework(request, app, status='live'):
with app.app_context():
framework = Framework.query.filter(
Framework.slug == 'digital-outcomes-and-specialists'
).first()
original_framework_status = framework.status
framework.status = status
db.session.add(framework)
db.session.commit()
def teardown():
with app.app_context():
framework = Framework.query.filter(
Framework.slug == 'digital-outcomes-and-specialists'
).first()
framework.status = original_framework_status
db.session.add(framework)
db.session.commit()
request.addfinalizer(teardown)
@pytest.fixture()
def expired_framework(request, app):
return live_framework(request, app, status='expired')
| import pytest
from .app import setup, teardown
@pytest.fixture(autouse=True, scope='session')
def db_migration(request):
setup()
request.addfinalizer(teardown)
Add live and expired framework pytest fixtures
This is a tiny attempt to move away from relying on database migrations
to set up framework records for tests. Using migration framework records
means we need to use actual (sometimes expired) frameworks to write tests
ties us to existing frameworks and require manual rollbacks for any record
changes since frameworks/lots tables are not reset between tests.
Using fixtures for frameworks makes it possible to depend explicitly
on database state required to run the test. This works better than
creating objects in setup/teardown since we can create more complicated
objects by using fixture dependencies (eg a service fixture relying on
framework and supplier fixtures etc.).
The fixture is still using DOS for now, but we could change it to create
a test framework with test lots as long as any additional test objects
are created using the fixture dependencies.from __future__ import absolute_import
import pytest
from .app import setup, teardown
from app import create_app
from app.models import db, Framework
@pytest.fixture(autouse=True, scope='session')
def db_migration(request):
setup()
request.addfinalizer(teardown)
@pytest.fixture(scope='session')
def app(request):
return create_app('test')
@pytest.fixture()
def live_framework(request, app, status='live'):
with app.app_context():
framework = Framework.query.filter(
Framework.slug == 'digital-outcomes-and-specialists'
).first()
original_framework_status = framework.status
framework.status = status
db.session.add(framework)
db.session.commit()
def teardown():
with app.app_context():
framework = Framework.query.filter(
Framework.slug == 'digital-outcomes-and-specialists'
).first()
framework.status = original_framework_status
db.session.add(framework)
db.session.commit()
request.addfinalizer(teardown)
@pytest.fixture()
def expired_framework(request, app):
return live_framework(request, app, status='expired')
| <commit_before>import pytest
from .app import setup, teardown
@pytest.fixture(autouse=True, scope='session')
def db_migration(request):
setup()
request.addfinalizer(teardown)
<commit_msg>Add live and expired framework pytest fixtures
This is a tiny attempt to move away from relying on database migrations
to set up framework records for tests. Using migration framework records
means we need to use actual (sometimes expired) frameworks to write tests
ties us to existing frameworks and require manual rollbacks for any record
changes since frameworks/lots tables are not reset between tests.
Using fixtures for frameworks makes it possible to depend explicitly
on database state required to run the test. This works better than
creating objects in setup/teardown since we can create more complicated
objects by using fixture dependencies (eg a service fixture relying on
framework and supplier fixtures etc.).
The fixture is still using DOS for now, but we could change it to create
a test framework with test lots as long as any additional test objects
are created using the fixture dependencies.<commit_after>from __future__ import absolute_import
import pytest
from .app import setup, teardown
from app import create_app
from app.models import db, Framework
@pytest.fixture(autouse=True, scope='session')
def db_migration(request):
setup()
request.addfinalizer(teardown)
@pytest.fixture(scope='session')
def app(request):
return create_app('test')
@pytest.fixture()
def live_framework(request, app, status='live'):
with app.app_context():
framework = Framework.query.filter(
Framework.slug == 'digital-outcomes-and-specialists'
).first()
original_framework_status = framework.status
framework.status = status
db.session.add(framework)
db.session.commit()
def teardown():
with app.app_context():
framework = Framework.query.filter(
Framework.slug == 'digital-outcomes-and-specialists'
).first()
framework.status = original_framework_status
db.session.add(framework)
db.session.commit()
request.addfinalizer(teardown)
@pytest.fixture()
def expired_framework(request, app):
return live_framework(request, app, status='expired')
|
95c23b465bc0e0ce0e1ae633ddd1573cfdc997e2 | unbound_legacy_api/blueprints/stats.py | unbound_legacy_api/blueprints/stats.py | from flask import Blueprint
from unbound_legacy_api.utils.response import create_response
stats_bp = Blueprint('stats', __name__, url_prefix='/stats')
@stats_bp.route('/ping')
def ping():
"""Generic ping route to check if api is up"""
return create_response(status='success')
| from flask import Blueprint
from unbound_legacy_api.utils.response import create_response
stats_bp = Blueprint('stats', __name__, url_prefix='/v1/stats')
@stats_bp.route('/ping')
def ping():
"""Generic ping route to check if api is up"""
return create_response(status='success')
| Add versioning to api routes | fix(): Add versioning to api routes
| Python | mit | UnboundLegacy/api | from flask import Blueprint
from unbound_legacy_api.utils.response import create_response
stats_bp = Blueprint('stats', __name__, url_prefix='/stats')
@stats_bp.route('/ping')
def ping():
"""Generic ping route to check if api is up"""
return create_response(status='success')
fix(): Add versioning to api routes | from flask import Blueprint
from unbound_legacy_api.utils.response import create_response
stats_bp = Blueprint('stats', __name__, url_prefix='/v1/stats')
@stats_bp.route('/ping')
def ping():
"""Generic ping route to check if api is up"""
return create_response(status='success')
| <commit_before>from flask import Blueprint
from unbound_legacy_api.utils.response import create_response
stats_bp = Blueprint('stats', __name__, url_prefix='/stats')
@stats_bp.route('/ping')
def ping():
"""Generic ping route to check if api is up"""
return create_response(status='success')
<commit_msg>fix(): Add versioning to api routes<commit_after> | from flask import Blueprint
from unbound_legacy_api.utils.response import create_response
stats_bp = Blueprint('stats', __name__, url_prefix='/v1/stats')
@stats_bp.route('/ping')
def ping():
"""Generic ping route to check if api is up"""
return create_response(status='success')
| from flask import Blueprint
from unbound_legacy_api.utils.response import create_response
stats_bp = Blueprint('stats', __name__, url_prefix='/stats')
@stats_bp.route('/ping')
def ping():
"""Generic ping route to check if api is up"""
return create_response(status='success')
fix(): Add versioning to api routesfrom flask import Blueprint
from unbound_legacy_api.utils.response import create_response
stats_bp = Blueprint('stats', __name__, url_prefix='/v1/stats')
@stats_bp.route('/ping')
def ping():
"""Generic ping route to check if api is up"""
return create_response(status='success')
| <commit_before>from flask import Blueprint
from unbound_legacy_api.utils.response import create_response
stats_bp = Blueprint('stats', __name__, url_prefix='/stats')
@stats_bp.route('/ping')
def ping():
"""Generic ping route to check if api is up"""
return create_response(status='success')
<commit_msg>fix(): Add versioning to api routes<commit_after>from flask import Blueprint
from unbound_legacy_api.utils.response import create_response
stats_bp = Blueprint('stats', __name__, url_prefix='/v1/stats')
@stats_bp.route('/ping')
def ping():
"""Generic ping route to check if api is up"""
return create_response(status='success')
|
9145be89c1a5ba1a2c47bfeef571d40b9eb060bc | test/integration/test_user_args.py | test/integration/test_user_args.py | from . import *
class TestUserArgs(IntegrationTest):
def __init__(self, *args, **kwargs):
IntegrationTest.__init__(
self, os.path.join(examples_dir, '10_custom_args'),
configure=False, *args, **kwargs
)
def test_build_default(self):
self.configure()
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from unnamed!\n')
def test_build_with_args(self):
self.configure(extra_args=['--name=foo'])
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from foo!\n')
| from six import assertRegex
from . import *
class TestUserArgs(IntegrationTest):
def __init__(self, *args, **kwargs):
IntegrationTest.__init__(
self, os.path.join(examples_dir, '10_custom_args'),
configure=False, *args, **kwargs
)
def test_build_default(self):
self.configure()
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from unnamed!\n')
def test_build_with_args(self):
self.configure(extra_args=['--name=foo'])
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from foo!\n')
def test_help(self):
os.chdir(self.srcdir)
output = self.assertPopen(
['bfg9000', 'help', 'configure']
)
assertRegex(self, output, r'(?m)^project-defined arguments:$')
assertRegex(self, output,
r'(?m)^\s+--name NAME\s+set the name to greet$')
def test_help_explicit_srcdir(self):
output = self.assertPopen(
['bfg9000', 'help', 'configure', self.srcdir]
)
assertRegex(self, output, r'(?m)^project-defined arguments:$')
assertRegex(self, output,
r'(?m)^\s+--name NAME\s+set the name to greet$')
| Add integration test for user-args help | Add integration test for user-args help
| Python | bsd-3-clause | jimporter/bfg9000,jimporter/bfg9000,jimporter/bfg9000,jimporter/bfg9000 | from . import *
class TestUserArgs(IntegrationTest):
def __init__(self, *args, **kwargs):
IntegrationTest.__init__(
self, os.path.join(examples_dir, '10_custom_args'),
configure=False, *args, **kwargs
)
def test_build_default(self):
self.configure()
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from unnamed!\n')
def test_build_with_args(self):
self.configure(extra_args=['--name=foo'])
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from foo!\n')
Add integration test for user-args help | from six import assertRegex
from . import *
class TestUserArgs(IntegrationTest):
def __init__(self, *args, **kwargs):
IntegrationTest.__init__(
self, os.path.join(examples_dir, '10_custom_args'),
configure=False, *args, **kwargs
)
def test_build_default(self):
self.configure()
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from unnamed!\n')
def test_build_with_args(self):
self.configure(extra_args=['--name=foo'])
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from foo!\n')
def test_help(self):
os.chdir(self.srcdir)
output = self.assertPopen(
['bfg9000', 'help', 'configure']
)
assertRegex(self, output, r'(?m)^project-defined arguments:$')
assertRegex(self, output,
r'(?m)^\s+--name NAME\s+set the name to greet$')
def test_help_explicit_srcdir(self):
output = self.assertPopen(
['bfg9000', 'help', 'configure', self.srcdir]
)
assertRegex(self, output, r'(?m)^project-defined arguments:$')
assertRegex(self, output,
r'(?m)^\s+--name NAME\s+set the name to greet$')
| <commit_before>from . import *
class TestUserArgs(IntegrationTest):
def __init__(self, *args, **kwargs):
IntegrationTest.__init__(
self, os.path.join(examples_dir, '10_custom_args'),
configure=False, *args, **kwargs
)
def test_build_default(self):
self.configure()
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from unnamed!\n')
def test_build_with_args(self):
self.configure(extra_args=['--name=foo'])
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from foo!\n')
<commit_msg>Add integration test for user-args help<commit_after> | from six import assertRegex
from . import *
class TestUserArgs(IntegrationTest):
def __init__(self, *args, **kwargs):
IntegrationTest.__init__(
self, os.path.join(examples_dir, '10_custom_args'),
configure=False, *args, **kwargs
)
def test_build_default(self):
self.configure()
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from unnamed!\n')
def test_build_with_args(self):
self.configure(extra_args=['--name=foo'])
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from foo!\n')
def test_help(self):
os.chdir(self.srcdir)
output = self.assertPopen(
['bfg9000', 'help', 'configure']
)
assertRegex(self, output, r'(?m)^project-defined arguments:$')
assertRegex(self, output,
r'(?m)^\s+--name NAME\s+set the name to greet$')
def test_help_explicit_srcdir(self):
output = self.assertPopen(
['bfg9000', 'help', 'configure', self.srcdir]
)
assertRegex(self, output, r'(?m)^project-defined arguments:$')
assertRegex(self, output,
r'(?m)^\s+--name NAME\s+set the name to greet$')
| from . import *
class TestUserArgs(IntegrationTest):
def __init__(self, *args, **kwargs):
IntegrationTest.__init__(
self, os.path.join(examples_dir, '10_custom_args'),
configure=False, *args, **kwargs
)
def test_build_default(self):
self.configure()
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from unnamed!\n')
def test_build_with_args(self):
self.configure(extra_args=['--name=foo'])
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from foo!\n')
Add integration test for user-args helpfrom six import assertRegex
from . import *
class TestUserArgs(IntegrationTest):
def __init__(self, *args, **kwargs):
IntegrationTest.__init__(
self, os.path.join(examples_dir, '10_custom_args'),
configure=False, *args, **kwargs
)
def test_build_default(self):
self.configure()
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from unnamed!\n')
def test_build_with_args(self):
self.configure(extra_args=['--name=foo'])
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from foo!\n')
def test_help(self):
os.chdir(self.srcdir)
output = self.assertPopen(
['bfg9000', 'help', 'configure']
)
assertRegex(self, output, r'(?m)^project-defined arguments:$')
assertRegex(self, output,
r'(?m)^\s+--name NAME\s+set the name to greet$')
def test_help_explicit_srcdir(self):
output = self.assertPopen(
['bfg9000', 'help', 'configure', self.srcdir]
)
assertRegex(self, output, r'(?m)^project-defined arguments:$')
assertRegex(self, output,
r'(?m)^\s+--name NAME\s+set the name to greet$')
| <commit_before>from . import *
class TestUserArgs(IntegrationTest):
def __init__(self, *args, **kwargs):
IntegrationTest.__init__(
self, os.path.join(examples_dir, '10_custom_args'),
configure=False, *args, **kwargs
)
def test_build_default(self):
self.configure()
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from unnamed!\n')
def test_build_with_args(self):
self.configure(extra_args=['--name=foo'])
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from foo!\n')
<commit_msg>Add integration test for user-args help<commit_after>from six import assertRegex
from . import *
class TestUserArgs(IntegrationTest):
def __init__(self, *args, **kwargs):
IntegrationTest.__init__(
self, os.path.join(examples_dir, '10_custom_args'),
configure=False, *args, **kwargs
)
def test_build_default(self):
self.configure()
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from unnamed!\n')
def test_build_with_args(self):
self.configure(extra_args=['--name=foo'])
self.build(executable('simple'))
self.assertOutput([executable('simple')], 'hello from foo!\n')
def test_help(self):
os.chdir(self.srcdir)
output = self.assertPopen(
['bfg9000', 'help', 'configure']
)
assertRegex(self, output, r'(?m)^project-defined arguments:$')
assertRegex(self, output,
r'(?m)^\s+--name NAME\s+set the name to greet$')
def test_help_explicit_srcdir(self):
output = self.assertPopen(
['bfg9000', 'help', 'configure', self.srcdir]
)
assertRegex(self, output, r'(?m)^project-defined arguments:$')
assertRegex(self, output,
r'(?m)^\s+--name NAME\s+set the name to greet$')
|
6a17674897bbb3a44fb2153967e3985dfdb3d5df | zounds/learn/graph.py | zounds/learn/graph.py | import featureflow as ff
from random_samples import ShuffledSamples
from random_samples import InfiniteSampler
from preprocess import PreprocessingPipeline
def learning_pipeline():
class LearningPipeline(ff.BaseModel):
samples = ff.PickleFeature(ff.IteratorNode)
shuffled = ff.PickleFeature(
ShuffledSamples,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
needs=samples)
return LearningPipeline
def infinite_streaming_learning_pipeline(cls):
roots = filter(lambda feature: feature.is_root, cls.features.itervalues())
if len(roots) != 1:
raise ValueError('cls must have a single root feature')
root = roots[0]
class InfiniteLearningPipeline(cls):
dataset = ff.Feature(
InfiniteSampler,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'))
pipeline = ff.ClobberPickleFeature(
PreprocessingPipeline,
needs=cls.features,
store=True)
@classmethod
def load_network(cls):
if not cls.exists():
raise RuntimeError('No network has been trained or saved')
instance = cls()
for p in instance.pipeline:
try:
return p.network
except AttributeError:
pass
raise RuntimeError('There is no network in the pipeline')
root.needs = InfiniteLearningPipeline.dataset
InfiniteLearningPipeline.__name__ = cls.__name__
InfiniteLearningPipeline.__module__ = cls.__module__
return InfiniteLearningPipeline
| import featureflow as ff
from random_samples import ShuffledSamples
from random_samples import InfiniteSampler
from preprocess import PreprocessingPipeline
def learning_pipeline():
class LearningPipeline(ff.BaseModel):
samples = ff.PickleFeature(ff.IteratorNode)
shuffled = ff.PickleFeature(
ShuffledSamples,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
needs=samples)
return LearningPipeline
def infinite_streaming_learning_pipeline(cls):
roots = filter(lambda feature: feature.is_root, cls.features.itervalues())
if len(roots) != 1:
raise ValueError('cls must have a single root feature')
root = roots[0]
class InfiniteLearningPipeline(cls):
dataset = ff.Feature(
InfiniteSampler,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
feature_filter=ff.Var('feature_filter'),
parallel=ff.Var('parallel'))
pipeline = ff.ClobberPickleFeature(
PreprocessingPipeline,
needs=cls.features,
store=True)
@classmethod
def load_network(cls):
if not cls.exists():
raise RuntimeError('No network has been trained or saved')
instance = cls()
for p in instance.pipeline:
try:
return p.network
except AttributeError:
pass
raise RuntimeError('There is no network in the pipeline')
root.needs = InfiniteLearningPipeline.dataset
InfiniteLearningPipeline.__name__ = cls.__name__
InfiniteLearningPipeline.__module__ = cls.__module__
return InfiniteLearningPipeline
| Add a new option allowing client code to turn off parallelism | Add a new option allowing client code to turn off parallelism
| Python | mit | JohnVinyard/zounds,JohnVinyard/zounds,JohnVinyard/zounds,JohnVinyard/zounds | import featureflow as ff
from random_samples import ShuffledSamples
from random_samples import InfiniteSampler
from preprocess import PreprocessingPipeline
def learning_pipeline():
class LearningPipeline(ff.BaseModel):
samples = ff.PickleFeature(ff.IteratorNode)
shuffled = ff.PickleFeature(
ShuffledSamples,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
needs=samples)
return LearningPipeline
def infinite_streaming_learning_pipeline(cls):
roots = filter(lambda feature: feature.is_root, cls.features.itervalues())
if len(roots) != 1:
raise ValueError('cls must have a single root feature')
root = roots[0]
class InfiniteLearningPipeline(cls):
dataset = ff.Feature(
InfiniteSampler,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'))
pipeline = ff.ClobberPickleFeature(
PreprocessingPipeline,
needs=cls.features,
store=True)
@classmethod
def load_network(cls):
if not cls.exists():
raise RuntimeError('No network has been trained or saved')
instance = cls()
for p in instance.pipeline:
try:
return p.network
except AttributeError:
pass
raise RuntimeError('There is no network in the pipeline')
root.needs = InfiniteLearningPipeline.dataset
InfiniteLearningPipeline.__name__ = cls.__name__
InfiniteLearningPipeline.__module__ = cls.__module__
return InfiniteLearningPipeline
Add a new option allowing client code to turn off parallelism | import featureflow as ff
from random_samples import ShuffledSamples
from random_samples import InfiniteSampler
from preprocess import PreprocessingPipeline
def learning_pipeline():
class LearningPipeline(ff.BaseModel):
samples = ff.PickleFeature(ff.IteratorNode)
shuffled = ff.PickleFeature(
ShuffledSamples,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
needs=samples)
return LearningPipeline
def infinite_streaming_learning_pipeline(cls):
roots = filter(lambda feature: feature.is_root, cls.features.itervalues())
if len(roots) != 1:
raise ValueError('cls must have a single root feature')
root = roots[0]
class InfiniteLearningPipeline(cls):
dataset = ff.Feature(
InfiniteSampler,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
feature_filter=ff.Var('feature_filter'),
parallel=ff.Var('parallel'))
pipeline = ff.ClobberPickleFeature(
PreprocessingPipeline,
needs=cls.features,
store=True)
@classmethod
def load_network(cls):
if not cls.exists():
raise RuntimeError('No network has been trained or saved')
instance = cls()
for p in instance.pipeline:
try:
return p.network
except AttributeError:
pass
raise RuntimeError('There is no network in the pipeline')
root.needs = InfiniteLearningPipeline.dataset
InfiniteLearningPipeline.__name__ = cls.__name__
InfiniteLearningPipeline.__module__ = cls.__module__
return InfiniteLearningPipeline
| <commit_before>import featureflow as ff
from random_samples import ShuffledSamples
from random_samples import InfiniteSampler
from preprocess import PreprocessingPipeline
def learning_pipeline():
class LearningPipeline(ff.BaseModel):
samples = ff.PickleFeature(ff.IteratorNode)
shuffled = ff.PickleFeature(
ShuffledSamples,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
needs=samples)
return LearningPipeline
def infinite_streaming_learning_pipeline(cls):
roots = filter(lambda feature: feature.is_root, cls.features.itervalues())
if len(roots) != 1:
raise ValueError('cls must have a single root feature')
root = roots[0]
class InfiniteLearningPipeline(cls):
dataset = ff.Feature(
InfiniteSampler,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'))
pipeline = ff.ClobberPickleFeature(
PreprocessingPipeline,
needs=cls.features,
store=True)
@classmethod
def load_network(cls):
if not cls.exists():
raise RuntimeError('No network has been trained or saved')
instance = cls()
for p in instance.pipeline:
try:
return p.network
except AttributeError:
pass
raise RuntimeError('There is no network in the pipeline')
root.needs = InfiniteLearningPipeline.dataset
InfiniteLearningPipeline.__name__ = cls.__name__
InfiniteLearningPipeline.__module__ = cls.__module__
return InfiniteLearningPipeline
<commit_msg>Add a new option allowing client code to turn off parallelism<commit_after> | import featureflow as ff
from random_samples import ShuffledSamples
from random_samples import InfiniteSampler
from preprocess import PreprocessingPipeline
def learning_pipeline():
class LearningPipeline(ff.BaseModel):
samples = ff.PickleFeature(ff.IteratorNode)
shuffled = ff.PickleFeature(
ShuffledSamples,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
needs=samples)
return LearningPipeline
def infinite_streaming_learning_pipeline(cls):
roots = filter(lambda feature: feature.is_root, cls.features.itervalues())
if len(roots) != 1:
raise ValueError('cls must have a single root feature')
root = roots[0]
class InfiniteLearningPipeline(cls):
dataset = ff.Feature(
InfiniteSampler,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
feature_filter=ff.Var('feature_filter'),
parallel=ff.Var('parallel'))
pipeline = ff.ClobberPickleFeature(
PreprocessingPipeline,
needs=cls.features,
store=True)
@classmethod
def load_network(cls):
if not cls.exists():
raise RuntimeError('No network has been trained or saved')
instance = cls()
for p in instance.pipeline:
try:
return p.network
except AttributeError:
pass
raise RuntimeError('There is no network in the pipeline')
root.needs = InfiniteLearningPipeline.dataset
InfiniteLearningPipeline.__name__ = cls.__name__
InfiniteLearningPipeline.__module__ = cls.__module__
return InfiniteLearningPipeline
| import featureflow as ff
from random_samples import ShuffledSamples
from random_samples import InfiniteSampler
from preprocess import PreprocessingPipeline
def learning_pipeline():
class LearningPipeline(ff.BaseModel):
samples = ff.PickleFeature(ff.IteratorNode)
shuffled = ff.PickleFeature(
ShuffledSamples,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
needs=samples)
return LearningPipeline
def infinite_streaming_learning_pipeline(cls):
roots = filter(lambda feature: feature.is_root, cls.features.itervalues())
if len(roots) != 1:
raise ValueError('cls must have a single root feature')
root = roots[0]
class InfiniteLearningPipeline(cls):
dataset = ff.Feature(
InfiniteSampler,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'))
pipeline = ff.ClobberPickleFeature(
PreprocessingPipeline,
needs=cls.features,
store=True)
@classmethod
def load_network(cls):
if not cls.exists():
raise RuntimeError('No network has been trained or saved')
instance = cls()
for p in instance.pipeline:
try:
return p.network
except AttributeError:
pass
raise RuntimeError('There is no network in the pipeline')
root.needs = InfiniteLearningPipeline.dataset
InfiniteLearningPipeline.__name__ = cls.__name__
InfiniteLearningPipeline.__module__ = cls.__module__
return InfiniteLearningPipeline
Add a new option allowing client code to turn off parallelismimport featureflow as ff
from random_samples import ShuffledSamples
from random_samples import InfiniteSampler
from preprocess import PreprocessingPipeline
def learning_pipeline():
class LearningPipeline(ff.BaseModel):
samples = ff.PickleFeature(ff.IteratorNode)
shuffled = ff.PickleFeature(
ShuffledSamples,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
needs=samples)
return LearningPipeline
def infinite_streaming_learning_pipeline(cls):
roots = filter(lambda feature: feature.is_root, cls.features.itervalues())
if len(roots) != 1:
raise ValueError('cls must have a single root feature')
root = roots[0]
class InfiniteLearningPipeline(cls):
dataset = ff.Feature(
InfiniteSampler,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
feature_filter=ff.Var('feature_filter'),
parallel=ff.Var('parallel'))
pipeline = ff.ClobberPickleFeature(
PreprocessingPipeline,
needs=cls.features,
store=True)
@classmethod
def load_network(cls):
if not cls.exists():
raise RuntimeError('No network has been trained or saved')
instance = cls()
for p in instance.pipeline:
try:
return p.network
except AttributeError:
pass
raise RuntimeError('There is no network in the pipeline')
root.needs = InfiniteLearningPipeline.dataset
InfiniteLearningPipeline.__name__ = cls.__name__
InfiniteLearningPipeline.__module__ = cls.__module__
return InfiniteLearningPipeline
| <commit_before>import featureflow as ff
from random_samples import ShuffledSamples
from random_samples import InfiniteSampler
from preprocess import PreprocessingPipeline
def learning_pipeline():
class LearningPipeline(ff.BaseModel):
samples = ff.PickleFeature(ff.IteratorNode)
shuffled = ff.PickleFeature(
ShuffledSamples,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
needs=samples)
return LearningPipeline
def infinite_streaming_learning_pipeline(cls):
roots = filter(lambda feature: feature.is_root, cls.features.itervalues())
if len(roots) != 1:
raise ValueError('cls must have a single root feature')
root = roots[0]
class InfiniteLearningPipeline(cls):
dataset = ff.Feature(
InfiniteSampler,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'))
pipeline = ff.ClobberPickleFeature(
PreprocessingPipeline,
needs=cls.features,
store=True)
@classmethod
def load_network(cls):
if not cls.exists():
raise RuntimeError('No network has been trained or saved')
instance = cls()
for p in instance.pipeline:
try:
return p.network
except AttributeError:
pass
raise RuntimeError('There is no network in the pipeline')
root.needs = InfiniteLearningPipeline.dataset
InfiniteLearningPipeline.__name__ = cls.__name__
InfiniteLearningPipeline.__module__ = cls.__module__
return InfiniteLearningPipeline
<commit_msg>Add a new option allowing client code to turn off parallelism<commit_after>import featureflow as ff
from random_samples import ShuffledSamples
from random_samples import InfiniteSampler
from preprocess import PreprocessingPipeline
def learning_pipeline():
class LearningPipeline(ff.BaseModel):
samples = ff.PickleFeature(ff.IteratorNode)
shuffled = ff.PickleFeature(
ShuffledSamples,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
needs=samples)
return LearningPipeline
def infinite_streaming_learning_pipeline(cls):
roots = filter(lambda feature: feature.is_root, cls.features.itervalues())
if len(roots) != 1:
raise ValueError('cls must have a single root feature')
root = roots[0]
class InfiniteLearningPipeline(cls):
dataset = ff.Feature(
InfiniteSampler,
nsamples=ff.Var('nsamples'),
dtype=ff.Var('dtype'),
feature_filter=ff.Var('feature_filter'),
parallel=ff.Var('parallel'))
pipeline = ff.ClobberPickleFeature(
PreprocessingPipeline,
needs=cls.features,
store=True)
@classmethod
def load_network(cls):
if not cls.exists():
raise RuntimeError('No network has been trained or saved')
instance = cls()
for p in instance.pipeline:
try:
return p.network
except AttributeError:
pass
raise RuntimeError('There is no network in the pipeline')
root.needs = InfiniteLearningPipeline.dataset
InfiniteLearningPipeline.__name__ = cls.__name__
InfiniteLearningPipeline.__module__ = cls.__module__
return InfiniteLearningPipeline
|
bfaf9d326fc0a2fc72a6f7b6ed92640c3fe9b87b | hirlite/__init__.py | hirlite/__init__.py | from .hirlite import Rlite, HirliteError
from .version import __version__
__all__ = ["Rlite", "HirliteError", "__version__"]
| import functools
from hirlite.hirlite import Rlite as RliteExtension, HirliteError
from hirlite.version import __version__
__all__ = ["Rlite", "HirliteError", "__version__"]
class Rlite(RliteExtension):
def __getattr__(self, command):
return functools.partial(self.command, command)
| Add support for calling commands by attr access | Add support for calling commands by attr access
| Python | bsd-2-clause | seppo0010/rlite-py,seppo0010/rlite-py,pombredanne/rlite-py,pombredanne/rlite-py | from .hirlite import Rlite, HirliteError
from .version import __version__
__all__ = ["Rlite", "HirliteError", "__version__"]
Add support for calling commands by attr access | import functools
from hirlite.hirlite import Rlite as RliteExtension, HirliteError
from hirlite.version import __version__
__all__ = ["Rlite", "HirliteError", "__version__"]
class Rlite(RliteExtension):
def __getattr__(self, command):
return functools.partial(self.command, command)
| <commit_before>from .hirlite import Rlite, HirliteError
from .version import __version__
__all__ = ["Rlite", "HirliteError", "__version__"]
<commit_msg>Add support for calling commands by attr access<commit_after> | import functools
from hirlite.hirlite import Rlite as RliteExtension, HirliteError
from hirlite.version import __version__
__all__ = ["Rlite", "HirliteError", "__version__"]
class Rlite(RliteExtension):
def __getattr__(self, command):
return functools.partial(self.command, command)
| from .hirlite import Rlite, HirliteError
from .version import __version__
__all__ = ["Rlite", "HirliteError", "__version__"]
Add support for calling commands by attr accessimport functools
from hirlite.hirlite import Rlite as RliteExtension, HirliteError
from hirlite.version import __version__
__all__ = ["Rlite", "HirliteError", "__version__"]
class Rlite(RliteExtension):
def __getattr__(self, command):
return functools.partial(self.command, command)
| <commit_before>from .hirlite import Rlite, HirliteError
from .version import __version__
__all__ = ["Rlite", "HirliteError", "__version__"]
<commit_msg>Add support for calling commands by attr access<commit_after>import functools
from hirlite.hirlite import Rlite as RliteExtension, HirliteError
from hirlite.version import __version__
__all__ = ["Rlite", "HirliteError", "__version__"]
class Rlite(RliteExtension):
def __getattr__(self, command):
return functools.partial(self.command, command)
|
9aef1f357a3319a31bd1995f462eb356011b6a93 | huxley/shortcuts.py | huxley/shortcuts.py | # Copyright (c) 2011-2013 Kunal Mehta. All rights reserved.
# Use of this source code is governed by a BSD License found in README.md.
from django.http import HttpResponse
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.utils import simplejson
from itertools import izip_longest
def render_template(request, template, context=None):
'''Wrap render_to_response with the context_instance argument set.'''
return render_to_response(template, context,
context_instance=RequestContext(request))
def render_json(data):
'''Return an HttpResponse object containing json-encoded data.'''
return HttpResponse(simplejson.dumps(data), mimetype='application/json')
def pairwise(iterable):
'''Group the elements of the given interable into 2-tuples.'''
i = iter(iterable)
return izip_longest(i, i)
| # Copyright (c) 2011-2013 Kunal Mehta. All rights reserved.
# Use of this source code is governed by a BSD License found in README.md.
from django.http import HttpResponse
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.utils import simplejson
from itertools import izip_longest
def render_template(request, template, context=None):
'''Wrap render_to_response with the context_instance argument set.'''
return render_to_response(template, context,
context_instance=RequestContext(request))
def render_json(data):
'''Return an HttpResponse object containing json-encoded data.'''
return HttpResponse(simplejson.dumps(data), content_type='application/json')
def pairwise(iterable):
'''Group the elements of the given interable into 2-tuples.'''
i = iter(iterable)
return izip_longest(i, i)
| Change mimetype kwarg to content_type. | Change mimetype kwarg to content_type.
| Python | bsd-3-clause | nathanielparke/huxley,bmun/huxley,bmun/huxley,nathanielparke/huxley,ctmunwebmaster/huxley,nathanielparke/huxley,ctmunwebmaster/huxley,bmun/huxley,bmun/huxley,ctmunwebmaster/huxley,ctmunwebmaster/huxley,nathanielparke/huxley | # Copyright (c) 2011-2013 Kunal Mehta. All rights reserved.
# Use of this source code is governed by a BSD License found in README.md.
from django.http import HttpResponse
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.utils import simplejson
from itertools import izip_longest
def render_template(request, template, context=None):
'''Wrap render_to_response with the context_instance argument set.'''
return render_to_response(template, context,
context_instance=RequestContext(request))
def render_json(data):
'''Return an HttpResponse object containing json-encoded data.'''
return HttpResponse(simplejson.dumps(data), mimetype='application/json')
def pairwise(iterable):
'''Group the elements of the given interable into 2-tuples.'''
i = iter(iterable)
return izip_longest(i, i)
Change mimetype kwarg to content_type. | # Copyright (c) 2011-2013 Kunal Mehta. All rights reserved.
# Use of this source code is governed by a BSD License found in README.md.
from django.http import HttpResponse
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.utils import simplejson
from itertools import izip_longest
def render_template(request, template, context=None):
'''Wrap render_to_response with the context_instance argument set.'''
return render_to_response(template, context,
context_instance=RequestContext(request))
def render_json(data):
'''Return an HttpResponse object containing json-encoded data.'''
return HttpResponse(simplejson.dumps(data), content_type='application/json')
def pairwise(iterable):
'''Group the elements of the given interable into 2-tuples.'''
i = iter(iterable)
return izip_longest(i, i)
| <commit_before># Copyright (c) 2011-2013 Kunal Mehta. All rights reserved.
# Use of this source code is governed by a BSD License found in README.md.
from django.http import HttpResponse
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.utils import simplejson
from itertools import izip_longest
def render_template(request, template, context=None):
'''Wrap render_to_response with the context_instance argument set.'''
return render_to_response(template, context,
context_instance=RequestContext(request))
def render_json(data):
'''Return an HttpResponse object containing json-encoded data.'''
return HttpResponse(simplejson.dumps(data), mimetype='application/json')
def pairwise(iterable):
'''Group the elements of the given interable into 2-tuples.'''
i = iter(iterable)
return izip_longest(i, i)
<commit_msg>Change mimetype kwarg to content_type.<commit_after> | # Copyright (c) 2011-2013 Kunal Mehta. All rights reserved.
# Use of this source code is governed by a BSD License found in README.md.
from django.http import HttpResponse
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.utils import simplejson
from itertools import izip_longest
def render_template(request, template, context=None):
'''Wrap render_to_response with the context_instance argument set.'''
return render_to_response(template, context,
context_instance=RequestContext(request))
def render_json(data):
'''Return an HttpResponse object containing json-encoded data.'''
return HttpResponse(simplejson.dumps(data), content_type='application/json')
def pairwise(iterable):
'''Group the elements of the given interable into 2-tuples.'''
i = iter(iterable)
return izip_longest(i, i)
| # Copyright (c) 2011-2013 Kunal Mehta. All rights reserved.
# Use of this source code is governed by a BSD License found in README.md.
from django.http import HttpResponse
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.utils import simplejson
from itertools import izip_longest
def render_template(request, template, context=None):
'''Wrap render_to_response with the context_instance argument set.'''
return render_to_response(template, context,
context_instance=RequestContext(request))
def render_json(data):
'''Return an HttpResponse object containing json-encoded data.'''
return HttpResponse(simplejson.dumps(data), mimetype='application/json')
def pairwise(iterable):
'''Group the elements of the given interable into 2-tuples.'''
i = iter(iterable)
return izip_longest(i, i)
Change mimetype kwarg to content_type.# Copyright (c) 2011-2013 Kunal Mehta. All rights reserved.
# Use of this source code is governed by a BSD License found in README.md.
from django.http import HttpResponse
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.utils import simplejson
from itertools import izip_longest
def render_template(request, template, context=None):
'''Wrap render_to_response with the context_instance argument set.'''
return render_to_response(template, context,
context_instance=RequestContext(request))
def render_json(data):
'''Return an HttpResponse object containing json-encoded data.'''
return HttpResponse(simplejson.dumps(data), content_type='application/json')
def pairwise(iterable):
'''Group the elements of the given interable into 2-tuples.'''
i = iter(iterable)
return izip_longest(i, i)
| <commit_before># Copyright (c) 2011-2013 Kunal Mehta. All rights reserved.
# Use of this source code is governed by a BSD License found in README.md.
from django.http import HttpResponse
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.utils import simplejson
from itertools import izip_longest
def render_template(request, template, context=None):
'''Wrap render_to_response with the context_instance argument set.'''
return render_to_response(template, context,
context_instance=RequestContext(request))
def render_json(data):
'''Return an HttpResponse object containing json-encoded data.'''
return HttpResponse(simplejson.dumps(data), mimetype='application/json')
def pairwise(iterable):
'''Group the elements of the given interable into 2-tuples.'''
i = iter(iterable)
return izip_longest(i, i)
<commit_msg>Change mimetype kwarg to content_type.<commit_after># Copyright (c) 2011-2013 Kunal Mehta. All rights reserved.
# Use of this source code is governed by a BSD License found in README.md.
from django.http import HttpResponse
from django.shortcuts import render_to_response
from django.template import RequestContext
from django.utils import simplejson
from itertools import izip_longest
def render_template(request, template, context=None):
'''Wrap render_to_response with the context_instance argument set.'''
return render_to_response(template, context,
context_instance=RequestContext(request))
def render_json(data):
'''Return an HttpResponse object containing json-encoded data.'''
return HttpResponse(simplejson.dumps(data), content_type='application/json')
def pairwise(iterable):
'''Group the elements of the given interable into 2-tuples.'''
i = iter(iterable)
return izip_longest(i, i)
|
700f6e6ef40a2d33e5678e260f03cd15148e0b3a | test/test_parse_file_guess_format.py | test/test_parse_file_guess_format.py | import unittest
import logging
from pathlib import Path
from shutil import copyfile
from tempfile import TemporaryDirectory
from rdflib.exceptions import ParserError
from rdflib import Graph
class FileParserGuessFormatTest(unittest.TestCase):
def test_ttl(self):
g = Graph()
self.assertIsInstance(g.parse("test/w3c/turtle/IRI_subject.ttl"), Graph)
def test_n3(self):
g = Graph()
self.assertIsInstance(g.parse("test/n3/example-lots_of_graphs.n3"), Graph)
def test_warning(self):
g = Graph()
graph_logger = logging.getLogger("rdflib")
with TemporaryDirectory() as tmpdirname:
newpath = Path(tmpdirname).joinpath("no_file_ext")
copyfile("test/rdf/Manifest.rdf", str(newpath))
with self.assertLogs(graph_logger, "WARNING"):
with self.assertRaises(ParserError):
g.parse(str(newpath))
if __name__ == "__main__":
unittest.main()
| import unittest
import logging
from pathlib import Path
from shutil import copyfile
from tempfile import TemporaryDirectory
from rdflib.exceptions import ParserError
from rdflib import Graph
class FileParserGuessFormatTest(unittest.TestCase):
def test_jsonld(self):
g = Graph()
self.assertIsInstance(g.parse("test/jsonld/1.1/manifest.jsonld"), Graph)
def test_ttl(self):
g = Graph()
self.assertIsInstance(g.parse("test/w3c/turtle/IRI_subject.ttl"), Graph)
def test_n3(self):
g = Graph()
self.assertIsInstance(g.parse("test/n3/example-lots_of_graphs.n3"), Graph)
def test_warning(self):
g = Graph()
graph_logger = logging.getLogger("rdflib")
with TemporaryDirectory() as tmpdirname:
newpath = Path(tmpdirname).joinpath("no_file_ext")
copyfile("test/rdf/Manifest.rdf", str(newpath))
with self.assertLogs(graph_logger, "WARNING"):
with self.assertRaises(ParserError):
g.parse(str(newpath))
if __name__ == "__main__":
unittest.main()
| Add test for adding JSON-LD to guess_format() | Add test for adding JSON-LD to guess_format()
This is a follow-on patch to:
e778e9413510721c2fedaae56d4ff826df265c30
Test was confirmed to pass by running this on the current `master`
branch, and confirmed to fail with e778e941 reverted.
nosetests test/test_parse_file_guess_format.py
Signed-off-by: Alex Nelson <d5aca716d6ed55a86fa350cb5231ca56b21085ca@nist.gov>
| Python | bsd-3-clause | RDFLib/rdflib,RDFLib/rdflib,RDFLib/rdflib,RDFLib/rdflib | import unittest
import logging
from pathlib import Path
from shutil import copyfile
from tempfile import TemporaryDirectory
from rdflib.exceptions import ParserError
from rdflib import Graph
class FileParserGuessFormatTest(unittest.TestCase):
def test_ttl(self):
g = Graph()
self.assertIsInstance(g.parse("test/w3c/turtle/IRI_subject.ttl"), Graph)
def test_n3(self):
g = Graph()
self.assertIsInstance(g.parse("test/n3/example-lots_of_graphs.n3"), Graph)
def test_warning(self):
g = Graph()
graph_logger = logging.getLogger("rdflib")
with TemporaryDirectory() as tmpdirname:
newpath = Path(tmpdirname).joinpath("no_file_ext")
copyfile("test/rdf/Manifest.rdf", str(newpath))
with self.assertLogs(graph_logger, "WARNING"):
with self.assertRaises(ParserError):
g.parse(str(newpath))
if __name__ == "__main__":
unittest.main()
Add test for adding JSON-LD to guess_format()
This is a follow-on patch to:
e778e9413510721c2fedaae56d4ff826df265c30
Test was confirmed to pass by running this on the current `master`
branch, and confirmed to fail with e778e941 reverted.
nosetests test/test_parse_file_guess_format.py
Signed-off-by: Alex Nelson <d5aca716d6ed55a86fa350cb5231ca56b21085ca@nist.gov> | import unittest
import logging
from pathlib import Path
from shutil import copyfile
from tempfile import TemporaryDirectory
from rdflib.exceptions import ParserError
from rdflib import Graph
class FileParserGuessFormatTest(unittest.TestCase):
def test_jsonld(self):
g = Graph()
self.assertIsInstance(g.parse("test/jsonld/1.1/manifest.jsonld"), Graph)
def test_ttl(self):
g = Graph()
self.assertIsInstance(g.parse("test/w3c/turtle/IRI_subject.ttl"), Graph)
def test_n3(self):
g = Graph()
self.assertIsInstance(g.parse("test/n3/example-lots_of_graphs.n3"), Graph)
def test_warning(self):
g = Graph()
graph_logger = logging.getLogger("rdflib")
with TemporaryDirectory() as tmpdirname:
newpath = Path(tmpdirname).joinpath("no_file_ext")
copyfile("test/rdf/Manifest.rdf", str(newpath))
with self.assertLogs(graph_logger, "WARNING"):
with self.assertRaises(ParserError):
g.parse(str(newpath))
if __name__ == "__main__":
unittest.main()
| <commit_before>import unittest
import logging
from pathlib import Path
from shutil import copyfile
from tempfile import TemporaryDirectory
from rdflib.exceptions import ParserError
from rdflib import Graph
class FileParserGuessFormatTest(unittest.TestCase):
def test_ttl(self):
g = Graph()
self.assertIsInstance(g.parse("test/w3c/turtle/IRI_subject.ttl"), Graph)
def test_n3(self):
g = Graph()
self.assertIsInstance(g.parse("test/n3/example-lots_of_graphs.n3"), Graph)
def test_warning(self):
g = Graph()
graph_logger = logging.getLogger("rdflib")
with TemporaryDirectory() as tmpdirname:
newpath = Path(tmpdirname).joinpath("no_file_ext")
copyfile("test/rdf/Manifest.rdf", str(newpath))
with self.assertLogs(graph_logger, "WARNING"):
with self.assertRaises(ParserError):
g.parse(str(newpath))
if __name__ == "__main__":
unittest.main()
<commit_msg>Add test for adding JSON-LD to guess_format()
This is a follow-on patch to:
e778e9413510721c2fedaae56d4ff826df265c30
Test was confirmed to pass by running this on the current `master`
branch, and confirmed to fail with e778e941 reverted.
nosetests test/test_parse_file_guess_format.py
Signed-off-by: Alex Nelson <d5aca716d6ed55a86fa350cb5231ca56b21085ca@nist.gov><commit_after> | import unittest
import logging
from pathlib import Path
from shutil import copyfile
from tempfile import TemporaryDirectory
from rdflib.exceptions import ParserError
from rdflib import Graph
class FileParserGuessFormatTest(unittest.TestCase):
def test_jsonld(self):
g = Graph()
self.assertIsInstance(g.parse("test/jsonld/1.1/manifest.jsonld"), Graph)
def test_ttl(self):
g = Graph()
self.assertIsInstance(g.parse("test/w3c/turtle/IRI_subject.ttl"), Graph)
def test_n3(self):
g = Graph()
self.assertIsInstance(g.parse("test/n3/example-lots_of_graphs.n3"), Graph)
def test_warning(self):
g = Graph()
graph_logger = logging.getLogger("rdflib")
with TemporaryDirectory() as tmpdirname:
newpath = Path(tmpdirname).joinpath("no_file_ext")
copyfile("test/rdf/Manifest.rdf", str(newpath))
with self.assertLogs(graph_logger, "WARNING"):
with self.assertRaises(ParserError):
g.parse(str(newpath))
if __name__ == "__main__":
unittest.main()
| import unittest
import logging
from pathlib import Path
from shutil import copyfile
from tempfile import TemporaryDirectory
from rdflib.exceptions import ParserError
from rdflib import Graph
class FileParserGuessFormatTest(unittest.TestCase):
def test_ttl(self):
g = Graph()
self.assertIsInstance(g.parse("test/w3c/turtle/IRI_subject.ttl"), Graph)
def test_n3(self):
g = Graph()
self.assertIsInstance(g.parse("test/n3/example-lots_of_graphs.n3"), Graph)
def test_warning(self):
g = Graph()
graph_logger = logging.getLogger("rdflib")
with TemporaryDirectory() as tmpdirname:
newpath = Path(tmpdirname).joinpath("no_file_ext")
copyfile("test/rdf/Manifest.rdf", str(newpath))
with self.assertLogs(graph_logger, "WARNING"):
with self.assertRaises(ParserError):
g.parse(str(newpath))
if __name__ == "__main__":
unittest.main()
Add test for adding JSON-LD to guess_format()
This is a follow-on patch to:
e778e9413510721c2fedaae56d4ff826df265c30
Test was confirmed to pass by running this on the current `master`
branch, and confirmed to fail with e778e941 reverted.
nosetests test/test_parse_file_guess_format.py
Signed-off-by: Alex Nelson <d5aca716d6ed55a86fa350cb5231ca56b21085ca@nist.gov>import unittest
import logging
from pathlib import Path
from shutil import copyfile
from tempfile import TemporaryDirectory
from rdflib.exceptions import ParserError
from rdflib import Graph
class FileParserGuessFormatTest(unittest.TestCase):
def test_jsonld(self):
g = Graph()
self.assertIsInstance(g.parse("test/jsonld/1.1/manifest.jsonld"), Graph)
def test_ttl(self):
g = Graph()
self.assertIsInstance(g.parse("test/w3c/turtle/IRI_subject.ttl"), Graph)
def test_n3(self):
g = Graph()
self.assertIsInstance(g.parse("test/n3/example-lots_of_graphs.n3"), Graph)
def test_warning(self):
g = Graph()
graph_logger = logging.getLogger("rdflib")
with TemporaryDirectory() as tmpdirname:
newpath = Path(tmpdirname).joinpath("no_file_ext")
copyfile("test/rdf/Manifest.rdf", str(newpath))
with self.assertLogs(graph_logger, "WARNING"):
with self.assertRaises(ParserError):
g.parse(str(newpath))
if __name__ == "__main__":
unittest.main()
| <commit_before>import unittest
import logging
from pathlib import Path
from shutil import copyfile
from tempfile import TemporaryDirectory
from rdflib.exceptions import ParserError
from rdflib import Graph
class FileParserGuessFormatTest(unittest.TestCase):
def test_ttl(self):
g = Graph()
self.assertIsInstance(g.parse("test/w3c/turtle/IRI_subject.ttl"), Graph)
def test_n3(self):
g = Graph()
self.assertIsInstance(g.parse("test/n3/example-lots_of_graphs.n3"), Graph)
def test_warning(self):
g = Graph()
graph_logger = logging.getLogger("rdflib")
with TemporaryDirectory() as tmpdirname:
newpath = Path(tmpdirname).joinpath("no_file_ext")
copyfile("test/rdf/Manifest.rdf", str(newpath))
with self.assertLogs(graph_logger, "WARNING"):
with self.assertRaises(ParserError):
g.parse(str(newpath))
if __name__ == "__main__":
unittest.main()
<commit_msg>Add test for adding JSON-LD to guess_format()
This is a follow-on patch to:
e778e9413510721c2fedaae56d4ff826df265c30
Test was confirmed to pass by running this on the current `master`
branch, and confirmed to fail with e778e941 reverted.
nosetests test/test_parse_file_guess_format.py
Signed-off-by: Alex Nelson <d5aca716d6ed55a86fa350cb5231ca56b21085ca@nist.gov><commit_after>import unittest
import logging
from pathlib import Path
from shutil import copyfile
from tempfile import TemporaryDirectory
from rdflib.exceptions import ParserError
from rdflib import Graph
class FileParserGuessFormatTest(unittest.TestCase):
def test_jsonld(self):
g = Graph()
self.assertIsInstance(g.parse("test/jsonld/1.1/manifest.jsonld"), Graph)
def test_ttl(self):
g = Graph()
self.assertIsInstance(g.parse("test/w3c/turtle/IRI_subject.ttl"), Graph)
def test_n3(self):
g = Graph()
self.assertIsInstance(g.parse("test/n3/example-lots_of_graphs.n3"), Graph)
def test_warning(self):
g = Graph()
graph_logger = logging.getLogger("rdflib")
with TemporaryDirectory() as tmpdirname:
newpath = Path(tmpdirname).joinpath("no_file_ext")
copyfile("test/rdf/Manifest.rdf", str(newpath))
with self.assertLogs(graph_logger, "WARNING"):
with self.assertRaises(ParserError):
g.parse(str(newpath))
if __name__ == "__main__":
unittest.main()
|
24c8122db0f38a1f798461a23d08535e4e6781d5 | photo/idxitem.py | photo/idxitem.py | """Provide the class IdxItem which represents an item in the index.
"""
import hashlib
def _md5file(fname):
"""Calculate the md5 hash for a file.
"""
m = hashlib.md5()
chunksize = 8192
with open(fname, 'rb') as f:
while True:
chunk = f.read(chunksize)
if not chunk:
break
m.update(chunk)
return m.hexdigest()
class IdxItem(object):
def __init__(self, filename=None, data=None):
self.filename = None
self.tags = []
if data is not None:
self.__dict__.update(data)
elif filename is not None:
self.filename = filename
self.md5 = _md5file(filename)
def as_dict(self):
return dict(self.__dict__)
| """Provide the class IdxItem which represents an item in the index.
"""
import hashlib
def _md5file(fname):
"""Calculate the md5 hash for a file.
"""
m = hashlib.md5()
chunksize = 8192
with open(fname, 'rb') as f:
while True:
chunk = f.read(chunksize)
if not chunk:
break
m.update(chunk)
return m.hexdigest()
class IdxItem(object):
def __init__(self, filename=None, data=None):
self.filename = None
self.tags = []
if data is not None:
self.__dict__.update(data)
elif filename is not None:
self.filename = filename
self.md5 = _md5file(filename)
self.tags = set(self.tags)
def as_dict(self):
d = self.__dict__.copy()
d['tags'] = list(d['tags'])
return d
| Convert tags to a set on init and back to a list on writing. | Convert tags to a set on init and back to a list on writing.
| Python | apache-2.0 | RKrahl/photo-tools | """Provide the class IdxItem which represents an item in the index.
"""
import hashlib
def _md5file(fname):
"""Calculate the md5 hash for a file.
"""
m = hashlib.md5()
chunksize = 8192
with open(fname, 'rb') as f:
while True:
chunk = f.read(chunksize)
if not chunk:
break
m.update(chunk)
return m.hexdigest()
class IdxItem(object):
def __init__(self, filename=None, data=None):
self.filename = None
self.tags = []
if data is not None:
self.__dict__.update(data)
elif filename is not None:
self.filename = filename
self.md5 = _md5file(filename)
def as_dict(self):
return dict(self.__dict__)
Convert tags to a set on init and back to a list on writing. | """Provide the class IdxItem which represents an item in the index.
"""
import hashlib
def _md5file(fname):
"""Calculate the md5 hash for a file.
"""
m = hashlib.md5()
chunksize = 8192
with open(fname, 'rb') as f:
while True:
chunk = f.read(chunksize)
if not chunk:
break
m.update(chunk)
return m.hexdigest()
class IdxItem(object):
def __init__(self, filename=None, data=None):
self.filename = None
self.tags = []
if data is not None:
self.__dict__.update(data)
elif filename is not None:
self.filename = filename
self.md5 = _md5file(filename)
self.tags = set(self.tags)
def as_dict(self):
d = self.__dict__.copy()
d['tags'] = list(d['tags'])
return d
| <commit_before>"""Provide the class IdxItem which represents an item in the index.
"""
import hashlib
def _md5file(fname):
"""Calculate the md5 hash for a file.
"""
m = hashlib.md5()
chunksize = 8192
with open(fname, 'rb') as f:
while True:
chunk = f.read(chunksize)
if not chunk:
break
m.update(chunk)
return m.hexdigest()
class IdxItem(object):
def __init__(self, filename=None, data=None):
self.filename = None
self.tags = []
if data is not None:
self.__dict__.update(data)
elif filename is not None:
self.filename = filename
self.md5 = _md5file(filename)
def as_dict(self):
return dict(self.__dict__)
<commit_msg>Convert tags to a set on init and back to a list on writing.<commit_after> | """Provide the class IdxItem which represents an item in the index.
"""
import hashlib
def _md5file(fname):
"""Calculate the md5 hash for a file.
"""
m = hashlib.md5()
chunksize = 8192
with open(fname, 'rb') as f:
while True:
chunk = f.read(chunksize)
if not chunk:
break
m.update(chunk)
return m.hexdigest()
class IdxItem(object):
def __init__(self, filename=None, data=None):
self.filename = None
self.tags = []
if data is not None:
self.__dict__.update(data)
elif filename is not None:
self.filename = filename
self.md5 = _md5file(filename)
self.tags = set(self.tags)
def as_dict(self):
d = self.__dict__.copy()
d['tags'] = list(d['tags'])
return d
| """Provide the class IdxItem which represents an item in the index.
"""
import hashlib
def _md5file(fname):
"""Calculate the md5 hash for a file.
"""
m = hashlib.md5()
chunksize = 8192
with open(fname, 'rb') as f:
while True:
chunk = f.read(chunksize)
if not chunk:
break
m.update(chunk)
return m.hexdigest()
class IdxItem(object):
def __init__(self, filename=None, data=None):
self.filename = None
self.tags = []
if data is not None:
self.__dict__.update(data)
elif filename is not None:
self.filename = filename
self.md5 = _md5file(filename)
def as_dict(self):
return dict(self.__dict__)
Convert tags to a set on init and back to a list on writing."""Provide the class IdxItem which represents an item in the index.
"""
import hashlib
def _md5file(fname):
"""Calculate the md5 hash for a file.
"""
m = hashlib.md5()
chunksize = 8192
with open(fname, 'rb') as f:
while True:
chunk = f.read(chunksize)
if not chunk:
break
m.update(chunk)
return m.hexdigest()
class IdxItem(object):
def __init__(self, filename=None, data=None):
self.filename = None
self.tags = []
if data is not None:
self.__dict__.update(data)
elif filename is not None:
self.filename = filename
self.md5 = _md5file(filename)
self.tags = set(self.tags)
def as_dict(self):
d = self.__dict__.copy()
d['tags'] = list(d['tags'])
return d
| <commit_before>"""Provide the class IdxItem which represents an item in the index.
"""
import hashlib
def _md5file(fname):
"""Calculate the md5 hash for a file.
"""
m = hashlib.md5()
chunksize = 8192
with open(fname, 'rb') as f:
while True:
chunk = f.read(chunksize)
if not chunk:
break
m.update(chunk)
return m.hexdigest()
class IdxItem(object):
def __init__(self, filename=None, data=None):
self.filename = None
self.tags = []
if data is not None:
self.__dict__.update(data)
elif filename is not None:
self.filename = filename
self.md5 = _md5file(filename)
def as_dict(self):
return dict(self.__dict__)
<commit_msg>Convert tags to a set on init and back to a list on writing.<commit_after>"""Provide the class IdxItem which represents an item in the index.
"""
import hashlib
def _md5file(fname):
"""Calculate the md5 hash for a file.
"""
m = hashlib.md5()
chunksize = 8192
with open(fname, 'rb') as f:
while True:
chunk = f.read(chunksize)
if not chunk:
break
m.update(chunk)
return m.hexdigest()
class IdxItem(object):
def __init__(self, filename=None, data=None):
self.filename = None
self.tags = []
if data is not None:
self.__dict__.update(data)
elif filename is not None:
self.filename = filename
self.md5 = _md5file(filename)
self.tags = set(self.tags)
def as_dict(self):
d = self.__dict__.copy()
d['tags'] = list(d['tags'])
return d
|
a75dbd5aa5e9b84d08919ea14743afb75182ee8b | steel/chunks/iff.py | steel/chunks/iff.py | import collections
import io
from steel.fields.numbers import BigEndian
from steel import fields
from steel.chunks import base
__all__ = ['Chunk', 'ChunkList', 'Form']
class Chunk(base.Chunk):
id = fields.String(size=4, encoding='ascii')
size = fields.Integer(size=4, endianness=BigEndian)
payload = base.Payload(size=size)
class ChunkList(base.ChunkList):
def __init__(self, *args, **kwargs):
# Just a simple override to default to a list of IFF chunks
return super(ChunkList, self).__init__(Chunk, *args, **kwargs)
class Form(base.Chunk, encoding='ascii'):
tag = fields.FixedString('FORM')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
| import collections
import io
from steel.fields.numbers import BigEndian
from steel import fields
from steel.chunks import base
__all__ = ['Chunk', 'ChunkList', 'List', 'Form', 'Prop']
class Chunk(base.Chunk):
id = fields.String(size=4, encoding='ascii')
size = fields.Integer(size=4, endianness=BigEndian)
payload = base.Payload(size=size)
class ChunkList(base.ChunkList):
def __init__(self, *args, **kwargs):
# Just a simple override to default to a list of IFF chunks
return super(ChunkList, self).__init__(Chunk, *args, **kwargs)
class List(base.Chunk, encoding='ascii'):
tag = fields.FixedString('LIST')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
class Form(base.Chunk, encoding='ascii'):
tag = fields.FixedString('FORM')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
class Prop(base.Chunk, encoding='ascii'):
tag = fields.FixedString('PROP')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
| Add a List and Prop for better IFF compliance | Add a List and Prop for better IFF compliance
| Python | bsd-3-clause | gulopine/steel | import collections
import io
from steel.fields.numbers import BigEndian
from steel import fields
from steel.chunks import base
__all__ = ['Chunk', 'ChunkList', 'Form']
class Chunk(base.Chunk):
id = fields.String(size=4, encoding='ascii')
size = fields.Integer(size=4, endianness=BigEndian)
payload = base.Payload(size=size)
class ChunkList(base.ChunkList):
def __init__(self, *args, **kwargs):
# Just a simple override to default to a list of IFF chunks
return super(ChunkList, self).__init__(Chunk, *args, **kwargs)
class Form(base.Chunk, encoding='ascii'):
tag = fields.FixedString('FORM')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
Add a List and Prop for better IFF compliance | import collections
import io
from steel.fields.numbers import BigEndian
from steel import fields
from steel.chunks import base
__all__ = ['Chunk', 'ChunkList', 'List', 'Form', 'Prop']
class Chunk(base.Chunk):
id = fields.String(size=4, encoding='ascii')
size = fields.Integer(size=4, endianness=BigEndian)
payload = base.Payload(size=size)
class ChunkList(base.ChunkList):
def __init__(self, *args, **kwargs):
# Just a simple override to default to a list of IFF chunks
return super(ChunkList, self).__init__(Chunk, *args, **kwargs)
class List(base.Chunk, encoding='ascii'):
tag = fields.FixedString('LIST')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
class Form(base.Chunk, encoding='ascii'):
tag = fields.FixedString('FORM')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
class Prop(base.Chunk, encoding='ascii'):
tag = fields.FixedString('PROP')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
| <commit_before>import collections
import io
from steel.fields.numbers import BigEndian
from steel import fields
from steel.chunks import base
__all__ = ['Chunk', 'ChunkList', 'Form']
class Chunk(base.Chunk):
id = fields.String(size=4, encoding='ascii')
size = fields.Integer(size=4, endianness=BigEndian)
payload = base.Payload(size=size)
class ChunkList(base.ChunkList):
def __init__(self, *args, **kwargs):
# Just a simple override to default to a list of IFF chunks
return super(ChunkList, self).__init__(Chunk, *args, **kwargs)
class Form(base.Chunk, encoding='ascii'):
tag = fields.FixedString('FORM')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
<commit_msg>Add a List and Prop for better IFF compliance<commit_after> | import collections
import io
from steel.fields.numbers import BigEndian
from steel import fields
from steel.chunks import base
__all__ = ['Chunk', 'ChunkList', 'List', 'Form', 'Prop']
class Chunk(base.Chunk):
id = fields.String(size=4, encoding='ascii')
size = fields.Integer(size=4, endianness=BigEndian)
payload = base.Payload(size=size)
class ChunkList(base.ChunkList):
def __init__(self, *args, **kwargs):
# Just a simple override to default to a list of IFF chunks
return super(ChunkList, self).__init__(Chunk, *args, **kwargs)
class List(base.Chunk, encoding='ascii'):
tag = fields.FixedString('LIST')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
class Form(base.Chunk, encoding='ascii'):
tag = fields.FixedString('FORM')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
class Prop(base.Chunk, encoding='ascii'):
tag = fields.FixedString('PROP')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
| import collections
import io
from steel.fields.numbers import BigEndian
from steel import fields
from steel.chunks import base
__all__ = ['Chunk', 'ChunkList', 'Form']
class Chunk(base.Chunk):
id = fields.String(size=4, encoding='ascii')
size = fields.Integer(size=4, endianness=BigEndian)
payload = base.Payload(size=size)
class ChunkList(base.ChunkList):
def __init__(self, *args, **kwargs):
# Just a simple override to default to a list of IFF chunks
return super(ChunkList, self).__init__(Chunk, *args, **kwargs)
class Form(base.Chunk, encoding='ascii'):
tag = fields.FixedString('FORM')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
Add a List and Prop for better IFF complianceimport collections
import io
from steel.fields.numbers import BigEndian
from steel import fields
from steel.chunks import base
__all__ = ['Chunk', 'ChunkList', 'List', 'Form', 'Prop']
class Chunk(base.Chunk):
id = fields.String(size=4, encoding='ascii')
size = fields.Integer(size=4, endianness=BigEndian)
payload = base.Payload(size=size)
class ChunkList(base.ChunkList):
def __init__(self, *args, **kwargs):
# Just a simple override to default to a list of IFF chunks
return super(ChunkList, self).__init__(Chunk, *args, **kwargs)
class List(base.Chunk, encoding='ascii'):
tag = fields.FixedString('LIST')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
class Form(base.Chunk, encoding='ascii'):
tag = fields.FixedString('FORM')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
class Prop(base.Chunk, encoding='ascii'):
tag = fields.FixedString('PROP')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
| <commit_before>import collections
import io
from steel.fields.numbers import BigEndian
from steel import fields
from steel.chunks import base
__all__ = ['Chunk', 'ChunkList', 'Form']
class Chunk(base.Chunk):
id = fields.String(size=4, encoding='ascii')
size = fields.Integer(size=4, endianness=BigEndian)
payload = base.Payload(size=size)
class ChunkList(base.ChunkList):
def __init__(self, *args, **kwargs):
# Just a simple override to default to a list of IFF chunks
return super(ChunkList, self).__init__(Chunk, *args, **kwargs)
class Form(base.Chunk, encoding='ascii'):
tag = fields.FixedString('FORM')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
<commit_msg>Add a List and Prop for better IFF compliance<commit_after>import collections
import io
from steel.fields.numbers import BigEndian
from steel import fields
from steel.chunks import base
__all__ = ['Chunk', 'ChunkList', 'List', 'Form', 'Prop']
class Chunk(base.Chunk):
id = fields.String(size=4, encoding='ascii')
size = fields.Integer(size=4, endianness=BigEndian)
payload = base.Payload(size=size)
class ChunkList(base.ChunkList):
def __init__(self, *args, **kwargs):
# Just a simple override to default to a list of IFF chunks
return super(ChunkList, self).__init__(Chunk, *args, **kwargs)
class List(base.Chunk, encoding='ascii'):
tag = fields.FixedString('LIST')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
class Form(base.Chunk, encoding='ascii'):
tag = fields.FixedString('FORM')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
class Prop(base.Chunk, encoding='ascii'):
tag = fields.FixedString('PROP')
size = fields.Integer(size=4, endianness=BigEndian)
id = fields.String(size=4)
payload = base.Payload(size=size)
|
376fa8dead817ae0b1e1e97547d7c95858b1fb0e | cairis/data/ImportDAO.py | cairis/data/ImportDAO.py | # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from cairis.core.ARM import *
from cairis.daemon.CairisHTTPError import CairisHTTPError, ARMHTTPError
from cairis.data.CairisDAO import CairisDAO
from cairis.bin.cimport import file_import
__author__ = 'Shamal Faily'
class ImportDAO(CairisDAO):
def __init__(self, session_id):
CairisDAO.__init__(self, session_id)
def file_import(self,importFile,mFormat,overwriteFlag):
try:
file_import(importFile,mFormat,overwriteFlag,self.session_id)
except DatabaseProxyException as ex:
self.close()
raise ARMHTTPError(ex)
| # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from cairis.core.ARM import *
from cairis.daemon.CairisHTTPError import CairisHTTPError, ARMHTTPError
from cairis.data.CairisDAO import CairisDAO
from cairis.bin.cimport import file_import
__author__ = 'Shamal Faily'
class ImportDAO(CairisDAO):
def __init__(self, session_id):
CairisDAO.__init__(self, session_id)
def file_import(self,importFile,mFormat,overwriteFlag):
try:
return file_import(importFile,mFormat,overwriteFlag,self.session_id)
except DatabaseProxyException as ex:
self.close()
raise ARMHTTPError(ex)
| Fix problems identified by broken test | Fix problems identified by broken test
| Python | apache-2.0 | nathanbjenx/cairis,failys/CAIRIS,failys/CAIRIS,nathanbjenx/cairis,nathanbjenx/cairis,nathanbjenx/cairis,failys/CAIRIS | # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from cairis.core.ARM import *
from cairis.daemon.CairisHTTPError import CairisHTTPError, ARMHTTPError
from cairis.data.CairisDAO import CairisDAO
from cairis.bin.cimport import file_import
__author__ = 'Shamal Faily'
class ImportDAO(CairisDAO):
def __init__(self, session_id):
CairisDAO.__init__(self, session_id)
def file_import(self,importFile,mFormat,overwriteFlag):
try:
file_import(importFile,mFormat,overwriteFlag,self.session_id)
except DatabaseProxyException as ex:
self.close()
raise ARMHTTPError(ex)
Fix problems identified by broken test | # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from cairis.core.ARM import *
from cairis.daemon.CairisHTTPError import CairisHTTPError, ARMHTTPError
from cairis.data.CairisDAO import CairisDAO
from cairis.bin.cimport import file_import
__author__ = 'Shamal Faily'
class ImportDAO(CairisDAO):
def __init__(self, session_id):
CairisDAO.__init__(self, session_id)
def file_import(self,importFile,mFormat,overwriteFlag):
try:
return file_import(importFile,mFormat,overwriteFlag,self.session_id)
except DatabaseProxyException as ex:
self.close()
raise ARMHTTPError(ex)
| <commit_before># Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from cairis.core.ARM import *
from cairis.daemon.CairisHTTPError import CairisHTTPError, ARMHTTPError
from cairis.data.CairisDAO import CairisDAO
from cairis.bin.cimport import file_import
__author__ = 'Shamal Faily'
class ImportDAO(CairisDAO):
def __init__(self, session_id):
CairisDAO.__init__(self, session_id)
def file_import(self,importFile,mFormat,overwriteFlag):
try:
file_import(importFile,mFormat,overwriteFlag,self.session_id)
except DatabaseProxyException as ex:
self.close()
raise ARMHTTPError(ex)
<commit_msg>Fix problems identified by broken test<commit_after> | # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from cairis.core.ARM import *
from cairis.daemon.CairisHTTPError import CairisHTTPError, ARMHTTPError
from cairis.data.CairisDAO import CairisDAO
from cairis.bin.cimport import file_import
__author__ = 'Shamal Faily'
class ImportDAO(CairisDAO):
def __init__(self, session_id):
CairisDAO.__init__(self, session_id)
def file_import(self,importFile,mFormat,overwriteFlag):
try:
return file_import(importFile,mFormat,overwriteFlag,self.session_id)
except DatabaseProxyException as ex:
self.close()
raise ARMHTTPError(ex)
| # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from cairis.core.ARM import *
from cairis.daemon.CairisHTTPError import CairisHTTPError, ARMHTTPError
from cairis.data.CairisDAO import CairisDAO
from cairis.bin.cimport import file_import
__author__ = 'Shamal Faily'
class ImportDAO(CairisDAO):
def __init__(self, session_id):
CairisDAO.__init__(self, session_id)
def file_import(self,importFile,mFormat,overwriteFlag):
try:
file_import(importFile,mFormat,overwriteFlag,self.session_id)
except DatabaseProxyException as ex:
self.close()
raise ARMHTTPError(ex)
Fix problems identified by broken test# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from cairis.core.ARM import *
from cairis.daemon.CairisHTTPError import CairisHTTPError, ARMHTTPError
from cairis.data.CairisDAO import CairisDAO
from cairis.bin.cimport import file_import
__author__ = 'Shamal Faily'
class ImportDAO(CairisDAO):
def __init__(self, session_id):
CairisDAO.__init__(self, session_id)
def file_import(self,importFile,mFormat,overwriteFlag):
try:
return file_import(importFile,mFormat,overwriteFlag,self.session_id)
except DatabaseProxyException as ex:
self.close()
raise ARMHTTPError(ex)
| <commit_before># Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from cairis.core.ARM import *
from cairis.daemon.CairisHTTPError import CairisHTTPError, ARMHTTPError
from cairis.data.CairisDAO import CairisDAO
from cairis.bin.cimport import file_import
__author__ = 'Shamal Faily'
class ImportDAO(CairisDAO):
def __init__(self, session_id):
CairisDAO.__init__(self, session_id)
def file_import(self,importFile,mFormat,overwriteFlag):
try:
file_import(importFile,mFormat,overwriteFlag,self.session_id)
except DatabaseProxyException as ex:
self.close()
raise ARMHTTPError(ex)
<commit_msg>Fix problems identified by broken test<commit_after># Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from cairis.core.ARM import *
from cairis.daemon.CairisHTTPError import CairisHTTPError, ARMHTTPError
from cairis.data.CairisDAO import CairisDAO
from cairis.bin.cimport import file_import
__author__ = 'Shamal Faily'
class ImportDAO(CairisDAO):
def __init__(self, session_id):
CairisDAO.__init__(self, session_id)
def file_import(self,importFile,mFormat,overwriteFlag):
try:
return file_import(importFile,mFormat,overwriteFlag,self.session_id)
except DatabaseProxyException as ex:
self.close()
raise ARMHTTPError(ex)
|
15ebd5a3509b20bad4cf0123dfac9be6878fa91c | app/models/bookmarks.py | app/models/bookmarks.py | from flask import current_app
from .. import db, login_manager
class Bookmarks(db.Model):
id = db.Column(db.Integer, primary_key=True)
listing_id = db.Column(db.Integer, unique=True)
merchant_id = db.Column(db.Integer, db.ForeignKey('user.id')) #also needs to be added to Hunter's Vendors model
def __init__(self, listing_id, merchant_id):
self.listing_id = listing_id
self.merchant_id = merchant_id
def __repr__(self):
return "<User: {} Bookmarked Listing: {}".format(self.merchant_id, self.listing_id)
| from flask import current_app
from .. import db, login_manager
class Bookmarks(db.Model):
id = db.Column(db.Integer, primary_key=True)
listing_id = db.Column(db.Integer, unique=True)
merchant_id = db.Column(db.Integer, db.ForeignKey('user.id'))
merchant = db.relationship('User', backref=db.backref('bookmarks', lazy='dynamic'))
def __init__(self, listing_id, merchant):
self.listing_id = listing_id
self.merchant = merchant
def __repr__(self):
return "<User: {} Bookmarked Listing: {}".format(self.merchant_id, self.listing_id)
| Set up the relationship between bookmark and merchant | Set up the relationship between bookmark and merchant
| Python | mit | hack4impact/reading-terminal-market,hack4impact/reading-terminal-market,hack4impact/reading-terminal-market | from flask import current_app
from .. import db, login_manager
class Bookmarks(db.Model):
id = db.Column(db.Integer, primary_key=True)
listing_id = db.Column(db.Integer, unique=True)
merchant_id = db.Column(db.Integer, db.ForeignKey('user.id')) #also needs to be added to Hunter's Vendors model
def __init__(self, listing_id, merchant_id):
self.listing_id = listing_id
self.merchant_id = merchant_id
def __repr__(self):
return "<User: {} Bookmarked Listing: {}".format(self.merchant_id, self.listing_id)
Set up the relationship between bookmark and merchant | from flask import current_app
from .. import db, login_manager
class Bookmarks(db.Model):
id = db.Column(db.Integer, primary_key=True)
listing_id = db.Column(db.Integer, unique=True)
merchant_id = db.Column(db.Integer, db.ForeignKey('user.id'))
merchant = db.relationship('User', backref=db.backref('bookmarks', lazy='dynamic'))
def __init__(self, listing_id, merchant):
self.listing_id = listing_id
self.merchant = merchant
def __repr__(self):
return "<User: {} Bookmarked Listing: {}".format(self.merchant_id, self.listing_id)
| <commit_before>from flask import current_app
from .. import db, login_manager
class Bookmarks(db.Model):
id = db.Column(db.Integer, primary_key=True)
listing_id = db.Column(db.Integer, unique=True)
merchant_id = db.Column(db.Integer, db.ForeignKey('user.id')) #also needs to be added to Hunter's Vendors model
def __init__(self, listing_id, merchant_id):
self.listing_id = listing_id
self.merchant_id = merchant_id
def __repr__(self):
return "<User: {} Bookmarked Listing: {}".format(self.merchant_id, self.listing_id)
<commit_msg>Set up the relationship between bookmark and merchant<commit_after> | from flask import current_app
from .. import db, login_manager
class Bookmarks(db.Model):
id = db.Column(db.Integer, primary_key=True)
listing_id = db.Column(db.Integer, unique=True)
merchant_id = db.Column(db.Integer, db.ForeignKey('user.id'))
merchant = db.relationship('User', backref=db.backref('bookmarks', lazy='dynamic'))
def __init__(self, listing_id, merchant):
self.listing_id = listing_id
self.merchant = merchant
def __repr__(self):
return "<User: {} Bookmarked Listing: {}".format(self.merchant_id, self.listing_id)
| from flask import current_app
from .. import db, login_manager
class Bookmarks(db.Model):
id = db.Column(db.Integer, primary_key=True)
listing_id = db.Column(db.Integer, unique=True)
merchant_id = db.Column(db.Integer, db.ForeignKey('user.id')) #also needs to be added to Hunter's Vendors model
def __init__(self, listing_id, merchant_id):
self.listing_id = listing_id
self.merchant_id = merchant_id
def __repr__(self):
return "<User: {} Bookmarked Listing: {}".format(self.merchant_id, self.listing_id)
Set up the relationship between bookmark and merchantfrom flask import current_app
from .. import db, login_manager
class Bookmarks(db.Model):
id = db.Column(db.Integer, primary_key=True)
listing_id = db.Column(db.Integer, unique=True)
merchant_id = db.Column(db.Integer, db.ForeignKey('user.id'))
merchant = db.relationship('User', backref=db.backref('bookmarks', lazy='dynamic'))
def __init__(self, listing_id, merchant):
self.listing_id = listing_id
self.merchant = merchant
def __repr__(self):
return "<User: {} Bookmarked Listing: {}".format(self.merchant_id, self.listing_id)
| <commit_before>from flask import current_app
from .. import db, login_manager
class Bookmarks(db.Model):
id = db.Column(db.Integer, primary_key=True)
listing_id = db.Column(db.Integer, unique=True)
merchant_id = db.Column(db.Integer, db.ForeignKey('user.id')) #also needs to be added to Hunter's Vendors model
def __init__(self, listing_id, merchant_id):
self.listing_id = listing_id
self.merchant_id = merchant_id
def __repr__(self):
return "<User: {} Bookmarked Listing: {}".format(self.merchant_id, self.listing_id)
<commit_msg>Set up the relationship between bookmark and merchant<commit_after>from flask import current_app
from .. import db, login_manager
class Bookmarks(db.Model):
id = db.Column(db.Integer, primary_key=True)
listing_id = db.Column(db.Integer, unique=True)
merchant_id = db.Column(db.Integer, db.ForeignKey('user.id'))
merchant = db.relationship('User', backref=db.backref('bookmarks', lazy='dynamic'))
def __init__(self, listing_id, merchant):
self.listing_id = listing_id
self.merchant = merchant
def __repr__(self):
return "<User: {} Bookmarked Listing: {}".format(self.merchant_id, self.listing_id)
|
c8aca84619493cd75cb12b2fc63dc4dccb158032 | common/lib/chem/setup.py | common/lib/chem/setup.py | from setuptools import setup
setup(
name="chem",
version="0.1.1",
packages=["chem"],
install_requires=[
"pyparsing==2.0.7",
"numpy==1.6.2",
"scipy==0.14.0",
"nltk==3.2.5",
],
)
| from setuptools import setup
setup(
name="chem",
version="0.1.2",
packages=["chem"],
install_requires=[
"pyparsing==2.0.7",
"numpy==1.6.2",
"scipy==0.14.0",
"nltk==3.2.5",
],
)
| Update chem version to force new nltk requirement to be picked up | Update chem version to force new nltk requirement to be picked up
| Python | agpl-3.0 | procangroup/edx-platform,teltek/edx-platform,kmoocdev2/edx-platform,eduNEXT/edunext-platform,a-parhom/edx-platform,edx-solutions/edx-platform,a-parhom/edx-platform,gymnasium/edx-platform,Edraak/edraak-platform,appsembler/edx-platform,edx-solutions/edx-platform,ahmedaljazzar/edx-platform,kmoocdev2/edx-platform,eduNEXT/edx-platform,proversity-org/edx-platform,eduNEXT/edunext-platform,arbrandes/edx-platform,cpennington/edx-platform,arbrandes/edx-platform,stvstnfrd/edx-platform,proversity-org/edx-platform,angelapper/edx-platform,cpennington/edx-platform,ahmedaljazzar/edx-platform,Edraak/edraak-platform,appsembler/edx-platform,msegado/edx-platform,CredoReference/edx-platform,Stanford-Online/edx-platform,Stanford-Online/edx-platform,jolyonb/edx-platform,ESOedX/edx-platform,stvstnfrd/edx-platform,stvstnfrd/edx-platform,Edraak/edraak-platform,TeachAtTUM/edx-platform,proversity-org/edx-platform,eduNEXT/edx-platform,proversity-org/edx-platform,arbrandes/edx-platform,a-parhom/edx-platform,philanthropy-u/edx-platform,procangroup/edx-platform,gsehub/edx-platform,Edraak/edraak-platform,eduNEXT/edunext-platform,BehavioralInsightsTeam/edx-platform,procangroup/edx-platform,gsehub/edx-platform,CredoReference/edx-platform,gymnasium/edx-platform,jolyonb/edx-platform,gymnasium/edx-platform,cpennington/edx-platform,edx/edx-platform,ahmedaljazzar/edx-platform,Stanford-Online/edx-platform,TeachAtTUM/edx-platform,EDUlib/edx-platform,appsembler/edx-platform,EDUlib/edx-platform,edx-solutions/edx-platform,philanthropy-u/edx-platform,procangroup/edx-platform,mitocw/edx-platform,eduNEXT/edunext-platform,msegado/edx-platform,gymnasium/edx-platform,mitocw/edx-platform,philanthropy-u/edx-platform,TeachAtTUM/edx-platform,ESOedX/edx-platform,arbrandes/edx-platform,edx-solutions/edx-platform,ESOedX/edx-platform,EDUlib/edx-platform,stvstnfrd/edx-platform,BehavioralInsightsTeam/edx-platform,gsehub/edx-platform,appsembler/edx-platform,cpennington/edx-platform,msegado/edx-platform,eduNEXT/edx-platform,kmoocdev2/edx-platform,angelapper/edx-platform,CredoReference/edx-platform,msegado/edx-platform,mitocw/edx-platform,TeachAtTUM/edx-platform,angelapper/edx-platform,EDUlib/edx-platform,a-parhom/edx-platform,teltek/edx-platform,BehavioralInsightsTeam/edx-platform,kmoocdev2/edx-platform,edx/edx-platform,jolyonb/edx-platform,teltek/edx-platform,edx/edx-platform,ahmedaljazzar/edx-platform,msegado/edx-platform,eduNEXT/edx-platform,CredoReference/edx-platform,BehavioralInsightsTeam/edx-platform,edx/edx-platform,ESOedX/edx-platform,Stanford-Online/edx-platform,mitocw/edx-platform,teltek/edx-platform,philanthropy-u/edx-platform,angelapper/edx-platform,gsehub/edx-platform,kmoocdev2/edx-platform,jolyonb/edx-platform | from setuptools import setup
setup(
name="chem",
version="0.1.1",
packages=["chem"],
install_requires=[
"pyparsing==2.0.7",
"numpy==1.6.2",
"scipy==0.14.0",
"nltk==3.2.5",
],
)
Update chem version to force new nltk requirement to be picked up | from setuptools import setup
setup(
name="chem",
version="0.1.2",
packages=["chem"],
install_requires=[
"pyparsing==2.0.7",
"numpy==1.6.2",
"scipy==0.14.0",
"nltk==3.2.5",
],
)
| <commit_before>from setuptools import setup
setup(
name="chem",
version="0.1.1",
packages=["chem"],
install_requires=[
"pyparsing==2.0.7",
"numpy==1.6.2",
"scipy==0.14.0",
"nltk==3.2.5",
],
)
<commit_msg>Update chem version to force new nltk requirement to be picked up<commit_after> | from setuptools import setup
setup(
name="chem",
version="0.1.2",
packages=["chem"],
install_requires=[
"pyparsing==2.0.7",
"numpy==1.6.2",
"scipy==0.14.0",
"nltk==3.2.5",
],
)
| from setuptools import setup
setup(
name="chem",
version="0.1.1",
packages=["chem"],
install_requires=[
"pyparsing==2.0.7",
"numpy==1.6.2",
"scipy==0.14.0",
"nltk==3.2.5",
],
)
Update chem version to force new nltk requirement to be picked upfrom setuptools import setup
setup(
name="chem",
version="0.1.2",
packages=["chem"],
install_requires=[
"pyparsing==2.0.7",
"numpy==1.6.2",
"scipy==0.14.0",
"nltk==3.2.5",
],
)
| <commit_before>from setuptools import setup
setup(
name="chem",
version="0.1.1",
packages=["chem"],
install_requires=[
"pyparsing==2.0.7",
"numpy==1.6.2",
"scipy==0.14.0",
"nltk==3.2.5",
],
)
<commit_msg>Update chem version to force new nltk requirement to be picked up<commit_after>from setuptools import setup
setup(
name="chem",
version="0.1.2",
packages=["chem"],
install_requires=[
"pyparsing==2.0.7",
"numpy==1.6.2",
"scipy==0.14.0",
"nltk==3.2.5",
],
)
|
013fa911c7b882a0b362549d4d9b1f9e1e688bc8 | violations/py_unittest.py | violations/py_unittest.py | import re
from django.template.loader import render_to_string
from tasks.const import STATUS_SUCCESS, STATUS_FAILED
from .base import library
@library.register('py_unittest')
def py_unittest_violation(data):
"""Python unittest violation parser"""
lines = data['raw'].split('\n')
line = ''
while len(lines) and not line.startswith('Ran'):
line = lines.pop(0)
summary = line
status = lines.pop(1)
data['status'] =\
STATUS_SUCCESS if status.find('OK') == 0 else STATUS_FAILED
data['preview'] = render_to_string('violations/py_tests/preview.html', {
'summary': summary,
'status': status,
})
data['prepared'] = render_to_string('violations/py_tests/prepared.html', {
'raw': data['raw'],
})
plot = {'failures': 0, 'errors': 0}
fail_match = re.match(r'.*failures=(\d*).*', status)
if fail_match:
plot['failures'] = int(fail_match.groups()[0])
error_match = re.match(r'.*errors=(\d*).*', status)
if error_match:
plot['errors'] = int(error_match.groups()[0])
data['plot'] = plot
return data
| import re
from django.template.loader import render_to_string
from tasks.const import STATUS_SUCCESS, STATUS_FAILED
from .base import library
@library.register('py_unittest')
def py_unittest_violation(data):
"""Python unittest violation parser"""
lines = data['raw'].split('\n')
line = ''
while len(lines) and not line.startswith('Ran'):
line = lines.pop(0)
summary = line
status = lines.pop(1)
data['status'] =\
STATUS_SUCCESS if status.find('OK') == 0 else STATUS_FAILED
data['preview'] = render_to_string('violations/py_tests/preview.html', {
'summary': summary,
'status': status,
})
data['prepared'] = render_to_string('violations/py_tests/prepared.html', {
'raw': data['raw'],
})
plot = {'failures': 0, 'errors': 0}
fail_match = re.match(r'.*failures=(\d*).*', status)
if fail_match:
plot['failures'] = int(fail_match.groups()[0])
error_match = re.match(r'.*errors=(\d*).*', status)
if error_match:
plot['errors'] = int(error_match.groups()[0])
total_match = re.match(r'Ran (\d*) tests .*', summary)
if total_match:
plot['test_count'] = int(total_match.groups()[0])
data['plot'] = plot
return data
| Add total tests count to py unittest graph | Add total tests count to py unittest graph
| Python | mit | nvbn/coviolations_web,nvbn/coviolations_web | import re
from django.template.loader import render_to_string
from tasks.const import STATUS_SUCCESS, STATUS_FAILED
from .base import library
@library.register('py_unittest')
def py_unittest_violation(data):
"""Python unittest violation parser"""
lines = data['raw'].split('\n')
line = ''
while len(lines) and not line.startswith('Ran'):
line = lines.pop(0)
summary = line
status = lines.pop(1)
data['status'] =\
STATUS_SUCCESS if status.find('OK') == 0 else STATUS_FAILED
data['preview'] = render_to_string('violations/py_tests/preview.html', {
'summary': summary,
'status': status,
})
data['prepared'] = render_to_string('violations/py_tests/prepared.html', {
'raw': data['raw'],
})
plot = {'failures': 0, 'errors': 0}
fail_match = re.match(r'.*failures=(\d*).*', status)
if fail_match:
plot['failures'] = int(fail_match.groups()[0])
error_match = re.match(r'.*errors=(\d*).*', status)
if error_match:
plot['errors'] = int(error_match.groups()[0])
data['plot'] = plot
return data
Add total tests count to py unittest graph | import re
from django.template.loader import render_to_string
from tasks.const import STATUS_SUCCESS, STATUS_FAILED
from .base import library
@library.register('py_unittest')
def py_unittest_violation(data):
"""Python unittest violation parser"""
lines = data['raw'].split('\n')
line = ''
while len(lines) and not line.startswith('Ran'):
line = lines.pop(0)
summary = line
status = lines.pop(1)
data['status'] =\
STATUS_SUCCESS if status.find('OK') == 0 else STATUS_FAILED
data['preview'] = render_to_string('violations/py_tests/preview.html', {
'summary': summary,
'status': status,
})
data['prepared'] = render_to_string('violations/py_tests/prepared.html', {
'raw': data['raw'],
})
plot = {'failures': 0, 'errors': 0}
fail_match = re.match(r'.*failures=(\d*).*', status)
if fail_match:
plot['failures'] = int(fail_match.groups()[0])
error_match = re.match(r'.*errors=(\d*).*', status)
if error_match:
plot['errors'] = int(error_match.groups()[0])
total_match = re.match(r'Ran (\d*) tests .*', summary)
if total_match:
plot['test_count'] = int(total_match.groups()[0])
data['plot'] = plot
return data
| <commit_before>import re
from django.template.loader import render_to_string
from tasks.const import STATUS_SUCCESS, STATUS_FAILED
from .base import library
@library.register('py_unittest')
def py_unittest_violation(data):
"""Python unittest violation parser"""
lines = data['raw'].split('\n')
line = ''
while len(lines) and not line.startswith('Ran'):
line = lines.pop(0)
summary = line
status = lines.pop(1)
data['status'] =\
STATUS_SUCCESS if status.find('OK') == 0 else STATUS_FAILED
data['preview'] = render_to_string('violations/py_tests/preview.html', {
'summary': summary,
'status': status,
})
data['prepared'] = render_to_string('violations/py_tests/prepared.html', {
'raw': data['raw'],
})
plot = {'failures': 0, 'errors': 0}
fail_match = re.match(r'.*failures=(\d*).*', status)
if fail_match:
plot['failures'] = int(fail_match.groups()[0])
error_match = re.match(r'.*errors=(\d*).*', status)
if error_match:
plot['errors'] = int(error_match.groups()[0])
data['plot'] = plot
return data
<commit_msg>Add total tests count to py unittest graph<commit_after> | import re
from django.template.loader import render_to_string
from tasks.const import STATUS_SUCCESS, STATUS_FAILED
from .base import library
@library.register('py_unittest')
def py_unittest_violation(data):
"""Python unittest violation parser"""
lines = data['raw'].split('\n')
line = ''
while len(lines) and not line.startswith('Ran'):
line = lines.pop(0)
summary = line
status = lines.pop(1)
data['status'] =\
STATUS_SUCCESS if status.find('OK') == 0 else STATUS_FAILED
data['preview'] = render_to_string('violations/py_tests/preview.html', {
'summary': summary,
'status': status,
})
data['prepared'] = render_to_string('violations/py_tests/prepared.html', {
'raw': data['raw'],
})
plot = {'failures': 0, 'errors': 0}
fail_match = re.match(r'.*failures=(\d*).*', status)
if fail_match:
plot['failures'] = int(fail_match.groups()[0])
error_match = re.match(r'.*errors=(\d*).*', status)
if error_match:
plot['errors'] = int(error_match.groups()[0])
total_match = re.match(r'Ran (\d*) tests .*', summary)
if total_match:
plot['test_count'] = int(total_match.groups()[0])
data['plot'] = plot
return data
| import re
from django.template.loader import render_to_string
from tasks.const import STATUS_SUCCESS, STATUS_FAILED
from .base import library
@library.register('py_unittest')
def py_unittest_violation(data):
"""Python unittest violation parser"""
lines = data['raw'].split('\n')
line = ''
while len(lines) and not line.startswith('Ran'):
line = lines.pop(0)
summary = line
status = lines.pop(1)
data['status'] =\
STATUS_SUCCESS if status.find('OK') == 0 else STATUS_FAILED
data['preview'] = render_to_string('violations/py_tests/preview.html', {
'summary': summary,
'status': status,
})
data['prepared'] = render_to_string('violations/py_tests/prepared.html', {
'raw': data['raw'],
})
plot = {'failures': 0, 'errors': 0}
fail_match = re.match(r'.*failures=(\d*).*', status)
if fail_match:
plot['failures'] = int(fail_match.groups()[0])
error_match = re.match(r'.*errors=(\d*).*', status)
if error_match:
plot['errors'] = int(error_match.groups()[0])
data['plot'] = plot
return data
Add total tests count to py unittest graphimport re
from django.template.loader import render_to_string
from tasks.const import STATUS_SUCCESS, STATUS_FAILED
from .base import library
@library.register('py_unittest')
def py_unittest_violation(data):
"""Python unittest violation parser"""
lines = data['raw'].split('\n')
line = ''
while len(lines) and not line.startswith('Ran'):
line = lines.pop(0)
summary = line
status = lines.pop(1)
data['status'] =\
STATUS_SUCCESS if status.find('OK') == 0 else STATUS_FAILED
data['preview'] = render_to_string('violations/py_tests/preview.html', {
'summary': summary,
'status': status,
})
data['prepared'] = render_to_string('violations/py_tests/prepared.html', {
'raw': data['raw'],
})
plot = {'failures': 0, 'errors': 0}
fail_match = re.match(r'.*failures=(\d*).*', status)
if fail_match:
plot['failures'] = int(fail_match.groups()[0])
error_match = re.match(r'.*errors=(\d*).*', status)
if error_match:
plot['errors'] = int(error_match.groups()[0])
total_match = re.match(r'Ran (\d*) tests .*', summary)
if total_match:
plot['test_count'] = int(total_match.groups()[0])
data['plot'] = plot
return data
| <commit_before>import re
from django.template.loader import render_to_string
from tasks.const import STATUS_SUCCESS, STATUS_FAILED
from .base import library
@library.register('py_unittest')
def py_unittest_violation(data):
"""Python unittest violation parser"""
lines = data['raw'].split('\n')
line = ''
while len(lines) and not line.startswith('Ran'):
line = lines.pop(0)
summary = line
status = lines.pop(1)
data['status'] =\
STATUS_SUCCESS if status.find('OK') == 0 else STATUS_FAILED
data['preview'] = render_to_string('violations/py_tests/preview.html', {
'summary': summary,
'status': status,
})
data['prepared'] = render_to_string('violations/py_tests/prepared.html', {
'raw': data['raw'],
})
plot = {'failures': 0, 'errors': 0}
fail_match = re.match(r'.*failures=(\d*).*', status)
if fail_match:
plot['failures'] = int(fail_match.groups()[0])
error_match = re.match(r'.*errors=(\d*).*', status)
if error_match:
plot['errors'] = int(error_match.groups()[0])
data['plot'] = plot
return data
<commit_msg>Add total tests count to py unittest graph<commit_after>import re
from django.template.loader import render_to_string
from tasks.const import STATUS_SUCCESS, STATUS_FAILED
from .base import library
@library.register('py_unittest')
def py_unittest_violation(data):
"""Python unittest violation parser"""
lines = data['raw'].split('\n')
line = ''
while len(lines) and not line.startswith('Ran'):
line = lines.pop(0)
summary = line
status = lines.pop(1)
data['status'] =\
STATUS_SUCCESS if status.find('OK') == 0 else STATUS_FAILED
data['preview'] = render_to_string('violations/py_tests/preview.html', {
'summary': summary,
'status': status,
})
data['prepared'] = render_to_string('violations/py_tests/prepared.html', {
'raw': data['raw'],
})
plot = {'failures': 0, 'errors': 0}
fail_match = re.match(r'.*failures=(\d*).*', status)
if fail_match:
plot['failures'] = int(fail_match.groups()[0])
error_match = re.match(r'.*errors=(\d*).*', status)
if error_match:
plot['errors'] = int(error_match.groups()[0])
total_match = re.match(r'Ran (\d*) tests .*', summary)
if total_match:
plot['test_count'] = int(total_match.groups()[0])
data['plot'] = plot
return data
|
1c627347a55faadc28cd975d313ec45a84fcba21 | freetalks/__init__.py | freetalks/__init__.py | import webapp2
from freetalks import handler
application = webapp2.WSGIApplication([
webapp2.Route(r'/', handler.general.Home, 'home'),
webapp2.Route(r'/talk/<talk:[a-z\d-]*>', handler.talk.Display, 'talk-display'),
])
| import webapp2
from freetalks import handler
application = webapp2.WSGIApplication([
webapp2.Route(r'/', handler.general.Home, 'home'),
webapp2.Route(r'/talk/<talk:[\w]+>', handler.talk.Display, 'talk-display'),
])
| Make generic more liberal although it doesn't match all valid keys | Make generic more liberal although it doesn't match all valid keys
| Python | mit | preichenberger/freetalks,preichenberger/freetalks | import webapp2
from freetalks import handler
application = webapp2.WSGIApplication([
webapp2.Route(r'/', handler.general.Home, 'home'),
webapp2.Route(r'/talk/<talk:[a-z\d-]*>', handler.talk.Display, 'talk-display'),
])
Make generic more liberal although it doesn't match all valid keys | import webapp2
from freetalks import handler
application = webapp2.WSGIApplication([
webapp2.Route(r'/', handler.general.Home, 'home'),
webapp2.Route(r'/talk/<talk:[\w]+>', handler.talk.Display, 'talk-display'),
])
| <commit_before>import webapp2
from freetalks import handler
application = webapp2.WSGIApplication([
webapp2.Route(r'/', handler.general.Home, 'home'),
webapp2.Route(r'/talk/<talk:[a-z\d-]*>', handler.talk.Display, 'talk-display'),
])
<commit_msg>Make generic more liberal although it doesn't match all valid keys<commit_after> | import webapp2
from freetalks import handler
application = webapp2.WSGIApplication([
webapp2.Route(r'/', handler.general.Home, 'home'),
webapp2.Route(r'/talk/<talk:[\w]+>', handler.talk.Display, 'talk-display'),
])
| import webapp2
from freetalks import handler
application = webapp2.WSGIApplication([
webapp2.Route(r'/', handler.general.Home, 'home'),
webapp2.Route(r'/talk/<talk:[a-z\d-]*>', handler.talk.Display, 'talk-display'),
])
Make generic more liberal although it doesn't match all valid keysimport webapp2
from freetalks import handler
application = webapp2.WSGIApplication([
webapp2.Route(r'/', handler.general.Home, 'home'),
webapp2.Route(r'/talk/<talk:[\w]+>', handler.talk.Display, 'talk-display'),
])
| <commit_before>import webapp2
from freetalks import handler
application = webapp2.WSGIApplication([
webapp2.Route(r'/', handler.general.Home, 'home'),
webapp2.Route(r'/talk/<talk:[a-z\d-]*>', handler.talk.Display, 'talk-display'),
])
<commit_msg>Make generic more liberal although it doesn't match all valid keys<commit_after>import webapp2
from freetalks import handler
application = webapp2.WSGIApplication([
webapp2.Route(r'/', handler.general.Home, 'home'),
webapp2.Route(r'/talk/<talk:[\w]+>', handler.talk.Display, 'talk-display'),
])
|
83c79251b5040e18c5c8ac65a5e140e59edc4d3f | test_readability.py | test_readability.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
import readability
import unittest
good_code = """
def this_is_some_good_code(var):
for i in range(10):
print(i)
"""
bad_code = """
tisgc = lambda var: [print(i) for i in range(10)]
"""
apl_code = u"""
life←{↑1 ⍵∨.∧3 4=+/,¯1 0 1∘.⊖¯1 0 1∘.⌽⊂⍵}
"""
class TestReadability(unittest.TestCase):
def setUp(self):
pass
def test_good_better_than_bad(self):
good_score = readability.score(good_code)
bad_score = readability.score(bad_code)
apl_score = readability.score(apl_code)
self.assertTrue(good_score < bad_score < apl_score)
def test_ignore_pattern(self):
self.assertFalse(readability.EXT_RE.match("abc.py"))
self.assertTrue(readability.EXT_RE.match("abc.pyc"))
if __name__ == '__main__':
unittest.main()
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
import readability
import unittest
good_code = """
def this_is_some_good_code(var):
for i in range(10):
print(i)
"""
bad_code = """
tisgc = lambda var: [print(i) for i in range(10)]
"""
# taken from http://en.wikipedia.org/wiki/APL_%28programming_language%29#Examples
apl_code = u"""
life←{↑1 ⍵∨.∧3 4=+/,¯1 0 1∘.⊖¯1 0 1∘.⌽⊂⍵}
"""
class TestReadability(unittest.TestCase):
def setUp(self):
pass
def test_good_better_than_bad(self):
good_score = readability.score(good_code)
bad_score = readability.score(bad_code)
apl_score = readability.score(apl_code)
self.assertTrue(good_score < bad_score < apl_score)
def test_ignore_pattern(self):
self.assertFalse(readability.EXT_RE.match("abc.py"))
self.assertTrue(readability.EXT_RE.match("abc.pyc"))
if __name__ == '__main__':
unittest.main()
| Test APL code should give credit to Wikipedia | Test APL code should give credit to Wikipedia
| Python | mit | swenson/readability | #!/usr/bin/env python
# -*- coding: utf-8 -*-
import readability
import unittest
good_code = """
def this_is_some_good_code(var):
for i in range(10):
print(i)
"""
bad_code = """
tisgc = lambda var: [print(i) for i in range(10)]
"""
apl_code = u"""
life←{↑1 ⍵∨.∧3 4=+/,¯1 0 1∘.⊖¯1 0 1∘.⌽⊂⍵}
"""
class TestReadability(unittest.TestCase):
def setUp(self):
pass
def test_good_better_than_bad(self):
good_score = readability.score(good_code)
bad_score = readability.score(bad_code)
apl_score = readability.score(apl_code)
self.assertTrue(good_score < bad_score < apl_score)
def test_ignore_pattern(self):
self.assertFalse(readability.EXT_RE.match("abc.py"))
self.assertTrue(readability.EXT_RE.match("abc.pyc"))
if __name__ == '__main__':
unittest.main()
Test APL code should give credit to Wikipedia | #!/usr/bin/env python
# -*- coding: utf-8 -*-
import readability
import unittest
good_code = """
def this_is_some_good_code(var):
for i in range(10):
print(i)
"""
bad_code = """
tisgc = lambda var: [print(i) for i in range(10)]
"""
# taken from http://en.wikipedia.org/wiki/APL_%28programming_language%29#Examples
apl_code = u"""
life←{↑1 ⍵∨.∧3 4=+/,¯1 0 1∘.⊖¯1 0 1∘.⌽⊂⍵}
"""
class TestReadability(unittest.TestCase):
def setUp(self):
pass
def test_good_better_than_bad(self):
good_score = readability.score(good_code)
bad_score = readability.score(bad_code)
apl_score = readability.score(apl_code)
self.assertTrue(good_score < bad_score < apl_score)
def test_ignore_pattern(self):
self.assertFalse(readability.EXT_RE.match("abc.py"))
self.assertTrue(readability.EXT_RE.match("abc.pyc"))
if __name__ == '__main__':
unittest.main()
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import readability
import unittest
good_code = """
def this_is_some_good_code(var):
for i in range(10):
print(i)
"""
bad_code = """
tisgc = lambda var: [print(i) for i in range(10)]
"""
apl_code = u"""
life←{↑1 ⍵∨.∧3 4=+/,¯1 0 1∘.⊖¯1 0 1∘.⌽⊂⍵}
"""
class TestReadability(unittest.TestCase):
def setUp(self):
pass
def test_good_better_than_bad(self):
good_score = readability.score(good_code)
bad_score = readability.score(bad_code)
apl_score = readability.score(apl_code)
self.assertTrue(good_score < bad_score < apl_score)
def test_ignore_pattern(self):
self.assertFalse(readability.EXT_RE.match("abc.py"))
self.assertTrue(readability.EXT_RE.match("abc.pyc"))
if __name__ == '__main__':
unittest.main()
<commit_msg>Test APL code should give credit to Wikipedia<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
import readability
import unittest
good_code = """
def this_is_some_good_code(var):
for i in range(10):
print(i)
"""
bad_code = """
tisgc = lambda var: [print(i) for i in range(10)]
"""
# taken from http://en.wikipedia.org/wiki/APL_%28programming_language%29#Examples
apl_code = u"""
life←{↑1 ⍵∨.∧3 4=+/,¯1 0 1∘.⊖¯1 0 1∘.⌽⊂⍵}
"""
class TestReadability(unittest.TestCase):
def setUp(self):
pass
def test_good_better_than_bad(self):
good_score = readability.score(good_code)
bad_score = readability.score(bad_code)
apl_score = readability.score(apl_code)
self.assertTrue(good_score < bad_score < apl_score)
def test_ignore_pattern(self):
self.assertFalse(readability.EXT_RE.match("abc.py"))
self.assertTrue(readability.EXT_RE.match("abc.pyc"))
if __name__ == '__main__':
unittest.main()
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
import readability
import unittest
good_code = """
def this_is_some_good_code(var):
for i in range(10):
print(i)
"""
bad_code = """
tisgc = lambda var: [print(i) for i in range(10)]
"""
apl_code = u"""
life←{↑1 ⍵∨.∧3 4=+/,¯1 0 1∘.⊖¯1 0 1∘.⌽⊂⍵}
"""
class TestReadability(unittest.TestCase):
def setUp(self):
pass
def test_good_better_than_bad(self):
good_score = readability.score(good_code)
bad_score = readability.score(bad_code)
apl_score = readability.score(apl_code)
self.assertTrue(good_score < bad_score < apl_score)
def test_ignore_pattern(self):
self.assertFalse(readability.EXT_RE.match("abc.py"))
self.assertTrue(readability.EXT_RE.match("abc.pyc"))
if __name__ == '__main__':
unittest.main()
Test APL code should give credit to Wikipedia#!/usr/bin/env python
# -*- coding: utf-8 -*-
import readability
import unittest
good_code = """
def this_is_some_good_code(var):
for i in range(10):
print(i)
"""
bad_code = """
tisgc = lambda var: [print(i) for i in range(10)]
"""
# taken from http://en.wikipedia.org/wiki/APL_%28programming_language%29#Examples
apl_code = u"""
life←{↑1 ⍵∨.∧3 4=+/,¯1 0 1∘.⊖¯1 0 1∘.⌽⊂⍵}
"""
class TestReadability(unittest.TestCase):
def setUp(self):
pass
def test_good_better_than_bad(self):
good_score = readability.score(good_code)
bad_score = readability.score(bad_code)
apl_score = readability.score(apl_code)
self.assertTrue(good_score < bad_score < apl_score)
def test_ignore_pattern(self):
self.assertFalse(readability.EXT_RE.match("abc.py"))
self.assertTrue(readability.EXT_RE.match("abc.pyc"))
if __name__ == '__main__':
unittest.main()
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import readability
import unittest
good_code = """
def this_is_some_good_code(var):
for i in range(10):
print(i)
"""
bad_code = """
tisgc = lambda var: [print(i) for i in range(10)]
"""
apl_code = u"""
life←{↑1 ⍵∨.∧3 4=+/,¯1 0 1∘.⊖¯1 0 1∘.⌽⊂⍵}
"""
class TestReadability(unittest.TestCase):
def setUp(self):
pass
def test_good_better_than_bad(self):
good_score = readability.score(good_code)
bad_score = readability.score(bad_code)
apl_score = readability.score(apl_code)
self.assertTrue(good_score < bad_score < apl_score)
def test_ignore_pattern(self):
self.assertFalse(readability.EXT_RE.match("abc.py"))
self.assertTrue(readability.EXT_RE.match("abc.pyc"))
if __name__ == '__main__':
unittest.main()
<commit_msg>Test APL code should give credit to Wikipedia<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import readability
import unittest
good_code = """
def this_is_some_good_code(var):
for i in range(10):
print(i)
"""
bad_code = """
tisgc = lambda var: [print(i) for i in range(10)]
"""
# taken from http://en.wikipedia.org/wiki/APL_%28programming_language%29#Examples
apl_code = u"""
life←{↑1 ⍵∨.∧3 4=+/,¯1 0 1∘.⊖¯1 0 1∘.⌽⊂⍵}
"""
class TestReadability(unittest.TestCase):
def setUp(self):
pass
def test_good_better_than_bad(self):
good_score = readability.score(good_code)
bad_score = readability.score(bad_code)
apl_score = readability.score(apl_code)
self.assertTrue(good_score < bad_score < apl_score)
def test_ignore_pattern(self):
self.assertFalse(readability.EXT_RE.match("abc.py"))
self.assertTrue(readability.EXT_RE.match("abc.pyc"))
if __name__ == '__main__':
unittest.main()
|
db03af21b3f46e5f5af89ccf224bc2bf4b9f6d9b | zephyr/projects/herobrine/BUILD.py | zephyr/projects/herobrine/BUILD.py | # Copyright 2021 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
def register_variant(project_name, extra_dts_overlays=(), extra_kconfig_files=()):
register_npcx_project(
project_name=project_name,
zephyr_board="npcx9",
dts_overlays=[
# Common to all projects.
here / "adc.dts",
here / "battery.dts",
here / "gpio.dts",
here / "common.dts",
here / "i2c.dts",
here / "interrupts.dts",
here / "motionsense.dts",
here / "pwm.dts",
here / "switchcap.dts",
here / "usbc.dts",
# Project-specific DTS customization.
*extra_dts_overlays,
],
kconfig_files=[
# Common to all projects.
here / "prj.conf",
# Project-specific KConfig customization.
*extra_kconfig_files,
],
)
register_variant(
project_name="herobrine_npcx9",
)
register_variant(
project_name="hoglin",
extra_kconfig_files=[here / "prj_hoglin.conf"],
)
| # Copyright 2021 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
def register_variant(project_name, extra_dts_overlays=(), extra_kconfig_files=()):
register_npcx_project(
project_name=project_name,
zephyr_board="npcx9",
dts_overlays=[
# Common to all projects.
here / "adc.dts",
here / "battery.dts",
here / "gpio.dts",
here / "common.dts",
here / "i2c.dts",
here / "interrupts.dts",
here / "motionsense.dts",
here / "pwm.dts",
here / "switchcap.dts",
here / "usbc.dts",
# Project-specific DTS customization.
*extra_dts_overlays,
],
kconfig_files=[
# Common to all projects.
here / "prj.conf",
# Project-specific KConfig customization.
*extra_kconfig_files,
],
)
register_variant(
project_name="herobrine_npcx9",
extra_kconfig_files=[here / "prj_herobrine_npcx9.conf"],
)
register_variant(
project_name="hoglin",
extra_kconfig_files=[here / "prj_hoglin.conf"],
)
| Include the missing project config | herobrine: Include the missing project config
Need to include the project config prj_herobrine_npcx9.conf to
the build. It defines CONFIG_BOARD_HEROBRINE_NPCX9=y. The board
specific alternative component code (alt_dev_replacement.c)
requires this Kconfig option.
BRANCH=None
BUG=b:216836197
TEST=Booted the herobrine_npcx9 image. No PPC access error and
PD message loop.
Change-Id: I73a9987ff49a4dfb346ebe7c994418278817c03b
Signed-off-by: Wai-Hong Tam <04b587fdf5845741a0c5a8b9cd59ca72d73ef8fc@google.com>
Reviewed-on: https://chromium-review.googlesource.com/c/chromiumos/platform/ec/+/3422771
Reviewed-by: Keith Short <f6d578f97033c25f1206cb57d52f358974c82954@chromium.org>
| Python | bsd-3-clause | coreboot/chrome-ec,coreboot/chrome-ec,coreboot/chrome-ec,coreboot/chrome-ec,coreboot/chrome-ec,coreboot/chrome-ec | # Copyright 2021 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
def register_variant(project_name, extra_dts_overlays=(), extra_kconfig_files=()):
register_npcx_project(
project_name=project_name,
zephyr_board="npcx9",
dts_overlays=[
# Common to all projects.
here / "adc.dts",
here / "battery.dts",
here / "gpio.dts",
here / "common.dts",
here / "i2c.dts",
here / "interrupts.dts",
here / "motionsense.dts",
here / "pwm.dts",
here / "switchcap.dts",
here / "usbc.dts",
# Project-specific DTS customization.
*extra_dts_overlays,
],
kconfig_files=[
# Common to all projects.
here / "prj.conf",
# Project-specific KConfig customization.
*extra_kconfig_files,
],
)
register_variant(
project_name="herobrine_npcx9",
)
register_variant(
project_name="hoglin",
extra_kconfig_files=[here / "prj_hoglin.conf"],
)
herobrine: Include the missing project config
Need to include the project config prj_herobrine_npcx9.conf to
the build. It defines CONFIG_BOARD_HEROBRINE_NPCX9=y. The board
specific alternative component code (alt_dev_replacement.c)
requires this Kconfig option.
BRANCH=None
BUG=b:216836197
TEST=Booted the herobrine_npcx9 image. No PPC access error and
PD message loop.
Change-Id: I73a9987ff49a4dfb346ebe7c994418278817c03b
Signed-off-by: Wai-Hong Tam <04b587fdf5845741a0c5a8b9cd59ca72d73ef8fc@google.com>
Reviewed-on: https://chromium-review.googlesource.com/c/chromiumos/platform/ec/+/3422771
Reviewed-by: Keith Short <f6d578f97033c25f1206cb57d52f358974c82954@chromium.org> | # Copyright 2021 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
def register_variant(project_name, extra_dts_overlays=(), extra_kconfig_files=()):
register_npcx_project(
project_name=project_name,
zephyr_board="npcx9",
dts_overlays=[
# Common to all projects.
here / "adc.dts",
here / "battery.dts",
here / "gpio.dts",
here / "common.dts",
here / "i2c.dts",
here / "interrupts.dts",
here / "motionsense.dts",
here / "pwm.dts",
here / "switchcap.dts",
here / "usbc.dts",
# Project-specific DTS customization.
*extra_dts_overlays,
],
kconfig_files=[
# Common to all projects.
here / "prj.conf",
# Project-specific KConfig customization.
*extra_kconfig_files,
],
)
register_variant(
project_name="herobrine_npcx9",
extra_kconfig_files=[here / "prj_herobrine_npcx9.conf"],
)
register_variant(
project_name="hoglin",
extra_kconfig_files=[here / "prj_hoglin.conf"],
)
| <commit_before># Copyright 2021 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
def register_variant(project_name, extra_dts_overlays=(), extra_kconfig_files=()):
register_npcx_project(
project_name=project_name,
zephyr_board="npcx9",
dts_overlays=[
# Common to all projects.
here / "adc.dts",
here / "battery.dts",
here / "gpio.dts",
here / "common.dts",
here / "i2c.dts",
here / "interrupts.dts",
here / "motionsense.dts",
here / "pwm.dts",
here / "switchcap.dts",
here / "usbc.dts",
# Project-specific DTS customization.
*extra_dts_overlays,
],
kconfig_files=[
# Common to all projects.
here / "prj.conf",
# Project-specific KConfig customization.
*extra_kconfig_files,
],
)
register_variant(
project_name="herobrine_npcx9",
)
register_variant(
project_name="hoglin",
extra_kconfig_files=[here / "prj_hoglin.conf"],
)
<commit_msg>herobrine: Include the missing project config
Need to include the project config prj_herobrine_npcx9.conf to
the build. It defines CONFIG_BOARD_HEROBRINE_NPCX9=y. The board
specific alternative component code (alt_dev_replacement.c)
requires this Kconfig option.
BRANCH=None
BUG=b:216836197
TEST=Booted the herobrine_npcx9 image. No PPC access error and
PD message loop.
Change-Id: I73a9987ff49a4dfb346ebe7c994418278817c03b
Signed-off-by: Wai-Hong Tam <04b587fdf5845741a0c5a8b9cd59ca72d73ef8fc@google.com>
Reviewed-on: https://chromium-review.googlesource.com/c/chromiumos/platform/ec/+/3422771
Reviewed-by: Keith Short <f6d578f97033c25f1206cb57d52f358974c82954@chromium.org><commit_after> | # Copyright 2021 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
def register_variant(project_name, extra_dts_overlays=(), extra_kconfig_files=()):
register_npcx_project(
project_name=project_name,
zephyr_board="npcx9",
dts_overlays=[
# Common to all projects.
here / "adc.dts",
here / "battery.dts",
here / "gpio.dts",
here / "common.dts",
here / "i2c.dts",
here / "interrupts.dts",
here / "motionsense.dts",
here / "pwm.dts",
here / "switchcap.dts",
here / "usbc.dts",
# Project-specific DTS customization.
*extra_dts_overlays,
],
kconfig_files=[
# Common to all projects.
here / "prj.conf",
# Project-specific KConfig customization.
*extra_kconfig_files,
],
)
register_variant(
project_name="herobrine_npcx9",
extra_kconfig_files=[here / "prj_herobrine_npcx9.conf"],
)
register_variant(
project_name="hoglin",
extra_kconfig_files=[here / "prj_hoglin.conf"],
)
| # Copyright 2021 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
def register_variant(project_name, extra_dts_overlays=(), extra_kconfig_files=()):
register_npcx_project(
project_name=project_name,
zephyr_board="npcx9",
dts_overlays=[
# Common to all projects.
here / "adc.dts",
here / "battery.dts",
here / "gpio.dts",
here / "common.dts",
here / "i2c.dts",
here / "interrupts.dts",
here / "motionsense.dts",
here / "pwm.dts",
here / "switchcap.dts",
here / "usbc.dts",
# Project-specific DTS customization.
*extra_dts_overlays,
],
kconfig_files=[
# Common to all projects.
here / "prj.conf",
# Project-specific KConfig customization.
*extra_kconfig_files,
],
)
register_variant(
project_name="herobrine_npcx9",
)
register_variant(
project_name="hoglin",
extra_kconfig_files=[here / "prj_hoglin.conf"],
)
herobrine: Include the missing project config
Need to include the project config prj_herobrine_npcx9.conf to
the build. It defines CONFIG_BOARD_HEROBRINE_NPCX9=y. The board
specific alternative component code (alt_dev_replacement.c)
requires this Kconfig option.
BRANCH=None
BUG=b:216836197
TEST=Booted the herobrine_npcx9 image. No PPC access error and
PD message loop.
Change-Id: I73a9987ff49a4dfb346ebe7c994418278817c03b
Signed-off-by: Wai-Hong Tam <04b587fdf5845741a0c5a8b9cd59ca72d73ef8fc@google.com>
Reviewed-on: https://chromium-review.googlesource.com/c/chromiumos/platform/ec/+/3422771
Reviewed-by: Keith Short <f6d578f97033c25f1206cb57d52f358974c82954@chromium.org># Copyright 2021 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
def register_variant(project_name, extra_dts_overlays=(), extra_kconfig_files=()):
register_npcx_project(
project_name=project_name,
zephyr_board="npcx9",
dts_overlays=[
# Common to all projects.
here / "adc.dts",
here / "battery.dts",
here / "gpio.dts",
here / "common.dts",
here / "i2c.dts",
here / "interrupts.dts",
here / "motionsense.dts",
here / "pwm.dts",
here / "switchcap.dts",
here / "usbc.dts",
# Project-specific DTS customization.
*extra_dts_overlays,
],
kconfig_files=[
# Common to all projects.
here / "prj.conf",
# Project-specific KConfig customization.
*extra_kconfig_files,
],
)
register_variant(
project_name="herobrine_npcx9",
extra_kconfig_files=[here / "prj_herobrine_npcx9.conf"],
)
register_variant(
project_name="hoglin",
extra_kconfig_files=[here / "prj_hoglin.conf"],
)
| <commit_before># Copyright 2021 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
def register_variant(project_name, extra_dts_overlays=(), extra_kconfig_files=()):
register_npcx_project(
project_name=project_name,
zephyr_board="npcx9",
dts_overlays=[
# Common to all projects.
here / "adc.dts",
here / "battery.dts",
here / "gpio.dts",
here / "common.dts",
here / "i2c.dts",
here / "interrupts.dts",
here / "motionsense.dts",
here / "pwm.dts",
here / "switchcap.dts",
here / "usbc.dts",
# Project-specific DTS customization.
*extra_dts_overlays,
],
kconfig_files=[
# Common to all projects.
here / "prj.conf",
# Project-specific KConfig customization.
*extra_kconfig_files,
],
)
register_variant(
project_name="herobrine_npcx9",
)
register_variant(
project_name="hoglin",
extra_kconfig_files=[here / "prj_hoglin.conf"],
)
<commit_msg>herobrine: Include the missing project config
Need to include the project config prj_herobrine_npcx9.conf to
the build. It defines CONFIG_BOARD_HEROBRINE_NPCX9=y. The board
specific alternative component code (alt_dev_replacement.c)
requires this Kconfig option.
BRANCH=None
BUG=b:216836197
TEST=Booted the herobrine_npcx9 image. No PPC access error and
PD message loop.
Change-Id: I73a9987ff49a4dfb346ebe7c994418278817c03b
Signed-off-by: Wai-Hong Tam <04b587fdf5845741a0c5a8b9cd59ca72d73ef8fc@google.com>
Reviewed-on: https://chromium-review.googlesource.com/c/chromiumos/platform/ec/+/3422771
Reviewed-by: Keith Short <f6d578f97033c25f1206cb57d52f358974c82954@chromium.org><commit_after># Copyright 2021 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
def register_variant(project_name, extra_dts_overlays=(), extra_kconfig_files=()):
register_npcx_project(
project_name=project_name,
zephyr_board="npcx9",
dts_overlays=[
# Common to all projects.
here / "adc.dts",
here / "battery.dts",
here / "gpio.dts",
here / "common.dts",
here / "i2c.dts",
here / "interrupts.dts",
here / "motionsense.dts",
here / "pwm.dts",
here / "switchcap.dts",
here / "usbc.dts",
# Project-specific DTS customization.
*extra_dts_overlays,
],
kconfig_files=[
# Common to all projects.
here / "prj.conf",
# Project-specific KConfig customization.
*extra_kconfig_files,
],
)
register_variant(
project_name="herobrine_npcx9",
extra_kconfig_files=[here / "prj_herobrine_npcx9.conf"],
)
register_variant(
project_name="hoglin",
extra_kconfig_files=[here / "prj_hoglin.conf"],
)
|
f2ca059d6e3e9e593b053e6483ad071d58cc99d2 | tests/core/admin.py | tests/core/admin.py | from django.contrib import admin
from import_export.admin import ImportExportMixin
from .models import Book, Category, Author
class BookAdmin(ImportExportMixin, admin.ModelAdmin):
pass
admin.site.register(Book, BookAdmin)
admin.site.register(Category)
admin.site.register(Author)
| from django.contrib import admin
from import_export.admin import ImportExportMixin
from .models import Book, Category, Author
class BookAdmin(ImportExportMixin, admin.ModelAdmin):
list_filter = ['categories', 'author']
admin.site.register(Book, BookAdmin)
admin.site.register(Category)
admin.site.register(Author)
| Add BookAdmin options in test app | Add BookAdmin options in test app
| Python | bsd-2-clause | PetrDlouhy/django-import-export,copperleaftech/django-import-export,Akoten/django-import-export,daniell/django-import-export,pajod/django-import-export,piran/django-import-export,bmihelac/django-import-export,sergei-maertens/django-import-export,bmihelac/django-import-export,PetrDlouhy/django-import-export,copperleaftech/django-import-export,SalahAdDin/django-import-export,piran/django-import-export,ericdwang/django-import-export,ericdwang/django-import-export,bmihelac/django-import-export,daniell/django-import-export,bmihelac/django-import-export,django-import-export/django-import-export,rhunwicks/django-import-export,manelclos/django-import-export,SalahAdDin/django-import-export,copperleaftech/django-import-export,luto/django-import-export,Akoten/django-import-export,PetrDlouhy/django-import-export,daniell/django-import-export,brillgen/django-import-export,PetrDlouhy/django-import-export,Apkawa/django-import-export,luto/django-import-export,ylteq/dj-import-export,pajod/django-import-export,brillgen/django-import-export,jnns/django-import-export,sergei-maertens/django-import-export,brillgen/django-import-export,copperleaftech/django-import-export,Apkawa/django-import-export,brillgen/django-import-export,piran/django-import-export,ylteq/dj-import-export,daniell/django-import-export,jnns/django-import-export,jnns/django-import-export,django-import-export/django-import-export,rhunwicks/django-import-export,SalahAdDin/django-import-export,pajod/django-import-export,ericdwang/django-import-export,Akoten/django-import-export,jnns/django-import-export,luto/django-import-export,manelclos/django-import-export,pajod/django-import-export,django-import-export/django-import-export,manelclos/django-import-export,rhunwicks/django-import-export,Apkawa/django-import-export,django-import-export/django-import-export,sergei-maertens/django-import-export,ylteq/dj-import-export | from django.contrib import admin
from import_export.admin import ImportExportMixin
from .models import Book, Category, Author
class BookAdmin(ImportExportMixin, admin.ModelAdmin):
pass
admin.site.register(Book, BookAdmin)
admin.site.register(Category)
admin.site.register(Author)
Add BookAdmin options in test app | from django.contrib import admin
from import_export.admin import ImportExportMixin
from .models import Book, Category, Author
class BookAdmin(ImportExportMixin, admin.ModelAdmin):
list_filter = ['categories', 'author']
admin.site.register(Book, BookAdmin)
admin.site.register(Category)
admin.site.register(Author)
| <commit_before>from django.contrib import admin
from import_export.admin import ImportExportMixin
from .models import Book, Category, Author
class BookAdmin(ImportExportMixin, admin.ModelAdmin):
pass
admin.site.register(Book, BookAdmin)
admin.site.register(Category)
admin.site.register(Author)
<commit_msg>Add BookAdmin options in test app<commit_after> | from django.contrib import admin
from import_export.admin import ImportExportMixin
from .models import Book, Category, Author
class BookAdmin(ImportExportMixin, admin.ModelAdmin):
list_filter = ['categories', 'author']
admin.site.register(Book, BookAdmin)
admin.site.register(Category)
admin.site.register(Author)
| from django.contrib import admin
from import_export.admin import ImportExportMixin
from .models import Book, Category, Author
class BookAdmin(ImportExportMixin, admin.ModelAdmin):
pass
admin.site.register(Book, BookAdmin)
admin.site.register(Category)
admin.site.register(Author)
Add BookAdmin options in test appfrom django.contrib import admin
from import_export.admin import ImportExportMixin
from .models import Book, Category, Author
class BookAdmin(ImportExportMixin, admin.ModelAdmin):
list_filter = ['categories', 'author']
admin.site.register(Book, BookAdmin)
admin.site.register(Category)
admin.site.register(Author)
| <commit_before>from django.contrib import admin
from import_export.admin import ImportExportMixin
from .models import Book, Category, Author
class BookAdmin(ImportExportMixin, admin.ModelAdmin):
pass
admin.site.register(Book, BookAdmin)
admin.site.register(Category)
admin.site.register(Author)
<commit_msg>Add BookAdmin options in test app<commit_after>from django.contrib import admin
from import_export.admin import ImportExportMixin
from .models import Book, Category, Author
class BookAdmin(ImportExportMixin, admin.ModelAdmin):
list_filter = ['categories', 'author']
admin.site.register(Book, BookAdmin)
admin.site.register(Category)
admin.site.register(Author)
|
25c56d4c68ec484b47ef320cfb46601c4435470b | tests/test_crc32.py | tests/test_crc32.py | import hmac
import unittest
from twoping import crc32
class TestCRC32(unittest.TestCase):
def test_crc32(self):
c = crc32.new()
c.update(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hmac(self):
h = hmac.new(b"Secret key", b"Data to hash", crc32)
self.assertEqual(h.digest(), b"\x3c\xe1\xb6\xb9")
if __name__ == "__main__":
unittest.main()
| import hmac
import unittest
from twoping import crc32
class TestCRC32(unittest.TestCase):
def test_crc32(self):
c = crc32.new(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hmac(self):
h = hmac.new(b"Secret key", b"Data to hash", crc32)
self.assertEqual(h.digest(), b"\x3c\xe1\xb6\xb9")
def test_update(self):
c = crc32.new()
c.update(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hexdigest(self):
c = crc32.new(b"Data to hash")
self.assertEqual(c.hexdigest(), "449e0a5c")
def test_clear(self):
c = crc32.new(b"Data to hash")
c.clear()
self.assertEqual(c.digest(), b"\x00\x00\x00\x00")
def test_zero_padding(self):
c = crc32.new(b"jade")
self.assertEqual(c.digest(), b"\x00\x83\x52\x18")
if __name__ == "__main__":
unittest.main()
| Increase test coverage on crc32 | Increase test coverage on crc32
| Python | mpl-2.0 | rfinnie/2ping,rfinnie/2ping | import hmac
import unittest
from twoping import crc32
class TestCRC32(unittest.TestCase):
def test_crc32(self):
c = crc32.new()
c.update(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hmac(self):
h = hmac.new(b"Secret key", b"Data to hash", crc32)
self.assertEqual(h.digest(), b"\x3c\xe1\xb6\xb9")
if __name__ == "__main__":
unittest.main()
Increase test coverage on crc32 | import hmac
import unittest
from twoping import crc32
class TestCRC32(unittest.TestCase):
def test_crc32(self):
c = crc32.new(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hmac(self):
h = hmac.new(b"Secret key", b"Data to hash", crc32)
self.assertEqual(h.digest(), b"\x3c\xe1\xb6\xb9")
def test_update(self):
c = crc32.new()
c.update(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hexdigest(self):
c = crc32.new(b"Data to hash")
self.assertEqual(c.hexdigest(), "449e0a5c")
def test_clear(self):
c = crc32.new(b"Data to hash")
c.clear()
self.assertEqual(c.digest(), b"\x00\x00\x00\x00")
def test_zero_padding(self):
c = crc32.new(b"jade")
self.assertEqual(c.digest(), b"\x00\x83\x52\x18")
if __name__ == "__main__":
unittest.main()
| <commit_before>import hmac
import unittest
from twoping import crc32
class TestCRC32(unittest.TestCase):
def test_crc32(self):
c = crc32.new()
c.update(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hmac(self):
h = hmac.new(b"Secret key", b"Data to hash", crc32)
self.assertEqual(h.digest(), b"\x3c\xe1\xb6\xb9")
if __name__ == "__main__":
unittest.main()
<commit_msg>Increase test coverage on crc32<commit_after> | import hmac
import unittest
from twoping import crc32
class TestCRC32(unittest.TestCase):
def test_crc32(self):
c = crc32.new(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hmac(self):
h = hmac.new(b"Secret key", b"Data to hash", crc32)
self.assertEqual(h.digest(), b"\x3c\xe1\xb6\xb9")
def test_update(self):
c = crc32.new()
c.update(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hexdigest(self):
c = crc32.new(b"Data to hash")
self.assertEqual(c.hexdigest(), "449e0a5c")
def test_clear(self):
c = crc32.new(b"Data to hash")
c.clear()
self.assertEqual(c.digest(), b"\x00\x00\x00\x00")
def test_zero_padding(self):
c = crc32.new(b"jade")
self.assertEqual(c.digest(), b"\x00\x83\x52\x18")
if __name__ == "__main__":
unittest.main()
| import hmac
import unittest
from twoping import crc32
class TestCRC32(unittest.TestCase):
def test_crc32(self):
c = crc32.new()
c.update(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hmac(self):
h = hmac.new(b"Secret key", b"Data to hash", crc32)
self.assertEqual(h.digest(), b"\x3c\xe1\xb6\xb9")
if __name__ == "__main__":
unittest.main()
Increase test coverage on crc32import hmac
import unittest
from twoping import crc32
class TestCRC32(unittest.TestCase):
def test_crc32(self):
c = crc32.new(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hmac(self):
h = hmac.new(b"Secret key", b"Data to hash", crc32)
self.assertEqual(h.digest(), b"\x3c\xe1\xb6\xb9")
def test_update(self):
c = crc32.new()
c.update(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hexdigest(self):
c = crc32.new(b"Data to hash")
self.assertEqual(c.hexdigest(), "449e0a5c")
def test_clear(self):
c = crc32.new(b"Data to hash")
c.clear()
self.assertEqual(c.digest(), b"\x00\x00\x00\x00")
def test_zero_padding(self):
c = crc32.new(b"jade")
self.assertEqual(c.digest(), b"\x00\x83\x52\x18")
if __name__ == "__main__":
unittest.main()
| <commit_before>import hmac
import unittest
from twoping import crc32
class TestCRC32(unittest.TestCase):
def test_crc32(self):
c = crc32.new()
c.update(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hmac(self):
h = hmac.new(b"Secret key", b"Data to hash", crc32)
self.assertEqual(h.digest(), b"\x3c\xe1\xb6\xb9")
if __name__ == "__main__":
unittest.main()
<commit_msg>Increase test coverage on crc32<commit_after>import hmac
import unittest
from twoping import crc32
class TestCRC32(unittest.TestCase):
def test_crc32(self):
c = crc32.new(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hmac(self):
h = hmac.new(b"Secret key", b"Data to hash", crc32)
self.assertEqual(h.digest(), b"\x3c\xe1\xb6\xb9")
def test_update(self):
c = crc32.new()
c.update(b"Data to hash")
self.assertEqual(c.digest(), b"\x44\x9e\x0a\x5c")
def test_hexdigest(self):
c = crc32.new(b"Data to hash")
self.assertEqual(c.hexdigest(), "449e0a5c")
def test_clear(self):
c = crc32.new(b"Data to hash")
c.clear()
self.assertEqual(c.digest(), b"\x00\x00\x00\x00")
def test_zero_padding(self):
c = crc32.new(b"jade")
self.assertEqual(c.digest(), b"\x00\x83\x52\x18")
if __name__ == "__main__":
unittest.main()
|
096e41266ac3686c1757fc4b5087e3b786287f91 | webapp/byceps/database.py | webapp/byceps/database.py | # -*- coding: utf-8 -*-
"""
byceps.database
~~~~~~~~~~~~~~~
Database utilities.
:Copyright: 2006-2014 Jochen Kupperschmidt
"""
import uuid
from flask.ext.sqlalchemy import BaseQuery, SQLAlchemy
from sqlalchemy.dialects.postgresql import UUID
db = SQLAlchemy()
db.Uuid = UUID
def generate_uuid():
"""Generate a random UUID (Universally Unique IDentifier)."""
return uuid.uuid4()
| # -*- coding: utf-8 -*-
"""
byceps.database
~~~~~~~~~~~~~~~
Database utilities.
:Copyright: 2006-2014 Jochen Kupperschmidt
"""
import uuid
from flask.ext.sqlalchemy import BaseQuery, SQLAlchemy
from sqlalchemy.dialects.postgresql import UUID
db = SQLAlchemy(session_options={'autoflush': False})
db.Uuid = UUID
def generate_uuid():
"""Generate a random UUID (Universally Unique IDentifier)."""
return uuid.uuid4()
| Disable autoflushing as introduced with Flask-SQLAlchemy 2.0. | Disable autoflushing as introduced with Flask-SQLAlchemy 2.0.
| Python | bsd-3-clause | homeworkprod/byceps,m-ober/byceps,m-ober/byceps,m-ober/byceps,homeworkprod/byceps,homeworkprod/byceps | # -*- coding: utf-8 -*-
"""
byceps.database
~~~~~~~~~~~~~~~
Database utilities.
:Copyright: 2006-2014 Jochen Kupperschmidt
"""
import uuid
from flask.ext.sqlalchemy import BaseQuery, SQLAlchemy
from sqlalchemy.dialects.postgresql import UUID
db = SQLAlchemy()
db.Uuid = UUID
def generate_uuid():
"""Generate a random UUID (Universally Unique IDentifier)."""
return uuid.uuid4()
Disable autoflushing as introduced with Flask-SQLAlchemy 2.0. | # -*- coding: utf-8 -*-
"""
byceps.database
~~~~~~~~~~~~~~~
Database utilities.
:Copyright: 2006-2014 Jochen Kupperschmidt
"""
import uuid
from flask.ext.sqlalchemy import BaseQuery, SQLAlchemy
from sqlalchemy.dialects.postgresql import UUID
db = SQLAlchemy(session_options={'autoflush': False})
db.Uuid = UUID
def generate_uuid():
"""Generate a random UUID (Universally Unique IDentifier)."""
return uuid.uuid4()
| <commit_before># -*- coding: utf-8 -*-
"""
byceps.database
~~~~~~~~~~~~~~~
Database utilities.
:Copyright: 2006-2014 Jochen Kupperschmidt
"""
import uuid
from flask.ext.sqlalchemy import BaseQuery, SQLAlchemy
from sqlalchemy.dialects.postgresql import UUID
db = SQLAlchemy()
db.Uuid = UUID
def generate_uuid():
"""Generate a random UUID (Universally Unique IDentifier)."""
return uuid.uuid4()
<commit_msg>Disable autoflushing as introduced with Flask-SQLAlchemy 2.0.<commit_after> | # -*- coding: utf-8 -*-
"""
byceps.database
~~~~~~~~~~~~~~~
Database utilities.
:Copyright: 2006-2014 Jochen Kupperschmidt
"""
import uuid
from flask.ext.sqlalchemy import BaseQuery, SQLAlchemy
from sqlalchemy.dialects.postgresql import UUID
db = SQLAlchemy(session_options={'autoflush': False})
db.Uuid = UUID
def generate_uuid():
"""Generate a random UUID (Universally Unique IDentifier)."""
return uuid.uuid4()
| # -*- coding: utf-8 -*-
"""
byceps.database
~~~~~~~~~~~~~~~
Database utilities.
:Copyright: 2006-2014 Jochen Kupperschmidt
"""
import uuid
from flask.ext.sqlalchemy import BaseQuery, SQLAlchemy
from sqlalchemy.dialects.postgresql import UUID
db = SQLAlchemy()
db.Uuid = UUID
def generate_uuid():
"""Generate a random UUID (Universally Unique IDentifier)."""
return uuid.uuid4()
Disable autoflushing as introduced with Flask-SQLAlchemy 2.0.# -*- coding: utf-8 -*-
"""
byceps.database
~~~~~~~~~~~~~~~
Database utilities.
:Copyright: 2006-2014 Jochen Kupperschmidt
"""
import uuid
from flask.ext.sqlalchemy import BaseQuery, SQLAlchemy
from sqlalchemy.dialects.postgresql import UUID
db = SQLAlchemy(session_options={'autoflush': False})
db.Uuid = UUID
def generate_uuid():
"""Generate a random UUID (Universally Unique IDentifier)."""
return uuid.uuid4()
| <commit_before># -*- coding: utf-8 -*-
"""
byceps.database
~~~~~~~~~~~~~~~
Database utilities.
:Copyright: 2006-2014 Jochen Kupperschmidt
"""
import uuid
from flask.ext.sqlalchemy import BaseQuery, SQLAlchemy
from sqlalchemy.dialects.postgresql import UUID
db = SQLAlchemy()
db.Uuid = UUID
def generate_uuid():
"""Generate a random UUID (Universally Unique IDentifier)."""
return uuid.uuid4()
<commit_msg>Disable autoflushing as introduced with Flask-SQLAlchemy 2.0.<commit_after># -*- coding: utf-8 -*-
"""
byceps.database
~~~~~~~~~~~~~~~
Database utilities.
:Copyright: 2006-2014 Jochen Kupperschmidt
"""
import uuid
from flask.ext.sqlalchemy import BaseQuery, SQLAlchemy
from sqlalchemy.dialects.postgresql import UUID
db = SQLAlchemy(session_options={'autoflush': False})
db.Uuid = UUID
def generate_uuid():
"""Generate a random UUID (Universally Unique IDentifier)."""
return uuid.uuid4()
|
a393881b4cf79a34101c7d4821ed0ccd78f117cb | zsh/zsh_concat.py | zsh/zsh_concat.py | #!/usr/bin/env python3
from os import scandir
from sys import argv
from platform import uname
from pathlib import Path
filename_template = """
# -------------------------------------------------------------------------------
# filename: {filename}
# -------------------------------------------------------------------------------
{data}
# -------------------------------------------------------------------------------
# END
# -------------------------------------------------------------------------------
"""
def read_and_format_data(filename, outbuf):
"""
Read file and format
Args:
filename
Returns:
str
:param outbuf:
"""
with open(filename, 'r') as inbuf:
data = inbuf.read()
data = filename_template.format(filename=filename,
data=data)
outbuf.write(data)
def main(args):
parent_dir = Path(args[0]).parent
lib_dir = parent_dir.joinpath('lib')
hostname = uname()[1]
local_dir = parent_dir.joinpath('local')
with open('zsh_plugins.zsh', 'w') as outbuf:
for filename in scandir(str(lib_dir)):
read_and_format_data(filename.path, outbuf)
for filename in scandir(str(local_dir)):
filename = Path(filename.path)
if filename.stem == hostname:
read_and_format_data(str(filename), outbuf)
if __name__ == "__main__":
main(argv)
| #!/usr/bin/env python3
from os import scandir
from sys import argv
from platform import uname
from pathlib import Path
filename_template = """
# -------------------------------------------------------------------------------
# filename: {filename}
# -------------------------------------------------------------------------------
{data}
# -------------------------------------------------------------------------------
# END
# -------------------------------------------------------------------------------
"""
def read_and_format_data(filename, outbuf):
"""
Read file and format
Args:
filename:
Returns:
str
"""
with open(filename, 'r') as inbuf:
data = inbuf.read()
data = filename_template.format(filename=filename,
data=data)
outbuf.write(data)
def main(args):
parent_dir = Path(args[0]).parent
lib_dir = parent_dir.joinpath('lib')
hostname = uname()[1]
local_dir = parent_dir.joinpath('local')
outfilename = parent_dir.joinpath("zsh_plugins.zsh")
with open(str(outfilename), 'w') as outbuf:
for filename in scandir(str(lib_dir)):
read_and_format_data(filename.path, outbuf)
for filename in scandir(str(local_dir)):
filename = Path(filename.path)
if filename.stem == hostname:
read_and_format_data(str(filename), outbuf)
if __name__ == "__main__":
main(argv)
| Fix script to save output to script’s directory. | Fix script to save output to script’s directory.
| Python | mit | skk/dotfiles,skk/dotfiles | #!/usr/bin/env python3
from os import scandir
from sys import argv
from platform import uname
from pathlib import Path
filename_template = """
# -------------------------------------------------------------------------------
# filename: {filename}
# -------------------------------------------------------------------------------
{data}
# -------------------------------------------------------------------------------
# END
# -------------------------------------------------------------------------------
"""
def read_and_format_data(filename, outbuf):
"""
Read file and format
Args:
filename
Returns:
str
:param outbuf:
"""
with open(filename, 'r') as inbuf:
data = inbuf.read()
data = filename_template.format(filename=filename,
data=data)
outbuf.write(data)
def main(args):
parent_dir = Path(args[0]).parent
lib_dir = parent_dir.joinpath('lib')
hostname = uname()[1]
local_dir = parent_dir.joinpath('local')
with open('zsh_plugins.zsh', 'w') as outbuf:
for filename in scandir(str(lib_dir)):
read_and_format_data(filename.path, outbuf)
for filename in scandir(str(local_dir)):
filename = Path(filename.path)
if filename.stem == hostname:
read_and_format_data(str(filename), outbuf)
if __name__ == "__main__":
main(argv)
Fix script to save output to script’s directory. | #!/usr/bin/env python3
from os import scandir
from sys import argv
from platform import uname
from pathlib import Path
filename_template = """
# -------------------------------------------------------------------------------
# filename: {filename}
# -------------------------------------------------------------------------------
{data}
# -------------------------------------------------------------------------------
# END
# -------------------------------------------------------------------------------
"""
def read_and_format_data(filename, outbuf):
"""
Read file and format
Args:
filename:
Returns:
str
"""
with open(filename, 'r') as inbuf:
data = inbuf.read()
data = filename_template.format(filename=filename,
data=data)
outbuf.write(data)
def main(args):
parent_dir = Path(args[0]).parent
lib_dir = parent_dir.joinpath('lib')
hostname = uname()[1]
local_dir = parent_dir.joinpath('local')
outfilename = parent_dir.joinpath("zsh_plugins.zsh")
with open(str(outfilename), 'w') as outbuf:
for filename in scandir(str(lib_dir)):
read_and_format_data(filename.path, outbuf)
for filename in scandir(str(local_dir)):
filename = Path(filename.path)
if filename.stem == hostname:
read_and_format_data(str(filename), outbuf)
if __name__ == "__main__":
main(argv)
| <commit_before>#!/usr/bin/env python3
from os import scandir
from sys import argv
from platform import uname
from pathlib import Path
filename_template = """
# -------------------------------------------------------------------------------
# filename: {filename}
# -------------------------------------------------------------------------------
{data}
# -------------------------------------------------------------------------------
# END
# -------------------------------------------------------------------------------
"""
def read_and_format_data(filename, outbuf):
"""
Read file and format
Args:
filename
Returns:
str
:param outbuf:
"""
with open(filename, 'r') as inbuf:
data = inbuf.read()
data = filename_template.format(filename=filename,
data=data)
outbuf.write(data)
def main(args):
parent_dir = Path(args[0]).parent
lib_dir = parent_dir.joinpath('lib')
hostname = uname()[1]
local_dir = parent_dir.joinpath('local')
with open('zsh_plugins.zsh', 'w') as outbuf:
for filename in scandir(str(lib_dir)):
read_and_format_data(filename.path, outbuf)
for filename in scandir(str(local_dir)):
filename = Path(filename.path)
if filename.stem == hostname:
read_and_format_data(str(filename), outbuf)
if __name__ == "__main__":
main(argv)
<commit_msg>Fix script to save output to script’s directory.<commit_after> | #!/usr/bin/env python3
from os import scandir
from sys import argv
from platform import uname
from pathlib import Path
filename_template = """
# -------------------------------------------------------------------------------
# filename: {filename}
# -------------------------------------------------------------------------------
{data}
# -------------------------------------------------------------------------------
# END
# -------------------------------------------------------------------------------
"""
def read_and_format_data(filename, outbuf):
"""
Read file and format
Args:
filename:
Returns:
str
"""
with open(filename, 'r') as inbuf:
data = inbuf.read()
data = filename_template.format(filename=filename,
data=data)
outbuf.write(data)
def main(args):
parent_dir = Path(args[0]).parent
lib_dir = parent_dir.joinpath('lib')
hostname = uname()[1]
local_dir = parent_dir.joinpath('local')
outfilename = parent_dir.joinpath("zsh_plugins.zsh")
with open(str(outfilename), 'w') as outbuf:
for filename in scandir(str(lib_dir)):
read_and_format_data(filename.path, outbuf)
for filename in scandir(str(local_dir)):
filename = Path(filename.path)
if filename.stem == hostname:
read_and_format_data(str(filename), outbuf)
if __name__ == "__main__":
main(argv)
| #!/usr/bin/env python3
from os import scandir
from sys import argv
from platform import uname
from pathlib import Path
filename_template = """
# -------------------------------------------------------------------------------
# filename: {filename}
# -------------------------------------------------------------------------------
{data}
# -------------------------------------------------------------------------------
# END
# -------------------------------------------------------------------------------
"""
def read_and_format_data(filename, outbuf):
"""
Read file and format
Args:
filename
Returns:
str
:param outbuf:
"""
with open(filename, 'r') as inbuf:
data = inbuf.read()
data = filename_template.format(filename=filename,
data=data)
outbuf.write(data)
def main(args):
parent_dir = Path(args[0]).parent
lib_dir = parent_dir.joinpath('lib')
hostname = uname()[1]
local_dir = parent_dir.joinpath('local')
with open('zsh_plugins.zsh', 'w') as outbuf:
for filename in scandir(str(lib_dir)):
read_and_format_data(filename.path, outbuf)
for filename in scandir(str(local_dir)):
filename = Path(filename.path)
if filename.stem == hostname:
read_and_format_data(str(filename), outbuf)
if __name__ == "__main__":
main(argv)
Fix script to save output to script’s directory.#!/usr/bin/env python3
from os import scandir
from sys import argv
from platform import uname
from pathlib import Path
filename_template = """
# -------------------------------------------------------------------------------
# filename: {filename}
# -------------------------------------------------------------------------------
{data}
# -------------------------------------------------------------------------------
# END
# -------------------------------------------------------------------------------
"""
def read_and_format_data(filename, outbuf):
"""
Read file and format
Args:
filename:
Returns:
str
"""
with open(filename, 'r') as inbuf:
data = inbuf.read()
data = filename_template.format(filename=filename,
data=data)
outbuf.write(data)
def main(args):
parent_dir = Path(args[0]).parent
lib_dir = parent_dir.joinpath('lib')
hostname = uname()[1]
local_dir = parent_dir.joinpath('local')
outfilename = parent_dir.joinpath("zsh_plugins.zsh")
with open(str(outfilename), 'w') as outbuf:
for filename in scandir(str(lib_dir)):
read_and_format_data(filename.path, outbuf)
for filename in scandir(str(local_dir)):
filename = Path(filename.path)
if filename.stem == hostname:
read_and_format_data(str(filename), outbuf)
if __name__ == "__main__":
main(argv)
| <commit_before>#!/usr/bin/env python3
from os import scandir
from sys import argv
from platform import uname
from pathlib import Path
filename_template = """
# -------------------------------------------------------------------------------
# filename: {filename}
# -------------------------------------------------------------------------------
{data}
# -------------------------------------------------------------------------------
# END
# -------------------------------------------------------------------------------
"""
def read_and_format_data(filename, outbuf):
"""
Read file and format
Args:
filename
Returns:
str
:param outbuf:
"""
with open(filename, 'r') as inbuf:
data = inbuf.read()
data = filename_template.format(filename=filename,
data=data)
outbuf.write(data)
def main(args):
parent_dir = Path(args[0]).parent
lib_dir = parent_dir.joinpath('lib')
hostname = uname()[1]
local_dir = parent_dir.joinpath('local')
with open('zsh_plugins.zsh', 'w') as outbuf:
for filename in scandir(str(lib_dir)):
read_and_format_data(filename.path, outbuf)
for filename in scandir(str(local_dir)):
filename = Path(filename.path)
if filename.stem == hostname:
read_and_format_data(str(filename), outbuf)
if __name__ == "__main__":
main(argv)
<commit_msg>Fix script to save output to script’s directory.<commit_after>#!/usr/bin/env python3
from os import scandir
from sys import argv
from platform import uname
from pathlib import Path
filename_template = """
# -------------------------------------------------------------------------------
# filename: {filename}
# -------------------------------------------------------------------------------
{data}
# -------------------------------------------------------------------------------
# END
# -------------------------------------------------------------------------------
"""
def read_and_format_data(filename, outbuf):
"""
Read file and format
Args:
filename:
Returns:
str
"""
with open(filename, 'r') as inbuf:
data = inbuf.read()
data = filename_template.format(filename=filename,
data=data)
outbuf.write(data)
def main(args):
parent_dir = Path(args[0]).parent
lib_dir = parent_dir.joinpath('lib')
hostname = uname()[1]
local_dir = parent_dir.joinpath('local')
outfilename = parent_dir.joinpath("zsh_plugins.zsh")
with open(str(outfilename), 'w') as outbuf:
for filename in scandir(str(lib_dir)):
read_and_format_data(filename.path, outbuf)
for filename in scandir(str(local_dir)):
filename = Path(filename.path)
if filename.stem == hostname:
read_and_format_data(str(filename), outbuf)
if __name__ == "__main__":
main(argv)
|
ee32d3746a9fa788a06931063a8242f936b6ed18 | src/data/meta.py | src/data/meta.py | import collections
class Meta(collections.OrderedDict):
def __init__(self, *args, **kwargs):
self._smallest = float('inf')
self._largest = 0
self._ordered = True
super(Meta, self).__init__(*args, **kwargs)
def __setitem__(self, key, value, *args, **kwargs):
if key in self and self[key] == value:
raise AssertionError('Redundant assignment')
if value > self._smallest:
self._ordered = False
else:
self._smallest = value
if value > self._largest:
self._largest = value
super(Meta, self).__setitem__(key, value, *args, **kwargs)
self._changed()
def items(self):
self._reorder()
return super(Meta, self).items()
def first(self):
self._reorder()
for k, v in self.items():
return k, v
def peek(self):
self._reorder()
for first in self:
return first
def magnitude(self):
return self._largest
def _reorder(self):
if self._ordered:
return
order = sorted(super(Meta, self).items(), key=lambda x: x[1], reverse=True)
for k, v in order:
self.move_to_end(k)
self._ordered = True
def _changed(self):
pass
| import collections
import typing
class Meta(collections.OrderedDict, typing.MutableMapping[str, float]):
def __init__(self, *args, **kwargs) -> None:
self._smallest = float('inf')
self._largest = 0
self._ordered = True
super(Meta, self).__init__(*args, **kwargs)
def __setitem__(self, key: str, value: float) -> None:
if key in self and self[key] == value:
raise AssertionError('Redundant assignment')
if value > self._smallest:
self._ordered = False
else:
self._smallest = value
if value > self._largest:
self._largest = value
super(Meta, self).__setitem__(key, value)
self._changed()
def items(self) -> typing.ItemsView[str, float]:
self._reorder()
return super(Meta, self).items()
def first(self) -> typing.Tuple[str, float]:
self._reorder()
for k, v in self.items():
return k, v
def peek(self) -> str:
self._reorder()
for first in self:
return first
def magnitude(self) -> float:
return self._largest
def _reorder(self) -> None:
if self._ordered:
return
order = sorted(super(Meta, self).items(), key=lambda x: x[1], reverse=True)
for k, v in order:
self.move_to_end(k)
self._ordered = True
def _changed(self):
pass
| Add typing information to Meta. | Add typing information to Meta.
| Python | mit | PhilHarnish/forge,PhilHarnish/forge,PhilHarnish/forge,PhilHarnish/forge,PhilHarnish/forge,PhilHarnish/forge | import collections
class Meta(collections.OrderedDict):
def __init__(self, *args, **kwargs):
self._smallest = float('inf')
self._largest = 0
self._ordered = True
super(Meta, self).__init__(*args, **kwargs)
def __setitem__(self, key, value, *args, **kwargs):
if key in self and self[key] == value:
raise AssertionError('Redundant assignment')
if value > self._smallest:
self._ordered = False
else:
self._smallest = value
if value > self._largest:
self._largest = value
super(Meta, self).__setitem__(key, value, *args, **kwargs)
self._changed()
def items(self):
self._reorder()
return super(Meta, self).items()
def first(self):
self._reorder()
for k, v in self.items():
return k, v
def peek(self):
self._reorder()
for first in self:
return first
def magnitude(self):
return self._largest
def _reorder(self):
if self._ordered:
return
order = sorted(super(Meta, self).items(), key=lambda x: x[1], reverse=True)
for k, v in order:
self.move_to_end(k)
self._ordered = True
def _changed(self):
pass
Add typing information to Meta. | import collections
import typing
class Meta(collections.OrderedDict, typing.MutableMapping[str, float]):
def __init__(self, *args, **kwargs) -> None:
self._smallest = float('inf')
self._largest = 0
self._ordered = True
super(Meta, self).__init__(*args, **kwargs)
def __setitem__(self, key: str, value: float) -> None:
if key in self and self[key] == value:
raise AssertionError('Redundant assignment')
if value > self._smallest:
self._ordered = False
else:
self._smallest = value
if value > self._largest:
self._largest = value
super(Meta, self).__setitem__(key, value)
self._changed()
def items(self) -> typing.ItemsView[str, float]:
self._reorder()
return super(Meta, self).items()
def first(self) -> typing.Tuple[str, float]:
self._reorder()
for k, v in self.items():
return k, v
def peek(self) -> str:
self._reorder()
for first in self:
return first
def magnitude(self) -> float:
return self._largest
def _reorder(self) -> None:
if self._ordered:
return
order = sorted(super(Meta, self).items(), key=lambda x: x[1], reverse=True)
for k, v in order:
self.move_to_end(k)
self._ordered = True
def _changed(self):
pass
| <commit_before>import collections
class Meta(collections.OrderedDict):
def __init__(self, *args, **kwargs):
self._smallest = float('inf')
self._largest = 0
self._ordered = True
super(Meta, self).__init__(*args, **kwargs)
def __setitem__(self, key, value, *args, **kwargs):
if key in self and self[key] == value:
raise AssertionError('Redundant assignment')
if value > self._smallest:
self._ordered = False
else:
self._smallest = value
if value > self._largest:
self._largest = value
super(Meta, self).__setitem__(key, value, *args, **kwargs)
self._changed()
def items(self):
self._reorder()
return super(Meta, self).items()
def first(self):
self._reorder()
for k, v in self.items():
return k, v
def peek(self):
self._reorder()
for first in self:
return first
def magnitude(self):
return self._largest
def _reorder(self):
if self._ordered:
return
order = sorted(super(Meta, self).items(), key=lambda x: x[1], reverse=True)
for k, v in order:
self.move_to_end(k)
self._ordered = True
def _changed(self):
pass
<commit_msg>Add typing information to Meta.<commit_after> | import collections
import typing
class Meta(collections.OrderedDict, typing.MutableMapping[str, float]):
def __init__(self, *args, **kwargs) -> None:
self._smallest = float('inf')
self._largest = 0
self._ordered = True
super(Meta, self).__init__(*args, **kwargs)
def __setitem__(self, key: str, value: float) -> None:
if key in self and self[key] == value:
raise AssertionError('Redundant assignment')
if value > self._smallest:
self._ordered = False
else:
self._smallest = value
if value > self._largest:
self._largest = value
super(Meta, self).__setitem__(key, value)
self._changed()
def items(self) -> typing.ItemsView[str, float]:
self._reorder()
return super(Meta, self).items()
def first(self) -> typing.Tuple[str, float]:
self._reorder()
for k, v in self.items():
return k, v
def peek(self) -> str:
self._reorder()
for first in self:
return first
def magnitude(self) -> float:
return self._largest
def _reorder(self) -> None:
if self._ordered:
return
order = sorted(super(Meta, self).items(), key=lambda x: x[1], reverse=True)
for k, v in order:
self.move_to_end(k)
self._ordered = True
def _changed(self):
pass
| import collections
class Meta(collections.OrderedDict):
def __init__(self, *args, **kwargs):
self._smallest = float('inf')
self._largest = 0
self._ordered = True
super(Meta, self).__init__(*args, **kwargs)
def __setitem__(self, key, value, *args, **kwargs):
if key in self and self[key] == value:
raise AssertionError('Redundant assignment')
if value > self._smallest:
self._ordered = False
else:
self._smallest = value
if value > self._largest:
self._largest = value
super(Meta, self).__setitem__(key, value, *args, **kwargs)
self._changed()
def items(self):
self._reorder()
return super(Meta, self).items()
def first(self):
self._reorder()
for k, v in self.items():
return k, v
def peek(self):
self._reorder()
for first in self:
return first
def magnitude(self):
return self._largest
def _reorder(self):
if self._ordered:
return
order = sorted(super(Meta, self).items(), key=lambda x: x[1], reverse=True)
for k, v in order:
self.move_to_end(k)
self._ordered = True
def _changed(self):
pass
Add typing information to Meta.import collections
import typing
class Meta(collections.OrderedDict, typing.MutableMapping[str, float]):
def __init__(self, *args, **kwargs) -> None:
self._smallest = float('inf')
self._largest = 0
self._ordered = True
super(Meta, self).__init__(*args, **kwargs)
def __setitem__(self, key: str, value: float) -> None:
if key in self and self[key] == value:
raise AssertionError('Redundant assignment')
if value > self._smallest:
self._ordered = False
else:
self._smallest = value
if value > self._largest:
self._largest = value
super(Meta, self).__setitem__(key, value)
self._changed()
def items(self) -> typing.ItemsView[str, float]:
self._reorder()
return super(Meta, self).items()
def first(self) -> typing.Tuple[str, float]:
self._reorder()
for k, v in self.items():
return k, v
def peek(self) -> str:
self._reorder()
for first in self:
return first
def magnitude(self) -> float:
return self._largest
def _reorder(self) -> None:
if self._ordered:
return
order = sorted(super(Meta, self).items(), key=lambda x: x[1], reverse=True)
for k, v in order:
self.move_to_end(k)
self._ordered = True
def _changed(self):
pass
| <commit_before>import collections
class Meta(collections.OrderedDict):
def __init__(self, *args, **kwargs):
self._smallest = float('inf')
self._largest = 0
self._ordered = True
super(Meta, self).__init__(*args, **kwargs)
def __setitem__(self, key, value, *args, **kwargs):
if key in self and self[key] == value:
raise AssertionError('Redundant assignment')
if value > self._smallest:
self._ordered = False
else:
self._smallest = value
if value > self._largest:
self._largest = value
super(Meta, self).__setitem__(key, value, *args, **kwargs)
self._changed()
def items(self):
self._reorder()
return super(Meta, self).items()
def first(self):
self._reorder()
for k, v in self.items():
return k, v
def peek(self):
self._reorder()
for first in self:
return first
def magnitude(self):
return self._largest
def _reorder(self):
if self._ordered:
return
order = sorted(super(Meta, self).items(), key=lambda x: x[1], reverse=True)
for k, v in order:
self.move_to_end(k)
self._ordered = True
def _changed(self):
pass
<commit_msg>Add typing information to Meta.<commit_after>import collections
import typing
class Meta(collections.OrderedDict, typing.MutableMapping[str, float]):
def __init__(self, *args, **kwargs) -> None:
self._smallest = float('inf')
self._largest = 0
self._ordered = True
super(Meta, self).__init__(*args, **kwargs)
def __setitem__(self, key: str, value: float) -> None:
if key in self and self[key] == value:
raise AssertionError('Redundant assignment')
if value > self._smallest:
self._ordered = False
else:
self._smallest = value
if value > self._largest:
self._largest = value
super(Meta, self).__setitem__(key, value)
self._changed()
def items(self) -> typing.ItemsView[str, float]:
self._reorder()
return super(Meta, self).items()
def first(self) -> typing.Tuple[str, float]:
self._reorder()
for k, v in self.items():
return k, v
def peek(self) -> str:
self._reorder()
for first in self:
return first
def magnitude(self) -> float:
return self._largest
def _reorder(self) -> None:
if self._ordered:
return
order = sorted(super(Meta, self).items(), key=lambda x: x[1], reverse=True)
for k, v in order:
self.move_to_end(k)
self._ordered = True
def _changed(self):
pass
|
213c25934aa15c9d607833f145f54647d17364ca | rnacentral/portal/templatetags/portal_extras.py | rnacentral/portal/templatetags/portal_extras.py | """
Copyright [2009-2014] EMBL-European Bioinformatics Institute
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import template
from portal.config.expert_databases import expert_dbs
register = template.Library()
@register.assignment_tag
def get_expert_databases_columns():
"""
Return expert databases grouped and order for the website footer.
"""
dbs = sorted(expert_dbs, key=lambda x: x['name'].lower())
return [
[dbs[x] for x in [1,0,2,3,4,5,6]],
dbs[7:13],
dbs[13:19],
dbs[19:]
]
@register.assignment_tag
def get_expert_databases_list():
"""
Get an alphabetically sorted list of imported expert databases.
"""
imported_dbs = [x for x in expert_dbs if x['imported']]
return sorted(imported_dbs, key=lambda x: x['name'].lower())
| """
Copyright [2009-2014] EMBL-European Bioinformatics Institute
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import template
from portal.config.expert_databases import expert_dbs
register = template.Library()
@register.assignment_tag
def get_expert_databases_columns():
"""
Return expert databases grouped and order for the website footer.
"""
dbs = sorted(expert_dbs, key=lambda x: x['name'].lower())
return [
[dbs[x] for x in [1,0,2,3,4,5,6,7]],
dbs[8:16],
dbs[16:23],
dbs[23:]
]
@register.assignment_tag
def get_expert_databases_list():
"""
Get an alphabetically sorted list of imported expert databases.
"""
imported_dbs = [x for x in expert_dbs if x['imported']]
return sorted(imported_dbs, key=lambda x: x['name'].lower())
| Update expert database list in the footer | Update expert database list in the footer
| Python | apache-2.0 | RNAcentral/rnacentral-webcode,RNAcentral/rnacentral-webcode,RNAcentral/rnacentral-webcode,RNAcentral/rnacentral-webcode | """
Copyright [2009-2014] EMBL-European Bioinformatics Institute
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import template
from portal.config.expert_databases import expert_dbs
register = template.Library()
@register.assignment_tag
def get_expert_databases_columns():
"""
Return expert databases grouped and order for the website footer.
"""
dbs = sorted(expert_dbs, key=lambda x: x['name'].lower())
return [
[dbs[x] for x in [1,0,2,3,4,5,6]],
dbs[7:13],
dbs[13:19],
dbs[19:]
]
@register.assignment_tag
def get_expert_databases_list():
"""
Get an alphabetically sorted list of imported expert databases.
"""
imported_dbs = [x for x in expert_dbs if x['imported']]
return sorted(imported_dbs, key=lambda x: x['name'].lower())
Update expert database list in the footer | """
Copyright [2009-2014] EMBL-European Bioinformatics Institute
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import template
from portal.config.expert_databases import expert_dbs
register = template.Library()
@register.assignment_tag
def get_expert_databases_columns():
"""
Return expert databases grouped and order for the website footer.
"""
dbs = sorted(expert_dbs, key=lambda x: x['name'].lower())
return [
[dbs[x] for x in [1,0,2,3,4,5,6,7]],
dbs[8:16],
dbs[16:23],
dbs[23:]
]
@register.assignment_tag
def get_expert_databases_list():
"""
Get an alphabetically sorted list of imported expert databases.
"""
imported_dbs = [x for x in expert_dbs if x['imported']]
return sorted(imported_dbs, key=lambda x: x['name'].lower())
| <commit_before>"""
Copyright [2009-2014] EMBL-European Bioinformatics Institute
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import template
from portal.config.expert_databases import expert_dbs
register = template.Library()
@register.assignment_tag
def get_expert_databases_columns():
"""
Return expert databases grouped and order for the website footer.
"""
dbs = sorted(expert_dbs, key=lambda x: x['name'].lower())
return [
[dbs[x] for x in [1,0,2,3,4,5,6]],
dbs[7:13],
dbs[13:19],
dbs[19:]
]
@register.assignment_tag
def get_expert_databases_list():
"""
Get an alphabetically sorted list of imported expert databases.
"""
imported_dbs = [x for x in expert_dbs if x['imported']]
return sorted(imported_dbs, key=lambda x: x['name'].lower())
<commit_msg>Update expert database list in the footer<commit_after> | """
Copyright [2009-2014] EMBL-European Bioinformatics Institute
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import template
from portal.config.expert_databases import expert_dbs
register = template.Library()
@register.assignment_tag
def get_expert_databases_columns():
"""
Return expert databases grouped and order for the website footer.
"""
dbs = sorted(expert_dbs, key=lambda x: x['name'].lower())
return [
[dbs[x] for x in [1,0,2,3,4,5,6,7]],
dbs[8:16],
dbs[16:23],
dbs[23:]
]
@register.assignment_tag
def get_expert_databases_list():
"""
Get an alphabetically sorted list of imported expert databases.
"""
imported_dbs = [x for x in expert_dbs if x['imported']]
return sorted(imported_dbs, key=lambda x: x['name'].lower())
| """
Copyright [2009-2014] EMBL-European Bioinformatics Institute
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import template
from portal.config.expert_databases import expert_dbs
register = template.Library()
@register.assignment_tag
def get_expert_databases_columns():
"""
Return expert databases grouped and order for the website footer.
"""
dbs = sorted(expert_dbs, key=lambda x: x['name'].lower())
return [
[dbs[x] for x in [1,0,2,3,4,5,6]],
dbs[7:13],
dbs[13:19],
dbs[19:]
]
@register.assignment_tag
def get_expert_databases_list():
"""
Get an alphabetically sorted list of imported expert databases.
"""
imported_dbs = [x for x in expert_dbs if x['imported']]
return sorted(imported_dbs, key=lambda x: x['name'].lower())
Update expert database list in the footer"""
Copyright [2009-2014] EMBL-European Bioinformatics Institute
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import template
from portal.config.expert_databases import expert_dbs
register = template.Library()
@register.assignment_tag
def get_expert_databases_columns():
"""
Return expert databases grouped and order for the website footer.
"""
dbs = sorted(expert_dbs, key=lambda x: x['name'].lower())
return [
[dbs[x] for x in [1,0,2,3,4,5,6,7]],
dbs[8:16],
dbs[16:23],
dbs[23:]
]
@register.assignment_tag
def get_expert_databases_list():
"""
Get an alphabetically sorted list of imported expert databases.
"""
imported_dbs = [x for x in expert_dbs if x['imported']]
return sorted(imported_dbs, key=lambda x: x['name'].lower())
| <commit_before>"""
Copyright [2009-2014] EMBL-European Bioinformatics Institute
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import template
from portal.config.expert_databases import expert_dbs
register = template.Library()
@register.assignment_tag
def get_expert_databases_columns():
"""
Return expert databases grouped and order for the website footer.
"""
dbs = sorted(expert_dbs, key=lambda x: x['name'].lower())
return [
[dbs[x] for x in [1,0,2,3,4,5,6]],
dbs[7:13],
dbs[13:19],
dbs[19:]
]
@register.assignment_tag
def get_expert_databases_list():
"""
Get an alphabetically sorted list of imported expert databases.
"""
imported_dbs = [x for x in expert_dbs if x['imported']]
return sorted(imported_dbs, key=lambda x: x['name'].lower())
<commit_msg>Update expert database list in the footer<commit_after>"""
Copyright [2009-2014] EMBL-European Bioinformatics Institute
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import template
from portal.config.expert_databases import expert_dbs
register = template.Library()
@register.assignment_tag
def get_expert_databases_columns():
"""
Return expert databases grouped and order for the website footer.
"""
dbs = sorted(expert_dbs, key=lambda x: x['name'].lower())
return [
[dbs[x] for x in [1,0,2,3,4,5,6,7]],
dbs[8:16],
dbs[16:23],
dbs[23:]
]
@register.assignment_tag
def get_expert_databases_list():
"""
Get an alphabetically sorted list of imported expert databases.
"""
imported_dbs = [x for x in expert_dbs if x['imported']]
return sorted(imported_dbs, key=lambda x: x['name'].lower())
|
fb27a53d0ea46e9012610eeccd90b50be07388d9 | doctr/tests/test_local.py | doctr/tests/test_local.py | from ..local import check_repo_exists
from pytest import raises
def test_bad_user():
with raises(RuntimeError):
check_repo_exists('---/invaliduser')
def test_bad_repo():
with raises(RuntimeError):
check_repo_exists('drdoctr/---')
def test_repo_exists():
assert check_repo_exists('drdoctr/doctr')
def test_invalid_repo():
with raises(RuntimeError):
check_repo_exists('fdsf')
with raises(RuntimeError):
check_repo_exists('fdsf/fdfs/fd')
| from ..local import check_repo_exists
from pytest import raises
def test_bad_user():
with raises(RuntimeError):
check_repo_exists('---/invaliduser')
def test_bad_repo():
with raises(RuntimeError):
check_repo_exists('drdoctr/---')
def test_repo_exists():
assert not check_repo_exists('drdoctr/doctr')
def test_invalid_repo():
with raises(RuntimeError):
check_repo_exists('fdsf')
with raises(RuntimeError):
check_repo_exists('fdsf/fdfs/fd')
| Update expected test result for test_repo_exists now that check_repo_exists returns whether the repo is private or not | Update expected test result for test_repo_exists now that check_repo_exists returns whether the repo is private or not
| Python | mit | gforsyth/doctr_testing,drdoctr/doctr | from ..local import check_repo_exists
from pytest import raises
def test_bad_user():
with raises(RuntimeError):
check_repo_exists('---/invaliduser')
def test_bad_repo():
with raises(RuntimeError):
check_repo_exists('drdoctr/---')
def test_repo_exists():
assert check_repo_exists('drdoctr/doctr')
def test_invalid_repo():
with raises(RuntimeError):
check_repo_exists('fdsf')
with raises(RuntimeError):
check_repo_exists('fdsf/fdfs/fd')
Update expected test result for test_repo_exists now that check_repo_exists returns whether the repo is private or not | from ..local import check_repo_exists
from pytest import raises
def test_bad_user():
with raises(RuntimeError):
check_repo_exists('---/invaliduser')
def test_bad_repo():
with raises(RuntimeError):
check_repo_exists('drdoctr/---')
def test_repo_exists():
assert not check_repo_exists('drdoctr/doctr')
def test_invalid_repo():
with raises(RuntimeError):
check_repo_exists('fdsf')
with raises(RuntimeError):
check_repo_exists('fdsf/fdfs/fd')
| <commit_before>from ..local import check_repo_exists
from pytest import raises
def test_bad_user():
with raises(RuntimeError):
check_repo_exists('---/invaliduser')
def test_bad_repo():
with raises(RuntimeError):
check_repo_exists('drdoctr/---')
def test_repo_exists():
assert check_repo_exists('drdoctr/doctr')
def test_invalid_repo():
with raises(RuntimeError):
check_repo_exists('fdsf')
with raises(RuntimeError):
check_repo_exists('fdsf/fdfs/fd')
<commit_msg>Update expected test result for test_repo_exists now that check_repo_exists returns whether the repo is private or not<commit_after> | from ..local import check_repo_exists
from pytest import raises
def test_bad_user():
with raises(RuntimeError):
check_repo_exists('---/invaliduser')
def test_bad_repo():
with raises(RuntimeError):
check_repo_exists('drdoctr/---')
def test_repo_exists():
assert not check_repo_exists('drdoctr/doctr')
def test_invalid_repo():
with raises(RuntimeError):
check_repo_exists('fdsf')
with raises(RuntimeError):
check_repo_exists('fdsf/fdfs/fd')
| from ..local import check_repo_exists
from pytest import raises
def test_bad_user():
with raises(RuntimeError):
check_repo_exists('---/invaliduser')
def test_bad_repo():
with raises(RuntimeError):
check_repo_exists('drdoctr/---')
def test_repo_exists():
assert check_repo_exists('drdoctr/doctr')
def test_invalid_repo():
with raises(RuntimeError):
check_repo_exists('fdsf')
with raises(RuntimeError):
check_repo_exists('fdsf/fdfs/fd')
Update expected test result for test_repo_exists now that check_repo_exists returns whether the repo is private or notfrom ..local import check_repo_exists
from pytest import raises
def test_bad_user():
with raises(RuntimeError):
check_repo_exists('---/invaliduser')
def test_bad_repo():
with raises(RuntimeError):
check_repo_exists('drdoctr/---')
def test_repo_exists():
assert not check_repo_exists('drdoctr/doctr')
def test_invalid_repo():
with raises(RuntimeError):
check_repo_exists('fdsf')
with raises(RuntimeError):
check_repo_exists('fdsf/fdfs/fd')
| <commit_before>from ..local import check_repo_exists
from pytest import raises
def test_bad_user():
with raises(RuntimeError):
check_repo_exists('---/invaliduser')
def test_bad_repo():
with raises(RuntimeError):
check_repo_exists('drdoctr/---')
def test_repo_exists():
assert check_repo_exists('drdoctr/doctr')
def test_invalid_repo():
with raises(RuntimeError):
check_repo_exists('fdsf')
with raises(RuntimeError):
check_repo_exists('fdsf/fdfs/fd')
<commit_msg>Update expected test result for test_repo_exists now that check_repo_exists returns whether the repo is private or not<commit_after>from ..local import check_repo_exists
from pytest import raises
def test_bad_user():
with raises(RuntimeError):
check_repo_exists('---/invaliduser')
def test_bad_repo():
with raises(RuntimeError):
check_repo_exists('drdoctr/---')
def test_repo_exists():
assert not check_repo_exists('drdoctr/doctr')
def test_invalid_repo():
with raises(RuntimeError):
check_repo_exists('fdsf')
with raises(RuntimeError):
check_repo_exists('fdsf/fdfs/fd')
|
30230cb6fcb29cd437d3ce71c3370da6d38cb622 | python/04-2.py | python/04-2.py | #!/usr/bin/env python
import hashlib
prefix = ''
number = 1
with open('../inputs/04.txt') as f:
prefix = f.readlines()
prefix = prefix[0].rstrip()
while True:
md5 = hashlib.md5()
md5.update('{0}{1}'.format(prefix, number))
if md5.hexdigest()[:6] == '000000':
#print md5.hexdigest()
print number
break
number += 1
| #!/usr/bin/env python
import hashlib
prefix = ''
number = 1
with open('../inputs/04.txt') as f:
prefix = f.readlines()
prefix = prefix[0].rstrip()
md5 = hashlib.md5()
md5.update(prefix)
while True:
m = md5.copy()
m.update(str(number))
if m.hexdigest()[:6] == '000000':
print number
break
number += 1
| Use md5.copy() to be more efficient. | Use md5.copy() to be more efficient.
The hash.copy() documentation says this is more efficient given a common
initial substring.
| Python | mit | opello/adventofcode | #!/usr/bin/env python
import hashlib
prefix = ''
number = 1
with open('../inputs/04.txt') as f:
prefix = f.readlines()
prefix = prefix[0].rstrip()
while True:
md5 = hashlib.md5()
md5.update('{0}{1}'.format(prefix, number))
if md5.hexdigest()[:6] == '000000':
#print md5.hexdigest()
print number
break
number += 1
Use md5.copy() to be more efficient.
The hash.copy() documentation says this is more efficient given a common
initial substring. | #!/usr/bin/env python
import hashlib
prefix = ''
number = 1
with open('../inputs/04.txt') as f:
prefix = f.readlines()
prefix = prefix[0].rstrip()
md5 = hashlib.md5()
md5.update(prefix)
while True:
m = md5.copy()
m.update(str(number))
if m.hexdigest()[:6] == '000000':
print number
break
number += 1
| <commit_before>#!/usr/bin/env python
import hashlib
prefix = ''
number = 1
with open('../inputs/04.txt') as f:
prefix = f.readlines()
prefix = prefix[0].rstrip()
while True:
md5 = hashlib.md5()
md5.update('{0}{1}'.format(prefix, number))
if md5.hexdigest()[:6] == '000000':
#print md5.hexdigest()
print number
break
number += 1
<commit_msg>Use md5.copy() to be more efficient.
The hash.copy() documentation says this is more efficient given a common
initial substring.<commit_after> | #!/usr/bin/env python
import hashlib
prefix = ''
number = 1
with open('../inputs/04.txt') as f:
prefix = f.readlines()
prefix = prefix[0].rstrip()
md5 = hashlib.md5()
md5.update(prefix)
while True:
m = md5.copy()
m.update(str(number))
if m.hexdigest()[:6] == '000000':
print number
break
number += 1
| #!/usr/bin/env python
import hashlib
prefix = ''
number = 1
with open('../inputs/04.txt') as f:
prefix = f.readlines()
prefix = prefix[0].rstrip()
while True:
md5 = hashlib.md5()
md5.update('{0}{1}'.format(prefix, number))
if md5.hexdigest()[:6] == '000000':
#print md5.hexdigest()
print number
break
number += 1
Use md5.copy() to be more efficient.
The hash.copy() documentation says this is more efficient given a common
initial substring.#!/usr/bin/env python
import hashlib
prefix = ''
number = 1
with open('../inputs/04.txt') as f:
prefix = f.readlines()
prefix = prefix[0].rstrip()
md5 = hashlib.md5()
md5.update(prefix)
while True:
m = md5.copy()
m.update(str(number))
if m.hexdigest()[:6] == '000000':
print number
break
number += 1
| <commit_before>#!/usr/bin/env python
import hashlib
prefix = ''
number = 1
with open('../inputs/04.txt') as f:
prefix = f.readlines()
prefix = prefix[0].rstrip()
while True:
md5 = hashlib.md5()
md5.update('{0}{1}'.format(prefix, number))
if md5.hexdigest()[:6] == '000000':
#print md5.hexdigest()
print number
break
number += 1
<commit_msg>Use md5.copy() to be more efficient.
The hash.copy() documentation says this is more efficient given a common
initial substring.<commit_after>#!/usr/bin/env python
import hashlib
prefix = ''
number = 1
with open('../inputs/04.txt') as f:
prefix = f.readlines()
prefix = prefix[0].rstrip()
md5 = hashlib.md5()
md5.update(prefix)
while True:
m = md5.copy()
m.update(str(number))
if m.hexdigest()[:6] == '000000':
print number
break
number += 1
|
1152e7a329ee20494a4856f7a83f5ab1e6c4390e | runtests.py | runtests.py | #!/usr/bin/env python
import os, sys
from django.conf import settings
import django
DEFAULT_SETTINGS = dict(
INSTALLED_APPS=(
'django.contrib.contenttypes',
'model_utils',
'model_utils.tests',
),
DATABASES={
"default": {
"ENGINE": "django.db.backends.sqlite3"
}
},
)
def runtests(*test_args):
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
# Compatibility with Django 1.7's stricter initialization
if hasattr(django, 'setup'):
django.setup()
if not test_args:
test_args = ['tests']
parent = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, parent)
from django.test.simple import DjangoTestSuiteRunner
failures = DjangoTestSuiteRunner(
verbosity=1, interactive=True, failfast=False).run_tests(test_args)
sys.exit(failures)
if __name__ == '__main__':
runtests()
| #!/usr/bin/env python
import os, sys
from django.conf import settings
import django
DEFAULT_SETTINGS = dict(
INSTALLED_APPS=(
'model_utils',
'model_utils.tests',
),
DATABASES={
"default": {
"ENGINE": "django.db.backends.sqlite3"
}
},
)
def runtests(*test_args):
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
# Compatibility with Django 1.7's stricter initialization
if hasattr(django, 'setup'):
django.setup()
if not test_args:
test_args = ['tests']
parent = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, parent)
from django.test.simple import DjangoTestSuiteRunner
failures = DjangoTestSuiteRunner(
verbosity=1, interactive=True, failfast=False).run_tests(test_args)
sys.exit(failures)
if __name__ == '__main__':
runtests()
| Remove contenttypes from INSTALLED_APPS for testing; no longer needed. | Remove contenttypes from INSTALLED_APPS for testing; no longer needed.
| Python | bsd-3-clause | nemesisdesign/django-model-utils,timmygee/django-model-utils,patrys/django-model-utils,yeago/django-model-utils,nemesisdesign/django-model-utils,timmygee/django-model-utils,patrys/django-model-utils,carljm/django-model-utils,yeago/django-model-utils,carljm/django-model-utils | #!/usr/bin/env python
import os, sys
from django.conf import settings
import django
DEFAULT_SETTINGS = dict(
INSTALLED_APPS=(
'django.contrib.contenttypes',
'model_utils',
'model_utils.tests',
),
DATABASES={
"default": {
"ENGINE": "django.db.backends.sqlite3"
}
},
)
def runtests(*test_args):
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
# Compatibility with Django 1.7's stricter initialization
if hasattr(django, 'setup'):
django.setup()
if not test_args:
test_args = ['tests']
parent = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, parent)
from django.test.simple import DjangoTestSuiteRunner
failures = DjangoTestSuiteRunner(
verbosity=1, interactive=True, failfast=False).run_tests(test_args)
sys.exit(failures)
if __name__ == '__main__':
runtests()
Remove contenttypes from INSTALLED_APPS for testing; no longer needed. | #!/usr/bin/env python
import os, sys
from django.conf import settings
import django
DEFAULT_SETTINGS = dict(
INSTALLED_APPS=(
'model_utils',
'model_utils.tests',
),
DATABASES={
"default": {
"ENGINE": "django.db.backends.sqlite3"
}
},
)
def runtests(*test_args):
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
# Compatibility with Django 1.7's stricter initialization
if hasattr(django, 'setup'):
django.setup()
if not test_args:
test_args = ['tests']
parent = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, parent)
from django.test.simple import DjangoTestSuiteRunner
failures = DjangoTestSuiteRunner(
verbosity=1, interactive=True, failfast=False).run_tests(test_args)
sys.exit(failures)
if __name__ == '__main__':
runtests()
| <commit_before>#!/usr/bin/env python
import os, sys
from django.conf import settings
import django
DEFAULT_SETTINGS = dict(
INSTALLED_APPS=(
'django.contrib.contenttypes',
'model_utils',
'model_utils.tests',
),
DATABASES={
"default": {
"ENGINE": "django.db.backends.sqlite3"
}
},
)
def runtests(*test_args):
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
# Compatibility with Django 1.7's stricter initialization
if hasattr(django, 'setup'):
django.setup()
if not test_args:
test_args = ['tests']
parent = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, parent)
from django.test.simple import DjangoTestSuiteRunner
failures = DjangoTestSuiteRunner(
verbosity=1, interactive=True, failfast=False).run_tests(test_args)
sys.exit(failures)
if __name__ == '__main__':
runtests()
<commit_msg>Remove contenttypes from INSTALLED_APPS for testing; no longer needed.<commit_after> | #!/usr/bin/env python
import os, sys
from django.conf import settings
import django
DEFAULT_SETTINGS = dict(
INSTALLED_APPS=(
'model_utils',
'model_utils.tests',
),
DATABASES={
"default": {
"ENGINE": "django.db.backends.sqlite3"
}
},
)
def runtests(*test_args):
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
# Compatibility with Django 1.7's stricter initialization
if hasattr(django, 'setup'):
django.setup()
if not test_args:
test_args = ['tests']
parent = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, parent)
from django.test.simple import DjangoTestSuiteRunner
failures = DjangoTestSuiteRunner(
verbosity=1, interactive=True, failfast=False).run_tests(test_args)
sys.exit(failures)
if __name__ == '__main__':
runtests()
| #!/usr/bin/env python
import os, sys
from django.conf import settings
import django
DEFAULT_SETTINGS = dict(
INSTALLED_APPS=(
'django.contrib.contenttypes',
'model_utils',
'model_utils.tests',
),
DATABASES={
"default": {
"ENGINE": "django.db.backends.sqlite3"
}
},
)
def runtests(*test_args):
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
# Compatibility with Django 1.7's stricter initialization
if hasattr(django, 'setup'):
django.setup()
if not test_args:
test_args = ['tests']
parent = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, parent)
from django.test.simple import DjangoTestSuiteRunner
failures = DjangoTestSuiteRunner(
verbosity=1, interactive=True, failfast=False).run_tests(test_args)
sys.exit(failures)
if __name__ == '__main__':
runtests()
Remove contenttypes from INSTALLED_APPS for testing; no longer needed.#!/usr/bin/env python
import os, sys
from django.conf import settings
import django
DEFAULT_SETTINGS = dict(
INSTALLED_APPS=(
'model_utils',
'model_utils.tests',
),
DATABASES={
"default": {
"ENGINE": "django.db.backends.sqlite3"
}
},
)
def runtests(*test_args):
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
# Compatibility with Django 1.7's stricter initialization
if hasattr(django, 'setup'):
django.setup()
if not test_args:
test_args = ['tests']
parent = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, parent)
from django.test.simple import DjangoTestSuiteRunner
failures = DjangoTestSuiteRunner(
verbosity=1, interactive=True, failfast=False).run_tests(test_args)
sys.exit(failures)
if __name__ == '__main__':
runtests()
| <commit_before>#!/usr/bin/env python
import os, sys
from django.conf import settings
import django
DEFAULT_SETTINGS = dict(
INSTALLED_APPS=(
'django.contrib.contenttypes',
'model_utils',
'model_utils.tests',
),
DATABASES={
"default": {
"ENGINE": "django.db.backends.sqlite3"
}
},
)
def runtests(*test_args):
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
# Compatibility with Django 1.7's stricter initialization
if hasattr(django, 'setup'):
django.setup()
if not test_args:
test_args = ['tests']
parent = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, parent)
from django.test.simple import DjangoTestSuiteRunner
failures = DjangoTestSuiteRunner(
verbosity=1, interactive=True, failfast=False).run_tests(test_args)
sys.exit(failures)
if __name__ == '__main__':
runtests()
<commit_msg>Remove contenttypes from INSTALLED_APPS for testing; no longer needed.<commit_after>#!/usr/bin/env python
import os, sys
from django.conf import settings
import django
DEFAULT_SETTINGS = dict(
INSTALLED_APPS=(
'model_utils',
'model_utils.tests',
),
DATABASES={
"default": {
"ENGINE": "django.db.backends.sqlite3"
}
},
)
def runtests(*test_args):
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
# Compatibility with Django 1.7's stricter initialization
if hasattr(django, 'setup'):
django.setup()
if not test_args:
test_args = ['tests']
parent = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, parent)
from django.test.simple import DjangoTestSuiteRunner
failures = DjangoTestSuiteRunner(
verbosity=1, interactive=True, failfast=False).run_tests(test_args)
sys.exit(failures)
if __name__ == '__main__':
runtests()
|
155b1e6b8d431f1169a3e71d08d93d76a3414c59 | turbustat/statistics/vca_vcs/slice_thickness.py | turbustat/statistics/vca_vcs/slice_thickness.py | # Licensed under an MIT open source license - see LICENSE
import numpy as np
def change_slice_thickness(cube, slice_thickness=1.0):
'''
Degrades the velocity resolution of a data cube. This is to avoid
shot noise by removing velocity fluctuations at small thicknesses.
Parameters
----------
cube : numpy.ndarray
3D data cube to degrade
slice_thickness : float, optional
Thicknesses of the new slices. Minimum is 1.0
Thickness must be integer multiple of the original cube size
Returns
-------
degraded_cube : numpy.ndarray
Data cube degraded to new slice thickness
'''
assert isinstance(slice_thickness, float)
if slice_thickness < 1:
slice_thickness == 1
print "Slice Thickness must be at least 1.0. Returning original cube."
if slice_thickness == 1:
return cube
if cube.shape[0] % slice_thickness != 0:
raise TypeError("Slice thickness must be integer multiple of dimension"
" size % s" % (cube.shape[0]))
slice_thickness = int(slice_thickness)
# Want to average over velocity channels
new_channel_indices = np.arange(0, cube.shape[0] / slice_thickness)
degraded_cube = np.ones(
(cube.shape[0] / slice_thickness, cube.shape[1], cube.shape[2]))
for channel in new_channel_indices:
old_index = int(channel * slice_thickness)
channel = int(channel)
degraded_cube[channel, :, :] = \
np.nanmean(cube[old_index:old_index + slice_thickness], axis=0)
return degraded_cube
| # Licensed under an MIT open source license - see LICENSE
import numpy as np
from astropy import units as u
from spectral_cube import SpectralCube
from astropy.convolution import Gaussian1DKernel
def spectral_regrid_cube(cube, channel_width):
fwhm_factor = np.sqrt(8 * np.log(2))
current_resolution = np.diff(cube.spectral_axis[:2])[0]
target_resolution = channel_width.to(current_resolution.unit)
diff_factor = np.abs(target_resolution / current_resolution).value
pixel_scale = np.abs(current_resolution)
gaussian_width = ((target_resolution**2 - current_resolution**2)**0.5 /
pixel_scale / fwhm_factor)
kernel = Gaussian1DKernel(gaussian_width)
new_cube = cube.spectral_smooth(kernel)
# Now define the new spectral axis at the new resolution
num_chan = int(np.floor_divide(cube.shape[0], diff_factor))
new_specaxis = np.linspace(cube.spectral_axis.min().value,
cube.spectral_axis.max().value,
num_chan) * current_resolution.unit
# Keep the same order (max to min or min to max)
if current_resolution.value < 0:
new_specaxis = new_specaxis[::-1]
return new_cube.spectral_interpolate(new_specaxis,
suppress_smooth_warning=True)
| Add a corrected spectral regridding function that smooths before interpolating to a new spectral axis | Add a corrected spectral regridding function that smooths before interpolating to a new spectral axis
| Python | mit | e-koch/TurbuStat,Astroua/TurbuStat | # Licensed under an MIT open source license - see LICENSE
import numpy as np
def change_slice_thickness(cube, slice_thickness=1.0):
'''
Degrades the velocity resolution of a data cube. This is to avoid
shot noise by removing velocity fluctuations at small thicknesses.
Parameters
----------
cube : numpy.ndarray
3D data cube to degrade
slice_thickness : float, optional
Thicknesses of the new slices. Minimum is 1.0
Thickness must be integer multiple of the original cube size
Returns
-------
degraded_cube : numpy.ndarray
Data cube degraded to new slice thickness
'''
assert isinstance(slice_thickness, float)
if slice_thickness < 1:
slice_thickness == 1
print "Slice Thickness must be at least 1.0. Returning original cube."
if slice_thickness == 1:
return cube
if cube.shape[0] % slice_thickness != 0:
raise TypeError("Slice thickness must be integer multiple of dimension"
" size % s" % (cube.shape[0]))
slice_thickness = int(slice_thickness)
# Want to average over velocity channels
new_channel_indices = np.arange(0, cube.shape[0] / slice_thickness)
degraded_cube = np.ones(
(cube.shape[0] / slice_thickness, cube.shape[1], cube.shape[2]))
for channel in new_channel_indices:
old_index = int(channel * slice_thickness)
channel = int(channel)
degraded_cube[channel, :, :] = \
np.nanmean(cube[old_index:old_index + slice_thickness], axis=0)
return degraded_cube
Add a corrected spectral regridding function that smooths before interpolating to a new spectral axis | # Licensed under an MIT open source license - see LICENSE
import numpy as np
from astropy import units as u
from spectral_cube import SpectralCube
from astropy.convolution import Gaussian1DKernel
def spectral_regrid_cube(cube, channel_width):
fwhm_factor = np.sqrt(8 * np.log(2))
current_resolution = np.diff(cube.spectral_axis[:2])[0]
target_resolution = channel_width.to(current_resolution.unit)
diff_factor = np.abs(target_resolution / current_resolution).value
pixel_scale = np.abs(current_resolution)
gaussian_width = ((target_resolution**2 - current_resolution**2)**0.5 /
pixel_scale / fwhm_factor)
kernel = Gaussian1DKernel(gaussian_width)
new_cube = cube.spectral_smooth(kernel)
# Now define the new spectral axis at the new resolution
num_chan = int(np.floor_divide(cube.shape[0], diff_factor))
new_specaxis = np.linspace(cube.spectral_axis.min().value,
cube.spectral_axis.max().value,
num_chan) * current_resolution.unit
# Keep the same order (max to min or min to max)
if current_resolution.value < 0:
new_specaxis = new_specaxis[::-1]
return new_cube.spectral_interpolate(new_specaxis,
suppress_smooth_warning=True)
| <commit_before># Licensed under an MIT open source license - see LICENSE
import numpy as np
def change_slice_thickness(cube, slice_thickness=1.0):
'''
Degrades the velocity resolution of a data cube. This is to avoid
shot noise by removing velocity fluctuations at small thicknesses.
Parameters
----------
cube : numpy.ndarray
3D data cube to degrade
slice_thickness : float, optional
Thicknesses of the new slices. Minimum is 1.0
Thickness must be integer multiple of the original cube size
Returns
-------
degraded_cube : numpy.ndarray
Data cube degraded to new slice thickness
'''
assert isinstance(slice_thickness, float)
if slice_thickness < 1:
slice_thickness == 1
print "Slice Thickness must be at least 1.0. Returning original cube."
if slice_thickness == 1:
return cube
if cube.shape[0] % slice_thickness != 0:
raise TypeError("Slice thickness must be integer multiple of dimension"
" size % s" % (cube.shape[0]))
slice_thickness = int(slice_thickness)
# Want to average over velocity channels
new_channel_indices = np.arange(0, cube.shape[0] / slice_thickness)
degraded_cube = np.ones(
(cube.shape[0] / slice_thickness, cube.shape[1], cube.shape[2]))
for channel in new_channel_indices:
old_index = int(channel * slice_thickness)
channel = int(channel)
degraded_cube[channel, :, :] = \
np.nanmean(cube[old_index:old_index + slice_thickness], axis=0)
return degraded_cube
<commit_msg>Add a corrected spectral regridding function that smooths before interpolating to a new spectral axis<commit_after> | # Licensed under an MIT open source license - see LICENSE
import numpy as np
from astropy import units as u
from spectral_cube import SpectralCube
from astropy.convolution import Gaussian1DKernel
def spectral_regrid_cube(cube, channel_width):
fwhm_factor = np.sqrt(8 * np.log(2))
current_resolution = np.diff(cube.spectral_axis[:2])[0]
target_resolution = channel_width.to(current_resolution.unit)
diff_factor = np.abs(target_resolution / current_resolution).value
pixel_scale = np.abs(current_resolution)
gaussian_width = ((target_resolution**2 - current_resolution**2)**0.5 /
pixel_scale / fwhm_factor)
kernel = Gaussian1DKernel(gaussian_width)
new_cube = cube.spectral_smooth(kernel)
# Now define the new spectral axis at the new resolution
num_chan = int(np.floor_divide(cube.shape[0], diff_factor))
new_specaxis = np.linspace(cube.spectral_axis.min().value,
cube.spectral_axis.max().value,
num_chan) * current_resolution.unit
# Keep the same order (max to min or min to max)
if current_resolution.value < 0:
new_specaxis = new_specaxis[::-1]
return new_cube.spectral_interpolate(new_specaxis,
suppress_smooth_warning=True)
| # Licensed under an MIT open source license - see LICENSE
import numpy as np
def change_slice_thickness(cube, slice_thickness=1.0):
'''
Degrades the velocity resolution of a data cube. This is to avoid
shot noise by removing velocity fluctuations at small thicknesses.
Parameters
----------
cube : numpy.ndarray
3D data cube to degrade
slice_thickness : float, optional
Thicknesses of the new slices. Minimum is 1.0
Thickness must be integer multiple of the original cube size
Returns
-------
degraded_cube : numpy.ndarray
Data cube degraded to new slice thickness
'''
assert isinstance(slice_thickness, float)
if slice_thickness < 1:
slice_thickness == 1
print "Slice Thickness must be at least 1.0. Returning original cube."
if slice_thickness == 1:
return cube
if cube.shape[0] % slice_thickness != 0:
raise TypeError("Slice thickness must be integer multiple of dimension"
" size % s" % (cube.shape[0]))
slice_thickness = int(slice_thickness)
# Want to average over velocity channels
new_channel_indices = np.arange(0, cube.shape[0] / slice_thickness)
degraded_cube = np.ones(
(cube.shape[0] / slice_thickness, cube.shape[1], cube.shape[2]))
for channel in new_channel_indices:
old_index = int(channel * slice_thickness)
channel = int(channel)
degraded_cube[channel, :, :] = \
np.nanmean(cube[old_index:old_index + slice_thickness], axis=0)
return degraded_cube
Add a corrected spectral regridding function that smooths before interpolating to a new spectral axis# Licensed under an MIT open source license - see LICENSE
import numpy as np
from astropy import units as u
from spectral_cube import SpectralCube
from astropy.convolution import Gaussian1DKernel
def spectral_regrid_cube(cube, channel_width):
fwhm_factor = np.sqrt(8 * np.log(2))
current_resolution = np.diff(cube.spectral_axis[:2])[0]
target_resolution = channel_width.to(current_resolution.unit)
diff_factor = np.abs(target_resolution / current_resolution).value
pixel_scale = np.abs(current_resolution)
gaussian_width = ((target_resolution**2 - current_resolution**2)**0.5 /
pixel_scale / fwhm_factor)
kernel = Gaussian1DKernel(gaussian_width)
new_cube = cube.spectral_smooth(kernel)
# Now define the new spectral axis at the new resolution
num_chan = int(np.floor_divide(cube.shape[0], diff_factor))
new_specaxis = np.linspace(cube.spectral_axis.min().value,
cube.spectral_axis.max().value,
num_chan) * current_resolution.unit
# Keep the same order (max to min or min to max)
if current_resolution.value < 0:
new_specaxis = new_specaxis[::-1]
return new_cube.spectral_interpolate(new_specaxis,
suppress_smooth_warning=True)
| <commit_before># Licensed under an MIT open source license - see LICENSE
import numpy as np
def change_slice_thickness(cube, slice_thickness=1.0):
'''
Degrades the velocity resolution of a data cube. This is to avoid
shot noise by removing velocity fluctuations at small thicknesses.
Parameters
----------
cube : numpy.ndarray
3D data cube to degrade
slice_thickness : float, optional
Thicknesses of the new slices. Minimum is 1.0
Thickness must be integer multiple of the original cube size
Returns
-------
degraded_cube : numpy.ndarray
Data cube degraded to new slice thickness
'''
assert isinstance(slice_thickness, float)
if slice_thickness < 1:
slice_thickness == 1
print "Slice Thickness must be at least 1.0. Returning original cube."
if slice_thickness == 1:
return cube
if cube.shape[0] % slice_thickness != 0:
raise TypeError("Slice thickness must be integer multiple of dimension"
" size % s" % (cube.shape[0]))
slice_thickness = int(slice_thickness)
# Want to average over velocity channels
new_channel_indices = np.arange(0, cube.shape[0] / slice_thickness)
degraded_cube = np.ones(
(cube.shape[0] / slice_thickness, cube.shape[1], cube.shape[2]))
for channel in new_channel_indices:
old_index = int(channel * slice_thickness)
channel = int(channel)
degraded_cube[channel, :, :] = \
np.nanmean(cube[old_index:old_index + slice_thickness], axis=0)
return degraded_cube
<commit_msg>Add a corrected spectral regridding function that smooths before interpolating to a new spectral axis<commit_after># Licensed under an MIT open source license - see LICENSE
import numpy as np
from astropy import units as u
from spectral_cube import SpectralCube
from astropy.convolution import Gaussian1DKernel
def spectral_regrid_cube(cube, channel_width):
fwhm_factor = np.sqrt(8 * np.log(2))
current_resolution = np.diff(cube.spectral_axis[:2])[0]
target_resolution = channel_width.to(current_resolution.unit)
diff_factor = np.abs(target_resolution / current_resolution).value
pixel_scale = np.abs(current_resolution)
gaussian_width = ((target_resolution**2 - current_resolution**2)**0.5 /
pixel_scale / fwhm_factor)
kernel = Gaussian1DKernel(gaussian_width)
new_cube = cube.spectral_smooth(kernel)
# Now define the new spectral axis at the new resolution
num_chan = int(np.floor_divide(cube.shape[0], diff_factor))
new_specaxis = np.linspace(cube.spectral_axis.min().value,
cube.spectral_axis.max().value,
num_chan) * current_resolution.unit
# Keep the same order (max to min or min to max)
if current_resolution.value < 0:
new_specaxis = new_specaxis[::-1]
return new_cube.spectral_interpolate(new_specaxis,
suppress_smooth_warning=True)
|
e39925db2834a7491f9b8b505e1e1cf181840035 | clowder_server/views.py | clowder_server/views.py | from braces.views import CsrfExemptMixin
from django.core.mail import send_mail
from django.http import HttpResponse
from django.views.generic import TemplateView, View
from clowder_server.models import Alert, Ping
class APIView(CsrfExemptMixin, View):
def post(self, request):
name = request.POST.get('name')
frequency = request.POST.get('frequency')
value = request.POST.get('value')
status = int(request.POST.get('status', 1))
if status == -1:
send_mail('Subject here', 'Here is the message.', 'admin@clowder.io',
['keith@parkme.com'], fail_silently=False)
if frequency:
expiration_date = datetime.datetime.now() + int(frequency)
Alert.objects.filter(name=name).delete()
Alert.objects.create(
name=name,
expire_at=expiration_date
)
Ping.objects.create(
name=name,
value=value,
)
return HttpResponse('ok')
class DashboardView(TemplateView):
template_name = "dashboard.html"
def get_context_data(self, **context):
context['pings'] = Ping.objects.all().order_by('name', 'create')
return context
| from braces.views import CsrfExemptMixin
from django.core.mail import send_mail
from django.http import HttpResponse
from django.views.generic import TemplateView, View
from clowder_server.models import Alert, Ping
class APIView(CsrfExemptMixin, View):
def post(self, request):
name = request.POST.get('name')
frequency = request.POST.get('frequency')
value = request.POST.get('value')
status = int(request.POST.get('status', 1))
if status == -1:
send_mail('Subject here', 'Here is the message.', 'admin@clowder.io',
['keith@parkme.com'], fail_silently=False)
if frequency:
expiration_date = datetime.datetime.now() + int(frequency)
Alert.objects.filter(name=name).delete()
Alert.objects.create(
name=name,
expire_at=expiration_date
)
return HttpResponse('test')
Ping.objects.create(
name=name,
value=value,
)
return HttpResponse('ok')
class DashboardView(TemplateView):
template_name = "dashboard.html"
def get_context_data(self, **context):
context['pings'] = Ping.objects.all().order_by('name', 'create')
return context
| Add test response to frequency | Add test response to frequency
| Python | agpl-3.0 | framewr/clowder_server,framewr/clowder_server,keithhackbarth/clowder_server,keithhackbarth/clowder_server,framewr/clowder_server,keithhackbarth/clowder_server,keithhackbarth/clowder_server,framewr/clowder_server | from braces.views import CsrfExemptMixin
from django.core.mail import send_mail
from django.http import HttpResponse
from django.views.generic import TemplateView, View
from clowder_server.models import Alert, Ping
class APIView(CsrfExemptMixin, View):
def post(self, request):
name = request.POST.get('name')
frequency = request.POST.get('frequency')
value = request.POST.get('value')
status = int(request.POST.get('status', 1))
if status == -1:
send_mail('Subject here', 'Here is the message.', 'admin@clowder.io',
['keith@parkme.com'], fail_silently=False)
if frequency:
expiration_date = datetime.datetime.now() + int(frequency)
Alert.objects.filter(name=name).delete()
Alert.objects.create(
name=name,
expire_at=expiration_date
)
Ping.objects.create(
name=name,
value=value,
)
return HttpResponse('ok')
class DashboardView(TemplateView):
template_name = "dashboard.html"
def get_context_data(self, **context):
context['pings'] = Ping.objects.all().order_by('name', 'create')
return context
Add test response to frequency | from braces.views import CsrfExemptMixin
from django.core.mail import send_mail
from django.http import HttpResponse
from django.views.generic import TemplateView, View
from clowder_server.models import Alert, Ping
class APIView(CsrfExemptMixin, View):
def post(self, request):
name = request.POST.get('name')
frequency = request.POST.get('frequency')
value = request.POST.get('value')
status = int(request.POST.get('status', 1))
if status == -1:
send_mail('Subject here', 'Here is the message.', 'admin@clowder.io',
['keith@parkme.com'], fail_silently=False)
if frequency:
expiration_date = datetime.datetime.now() + int(frequency)
Alert.objects.filter(name=name).delete()
Alert.objects.create(
name=name,
expire_at=expiration_date
)
return HttpResponse('test')
Ping.objects.create(
name=name,
value=value,
)
return HttpResponse('ok')
class DashboardView(TemplateView):
template_name = "dashboard.html"
def get_context_data(self, **context):
context['pings'] = Ping.objects.all().order_by('name', 'create')
return context
| <commit_before>from braces.views import CsrfExemptMixin
from django.core.mail import send_mail
from django.http import HttpResponse
from django.views.generic import TemplateView, View
from clowder_server.models import Alert, Ping
class APIView(CsrfExemptMixin, View):
def post(self, request):
name = request.POST.get('name')
frequency = request.POST.get('frequency')
value = request.POST.get('value')
status = int(request.POST.get('status', 1))
if status == -1:
send_mail('Subject here', 'Here is the message.', 'admin@clowder.io',
['keith@parkme.com'], fail_silently=False)
if frequency:
expiration_date = datetime.datetime.now() + int(frequency)
Alert.objects.filter(name=name).delete()
Alert.objects.create(
name=name,
expire_at=expiration_date
)
Ping.objects.create(
name=name,
value=value,
)
return HttpResponse('ok')
class DashboardView(TemplateView):
template_name = "dashboard.html"
def get_context_data(self, **context):
context['pings'] = Ping.objects.all().order_by('name', 'create')
return context
<commit_msg>Add test response to frequency<commit_after> | from braces.views import CsrfExemptMixin
from django.core.mail import send_mail
from django.http import HttpResponse
from django.views.generic import TemplateView, View
from clowder_server.models import Alert, Ping
class APIView(CsrfExemptMixin, View):
def post(self, request):
name = request.POST.get('name')
frequency = request.POST.get('frequency')
value = request.POST.get('value')
status = int(request.POST.get('status', 1))
if status == -1:
send_mail('Subject here', 'Here is the message.', 'admin@clowder.io',
['keith@parkme.com'], fail_silently=False)
if frequency:
expiration_date = datetime.datetime.now() + int(frequency)
Alert.objects.filter(name=name).delete()
Alert.objects.create(
name=name,
expire_at=expiration_date
)
return HttpResponse('test')
Ping.objects.create(
name=name,
value=value,
)
return HttpResponse('ok')
class DashboardView(TemplateView):
template_name = "dashboard.html"
def get_context_data(self, **context):
context['pings'] = Ping.objects.all().order_by('name', 'create')
return context
| from braces.views import CsrfExemptMixin
from django.core.mail import send_mail
from django.http import HttpResponse
from django.views.generic import TemplateView, View
from clowder_server.models import Alert, Ping
class APIView(CsrfExemptMixin, View):
def post(self, request):
name = request.POST.get('name')
frequency = request.POST.get('frequency')
value = request.POST.get('value')
status = int(request.POST.get('status', 1))
if status == -1:
send_mail('Subject here', 'Here is the message.', 'admin@clowder.io',
['keith@parkme.com'], fail_silently=False)
if frequency:
expiration_date = datetime.datetime.now() + int(frequency)
Alert.objects.filter(name=name).delete()
Alert.objects.create(
name=name,
expire_at=expiration_date
)
Ping.objects.create(
name=name,
value=value,
)
return HttpResponse('ok')
class DashboardView(TemplateView):
template_name = "dashboard.html"
def get_context_data(self, **context):
context['pings'] = Ping.objects.all().order_by('name', 'create')
return context
Add test response to frequencyfrom braces.views import CsrfExemptMixin
from django.core.mail import send_mail
from django.http import HttpResponse
from django.views.generic import TemplateView, View
from clowder_server.models import Alert, Ping
class APIView(CsrfExemptMixin, View):
def post(self, request):
name = request.POST.get('name')
frequency = request.POST.get('frequency')
value = request.POST.get('value')
status = int(request.POST.get('status', 1))
if status == -1:
send_mail('Subject here', 'Here is the message.', 'admin@clowder.io',
['keith@parkme.com'], fail_silently=False)
if frequency:
expiration_date = datetime.datetime.now() + int(frequency)
Alert.objects.filter(name=name).delete()
Alert.objects.create(
name=name,
expire_at=expiration_date
)
return HttpResponse('test')
Ping.objects.create(
name=name,
value=value,
)
return HttpResponse('ok')
class DashboardView(TemplateView):
template_name = "dashboard.html"
def get_context_data(self, **context):
context['pings'] = Ping.objects.all().order_by('name', 'create')
return context
| <commit_before>from braces.views import CsrfExemptMixin
from django.core.mail import send_mail
from django.http import HttpResponse
from django.views.generic import TemplateView, View
from clowder_server.models import Alert, Ping
class APIView(CsrfExemptMixin, View):
def post(self, request):
name = request.POST.get('name')
frequency = request.POST.get('frequency')
value = request.POST.get('value')
status = int(request.POST.get('status', 1))
if status == -1:
send_mail('Subject here', 'Here is the message.', 'admin@clowder.io',
['keith@parkme.com'], fail_silently=False)
if frequency:
expiration_date = datetime.datetime.now() + int(frequency)
Alert.objects.filter(name=name).delete()
Alert.objects.create(
name=name,
expire_at=expiration_date
)
Ping.objects.create(
name=name,
value=value,
)
return HttpResponse('ok')
class DashboardView(TemplateView):
template_name = "dashboard.html"
def get_context_data(self, **context):
context['pings'] = Ping.objects.all().order_by('name', 'create')
return context
<commit_msg>Add test response to frequency<commit_after>from braces.views import CsrfExemptMixin
from django.core.mail import send_mail
from django.http import HttpResponse
from django.views.generic import TemplateView, View
from clowder_server.models import Alert, Ping
class APIView(CsrfExemptMixin, View):
def post(self, request):
name = request.POST.get('name')
frequency = request.POST.get('frequency')
value = request.POST.get('value')
status = int(request.POST.get('status', 1))
if status == -1:
send_mail('Subject here', 'Here is the message.', 'admin@clowder.io',
['keith@parkme.com'], fail_silently=False)
if frequency:
expiration_date = datetime.datetime.now() + int(frequency)
Alert.objects.filter(name=name).delete()
Alert.objects.create(
name=name,
expire_at=expiration_date
)
return HttpResponse('test')
Ping.objects.create(
name=name,
value=value,
)
return HttpResponse('ok')
class DashboardView(TemplateView):
template_name = "dashboard.html"
def get_context_data(self, **context):
context['pings'] = Ping.objects.all().order_by('name', 'create')
return context
|
855434523df57183c31ed9b10e7458232b79046a | aclarknet/aclarknet/aclarknet/models.py | aclarknet/aclarknet/aclarknet/models.py | from django.db import models
class Client(models.Model):
client_name = models.CharField(max_length=60)
class Service(models.Model):
name = models.CharField(max_length=60)
class TeamMember(models.Model):
name = models.CharField(max_length=60)
| from django.db import models
class Client(models.Model):
client_name = models.CharField(max_length=60)
def __unicode__(self):
return self.client_name
class Service(models.Model):
name = models.CharField(max_length=60)
def __unicode__(self):
return self.name
class TeamMember(models.Model):
name = models.CharField(max_length=60)
def __unicode__(self):
return self.name
| Fix object name in Django Admin | Fix object name in Django Admin
http://stackoverflow.com/questions/9336463/django-xxxxxx-object-display-customization-in-admin-action-sidebar
| Python | mit | ACLARKNET/aclarknet-django,ACLARKNET/aclarknet-django | from django.db import models
class Client(models.Model):
client_name = models.CharField(max_length=60)
class Service(models.Model):
name = models.CharField(max_length=60)
class TeamMember(models.Model):
name = models.CharField(max_length=60)
Fix object name in Django Admin
http://stackoverflow.com/questions/9336463/django-xxxxxx-object-display-customization-in-admin-action-sidebar | from django.db import models
class Client(models.Model):
client_name = models.CharField(max_length=60)
def __unicode__(self):
return self.client_name
class Service(models.Model):
name = models.CharField(max_length=60)
def __unicode__(self):
return self.name
class TeamMember(models.Model):
name = models.CharField(max_length=60)
def __unicode__(self):
return self.name
| <commit_before>from django.db import models
class Client(models.Model):
client_name = models.CharField(max_length=60)
class Service(models.Model):
name = models.CharField(max_length=60)
class TeamMember(models.Model):
name = models.CharField(max_length=60)
<commit_msg>Fix object name in Django Admin
http://stackoverflow.com/questions/9336463/django-xxxxxx-object-display-customization-in-admin-action-sidebar<commit_after> | from django.db import models
class Client(models.Model):
client_name = models.CharField(max_length=60)
def __unicode__(self):
return self.client_name
class Service(models.Model):
name = models.CharField(max_length=60)
def __unicode__(self):
return self.name
class TeamMember(models.Model):
name = models.CharField(max_length=60)
def __unicode__(self):
return self.name
| from django.db import models
class Client(models.Model):
client_name = models.CharField(max_length=60)
class Service(models.Model):
name = models.CharField(max_length=60)
class TeamMember(models.Model):
name = models.CharField(max_length=60)
Fix object name in Django Admin
http://stackoverflow.com/questions/9336463/django-xxxxxx-object-display-customization-in-admin-action-sidebarfrom django.db import models
class Client(models.Model):
client_name = models.CharField(max_length=60)
def __unicode__(self):
return self.client_name
class Service(models.Model):
name = models.CharField(max_length=60)
def __unicode__(self):
return self.name
class TeamMember(models.Model):
name = models.CharField(max_length=60)
def __unicode__(self):
return self.name
| <commit_before>from django.db import models
class Client(models.Model):
client_name = models.CharField(max_length=60)
class Service(models.Model):
name = models.CharField(max_length=60)
class TeamMember(models.Model):
name = models.CharField(max_length=60)
<commit_msg>Fix object name in Django Admin
http://stackoverflow.com/questions/9336463/django-xxxxxx-object-display-customization-in-admin-action-sidebar<commit_after>from django.db import models
class Client(models.Model):
client_name = models.CharField(max_length=60)
def __unicode__(self):
return self.client_name
class Service(models.Model):
name = models.CharField(max_length=60)
def __unicode__(self):
return self.name
class TeamMember(models.Model):
name = models.CharField(max_length=60)
def __unicode__(self):
return self.name
|
6b06ff67097d0a2ef639df4a3d9baf4f6677b5fd | lmj/sim/__init__.py | lmj/sim/__init__.py | # Copyright (c) 2013 Leif Johnson <leif@leifjohnson.net>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
'''Yet another OpenGL-and-physics simulator framework !'''
from . import ode
from .log import enable_default_logging, get_logger
from .ode import make_quaternion
from .world import Viewer, World
import plac
def call(main):
'''Enable logging and start up a main method.'''
enable_default_logging()
plac.call(main)
args = plac.annotations
| # Copyright (c) 2013 Leif Johnson <leif@leifjohnson.net>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
'''Yet another OpenGL-and-physics simulator framework !'''
from . import physics
from .log import enable_default_logging, get_logger
from .world import Viewer, World
import plac
def call(main):
'''Enable logging and start up a main method.'''
enable_default_logging()
plac.call(main)
args = plac.annotations
| Update package imports for module name change. | Update package imports for module name change.
| Python | mit | EmbodiedCognition/pagoda,EmbodiedCognition/pagoda | # Copyright (c) 2013 Leif Johnson <leif@leifjohnson.net>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
'''Yet another OpenGL-and-physics simulator framework !'''
from . import ode
from .log import enable_default_logging, get_logger
from .ode import make_quaternion
from .world import Viewer, World
import plac
def call(main):
'''Enable logging and start up a main method.'''
enable_default_logging()
plac.call(main)
args = plac.annotations
Update package imports for module name change. | # Copyright (c) 2013 Leif Johnson <leif@leifjohnson.net>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
'''Yet another OpenGL-and-physics simulator framework !'''
from . import physics
from .log import enable_default_logging, get_logger
from .world import Viewer, World
import plac
def call(main):
'''Enable logging and start up a main method.'''
enable_default_logging()
plac.call(main)
args = plac.annotations
| <commit_before># Copyright (c) 2013 Leif Johnson <leif@leifjohnson.net>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
'''Yet another OpenGL-and-physics simulator framework !'''
from . import ode
from .log import enable_default_logging, get_logger
from .ode import make_quaternion
from .world import Viewer, World
import plac
def call(main):
'''Enable logging and start up a main method.'''
enable_default_logging()
plac.call(main)
args = plac.annotations
<commit_msg>Update package imports for module name change.<commit_after> | # Copyright (c) 2013 Leif Johnson <leif@leifjohnson.net>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
'''Yet another OpenGL-and-physics simulator framework !'''
from . import physics
from .log import enable_default_logging, get_logger
from .world import Viewer, World
import plac
def call(main):
'''Enable logging and start up a main method.'''
enable_default_logging()
plac.call(main)
args = plac.annotations
| # Copyright (c) 2013 Leif Johnson <leif@leifjohnson.net>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
'''Yet another OpenGL-and-physics simulator framework !'''
from . import ode
from .log import enable_default_logging, get_logger
from .ode import make_quaternion
from .world import Viewer, World
import plac
def call(main):
'''Enable logging and start up a main method.'''
enable_default_logging()
plac.call(main)
args = plac.annotations
Update package imports for module name change.# Copyright (c) 2013 Leif Johnson <leif@leifjohnson.net>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
'''Yet another OpenGL-and-physics simulator framework !'''
from . import physics
from .log import enable_default_logging, get_logger
from .world import Viewer, World
import plac
def call(main):
'''Enable logging and start up a main method.'''
enable_default_logging()
plac.call(main)
args = plac.annotations
| <commit_before># Copyright (c) 2013 Leif Johnson <leif@leifjohnson.net>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
'''Yet another OpenGL-and-physics simulator framework !'''
from . import ode
from .log import enable_default_logging, get_logger
from .ode import make_quaternion
from .world import Viewer, World
import plac
def call(main):
'''Enable logging and start up a main method.'''
enable_default_logging()
plac.call(main)
args = plac.annotations
<commit_msg>Update package imports for module name change.<commit_after># Copyright (c) 2013 Leif Johnson <leif@leifjohnson.net>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
'''Yet another OpenGL-and-physics simulator framework !'''
from . import physics
from .log import enable_default_logging, get_logger
from .world import Viewer, World
import plac
def call(main):
'''Enable logging and start up a main method.'''
enable_default_logging()
plac.call(main)
args = plac.annotations
|
fd907ac07d5d20dcf8964dbef324bfa2da93ed44 | armstrong/core/arm_sections/managers.py | armstrong/core/arm_sections/managers.py | from django.db import models
class SectionSlugManager(models.Manager):
def __init__(self, section_field="primary_section", slug_field="slug",
*args, **kwargs):
super(SectionSlugManager, self).__init__(*args, **kwargs)
self.section_field = section_field
self.slug_field = slug_field
def get_by_slug(self, slug):
if slug[-1] == "/":
slug = slug[0:-1]
if slug[0] == "/":
slug = slug[1:]
section_slug, content_slug = slug.rsplit("/", 1)
section_slug += "/"
kwargs = {
"%s__full_slug" % self.section_field: section_slug,
self.slug_field: content_slug,
}
qs = self.model.objects.filter(**kwargs)
if hasattr(qs, "select_subclasses"):
qs = qs.select_subclasses()
try:
return qs[0]
except IndexError:
raise self.model.DoesNotExist
| from django.db import models
class SectionSlugManager(models.Manager):
def __init__(self, section_field="primary_section", slug_field="slug",
*args, **kwargs):
super(SectionSlugManager, self).__init__(*args, **kwargs)
self.section_field = section_field
self.slug_field = slug_field
def get_by_slug(self, slug):
if slug[-1] == "/":
slug = slug[0:-1]
if slug[0] == "/":
slug = slug[1:]
try:
section_slug, content_slug = slug.rsplit("/", 1)
section_slug += "/"
except ValueError:
raise self.model.DoesNotExist
kwargs = {
"%s__full_slug" % self.section_field: section_slug,
self.slug_field: content_slug,
}
qs = self.model.objects.filter(**kwargs)
if hasattr(qs, "select_subclasses"):
qs = qs.select_subclasses()
try:
return qs[0]
except IndexError:
raise self.model.DoesNotExist
| Handle situation where only a root full_slug is passed. | Handle situation where only a root full_slug is passed.
i.e. "section/" would break the rsplit(). If there is no slug, we can safely raise a DoesNotExist.
| Python | apache-2.0 | armstrong/armstrong.core.arm_sections,texastribune/armstrong.core.tt_sections,armstrong/armstrong.core.arm_sections,texastribune/armstrong.core.tt_sections,texastribune/armstrong.core.tt_sections | from django.db import models
class SectionSlugManager(models.Manager):
def __init__(self, section_field="primary_section", slug_field="slug",
*args, **kwargs):
super(SectionSlugManager, self).__init__(*args, **kwargs)
self.section_field = section_field
self.slug_field = slug_field
def get_by_slug(self, slug):
if slug[-1] == "/":
slug = slug[0:-1]
if slug[0] == "/":
slug = slug[1:]
section_slug, content_slug = slug.rsplit("/", 1)
section_slug += "/"
kwargs = {
"%s__full_slug" % self.section_field: section_slug,
self.slug_field: content_slug,
}
qs = self.model.objects.filter(**kwargs)
if hasattr(qs, "select_subclasses"):
qs = qs.select_subclasses()
try:
return qs[0]
except IndexError:
raise self.model.DoesNotExist
Handle situation where only a root full_slug is passed.
i.e. "section/" would break the rsplit(). If there is no slug, we can safely raise a DoesNotExist. | from django.db import models
class SectionSlugManager(models.Manager):
def __init__(self, section_field="primary_section", slug_field="slug",
*args, **kwargs):
super(SectionSlugManager, self).__init__(*args, **kwargs)
self.section_field = section_field
self.slug_field = slug_field
def get_by_slug(self, slug):
if slug[-1] == "/":
slug = slug[0:-1]
if slug[0] == "/":
slug = slug[1:]
try:
section_slug, content_slug = slug.rsplit("/", 1)
section_slug += "/"
except ValueError:
raise self.model.DoesNotExist
kwargs = {
"%s__full_slug" % self.section_field: section_slug,
self.slug_field: content_slug,
}
qs = self.model.objects.filter(**kwargs)
if hasattr(qs, "select_subclasses"):
qs = qs.select_subclasses()
try:
return qs[0]
except IndexError:
raise self.model.DoesNotExist
| <commit_before>from django.db import models
class SectionSlugManager(models.Manager):
def __init__(self, section_field="primary_section", slug_field="slug",
*args, **kwargs):
super(SectionSlugManager, self).__init__(*args, **kwargs)
self.section_field = section_field
self.slug_field = slug_field
def get_by_slug(self, slug):
if slug[-1] == "/":
slug = slug[0:-1]
if slug[0] == "/":
slug = slug[1:]
section_slug, content_slug = slug.rsplit("/", 1)
section_slug += "/"
kwargs = {
"%s__full_slug" % self.section_field: section_slug,
self.slug_field: content_slug,
}
qs = self.model.objects.filter(**kwargs)
if hasattr(qs, "select_subclasses"):
qs = qs.select_subclasses()
try:
return qs[0]
except IndexError:
raise self.model.DoesNotExist
<commit_msg>Handle situation where only a root full_slug is passed.
i.e. "section/" would break the rsplit(). If there is no slug, we can safely raise a DoesNotExist.<commit_after> | from django.db import models
class SectionSlugManager(models.Manager):
def __init__(self, section_field="primary_section", slug_field="slug",
*args, **kwargs):
super(SectionSlugManager, self).__init__(*args, **kwargs)
self.section_field = section_field
self.slug_field = slug_field
def get_by_slug(self, slug):
if slug[-1] == "/":
slug = slug[0:-1]
if slug[0] == "/":
slug = slug[1:]
try:
section_slug, content_slug = slug.rsplit("/", 1)
section_slug += "/"
except ValueError:
raise self.model.DoesNotExist
kwargs = {
"%s__full_slug" % self.section_field: section_slug,
self.slug_field: content_slug,
}
qs = self.model.objects.filter(**kwargs)
if hasattr(qs, "select_subclasses"):
qs = qs.select_subclasses()
try:
return qs[0]
except IndexError:
raise self.model.DoesNotExist
| from django.db import models
class SectionSlugManager(models.Manager):
def __init__(self, section_field="primary_section", slug_field="slug",
*args, **kwargs):
super(SectionSlugManager, self).__init__(*args, **kwargs)
self.section_field = section_field
self.slug_field = slug_field
def get_by_slug(self, slug):
if slug[-1] == "/":
slug = slug[0:-1]
if slug[0] == "/":
slug = slug[1:]
section_slug, content_slug = slug.rsplit("/", 1)
section_slug += "/"
kwargs = {
"%s__full_slug" % self.section_field: section_slug,
self.slug_field: content_slug,
}
qs = self.model.objects.filter(**kwargs)
if hasattr(qs, "select_subclasses"):
qs = qs.select_subclasses()
try:
return qs[0]
except IndexError:
raise self.model.DoesNotExist
Handle situation where only a root full_slug is passed.
i.e. "section/" would break the rsplit(). If there is no slug, we can safely raise a DoesNotExist.from django.db import models
class SectionSlugManager(models.Manager):
def __init__(self, section_field="primary_section", slug_field="slug",
*args, **kwargs):
super(SectionSlugManager, self).__init__(*args, **kwargs)
self.section_field = section_field
self.slug_field = slug_field
def get_by_slug(self, slug):
if slug[-1] == "/":
slug = slug[0:-1]
if slug[0] == "/":
slug = slug[1:]
try:
section_slug, content_slug = slug.rsplit("/", 1)
section_slug += "/"
except ValueError:
raise self.model.DoesNotExist
kwargs = {
"%s__full_slug" % self.section_field: section_slug,
self.slug_field: content_slug,
}
qs = self.model.objects.filter(**kwargs)
if hasattr(qs, "select_subclasses"):
qs = qs.select_subclasses()
try:
return qs[0]
except IndexError:
raise self.model.DoesNotExist
| <commit_before>from django.db import models
class SectionSlugManager(models.Manager):
def __init__(self, section_field="primary_section", slug_field="slug",
*args, **kwargs):
super(SectionSlugManager, self).__init__(*args, **kwargs)
self.section_field = section_field
self.slug_field = slug_field
def get_by_slug(self, slug):
if slug[-1] == "/":
slug = slug[0:-1]
if slug[0] == "/":
slug = slug[1:]
section_slug, content_slug = slug.rsplit("/", 1)
section_slug += "/"
kwargs = {
"%s__full_slug" % self.section_field: section_slug,
self.slug_field: content_slug,
}
qs = self.model.objects.filter(**kwargs)
if hasattr(qs, "select_subclasses"):
qs = qs.select_subclasses()
try:
return qs[0]
except IndexError:
raise self.model.DoesNotExist
<commit_msg>Handle situation where only a root full_slug is passed.
i.e. "section/" would break the rsplit(). If there is no slug, we can safely raise a DoesNotExist.<commit_after>from django.db import models
class SectionSlugManager(models.Manager):
def __init__(self, section_field="primary_section", slug_field="slug",
*args, **kwargs):
super(SectionSlugManager, self).__init__(*args, **kwargs)
self.section_field = section_field
self.slug_field = slug_field
def get_by_slug(self, slug):
if slug[-1] == "/":
slug = slug[0:-1]
if slug[0] == "/":
slug = slug[1:]
try:
section_slug, content_slug = slug.rsplit("/", 1)
section_slug += "/"
except ValueError:
raise self.model.DoesNotExist
kwargs = {
"%s__full_slug" % self.section_field: section_slug,
self.slug_field: content_slug,
}
qs = self.model.objects.filter(**kwargs)
if hasattr(qs, "select_subclasses"):
qs = qs.select_subclasses()
try:
return qs[0]
except IndexError:
raise self.model.DoesNotExist
|
03340917e96b7076ca420bea4e121f89c05935f6 | censusreporter/config/prod/settings.py | censusreporter/config/prod/settings.py | from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = [
'censusreporter.org',
'www.censusreporter.org',
'censusreporter.dokku.censusreporter.org',
]
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
| from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = ['*']
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
| Allow all hosts to support Dokku's healthcheck | Allow all hosts to support Dokku's healthcheck
| Python | mit | censusreporter/censusreporter,censusreporter/censusreporter,censusreporter/censusreporter,censusreporter/censusreporter | from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = [
'censusreporter.org',
'www.censusreporter.org',
'censusreporter.dokku.censusreporter.org',
]
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
Allow all hosts to support Dokku's healthcheck | from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = ['*']
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
| <commit_before>from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = [
'censusreporter.org',
'www.censusreporter.org',
'censusreporter.dokku.censusreporter.org',
]
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
<commit_msg>Allow all hosts to support Dokku's healthcheck<commit_after> | from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = ['*']
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
| from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = [
'censusreporter.org',
'www.censusreporter.org',
'censusreporter.dokku.censusreporter.org',
]
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
Allow all hosts to support Dokku's healthcheckfrom censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = ['*']
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
| <commit_before>from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = [
'censusreporter.org',
'www.censusreporter.org',
'censusreporter.dokku.censusreporter.org',
]
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
<commit_msg>Allow all hosts to support Dokku's healthcheck<commit_after>from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = ['*']
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
|
20b4c81137d4abdd4db0dc80ae9a2e38cca4e8eb | examples/hello_twisted.py | examples/hello_twisted.py | """A simple example of Pyglet/Twisted integration. A Pyglet window
is displayed, and both Pyglet and Twisted are making scheduled calls
and regular intervals. Interacting with the window doesn't interfere
with either calls.
"""
import pyglet
import pygletreactor
pygletreactor.install() # <- this must come before...
from twisted.internet import reactor, task # <- ...importing this reactor!
# Create a Pyglet window with a simple message
window = pyglet.window.Window()
label = pyglet.text.Label('hello world',
x = window.width / 2,
y = window.height / 2,
anchor_x = 'center',
anchor_y = 'center')
@window.event
def on_draw():
window.clear()
label.draw()
@window.event
def on_close():
reactor.stop()
# Return true to ensure that no other handlers
# on the stack receive the on_close event
return True
# Schedule a function call in Pyglet
def runEverySecondPyglet(dt):
print "pyglet call: one second has passed"
pyglet.clock.schedule_interval(runEverySecondPyglet, 1)
# Schedule a function call in Twisted
def runEverySecondTwisted():
print "twisted call: 1.5 seconds have passed"
l = task.LoopingCall(runEverySecondTwisted)
l.start(1.5)
# Start the reactor
reactor.run()
| """A simple example of Pyglet/Twisted integration. A Pyglet window
is displayed, and both Pyglet and Twisted are making scheduled calls
and regular intervals. Interacting with the window doesn't interfere
with either calls.
"""
import pyglet
import pygletreactor
pygletreactor.install() # <- this must come before...
from twisted.internet import reactor, task # <- ...importing this reactor!
# Create a Pyglet window with a simple message
window = pyglet.window.Window()
label = pyglet.text.Label('hello world',
x = window.width / 2,
y = window.height / 2,
anchor_x = 'center',
anchor_y = 'center')
@window.event
def on_draw():
window.clear()
label.draw()
@window.event
def on_close():
reactor.callFromThread(reactor.stop)
# Return true to ensure that no other handlers
# on the stack receive the on_close event
return True
# Schedule a function call in Pyglet
def runEverySecondPyglet(dt):
print "pyglet call: one second has passed"
pyglet.clock.schedule_interval(runEverySecondPyglet, 1)
# Schedule a function call in Twisted
def runEverySecondTwisted():
print "twisted call: 1.5 seconds have passed"
l = task.LoopingCall(runEverySecondTwisted)
l.start(1.5)
# Start the reactor
reactor.run()
| Call to stop the reactor now uses the correct convention when closing from a thread other than the main reactor thread. | Call to stop the reactor now uses the correct convention when closing from a thread other than the main reactor thread.
Fixes Issue 5.
git-svn-id: a0251d2471cc357dbf602d275638891bc89eba80@9 4515f058-c067-11dd-9cb5-179210ed59e1
| Python | mit | padraigkitterick/pyglet-twisted | """A simple example of Pyglet/Twisted integration. A Pyglet window
is displayed, and both Pyglet and Twisted are making scheduled calls
and regular intervals. Interacting with the window doesn't interfere
with either calls.
"""
import pyglet
import pygletreactor
pygletreactor.install() # <- this must come before...
from twisted.internet import reactor, task # <- ...importing this reactor!
# Create a Pyglet window with a simple message
window = pyglet.window.Window()
label = pyglet.text.Label('hello world',
x = window.width / 2,
y = window.height / 2,
anchor_x = 'center',
anchor_y = 'center')
@window.event
def on_draw():
window.clear()
label.draw()
@window.event
def on_close():
reactor.stop()
# Return true to ensure that no other handlers
# on the stack receive the on_close event
return True
# Schedule a function call in Pyglet
def runEverySecondPyglet(dt):
print "pyglet call: one second has passed"
pyglet.clock.schedule_interval(runEverySecondPyglet, 1)
# Schedule a function call in Twisted
def runEverySecondTwisted():
print "twisted call: 1.5 seconds have passed"
l = task.LoopingCall(runEverySecondTwisted)
l.start(1.5)
# Start the reactor
reactor.run()
Call to stop the reactor now uses the correct convention when closing from a thread other than the main reactor thread.
Fixes Issue 5.
git-svn-id: a0251d2471cc357dbf602d275638891bc89eba80@9 4515f058-c067-11dd-9cb5-179210ed59e1 | """A simple example of Pyglet/Twisted integration. A Pyglet window
is displayed, and both Pyglet and Twisted are making scheduled calls
and regular intervals. Interacting with the window doesn't interfere
with either calls.
"""
import pyglet
import pygletreactor
pygletreactor.install() # <- this must come before...
from twisted.internet import reactor, task # <- ...importing this reactor!
# Create a Pyglet window with a simple message
window = pyglet.window.Window()
label = pyglet.text.Label('hello world',
x = window.width / 2,
y = window.height / 2,
anchor_x = 'center',
anchor_y = 'center')
@window.event
def on_draw():
window.clear()
label.draw()
@window.event
def on_close():
reactor.callFromThread(reactor.stop)
# Return true to ensure that no other handlers
# on the stack receive the on_close event
return True
# Schedule a function call in Pyglet
def runEverySecondPyglet(dt):
print "pyglet call: one second has passed"
pyglet.clock.schedule_interval(runEverySecondPyglet, 1)
# Schedule a function call in Twisted
def runEverySecondTwisted():
print "twisted call: 1.5 seconds have passed"
l = task.LoopingCall(runEverySecondTwisted)
l.start(1.5)
# Start the reactor
reactor.run()
| <commit_before>"""A simple example of Pyglet/Twisted integration. A Pyglet window
is displayed, and both Pyglet and Twisted are making scheduled calls
and regular intervals. Interacting with the window doesn't interfere
with either calls.
"""
import pyglet
import pygletreactor
pygletreactor.install() # <- this must come before...
from twisted.internet import reactor, task # <- ...importing this reactor!
# Create a Pyglet window with a simple message
window = pyglet.window.Window()
label = pyglet.text.Label('hello world',
x = window.width / 2,
y = window.height / 2,
anchor_x = 'center',
anchor_y = 'center')
@window.event
def on_draw():
window.clear()
label.draw()
@window.event
def on_close():
reactor.stop()
# Return true to ensure that no other handlers
# on the stack receive the on_close event
return True
# Schedule a function call in Pyglet
def runEverySecondPyglet(dt):
print "pyglet call: one second has passed"
pyglet.clock.schedule_interval(runEverySecondPyglet, 1)
# Schedule a function call in Twisted
def runEverySecondTwisted():
print "twisted call: 1.5 seconds have passed"
l = task.LoopingCall(runEverySecondTwisted)
l.start(1.5)
# Start the reactor
reactor.run()
<commit_msg>Call to stop the reactor now uses the correct convention when closing from a thread other than the main reactor thread.
Fixes Issue 5.
git-svn-id: a0251d2471cc357dbf602d275638891bc89eba80@9 4515f058-c067-11dd-9cb5-179210ed59e1<commit_after> | """A simple example of Pyglet/Twisted integration. A Pyglet window
is displayed, and both Pyglet and Twisted are making scheduled calls
and regular intervals. Interacting with the window doesn't interfere
with either calls.
"""
import pyglet
import pygletreactor
pygletreactor.install() # <- this must come before...
from twisted.internet import reactor, task # <- ...importing this reactor!
# Create a Pyglet window with a simple message
window = pyglet.window.Window()
label = pyglet.text.Label('hello world',
x = window.width / 2,
y = window.height / 2,
anchor_x = 'center',
anchor_y = 'center')
@window.event
def on_draw():
window.clear()
label.draw()
@window.event
def on_close():
reactor.callFromThread(reactor.stop)
# Return true to ensure that no other handlers
# on the stack receive the on_close event
return True
# Schedule a function call in Pyglet
def runEverySecondPyglet(dt):
print "pyglet call: one second has passed"
pyglet.clock.schedule_interval(runEverySecondPyglet, 1)
# Schedule a function call in Twisted
def runEverySecondTwisted():
print "twisted call: 1.5 seconds have passed"
l = task.LoopingCall(runEverySecondTwisted)
l.start(1.5)
# Start the reactor
reactor.run()
| """A simple example of Pyglet/Twisted integration. A Pyglet window
is displayed, and both Pyglet and Twisted are making scheduled calls
and regular intervals. Interacting with the window doesn't interfere
with either calls.
"""
import pyglet
import pygletreactor
pygletreactor.install() # <- this must come before...
from twisted.internet import reactor, task # <- ...importing this reactor!
# Create a Pyglet window with a simple message
window = pyglet.window.Window()
label = pyglet.text.Label('hello world',
x = window.width / 2,
y = window.height / 2,
anchor_x = 'center',
anchor_y = 'center')
@window.event
def on_draw():
window.clear()
label.draw()
@window.event
def on_close():
reactor.stop()
# Return true to ensure that no other handlers
# on the stack receive the on_close event
return True
# Schedule a function call in Pyglet
def runEverySecondPyglet(dt):
print "pyglet call: one second has passed"
pyglet.clock.schedule_interval(runEverySecondPyglet, 1)
# Schedule a function call in Twisted
def runEverySecondTwisted():
print "twisted call: 1.5 seconds have passed"
l = task.LoopingCall(runEverySecondTwisted)
l.start(1.5)
# Start the reactor
reactor.run()
Call to stop the reactor now uses the correct convention when closing from a thread other than the main reactor thread.
Fixes Issue 5.
git-svn-id: a0251d2471cc357dbf602d275638891bc89eba80@9 4515f058-c067-11dd-9cb5-179210ed59e1"""A simple example of Pyglet/Twisted integration. A Pyglet window
is displayed, and both Pyglet and Twisted are making scheduled calls
and regular intervals. Interacting with the window doesn't interfere
with either calls.
"""
import pyglet
import pygletreactor
pygletreactor.install() # <- this must come before...
from twisted.internet import reactor, task # <- ...importing this reactor!
# Create a Pyglet window with a simple message
window = pyglet.window.Window()
label = pyglet.text.Label('hello world',
x = window.width / 2,
y = window.height / 2,
anchor_x = 'center',
anchor_y = 'center')
@window.event
def on_draw():
window.clear()
label.draw()
@window.event
def on_close():
reactor.callFromThread(reactor.stop)
# Return true to ensure that no other handlers
# on the stack receive the on_close event
return True
# Schedule a function call in Pyglet
def runEverySecondPyglet(dt):
print "pyglet call: one second has passed"
pyglet.clock.schedule_interval(runEverySecondPyglet, 1)
# Schedule a function call in Twisted
def runEverySecondTwisted():
print "twisted call: 1.5 seconds have passed"
l = task.LoopingCall(runEverySecondTwisted)
l.start(1.5)
# Start the reactor
reactor.run()
| <commit_before>"""A simple example of Pyglet/Twisted integration. A Pyglet window
is displayed, and both Pyglet and Twisted are making scheduled calls
and regular intervals. Interacting with the window doesn't interfere
with either calls.
"""
import pyglet
import pygletreactor
pygletreactor.install() # <- this must come before...
from twisted.internet import reactor, task # <- ...importing this reactor!
# Create a Pyglet window with a simple message
window = pyglet.window.Window()
label = pyglet.text.Label('hello world',
x = window.width / 2,
y = window.height / 2,
anchor_x = 'center',
anchor_y = 'center')
@window.event
def on_draw():
window.clear()
label.draw()
@window.event
def on_close():
reactor.stop()
# Return true to ensure that no other handlers
# on the stack receive the on_close event
return True
# Schedule a function call in Pyglet
def runEverySecondPyglet(dt):
print "pyglet call: one second has passed"
pyglet.clock.schedule_interval(runEverySecondPyglet, 1)
# Schedule a function call in Twisted
def runEverySecondTwisted():
print "twisted call: 1.5 seconds have passed"
l = task.LoopingCall(runEverySecondTwisted)
l.start(1.5)
# Start the reactor
reactor.run()
<commit_msg>Call to stop the reactor now uses the correct convention when closing from a thread other than the main reactor thread.
Fixes Issue 5.
git-svn-id: a0251d2471cc357dbf602d275638891bc89eba80@9 4515f058-c067-11dd-9cb5-179210ed59e1<commit_after>"""A simple example of Pyglet/Twisted integration. A Pyglet window
is displayed, and both Pyglet and Twisted are making scheduled calls
and regular intervals. Interacting with the window doesn't interfere
with either calls.
"""
import pyglet
import pygletreactor
pygletreactor.install() # <- this must come before...
from twisted.internet import reactor, task # <- ...importing this reactor!
# Create a Pyglet window with a simple message
window = pyglet.window.Window()
label = pyglet.text.Label('hello world',
x = window.width / 2,
y = window.height / 2,
anchor_x = 'center',
anchor_y = 'center')
@window.event
def on_draw():
window.clear()
label.draw()
@window.event
def on_close():
reactor.callFromThread(reactor.stop)
# Return true to ensure that no other handlers
# on the stack receive the on_close event
return True
# Schedule a function call in Pyglet
def runEverySecondPyglet(dt):
print "pyglet call: one second has passed"
pyglet.clock.schedule_interval(runEverySecondPyglet, 1)
# Schedule a function call in Twisted
def runEverySecondTwisted():
print "twisted call: 1.5 seconds have passed"
l = task.LoopingCall(runEverySecondTwisted)
l.start(1.5)
# Start the reactor
reactor.run()
|
7778b98e1a0d0ac7b9c14e4536e62de4db7debc9 | tests/integration/suite/test_global_role_bindings.py | tests/integration/suite/test_global_role_bindings.py | from .common import random_str
def test_cannot_update_global_role(admin_mc, remove_resource):
"""Asserts that globalRoleId field cannot be changed"""
admin_client = admin_mc.client
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
userId=admin_mc.user.id,
globalRoleId="nodedrivers-manage")
remove_resource(grb)
grb = admin_client.update_by_id_global_role_binding(
id=grb.id,
globalRoleId="settings-manage")
assert grb.globalRoleId == "nodedrivers-manage"
| import pytest
from rancher import ApiError
from .common import random_str
def test_cannot_update_global_role(admin_mc, remove_resource):
"""Asserts that globalRoleId field cannot be changed"""
admin_client = admin_mc.client
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
userId=admin_mc.user.id,
globalRoleId="nodedrivers-manage")
remove_resource(grb)
grb = admin_client.update_by_id_global_role_binding(
id=grb.id,
globalRoleId="settings-manage")
assert grb.globalRoleId == "nodedrivers-manage"
def test_globalrole_must_exist(admin_mc, remove_resource):
"""Asserts that globalRoleId must reference an existing role"""
admin_client = admin_mc.client
with pytest.raises(ApiError) as e:
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
globalRoleId="somefakerole",
userId=admin_mc.user.id
)
remove_resource(grb)
assert e.value.error.status == 404
assert "globalRole.management.cattle.io \"somefakerole\" not found" in \
e.value.error.message
| Add test for GRB validator | Add test for GRB validator
| Python | apache-2.0 | cjellick/rancher,rancherio/rancher,cjellick/rancher,rancher/rancher,cjellick/rancher,rancherio/rancher,rancher/rancher,rancher/rancher,rancher/rancher | from .common import random_str
def test_cannot_update_global_role(admin_mc, remove_resource):
"""Asserts that globalRoleId field cannot be changed"""
admin_client = admin_mc.client
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
userId=admin_mc.user.id,
globalRoleId="nodedrivers-manage")
remove_resource(grb)
grb = admin_client.update_by_id_global_role_binding(
id=grb.id,
globalRoleId="settings-manage")
assert grb.globalRoleId == "nodedrivers-manage"
Add test for GRB validator | import pytest
from rancher import ApiError
from .common import random_str
def test_cannot_update_global_role(admin_mc, remove_resource):
"""Asserts that globalRoleId field cannot be changed"""
admin_client = admin_mc.client
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
userId=admin_mc.user.id,
globalRoleId="nodedrivers-manage")
remove_resource(grb)
grb = admin_client.update_by_id_global_role_binding(
id=grb.id,
globalRoleId="settings-manage")
assert grb.globalRoleId == "nodedrivers-manage"
def test_globalrole_must_exist(admin_mc, remove_resource):
"""Asserts that globalRoleId must reference an existing role"""
admin_client = admin_mc.client
with pytest.raises(ApiError) as e:
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
globalRoleId="somefakerole",
userId=admin_mc.user.id
)
remove_resource(grb)
assert e.value.error.status == 404
assert "globalRole.management.cattle.io \"somefakerole\" not found" in \
e.value.error.message
| <commit_before>from .common import random_str
def test_cannot_update_global_role(admin_mc, remove_resource):
"""Asserts that globalRoleId field cannot be changed"""
admin_client = admin_mc.client
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
userId=admin_mc.user.id,
globalRoleId="nodedrivers-manage")
remove_resource(grb)
grb = admin_client.update_by_id_global_role_binding(
id=grb.id,
globalRoleId="settings-manage")
assert grb.globalRoleId == "nodedrivers-manage"
<commit_msg>Add test for GRB validator<commit_after> | import pytest
from rancher import ApiError
from .common import random_str
def test_cannot_update_global_role(admin_mc, remove_resource):
"""Asserts that globalRoleId field cannot be changed"""
admin_client = admin_mc.client
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
userId=admin_mc.user.id,
globalRoleId="nodedrivers-manage")
remove_resource(grb)
grb = admin_client.update_by_id_global_role_binding(
id=grb.id,
globalRoleId="settings-manage")
assert grb.globalRoleId == "nodedrivers-manage"
def test_globalrole_must_exist(admin_mc, remove_resource):
"""Asserts that globalRoleId must reference an existing role"""
admin_client = admin_mc.client
with pytest.raises(ApiError) as e:
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
globalRoleId="somefakerole",
userId=admin_mc.user.id
)
remove_resource(grb)
assert e.value.error.status == 404
assert "globalRole.management.cattle.io \"somefakerole\" not found" in \
e.value.error.message
| from .common import random_str
def test_cannot_update_global_role(admin_mc, remove_resource):
"""Asserts that globalRoleId field cannot be changed"""
admin_client = admin_mc.client
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
userId=admin_mc.user.id,
globalRoleId="nodedrivers-manage")
remove_resource(grb)
grb = admin_client.update_by_id_global_role_binding(
id=grb.id,
globalRoleId="settings-manage")
assert grb.globalRoleId == "nodedrivers-manage"
Add test for GRB validatorimport pytest
from rancher import ApiError
from .common import random_str
def test_cannot_update_global_role(admin_mc, remove_resource):
"""Asserts that globalRoleId field cannot be changed"""
admin_client = admin_mc.client
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
userId=admin_mc.user.id,
globalRoleId="nodedrivers-manage")
remove_resource(grb)
grb = admin_client.update_by_id_global_role_binding(
id=grb.id,
globalRoleId="settings-manage")
assert grb.globalRoleId == "nodedrivers-manage"
def test_globalrole_must_exist(admin_mc, remove_resource):
"""Asserts that globalRoleId must reference an existing role"""
admin_client = admin_mc.client
with pytest.raises(ApiError) as e:
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
globalRoleId="somefakerole",
userId=admin_mc.user.id
)
remove_resource(grb)
assert e.value.error.status == 404
assert "globalRole.management.cattle.io \"somefakerole\" not found" in \
e.value.error.message
| <commit_before>from .common import random_str
def test_cannot_update_global_role(admin_mc, remove_resource):
"""Asserts that globalRoleId field cannot be changed"""
admin_client = admin_mc.client
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
userId=admin_mc.user.id,
globalRoleId="nodedrivers-manage")
remove_resource(grb)
grb = admin_client.update_by_id_global_role_binding(
id=grb.id,
globalRoleId="settings-manage")
assert grb.globalRoleId == "nodedrivers-manage"
<commit_msg>Add test for GRB validator<commit_after>import pytest
from rancher import ApiError
from .common import random_str
def test_cannot_update_global_role(admin_mc, remove_resource):
"""Asserts that globalRoleId field cannot be changed"""
admin_client = admin_mc.client
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
userId=admin_mc.user.id,
globalRoleId="nodedrivers-manage")
remove_resource(grb)
grb = admin_client.update_by_id_global_role_binding(
id=grb.id,
globalRoleId="settings-manage")
assert grb.globalRoleId == "nodedrivers-manage"
def test_globalrole_must_exist(admin_mc, remove_resource):
"""Asserts that globalRoleId must reference an existing role"""
admin_client = admin_mc.client
with pytest.raises(ApiError) as e:
grb = admin_client.create_global_role_binding(
name="gr-" + random_str(),
globalRoleId="somefakerole",
userId=admin_mc.user.id
)
remove_resource(grb)
assert e.value.error.status == 404
assert "globalRole.management.cattle.io \"somefakerole\" not found" in \
e.value.error.message
|
619253a51d0b79f170065e6023530937d7111102 | awscfncli/config/config.py | awscfncli/config/config.py | # -*- encoding: utf-8 -*-
import logging
import yaml
from collections import namedtuple
log = logging.getLogger(__name__)
def load(filename):
with open(filename) as fp:
config = yaml.safe_load(fp)
return CfnCliConfig.load(config)
class CfnCliConfig(namedtuple('CfnCliConfig', 'version blueprints environments')):
CFNFILE_V1 = 1
CFNFILE_V2 = 2
VERSION_SECTION = 'Version'
BLUEPRINT_SECTION = 'Blueprints'
ENVIRONMENT_SECTION = 'Environments'
@classmethod
def load(cls, config):
# load version
version = config.get(cls.VERSION_SECTION, cls.CFNFILE_V1)
# load blueprint into dict
blueprint_section = config.get(cls.BLUEPRINT_SECTION, {})
blueprints = {}
for key, val in blueprint_section:
blueprints[key] = Blueprint.load(val)
# load environment into dict
environment_section = config.get(cls.ENVIRONMENT_SECTION, {})
environments = {}
for key, val in environment_section:
environments[key] = Environment.load(val)
return cls(version, blueprints, environments)
class Stack(namedtuple('Stack', '')):
@classmethod
def load(cls, config):
return cls()
class Environment(namedtuple('Environment', '')):
@classmethod
def load(cls, config):
return cls()
class Blueprint(namedtuple('Blueprint', '')):
@classmethod
def load(cls, config):
return cls()
| # -*- encoding: utf-8 -*-
import logging
import yaml
from collections import namedtuple
log = logging.getLogger(__name__)
def load(filename):
with open(filename) as fp:
config = yaml.safe_load(fp)
return CfnCliConfig.load(config)
class CfnCliConfig(namedtuple('CfnCliConfig', 'version blueprints environments')):
CFNFILE_V1 = 1
CFNFILE_V2 = 2
VERSION_SECTION = 'Version'
BLUEPRINT_SECTION = 'Blueprints'
ENVIRONMENT_SECTION = 'Environments'
@staticmethod
def load(config):
# load version
version = config.get(CfnCliConfig.VERSION_SECTION, CfnCliConfig.CFNFILE_V1)
# load blueprint into dict
blueprint_section = config.get(CfnCliConfig.BLUEPRINT_SECTION, {})
blueprints = {}
for key, val in blueprint_section:
blueprints[key] = Blueprint.load(val)
# load environment into dict
environment_section = config.get(CfnCliConfig.ENVIRONMENT_SECTION, {})
environments = {}
for key, val in environment_section:
environments[key] = Environment.load(val)
return CfnCliConfig(version, blueprints, environments)
class Stack(namedtuple('Stack', '')):
@staticmethod
def load(config):
return Stack()
class Environment(namedtuple('Environment', '')):
@staticmethod
def load(config):
return Environment()
class Blueprint(namedtuple('Blueprint', '')):
@staticmethod
def load(config):
return Blueprint()
| Use static method instead of classmethod | Use static method instead of classmethod
| Python | mit | Kotaimen/awscfncli,Kotaimen/awscfncli | # -*- encoding: utf-8 -*-
import logging
import yaml
from collections import namedtuple
log = logging.getLogger(__name__)
def load(filename):
with open(filename) as fp:
config = yaml.safe_load(fp)
return CfnCliConfig.load(config)
class CfnCliConfig(namedtuple('CfnCliConfig', 'version blueprints environments')):
CFNFILE_V1 = 1
CFNFILE_V2 = 2
VERSION_SECTION = 'Version'
BLUEPRINT_SECTION = 'Blueprints'
ENVIRONMENT_SECTION = 'Environments'
@classmethod
def load(cls, config):
# load version
version = config.get(cls.VERSION_SECTION, cls.CFNFILE_V1)
# load blueprint into dict
blueprint_section = config.get(cls.BLUEPRINT_SECTION, {})
blueprints = {}
for key, val in blueprint_section:
blueprints[key] = Blueprint.load(val)
# load environment into dict
environment_section = config.get(cls.ENVIRONMENT_SECTION, {})
environments = {}
for key, val in environment_section:
environments[key] = Environment.load(val)
return cls(version, blueprints, environments)
class Stack(namedtuple('Stack', '')):
@classmethod
def load(cls, config):
return cls()
class Environment(namedtuple('Environment', '')):
@classmethod
def load(cls, config):
return cls()
class Blueprint(namedtuple('Blueprint', '')):
@classmethod
def load(cls, config):
return cls()
Use static method instead of classmethod | # -*- encoding: utf-8 -*-
import logging
import yaml
from collections import namedtuple
log = logging.getLogger(__name__)
def load(filename):
with open(filename) as fp:
config = yaml.safe_load(fp)
return CfnCliConfig.load(config)
class CfnCliConfig(namedtuple('CfnCliConfig', 'version blueprints environments')):
CFNFILE_V1 = 1
CFNFILE_V2 = 2
VERSION_SECTION = 'Version'
BLUEPRINT_SECTION = 'Blueprints'
ENVIRONMENT_SECTION = 'Environments'
@staticmethod
def load(config):
# load version
version = config.get(CfnCliConfig.VERSION_SECTION, CfnCliConfig.CFNFILE_V1)
# load blueprint into dict
blueprint_section = config.get(CfnCliConfig.BLUEPRINT_SECTION, {})
blueprints = {}
for key, val in blueprint_section:
blueprints[key] = Blueprint.load(val)
# load environment into dict
environment_section = config.get(CfnCliConfig.ENVIRONMENT_SECTION, {})
environments = {}
for key, val in environment_section:
environments[key] = Environment.load(val)
return CfnCliConfig(version, blueprints, environments)
class Stack(namedtuple('Stack', '')):
@staticmethod
def load(config):
return Stack()
class Environment(namedtuple('Environment', '')):
@staticmethod
def load(config):
return Environment()
class Blueprint(namedtuple('Blueprint', '')):
@staticmethod
def load(config):
return Blueprint()
| <commit_before># -*- encoding: utf-8 -*-
import logging
import yaml
from collections import namedtuple
log = logging.getLogger(__name__)
def load(filename):
with open(filename) as fp:
config = yaml.safe_load(fp)
return CfnCliConfig.load(config)
class CfnCliConfig(namedtuple('CfnCliConfig', 'version blueprints environments')):
CFNFILE_V1 = 1
CFNFILE_V2 = 2
VERSION_SECTION = 'Version'
BLUEPRINT_SECTION = 'Blueprints'
ENVIRONMENT_SECTION = 'Environments'
@classmethod
def load(cls, config):
# load version
version = config.get(cls.VERSION_SECTION, cls.CFNFILE_V1)
# load blueprint into dict
blueprint_section = config.get(cls.BLUEPRINT_SECTION, {})
blueprints = {}
for key, val in blueprint_section:
blueprints[key] = Blueprint.load(val)
# load environment into dict
environment_section = config.get(cls.ENVIRONMENT_SECTION, {})
environments = {}
for key, val in environment_section:
environments[key] = Environment.load(val)
return cls(version, blueprints, environments)
class Stack(namedtuple('Stack', '')):
@classmethod
def load(cls, config):
return cls()
class Environment(namedtuple('Environment', '')):
@classmethod
def load(cls, config):
return cls()
class Blueprint(namedtuple('Blueprint', '')):
@classmethod
def load(cls, config):
return cls()
<commit_msg>Use static method instead of classmethod<commit_after> | # -*- encoding: utf-8 -*-
import logging
import yaml
from collections import namedtuple
log = logging.getLogger(__name__)
def load(filename):
with open(filename) as fp:
config = yaml.safe_load(fp)
return CfnCliConfig.load(config)
class CfnCliConfig(namedtuple('CfnCliConfig', 'version blueprints environments')):
CFNFILE_V1 = 1
CFNFILE_V2 = 2
VERSION_SECTION = 'Version'
BLUEPRINT_SECTION = 'Blueprints'
ENVIRONMENT_SECTION = 'Environments'
@staticmethod
def load(config):
# load version
version = config.get(CfnCliConfig.VERSION_SECTION, CfnCliConfig.CFNFILE_V1)
# load blueprint into dict
blueprint_section = config.get(CfnCliConfig.BLUEPRINT_SECTION, {})
blueprints = {}
for key, val in blueprint_section:
blueprints[key] = Blueprint.load(val)
# load environment into dict
environment_section = config.get(CfnCliConfig.ENVIRONMENT_SECTION, {})
environments = {}
for key, val in environment_section:
environments[key] = Environment.load(val)
return CfnCliConfig(version, blueprints, environments)
class Stack(namedtuple('Stack', '')):
@staticmethod
def load(config):
return Stack()
class Environment(namedtuple('Environment', '')):
@staticmethod
def load(config):
return Environment()
class Blueprint(namedtuple('Blueprint', '')):
@staticmethod
def load(config):
return Blueprint()
| # -*- encoding: utf-8 -*-
import logging
import yaml
from collections import namedtuple
log = logging.getLogger(__name__)
def load(filename):
with open(filename) as fp:
config = yaml.safe_load(fp)
return CfnCliConfig.load(config)
class CfnCliConfig(namedtuple('CfnCliConfig', 'version blueprints environments')):
CFNFILE_V1 = 1
CFNFILE_V2 = 2
VERSION_SECTION = 'Version'
BLUEPRINT_SECTION = 'Blueprints'
ENVIRONMENT_SECTION = 'Environments'
@classmethod
def load(cls, config):
# load version
version = config.get(cls.VERSION_SECTION, cls.CFNFILE_V1)
# load blueprint into dict
blueprint_section = config.get(cls.BLUEPRINT_SECTION, {})
blueprints = {}
for key, val in blueprint_section:
blueprints[key] = Blueprint.load(val)
# load environment into dict
environment_section = config.get(cls.ENVIRONMENT_SECTION, {})
environments = {}
for key, val in environment_section:
environments[key] = Environment.load(val)
return cls(version, blueprints, environments)
class Stack(namedtuple('Stack', '')):
@classmethod
def load(cls, config):
return cls()
class Environment(namedtuple('Environment', '')):
@classmethod
def load(cls, config):
return cls()
class Blueprint(namedtuple('Blueprint', '')):
@classmethod
def load(cls, config):
return cls()
Use static method instead of classmethod# -*- encoding: utf-8 -*-
import logging
import yaml
from collections import namedtuple
log = logging.getLogger(__name__)
def load(filename):
with open(filename) as fp:
config = yaml.safe_load(fp)
return CfnCliConfig.load(config)
class CfnCliConfig(namedtuple('CfnCliConfig', 'version blueprints environments')):
CFNFILE_V1 = 1
CFNFILE_V2 = 2
VERSION_SECTION = 'Version'
BLUEPRINT_SECTION = 'Blueprints'
ENVIRONMENT_SECTION = 'Environments'
@staticmethod
def load(config):
# load version
version = config.get(CfnCliConfig.VERSION_SECTION, CfnCliConfig.CFNFILE_V1)
# load blueprint into dict
blueprint_section = config.get(CfnCliConfig.BLUEPRINT_SECTION, {})
blueprints = {}
for key, val in blueprint_section:
blueprints[key] = Blueprint.load(val)
# load environment into dict
environment_section = config.get(CfnCliConfig.ENVIRONMENT_SECTION, {})
environments = {}
for key, val in environment_section:
environments[key] = Environment.load(val)
return CfnCliConfig(version, blueprints, environments)
class Stack(namedtuple('Stack', '')):
@staticmethod
def load(config):
return Stack()
class Environment(namedtuple('Environment', '')):
@staticmethod
def load(config):
return Environment()
class Blueprint(namedtuple('Blueprint', '')):
@staticmethod
def load(config):
return Blueprint()
| <commit_before># -*- encoding: utf-8 -*-
import logging
import yaml
from collections import namedtuple
log = logging.getLogger(__name__)
def load(filename):
with open(filename) as fp:
config = yaml.safe_load(fp)
return CfnCliConfig.load(config)
class CfnCliConfig(namedtuple('CfnCliConfig', 'version blueprints environments')):
CFNFILE_V1 = 1
CFNFILE_V2 = 2
VERSION_SECTION = 'Version'
BLUEPRINT_SECTION = 'Blueprints'
ENVIRONMENT_SECTION = 'Environments'
@classmethod
def load(cls, config):
# load version
version = config.get(cls.VERSION_SECTION, cls.CFNFILE_V1)
# load blueprint into dict
blueprint_section = config.get(cls.BLUEPRINT_SECTION, {})
blueprints = {}
for key, val in blueprint_section:
blueprints[key] = Blueprint.load(val)
# load environment into dict
environment_section = config.get(cls.ENVIRONMENT_SECTION, {})
environments = {}
for key, val in environment_section:
environments[key] = Environment.load(val)
return cls(version, blueprints, environments)
class Stack(namedtuple('Stack', '')):
@classmethod
def load(cls, config):
return cls()
class Environment(namedtuple('Environment', '')):
@classmethod
def load(cls, config):
return cls()
class Blueprint(namedtuple('Blueprint', '')):
@classmethod
def load(cls, config):
return cls()
<commit_msg>Use static method instead of classmethod<commit_after># -*- encoding: utf-8 -*-
import logging
import yaml
from collections import namedtuple
log = logging.getLogger(__name__)
def load(filename):
with open(filename) as fp:
config = yaml.safe_load(fp)
return CfnCliConfig.load(config)
class CfnCliConfig(namedtuple('CfnCliConfig', 'version blueprints environments')):
CFNFILE_V1 = 1
CFNFILE_V2 = 2
VERSION_SECTION = 'Version'
BLUEPRINT_SECTION = 'Blueprints'
ENVIRONMENT_SECTION = 'Environments'
@staticmethod
def load(config):
# load version
version = config.get(CfnCliConfig.VERSION_SECTION, CfnCliConfig.CFNFILE_V1)
# load blueprint into dict
blueprint_section = config.get(CfnCliConfig.BLUEPRINT_SECTION, {})
blueprints = {}
for key, val in blueprint_section:
blueprints[key] = Blueprint.load(val)
# load environment into dict
environment_section = config.get(CfnCliConfig.ENVIRONMENT_SECTION, {})
environments = {}
for key, val in environment_section:
environments[key] = Environment.load(val)
return CfnCliConfig(version, blueprints, environments)
class Stack(namedtuple('Stack', '')):
@staticmethod
def load(config):
return Stack()
class Environment(namedtuple('Environment', '')):
@staticmethod
def load(config):
return Environment()
class Blueprint(namedtuple('Blueprint', '')):
@staticmethod
def load(config):
return Blueprint()
|
1285e4bcbdbcf3c28eced497c8585892f3ae1239 | django_summernote/admin.py | django_summernote/admin.py | from django.contrib import admin
from django.db import models
from django_summernote.widgets import SummernoteWidget, SummernoteInplaceWidget
from django_summernote.models import Attachment
from django_summernote.settings import summernote_config, get_attachment_model
__widget__ = SummernoteWidget if summernote_config['iframe'] \
else SummernoteInplaceWidget
class SummernoteInlineModelAdmin(admin.options.InlineModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class SummernoteModelAdmin(admin.ModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class AttachmentAdmin(admin.ModelAdmin):
list_display = ['name', 'file', 'uploaded']
search_fields = ['name']
ordering = ('-id',)
def save_model(self, request, obj, form, change):
obj.name = obj.file.name if (not obj.name) else obj.name
super(AttachmentAdmin, self).save_model(request, obj, form, change)
admin.site.register(get_attachment_model(), AttachmentAdmin)
| from django.contrib import admin
from django.db import models
from django_summernote.widgets import SummernoteWidget, SummernoteInplaceWidget
from django_summernote.settings import summernote_config, get_attachment_model
__widget__ = SummernoteWidget if summernote_config['iframe'] \
else SummernoteInplaceWidget
class SummernoteInlineModelAdmin(admin.options.InlineModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class SummernoteModelAdmin(admin.ModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class AttachmentAdmin(admin.ModelAdmin):
list_display = ['name', 'file', 'uploaded']
search_fields = ['name']
ordering = ('-id',)
def save_model(self, request, obj, form, change):
obj.name = obj.file.name if (not obj.name) else obj.name
super(AttachmentAdmin, self).save_model(request, obj, form, change)
admin.site.register(get_attachment_model(), AttachmentAdmin)
| Remove a non-used module importing | Remove a non-used module importing
| Python | mit | lqez/django-summernote,summernote/django-summernote,lqez/django-summernote,lqez/django-summernote,summernote/django-summernote,summernote/django-summernote | from django.contrib import admin
from django.db import models
from django_summernote.widgets import SummernoteWidget, SummernoteInplaceWidget
from django_summernote.models import Attachment
from django_summernote.settings import summernote_config, get_attachment_model
__widget__ = SummernoteWidget if summernote_config['iframe'] \
else SummernoteInplaceWidget
class SummernoteInlineModelAdmin(admin.options.InlineModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class SummernoteModelAdmin(admin.ModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class AttachmentAdmin(admin.ModelAdmin):
list_display = ['name', 'file', 'uploaded']
search_fields = ['name']
ordering = ('-id',)
def save_model(self, request, obj, form, change):
obj.name = obj.file.name if (not obj.name) else obj.name
super(AttachmentAdmin, self).save_model(request, obj, form, change)
admin.site.register(get_attachment_model(), AttachmentAdmin)
Remove a non-used module importing | from django.contrib import admin
from django.db import models
from django_summernote.widgets import SummernoteWidget, SummernoteInplaceWidget
from django_summernote.settings import summernote_config, get_attachment_model
__widget__ = SummernoteWidget if summernote_config['iframe'] \
else SummernoteInplaceWidget
class SummernoteInlineModelAdmin(admin.options.InlineModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class SummernoteModelAdmin(admin.ModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class AttachmentAdmin(admin.ModelAdmin):
list_display = ['name', 'file', 'uploaded']
search_fields = ['name']
ordering = ('-id',)
def save_model(self, request, obj, form, change):
obj.name = obj.file.name if (not obj.name) else obj.name
super(AttachmentAdmin, self).save_model(request, obj, form, change)
admin.site.register(get_attachment_model(), AttachmentAdmin)
| <commit_before>from django.contrib import admin
from django.db import models
from django_summernote.widgets import SummernoteWidget, SummernoteInplaceWidget
from django_summernote.models import Attachment
from django_summernote.settings import summernote_config, get_attachment_model
__widget__ = SummernoteWidget if summernote_config['iframe'] \
else SummernoteInplaceWidget
class SummernoteInlineModelAdmin(admin.options.InlineModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class SummernoteModelAdmin(admin.ModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class AttachmentAdmin(admin.ModelAdmin):
list_display = ['name', 'file', 'uploaded']
search_fields = ['name']
ordering = ('-id',)
def save_model(self, request, obj, form, change):
obj.name = obj.file.name if (not obj.name) else obj.name
super(AttachmentAdmin, self).save_model(request, obj, form, change)
admin.site.register(get_attachment_model(), AttachmentAdmin)
<commit_msg>Remove a non-used module importing<commit_after> | from django.contrib import admin
from django.db import models
from django_summernote.widgets import SummernoteWidget, SummernoteInplaceWidget
from django_summernote.settings import summernote_config, get_attachment_model
__widget__ = SummernoteWidget if summernote_config['iframe'] \
else SummernoteInplaceWidget
class SummernoteInlineModelAdmin(admin.options.InlineModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class SummernoteModelAdmin(admin.ModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class AttachmentAdmin(admin.ModelAdmin):
list_display = ['name', 'file', 'uploaded']
search_fields = ['name']
ordering = ('-id',)
def save_model(self, request, obj, form, change):
obj.name = obj.file.name if (not obj.name) else obj.name
super(AttachmentAdmin, self).save_model(request, obj, form, change)
admin.site.register(get_attachment_model(), AttachmentAdmin)
| from django.contrib import admin
from django.db import models
from django_summernote.widgets import SummernoteWidget, SummernoteInplaceWidget
from django_summernote.models import Attachment
from django_summernote.settings import summernote_config, get_attachment_model
__widget__ = SummernoteWidget if summernote_config['iframe'] \
else SummernoteInplaceWidget
class SummernoteInlineModelAdmin(admin.options.InlineModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class SummernoteModelAdmin(admin.ModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class AttachmentAdmin(admin.ModelAdmin):
list_display = ['name', 'file', 'uploaded']
search_fields = ['name']
ordering = ('-id',)
def save_model(self, request, obj, form, change):
obj.name = obj.file.name if (not obj.name) else obj.name
super(AttachmentAdmin, self).save_model(request, obj, form, change)
admin.site.register(get_attachment_model(), AttachmentAdmin)
Remove a non-used module importingfrom django.contrib import admin
from django.db import models
from django_summernote.widgets import SummernoteWidget, SummernoteInplaceWidget
from django_summernote.settings import summernote_config, get_attachment_model
__widget__ = SummernoteWidget if summernote_config['iframe'] \
else SummernoteInplaceWidget
class SummernoteInlineModelAdmin(admin.options.InlineModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class SummernoteModelAdmin(admin.ModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class AttachmentAdmin(admin.ModelAdmin):
list_display = ['name', 'file', 'uploaded']
search_fields = ['name']
ordering = ('-id',)
def save_model(self, request, obj, form, change):
obj.name = obj.file.name if (not obj.name) else obj.name
super(AttachmentAdmin, self).save_model(request, obj, form, change)
admin.site.register(get_attachment_model(), AttachmentAdmin)
| <commit_before>from django.contrib import admin
from django.db import models
from django_summernote.widgets import SummernoteWidget, SummernoteInplaceWidget
from django_summernote.models import Attachment
from django_summernote.settings import summernote_config, get_attachment_model
__widget__ = SummernoteWidget if summernote_config['iframe'] \
else SummernoteInplaceWidget
class SummernoteInlineModelAdmin(admin.options.InlineModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class SummernoteModelAdmin(admin.ModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class AttachmentAdmin(admin.ModelAdmin):
list_display = ['name', 'file', 'uploaded']
search_fields = ['name']
ordering = ('-id',)
def save_model(self, request, obj, form, change):
obj.name = obj.file.name if (not obj.name) else obj.name
super(AttachmentAdmin, self).save_model(request, obj, form, change)
admin.site.register(get_attachment_model(), AttachmentAdmin)
<commit_msg>Remove a non-used module importing<commit_after>from django.contrib import admin
from django.db import models
from django_summernote.widgets import SummernoteWidget, SummernoteInplaceWidget
from django_summernote.settings import summernote_config, get_attachment_model
__widget__ = SummernoteWidget if summernote_config['iframe'] \
else SummernoteInplaceWidget
class SummernoteInlineModelAdmin(admin.options.InlineModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class SummernoteModelAdmin(admin.ModelAdmin):
formfield_overrides = {models.TextField: {'widget': __widget__}}
class AttachmentAdmin(admin.ModelAdmin):
list_display = ['name', 'file', 'uploaded']
search_fields = ['name']
ordering = ('-id',)
def save_model(self, request, obj, form, change):
obj.name = obj.file.name if (not obj.name) else obj.name
super(AttachmentAdmin, self).save_model(request, obj, form, change)
admin.site.register(get_attachment_model(), AttachmentAdmin)
|
7e638636606a4f7f7b5b6a09ec508746c8ca8f32 | Importacions_F1_Q1/Fact_impF1_eliminar_Ja_existeix.py | Importacions_F1_Q1/Fact_impF1_eliminar_Ja_existeix.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from ooop import OOOP
import configdb
O = OOOP(**configdb.ooop)
imp_obj = O.GiscedataFacturacioImportacioLinia
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','like','Ja existeix una factura')])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like','XML erroni')])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"XML no es correspon al tipus F1")])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"Document invàlid")])
total = len(imp_del_ids)
n = 0
for imp_del_id in imp_del_ids:
try:
imp_obj.unlink([imp_del_id])
n +=1
print "%d/%d" % (n,total)
except Exception, e:
print e
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from ooop import OOOP
import configdb
O = OOOP(**configdb.ooop)
imp_obj = O.GiscedataFacturacioImportacioLinia
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','=','Aquest fitxer XML ja s\'ha processat en els següents IDs')])
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','=','Ja existeix una factura amb el mateix origen')])
#imp_del_ids = imp_obj.search([('state','=','erroni'),('info','like','Ja existeix una factura')])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like','XML erroni')])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"XML no es correspon al tipus F1")])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"Document invàlid")])
total = len(imp_del_ids)
n = 0
for imp_del_id in imp_del_ids:
try:
imp_obj.unlink([imp_del_id])
n +=1
print "%d/%d" % (n,total)
except Exception, e:
print e
| Fix Cannot delete invoice(s) that are already opened or paid | Fix Cannot delete invoice(s) that are already opened or paid
| Python | agpl-3.0 | Som-Energia/invoice-janitor | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from ooop import OOOP
import configdb
O = OOOP(**configdb.ooop)
imp_obj = O.GiscedataFacturacioImportacioLinia
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','like','Ja existeix una factura')])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like','XML erroni')])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"XML no es correspon al tipus F1")])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"Document invàlid")])
total = len(imp_del_ids)
n = 0
for imp_del_id in imp_del_ids:
try:
imp_obj.unlink([imp_del_id])
n +=1
print "%d/%d" % (n,total)
except Exception, e:
print e
Fix Cannot delete invoice(s) that are already opened or paid | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from ooop import OOOP
import configdb
O = OOOP(**configdb.ooop)
imp_obj = O.GiscedataFacturacioImportacioLinia
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','=','Aquest fitxer XML ja s\'ha processat en els següents IDs')])
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','=','Ja existeix una factura amb el mateix origen')])
#imp_del_ids = imp_obj.search([('state','=','erroni'),('info','like','Ja existeix una factura')])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like','XML erroni')])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"XML no es correspon al tipus F1")])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"Document invàlid")])
total = len(imp_del_ids)
n = 0
for imp_del_id in imp_del_ids:
try:
imp_obj.unlink([imp_del_id])
n +=1
print "%d/%d" % (n,total)
except Exception, e:
print e
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from ooop import OOOP
import configdb
O = OOOP(**configdb.ooop)
imp_obj = O.GiscedataFacturacioImportacioLinia
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','like','Ja existeix una factura')])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like','XML erroni')])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"XML no es correspon al tipus F1")])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"Document invàlid")])
total = len(imp_del_ids)
n = 0
for imp_del_id in imp_del_ids:
try:
imp_obj.unlink([imp_del_id])
n +=1
print "%d/%d" % (n,total)
except Exception, e:
print e
<commit_msg>Fix Cannot delete invoice(s) that are already opened or paid<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from ooop import OOOP
import configdb
O = OOOP(**configdb.ooop)
imp_obj = O.GiscedataFacturacioImportacioLinia
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','=','Aquest fitxer XML ja s\'ha processat en els següents IDs')])
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','=','Ja existeix una factura amb el mateix origen')])
#imp_del_ids = imp_obj.search([('state','=','erroni'),('info','like','Ja existeix una factura')])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like','XML erroni')])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"XML no es correspon al tipus F1")])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"Document invàlid")])
total = len(imp_del_ids)
n = 0
for imp_del_id in imp_del_ids:
try:
imp_obj.unlink([imp_del_id])
n +=1
print "%d/%d" % (n,total)
except Exception, e:
print e
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from ooop import OOOP
import configdb
O = OOOP(**configdb.ooop)
imp_obj = O.GiscedataFacturacioImportacioLinia
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','like','Ja existeix una factura')])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like','XML erroni')])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"XML no es correspon al tipus F1")])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"Document invàlid")])
total = len(imp_del_ids)
n = 0
for imp_del_id in imp_del_ids:
try:
imp_obj.unlink([imp_del_id])
n +=1
print "%d/%d" % (n,total)
except Exception, e:
print e
Fix Cannot delete invoice(s) that are already opened or paid#!/usr/bin/env python
# -*- coding: utf-8 -*-
from ooop import OOOP
import configdb
O = OOOP(**configdb.ooop)
imp_obj = O.GiscedataFacturacioImportacioLinia
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','=','Aquest fitxer XML ja s\'ha processat en els següents IDs')])
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','=','Ja existeix una factura amb el mateix origen')])
#imp_del_ids = imp_obj.search([('state','=','erroni'),('info','like','Ja existeix una factura')])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like','XML erroni')])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"XML no es correspon al tipus F1")])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"Document invàlid")])
total = len(imp_del_ids)
n = 0
for imp_del_id in imp_del_ids:
try:
imp_obj.unlink([imp_del_id])
n +=1
print "%d/%d" % (n,total)
except Exception, e:
print e
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from ooop import OOOP
import configdb
O = OOOP(**configdb.ooop)
imp_obj = O.GiscedataFacturacioImportacioLinia
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','like','Ja existeix una factura')])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like','XML erroni')])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"XML no es correspon al tipus F1")])
imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"Document invàlid")])
total = len(imp_del_ids)
n = 0
for imp_del_id in imp_del_ids:
try:
imp_obj.unlink([imp_del_id])
n +=1
print "%d/%d" % (n,total)
except Exception, e:
print e
<commit_msg>Fix Cannot delete invoice(s) that are already opened or paid<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from ooop import OOOP
import configdb
O = OOOP(**configdb.ooop)
imp_obj = O.GiscedataFacturacioImportacioLinia
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','=','Aquest fitxer XML ja s\'ha processat en els següents IDs')])
imp_del_ids = imp_obj.search([('state','=','erroni'),('info','=','Ja existeix una factura amb el mateix origen')])
#imp_del_ids = imp_obj.search([('state','=','erroni'),('info','like','Ja existeix una factura')])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like','XML erroni')])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"XML no es correspon al tipus F1")])
#imp_del_ids += imp_obj.search([('state','=','erroni'),('info','like',"Document invàlid")])
total = len(imp_del_ids)
n = 0
for imp_del_id in imp_del_ids:
try:
imp_obj.unlink([imp_del_id])
n +=1
print "%d/%d" % (n,total)
except Exception, e:
print e
|
8cee7d5478cde2b188da4dc93f844073be729a48 | src/gerobak/apps/profile/models.py | src/gerobak/apps/profile/models.py | from django.db import models
from django.contrib.auth.models import User
from django.conf import settings
from django.db.models.signals import post_save
from gerobak import utils
class Profile(models.Model):
pid = models.CharField(max_length=8)
user = models.ForeignKey(User)
added = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
name = models.CharField(max_length=50)
desc = models.CharField(max_length=250, null=True, blank=True)
arch = models.CharField(max_length=10, choices=settings.GEROBAK_ARCHS,
default=settings.GEROBAK_DEFAULT_ARCH)
repo_updated = models.DateTimeField(null=True, default=None)
sources_updated = models.DateTimeField(null=True, default=None)
sources_total = models.IntegerField(null=True, default=None)
status_updated = models.DateTimeField(null=True, default=None)
status_hash = models.CharField(max_length=32, null=True, default=None)
status_size = models.IntegerField(default=0)
def is_ready(self):
return self.repo_updated is not None and \
self.status_updated is not None
def generate_pid(self):
import uuid
return str(uuid.uuid4()).split('-')[0]
| from django.db import models
from django.contrib.auth.models import User
from django.conf import settings
from django.db.models.signals import post_save
from gerobak import utils
class Profile(models.Model):
pid = models.CharField(max_length=8)
user = models.ForeignKey(User)
added = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
name = models.CharField(max_length=50)
desc = models.CharField(max_length=250, null=True, blank=True)
arch = models.CharField(max_length=10, choices=settings.GEROBAK_ARCHS,
default=settings.GEROBAK_DEFAULT_ARCH)
repo_updated = models.DateTimeField(null=True, default=None)
sources_updated = models.DateTimeField(null=True, default=None)
sources_total = models.IntegerField(null=True, default=None)
status_updated = models.DateTimeField(null=True, default=None)
status_hash = models.CharField(max_length=32, null=True, default=None)
status_size = models.IntegerField(default=0)
tid_update = models.CharField(max_length=36, null=True, default=None)
tid_install = models.CharField(max_length=36, null=True, default=None)
tid_upgrade = models.CharField(max_length=36, null=True, default=None)
def is_ready(self):
return self.repo_updated is not None and \
self.status_updated is not None
def generate_pid(self):
import uuid
return str(uuid.uuid4()).split('-')[0]
| Store task_id for update, install, and upgrade processes in the database. | Store task_id for update, install, and upgrade processes in the database.
| Python | agpl-3.0 | fajran/gerobak,fajran/gerobak | from django.db import models
from django.contrib.auth.models import User
from django.conf import settings
from django.db.models.signals import post_save
from gerobak import utils
class Profile(models.Model):
pid = models.CharField(max_length=8)
user = models.ForeignKey(User)
added = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
name = models.CharField(max_length=50)
desc = models.CharField(max_length=250, null=True, blank=True)
arch = models.CharField(max_length=10, choices=settings.GEROBAK_ARCHS,
default=settings.GEROBAK_DEFAULT_ARCH)
repo_updated = models.DateTimeField(null=True, default=None)
sources_updated = models.DateTimeField(null=True, default=None)
sources_total = models.IntegerField(null=True, default=None)
status_updated = models.DateTimeField(null=True, default=None)
status_hash = models.CharField(max_length=32, null=True, default=None)
status_size = models.IntegerField(default=0)
def is_ready(self):
return self.repo_updated is not None and \
self.status_updated is not None
def generate_pid(self):
import uuid
return str(uuid.uuid4()).split('-')[0]
Store task_id for update, install, and upgrade processes in the database. | from django.db import models
from django.contrib.auth.models import User
from django.conf import settings
from django.db.models.signals import post_save
from gerobak import utils
class Profile(models.Model):
pid = models.CharField(max_length=8)
user = models.ForeignKey(User)
added = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
name = models.CharField(max_length=50)
desc = models.CharField(max_length=250, null=True, blank=True)
arch = models.CharField(max_length=10, choices=settings.GEROBAK_ARCHS,
default=settings.GEROBAK_DEFAULT_ARCH)
repo_updated = models.DateTimeField(null=True, default=None)
sources_updated = models.DateTimeField(null=True, default=None)
sources_total = models.IntegerField(null=True, default=None)
status_updated = models.DateTimeField(null=True, default=None)
status_hash = models.CharField(max_length=32, null=True, default=None)
status_size = models.IntegerField(default=0)
tid_update = models.CharField(max_length=36, null=True, default=None)
tid_install = models.CharField(max_length=36, null=True, default=None)
tid_upgrade = models.CharField(max_length=36, null=True, default=None)
def is_ready(self):
return self.repo_updated is not None and \
self.status_updated is not None
def generate_pid(self):
import uuid
return str(uuid.uuid4()).split('-')[0]
| <commit_before>from django.db import models
from django.contrib.auth.models import User
from django.conf import settings
from django.db.models.signals import post_save
from gerobak import utils
class Profile(models.Model):
pid = models.CharField(max_length=8)
user = models.ForeignKey(User)
added = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
name = models.CharField(max_length=50)
desc = models.CharField(max_length=250, null=True, blank=True)
arch = models.CharField(max_length=10, choices=settings.GEROBAK_ARCHS,
default=settings.GEROBAK_DEFAULT_ARCH)
repo_updated = models.DateTimeField(null=True, default=None)
sources_updated = models.DateTimeField(null=True, default=None)
sources_total = models.IntegerField(null=True, default=None)
status_updated = models.DateTimeField(null=True, default=None)
status_hash = models.CharField(max_length=32, null=True, default=None)
status_size = models.IntegerField(default=0)
def is_ready(self):
return self.repo_updated is not None and \
self.status_updated is not None
def generate_pid(self):
import uuid
return str(uuid.uuid4()).split('-')[0]
<commit_msg>Store task_id for update, install, and upgrade processes in the database.<commit_after> | from django.db import models
from django.contrib.auth.models import User
from django.conf import settings
from django.db.models.signals import post_save
from gerobak import utils
class Profile(models.Model):
pid = models.CharField(max_length=8)
user = models.ForeignKey(User)
added = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
name = models.CharField(max_length=50)
desc = models.CharField(max_length=250, null=True, blank=True)
arch = models.CharField(max_length=10, choices=settings.GEROBAK_ARCHS,
default=settings.GEROBAK_DEFAULT_ARCH)
repo_updated = models.DateTimeField(null=True, default=None)
sources_updated = models.DateTimeField(null=True, default=None)
sources_total = models.IntegerField(null=True, default=None)
status_updated = models.DateTimeField(null=True, default=None)
status_hash = models.CharField(max_length=32, null=True, default=None)
status_size = models.IntegerField(default=0)
tid_update = models.CharField(max_length=36, null=True, default=None)
tid_install = models.CharField(max_length=36, null=True, default=None)
tid_upgrade = models.CharField(max_length=36, null=True, default=None)
def is_ready(self):
return self.repo_updated is not None and \
self.status_updated is not None
def generate_pid(self):
import uuid
return str(uuid.uuid4()).split('-')[0]
| from django.db import models
from django.contrib.auth.models import User
from django.conf import settings
from django.db.models.signals import post_save
from gerobak import utils
class Profile(models.Model):
pid = models.CharField(max_length=8)
user = models.ForeignKey(User)
added = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
name = models.CharField(max_length=50)
desc = models.CharField(max_length=250, null=True, blank=True)
arch = models.CharField(max_length=10, choices=settings.GEROBAK_ARCHS,
default=settings.GEROBAK_DEFAULT_ARCH)
repo_updated = models.DateTimeField(null=True, default=None)
sources_updated = models.DateTimeField(null=True, default=None)
sources_total = models.IntegerField(null=True, default=None)
status_updated = models.DateTimeField(null=True, default=None)
status_hash = models.CharField(max_length=32, null=True, default=None)
status_size = models.IntegerField(default=0)
def is_ready(self):
return self.repo_updated is not None and \
self.status_updated is not None
def generate_pid(self):
import uuid
return str(uuid.uuid4()).split('-')[0]
Store task_id for update, install, and upgrade processes in the database.from django.db import models
from django.contrib.auth.models import User
from django.conf import settings
from django.db.models.signals import post_save
from gerobak import utils
class Profile(models.Model):
pid = models.CharField(max_length=8)
user = models.ForeignKey(User)
added = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
name = models.CharField(max_length=50)
desc = models.CharField(max_length=250, null=True, blank=True)
arch = models.CharField(max_length=10, choices=settings.GEROBAK_ARCHS,
default=settings.GEROBAK_DEFAULT_ARCH)
repo_updated = models.DateTimeField(null=True, default=None)
sources_updated = models.DateTimeField(null=True, default=None)
sources_total = models.IntegerField(null=True, default=None)
status_updated = models.DateTimeField(null=True, default=None)
status_hash = models.CharField(max_length=32, null=True, default=None)
status_size = models.IntegerField(default=0)
tid_update = models.CharField(max_length=36, null=True, default=None)
tid_install = models.CharField(max_length=36, null=True, default=None)
tid_upgrade = models.CharField(max_length=36, null=True, default=None)
def is_ready(self):
return self.repo_updated is not None and \
self.status_updated is not None
def generate_pid(self):
import uuid
return str(uuid.uuid4()).split('-')[0]
| <commit_before>from django.db import models
from django.contrib.auth.models import User
from django.conf import settings
from django.db.models.signals import post_save
from gerobak import utils
class Profile(models.Model):
pid = models.CharField(max_length=8)
user = models.ForeignKey(User)
added = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
name = models.CharField(max_length=50)
desc = models.CharField(max_length=250, null=True, blank=True)
arch = models.CharField(max_length=10, choices=settings.GEROBAK_ARCHS,
default=settings.GEROBAK_DEFAULT_ARCH)
repo_updated = models.DateTimeField(null=True, default=None)
sources_updated = models.DateTimeField(null=True, default=None)
sources_total = models.IntegerField(null=True, default=None)
status_updated = models.DateTimeField(null=True, default=None)
status_hash = models.CharField(max_length=32, null=True, default=None)
status_size = models.IntegerField(default=0)
def is_ready(self):
return self.repo_updated is not None and \
self.status_updated is not None
def generate_pid(self):
import uuid
return str(uuid.uuid4()).split('-')[0]
<commit_msg>Store task_id for update, install, and upgrade processes in the database.<commit_after>from django.db import models
from django.contrib.auth.models import User
from django.conf import settings
from django.db.models.signals import post_save
from gerobak import utils
class Profile(models.Model):
pid = models.CharField(max_length=8)
user = models.ForeignKey(User)
added = models.DateTimeField(auto_now_add=True)
updated = models.DateTimeField(auto_now=True)
name = models.CharField(max_length=50)
desc = models.CharField(max_length=250, null=True, blank=True)
arch = models.CharField(max_length=10, choices=settings.GEROBAK_ARCHS,
default=settings.GEROBAK_DEFAULT_ARCH)
repo_updated = models.DateTimeField(null=True, default=None)
sources_updated = models.DateTimeField(null=True, default=None)
sources_total = models.IntegerField(null=True, default=None)
status_updated = models.DateTimeField(null=True, default=None)
status_hash = models.CharField(max_length=32, null=True, default=None)
status_size = models.IntegerField(default=0)
tid_update = models.CharField(max_length=36, null=True, default=None)
tid_install = models.CharField(max_length=36, null=True, default=None)
tid_upgrade = models.CharField(max_length=36, null=True, default=None)
def is_ready(self):
return self.repo_updated is not None and \
self.status_updated is not None
def generate_pid(self):
import uuid
return str(uuid.uuid4()).split('-')[0]
|
6021f5af785e5234b9f83ea4ac740571b9308ae4 | Communication/mavtester.py | Communication/mavtester.py | #!/usr/bin/env python
'''
test mavlink messages
Do not forget to precise the baudrate (default 115200)
'''
import sys, struct, time, os
from curses import ascii
from pymavlink import mavutil
from argparse import ArgumentParser
parser = ArgumentParser(description=__doc__)
parser.add_argument("--baudrate", type=int,
help="master port baud rate", default=115200)
parser.add_argument("--device", required=True, help="serial device")
parser.add_argument("--source-system", dest='SOURCE_SYSTEM', type=int,
default=255, help='MAVLink source system for this GCS')
args = parser.parse_args()
def wait_heartbeat(m):
'''wait for a heartbeat so we know the target system IDs'''
print("Waiting for APM heartbeat")
msg = m.recv_match(type='HEARTBEAT', blocking=True)
print("Heartbeat from APM (system %u component %u)" % (m.target_system, m.target_component))
# create a mavlink serial instance
master = mavutil.mavlink_connection(args.device, baud=args.baudrate, source_system=args.SOURCE_SYSTEM)
# wait for the heartbeat msg to find the system ID
while True:
wait_heartbeat(master)
| #!/usr/bin/env python
'''
test mavlink messages
Do not forget to precise the baudrate (default 115200)
'''
import sys, struct, time, os
from curses import ascii
from pymavlink import mavutil
from argparse import ArgumentParser
parser = ArgumentParser(description=__doc__)
parser.add_argument("--baudrate", type=int,
help="master port baud rate", default=115200)
parser.add_argument("--device", required=True, help="serial device")
parser.add_argument("--source-system", dest='SOURCE_SYSTEM', type=int,
default=255, help='MAVLink source system for this GCS')
args = parser.parse_args()
def wait_heartbeat(m):
'''wait for a heartbeat so we know the target system IDs'''
print("Waiting for APM heartbeat")
msg = m.recv_match(type='HEARTBEAT', blocking=True)
print("Heartbeat from APM (system %u component %u)" % (m.target_system, m.target_component))
# create a mavlink serial instance
master = mavutil.mavlink_connection(args.device, baud=args.baudrate, source_system=args.SOURCE_SYSTEM)
# wait for the heartbeat msg to find the system ID
while True:
wait_heartbeat(master)
msg = master.recv_match(type='GPS_RAW_INT', blocking=False)
print msg
| Add reception of GPS_RAW_INT messages as demo | Add reception of GPS_RAW_INT messages as demo
| Python | mit | baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite,baptistelabat/robokite | #!/usr/bin/env python
'''
test mavlink messages
Do not forget to precise the baudrate (default 115200)
'''
import sys, struct, time, os
from curses import ascii
from pymavlink import mavutil
from argparse import ArgumentParser
parser = ArgumentParser(description=__doc__)
parser.add_argument("--baudrate", type=int,
help="master port baud rate", default=115200)
parser.add_argument("--device", required=True, help="serial device")
parser.add_argument("--source-system", dest='SOURCE_SYSTEM', type=int,
default=255, help='MAVLink source system for this GCS')
args = parser.parse_args()
def wait_heartbeat(m):
'''wait for a heartbeat so we know the target system IDs'''
print("Waiting for APM heartbeat")
msg = m.recv_match(type='HEARTBEAT', blocking=True)
print("Heartbeat from APM (system %u component %u)" % (m.target_system, m.target_component))
# create a mavlink serial instance
master = mavutil.mavlink_connection(args.device, baud=args.baudrate, source_system=args.SOURCE_SYSTEM)
# wait for the heartbeat msg to find the system ID
while True:
wait_heartbeat(master)
Add reception of GPS_RAW_INT messages as demo | #!/usr/bin/env python
'''
test mavlink messages
Do not forget to precise the baudrate (default 115200)
'''
import sys, struct, time, os
from curses import ascii
from pymavlink import mavutil
from argparse import ArgumentParser
parser = ArgumentParser(description=__doc__)
parser.add_argument("--baudrate", type=int,
help="master port baud rate", default=115200)
parser.add_argument("--device", required=True, help="serial device")
parser.add_argument("--source-system", dest='SOURCE_SYSTEM', type=int,
default=255, help='MAVLink source system for this GCS')
args = parser.parse_args()
def wait_heartbeat(m):
'''wait for a heartbeat so we know the target system IDs'''
print("Waiting for APM heartbeat")
msg = m.recv_match(type='HEARTBEAT', blocking=True)
print("Heartbeat from APM (system %u component %u)" % (m.target_system, m.target_component))
# create a mavlink serial instance
master = mavutil.mavlink_connection(args.device, baud=args.baudrate, source_system=args.SOURCE_SYSTEM)
# wait for the heartbeat msg to find the system ID
while True:
wait_heartbeat(master)
msg = master.recv_match(type='GPS_RAW_INT', blocking=False)
print msg
| <commit_before>#!/usr/bin/env python
'''
test mavlink messages
Do not forget to precise the baudrate (default 115200)
'''
import sys, struct, time, os
from curses import ascii
from pymavlink import mavutil
from argparse import ArgumentParser
parser = ArgumentParser(description=__doc__)
parser.add_argument("--baudrate", type=int,
help="master port baud rate", default=115200)
parser.add_argument("--device", required=True, help="serial device")
parser.add_argument("--source-system", dest='SOURCE_SYSTEM', type=int,
default=255, help='MAVLink source system for this GCS')
args = parser.parse_args()
def wait_heartbeat(m):
'''wait for a heartbeat so we know the target system IDs'''
print("Waiting for APM heartbeat")
msg = m.recv_match(type='HEARTBEAT', blocking=True)
print("Heartbeat from APM (system %u component %u)" % (m.target_system, m.target_component))
# create a mavlink serial instance
master = mavutil.mavlink_connection(args.device, baud=args.baudrate, source_system=args.SOURCE_SYSTEM)
# wait for the heartbeat msg to find the system ID
while True:
wait_heartbeat(master)
<commit_msg>Add reception of GPS_RAW_INT messages as demo<commit_after> | #!/usr/bin/env python
'''
test mavlink messages
Do not forget to precise the baudrate (default 115200)
'''
import sys, struct, time, os
from curses import ascii
from pymavlink import mavutil
from argparse import ArgumentParser
parser = ArgumentParser(description=__doc__)
parser.add_argument("--baudrate", type=int,
help="master port baud rate", default=115200)
parser.add_argument("--device", required=True, help="serial device")
parser.add_argument("--source-system", dest='SOURCE_SYSTEM', type=int,
default=255, help='MAVLink source system for this GCS')
args = parser.parse_args()
def wait_heartbeat(m):
'''wait for a heartbeat so we know the target system IDs'''
print("Waiting for APM heartbeat")
msg = m.recv_match(type='HEARTBEAT', blocking=True)
print("Heartbeat from APM (system %u component %u)" % (m.target_system, m.target_component))
# create a mavlink serial instance
master = mavutil.mavlink_connection(args.device, baud=args.baudrate, source_system=args.SOURCE_SYSTEM)
# wait for the heartbeat msg to find the system ID
while True:
wait_heartbeat(master)
msg = master.recv_match(type='GPS_RAW_INT', blocking=False)
print msg
| #!/usr/bin/env python
'''
test mavlink messages
Do not forget to precise the baudrate (default 115200)
'''
import sys, struct, time, os
from curses import ascii
from pymavlink import mavutil
from argparse import ArgumentParser
parser = ArgumentParser(description=__doc__)
parser.add_argument("--baudrate", type=int,
help="master port baud rate", default=115200)
parser.add_argument("--device", required=True, help="serial device")
parser.add_argument("--source-system", dest='SOURCE_SYSTEM', type=int,
default=255, help='MAVLink source system for this GCS')
args = parser.parse_args()
def wait_heartbeat(m):
'''wait for a heartbeat so we know the target system IDs'''
print("Waiting for APM heartbeat")
msg = m.recv_match(type='HEARTBEAT', blocking=True)
print("Heartbeat from APM (system %u component %u)" % (m.target_system, m.target_component))
# create a mavlink serial instance
master = mavutil.mavlink_connection(args.device, baud=args.baudrate, source_system=args.SOURCE_SYSTEM)
# wait for the heartbeat msg to find the system ID
while True:
wait_heartbeat(master)
Add reception of GPS_RAW_INT messages as demo#!/usr/bin/env python
'''
test mavlink messages
Do not forget to precise the baudrate (default 115200)
'''
import sys, struct, time, os
from curses import ascii
from pymavlink import mavutil
from argparse import ArgumentParser
parser = ArgumentParser(description=__doc__)
parser.add_argument("--baudrate", type=int,
help="master port baud rate", default=115200)
parser.add_argument("--device", required=True, help="serial device")
parser.add_argument("--source-system", dest='SOURCE_SYSTEM', type=int,
default=255, help='MAVLink source system for this GCS')
args = parser.parse_args()
def wait_heartbeat(m):
'''wait for a heartbeat so we know the target system IDs'''
print("Waiting for APM heartbeat")
msg = m.recv_match(type='HEARTBEAT', blocking=True)
print("Heartbeat from APM (system %u component %u)" % (m.target_system, m.target_component))
# create a mavlink serial instance
master = mavutil.mavlink_connection(args.device, baud=args.baudrate, source_system=args.SOURCE_SYSTEM)
# wait for the heartbeat msg to find the system ID
while True:
wait_heartbeat(master)
msg = master.recv_match(type='GPS_RAW_INT', blocking=False)
print msg
| <commit_before>#!/usr/bin/env python
'''
test mavlink messages
Do not forget to precise the baudrate (default 115200)
'''
import sys, struct, time, os
from curses import ascii
from pymavlink import mavutil
from argparse import ArgumentParser
parser = ArgumentParser(description=__doc__)
parser.add_argument("--baudrate", type=int,
help="master port baud rate", default=115200)
parser.add_argument("--device", required=True, help="serial device")
parser.add_argument("--source-system", dest='SOURCE_SYSTEM', type=int,
default=255, help='MAVLink source system for this GCS')
args = parser.parse_args()
def wait_heartbeat(m):
'''wait for a heartbeat so we know the target system IDs'''
print("Waiting for APM heartbeat")
msg = m.recv_match(type='HEARTBEAT', blocking=True)
print("Heartbeat from APM (system %u component %u)" % (m.target_system, m.target_component))
# create a mavlink serial instance
master = mavutil.mavlink_connection(args.device, baud=args.baudrate, source_system=args.SOURCE_SYSTEM)
# wait for the heartbeat msg to find the system ID
while True:
wait_heartbeat(master)
<commit_msg>Add reception of GPS_RAW_INT messages as demo<commit_after>#!/usr/bin/env python
'''
test mavlink messages
Do not forget to precise the baudrate (default 115200)
'''
import sys, struct, time, os
from curses import ascii
from pymavlink import mavutil
from argparse import ArgumentParser
parser = ArgumentParser(description=__doc__)
parser.add_argument("--baudrate", type=int,
help="master port baud rate", default=115200)
parser.add_argument("--device", required=True, help="serial device")
parser.add_argument("--source-system", dest='SOURCE_SYSTEM', type=int,
default=255, help='MAVLink source system for this GCS')
args = parser.parse_args()
def wait_heartbeat(m):
'''wait for a heartbeat so we know the target system IDs'''
print("Waiting for APM heartbeat")
msg = m.recv_match(type='HEARTBEAT', blocking=True)
print("Heartbeat from APM (system %u component %u)" % (m.target_system, m.target_component))
# create a mavlink serial instance
master = mavutil.mavlink_connection(args.device, baud=args.baudrate, source_system=args.SOURCE_SYSTEM)
# wait for the heartbeat msg to find the system ID
while True:
wait_heartbeat(master)
msg = master.recv_match(type='GPS_RAW_INT', blocking=False)
print msg
|
3b8811af898ec8cbaa93c69c6b702b92756713dc | vumi/persist/tests/test_riak_manager.py | vumi/persist/tests/test_riak_manager.py | """Tests for vumi.persist.riak_manager."""
from twisted.trial.unittest import TestCase
from vumi.persist.riak_manager import RiakManager
class TestRiakManager(TestCase):
pass
| """Tests for vumi.persist.riak_manager."""
from itertools import count
from twisted.trial.unittest import TestCase
from twisted.internet.defer import returnValue
from vumi.persist.riak_manager import RiakManager, flatten_generator
from vumi.persist.tests.test_txriak_manager import CommonRiakManagerTests
class TestRiakManager(CommonRiakManagerTests, TestCase):
"""Most tests are inherited from the CommonRiakManagerTests mixin."""
def setUp(self):
self.manager = RiakManager.from_config({'bucket_prefix': 'test.'})
self.manager.purge_all()
def tearDown(self):
self.manager.purge_all()
def test_call_decorator(self):
self.assertEqual(RiakManager.call_decorator, flatten_generator)
def test_flatten_generator(self):
results = []
counter = count()
@flatten_generator
def f():
for i in range(3):
a = yield counter.next()
results.append(a)
ret = f()
self.assertEqual(ret, None)
self.assertEqual(results, list(range(3)))
def test_flatter_generator_with_return_value(self):
@flatten_generator
def f():
yield None
returnValue("foo")
ret = f()
self.assertEqual(ret, "foo")
| Add tests for (nottx)riak manager. | Add tests for (nottx)riak manager.
| Python | bsd-3-clause | vishwaprakashmishra/xmatrix,harrissoerja/vumi,TouK/vumi,vishwaprakashmishra/xmatrix,vishwaprakashmishra/xmatrix,TouK/vumi,TouK/vumi,harrissoerja/vumi,harrissoerja/vumi | """Tests for vumi.persist.riak_manager."""
from twisted.trial.unittest import TestCase
from vumi.persist.riak_manager import RiakManager
class TestRiakManager(TestCase):
pass
Add tests for (nottx)riak manager. | """Tests for vumi.persist.riak_manager."""
from itertools import count
from twisted.trial.unittest import TestCase
from twisted.internet.defer import returnValue
from vumi.persist.riak_manager import RiakManager, flatten_generator
from vumi.persist.tests.test_txriak_manager import CommonRiakManagerTests
class TestRiakManager(CommonRiakManagerTests, TestCase):
"""Most tests are inherited from the CommonRiakManagerTests mixin."""
def setUp(self):
self.manager = RiakManager.from_config({'bucket_prefix': 'test.'})
self.manager.purge_all()
def tearDown(self):
self.manager.purge_all()
def test_call_decorator(self):
self.assertEqual(RiakManager.call_decorator, flatten_generator)
def test_flatten_generator(self):
results = []
counter = count()
@flatten_generator
def f():
for i in range(3):
a = yield counter.next()
results.append(a)
ret = f()
self.assertEqual(ret, None)
self.assertEqual(results, list(range(3)))
def test_flatter_generator_with_return_value(self):
@flatten_generator
def f():
yield None
returnValue("foo")
ret = f()
self.assertEqual(ret, "foo")
| <commit_before>"""Tests for vumi.persist.riak_manager."""
from twisted.trial.unittest import TestCase
from vumi.persist.riak_manager import RiakManager
class TestRiakManager(TestCase):
pass
<commit_msg>Add tests for (nottx)riak manager.<commit_after> | """Tests for vumi.persist.riak_manager."""
from itertools import count
from twisted.trial.unittest import TestCase
from twisted.internet.defer import returnValue
from vumi.persist.riak_manager import RiakManager, flatten_generator
from vumi.persist.tests.test_txriak_manager import CommonRiakManagerTests
class TestRiakManager(CommonRiakManagerTests, TestCase):
"""Most tests are inherited from the CommonRiakManagerTests mixin."""
def setUp(self):
self.manager = RiakManager.from_config({'bucket_prefix': 'test.'})
self.manager.purge_all()
def tearDown(self):
self.manager.purge_all()
def test_call_decorator(self):
self.assertEqual(RiakManager.call_decorator, flatten_generator)
def test_flatten_generator(self):
results = []
counter = count()
@flatten_generator
def f():
for i in range(3):
a = yield counter.next()
results.append(a)
ret = f()
self.assertEqual(ret, None)
self.assertEqual(results, list(range(3)))
def test_flatter_generator_with_return_value(self):
@flatten_generator
def f():
yield None
returnValue("foo")
ret = f()
self.assertEqual(ret, "foo")
| """Tests for vumi.persist.riak_manager."""
from twisted.trial.unittest import TestCase
from vumi.persist.riak_manager import RiakManager
class TestRiakManager(TestCase):
pass
Add tests for (nottx)riak manager."""Tests for vumi.persist.riak_manager."""
from itertools import count
from twisted.trial.unittest import TestCase
from twisted.internet.defer import returnValue
from vumi.persist.riak_manager import RiakManager, flatten_generator
from vumi.persist.tests.test_txriak_manager import CommonRiakManagerTests
class TestRiakManager(CommonRiakManagerTests, TestCase):
"""Most tests are inherited from the CommonRiakManagerTests mixin."""
def setUp(self):
self.manager = RiakManager.from_config({'bucket_prefix': 'test.'})
self.manager.purge_all()
def tearDown(self):
self.manager.purge_all()
def test_call_decorator(self):
self.assertEqual(RiakManager.call_decorator, flatten_generator)
def test_flatten_generator(self):
results = []
counter = count()
@flatten_generator
def f():
for i in range(3):
a = yield counter.next()
results.append(a)
ret = f()
self.assertEqual(ret, None)
self.assertEqual(results, list(range(3)))
def test_flatter_generator_with_return_value(self):
@flatten_generator
def f():
yield None
returnValue("foo")
ret = f()
self.assertEqual(ret, "foo")
| <commit_before>"""Tests for vumi.persist.riak_manager."""
from twisted.trial.unittest import TestCase
from vumi.persist.riak_manager import RiakManager
class TestRiakManager(TestCase):
pass
<commit_msg>Add tests for (nottx)riak manager.<commit_after>"""Tests for vumi.persist.riak_manager."""
from itertools import count
from twisted.trial.unittest import TestCase
from twisted.internet.defer import returnValue
from vumi.persist.riak_manager import RiakManager, flatten_generator
from vumi.persist.tests.test_txriak_manager import CommonRiakManagerTests
class TestRiakManager(CommonRiakManagerTests, TestCase):
"""Most tests are inherited from the CommonRiakManagerTests mixin."""
def setUp(self):
self.manager = RiakManager.from_config({'bucket_prefix': 'test.'})
self.manager.purge_all()
def tearDown(self):
self.manager.purge_all()
def test_call_decorator(self):
self.assertEqual(RiakManager.call_decorator, flatten_generator)
def test_flatten_generator(self):
results = []
counter = count()
@flatten_generator
def f():
for i in range(3):
a = yield counter.next()
results.append(a)
ret = f()
self.assertEqual(ret, None)
self.assertEqual(results, list(range(3)))
def test_flatter_generator_with_return_value(self):
@flatten_generator
def f():
yield None
returnValue("foo")
ret = f()
self.assertEqual(ret, "foo")
|
c4ea39ab8666a2872b25c9b8619f1b0feb823d9f | server.py | server.py | import os
from app import create_app, db
from flask.ext.script import Manager, Shell, Server
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
manager.add_command("runserver", Server(host="0.0.0.0"))
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db)
manager.add_command('shell', Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if __name__ == '__main__':
manager.run()
| import os
from app import create_app, db
from flask.ext.script import Manager, Shell, Server
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
manager.add_command("runserver", Server(host="0.0.0.0"))
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db)
manager.add_command('shell', Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
successful = unittest.TextTestRunner(verbosity=2).run(tests).wasSuccessful()
import sys
sys.exit(successful is False)
if __name__ == '__main__':
manager.run()
| Change system return exit on failed tests | Change system return exit on failed tests
| Python | mit | luisfcofv/Superhero | import os
from app import create_app, db
from flask.ext.script import Manager, Shell, Server
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
manager.add_command("runserver", Server(host="0.0.0.0"))
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db)
manager.add_command('shell', Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if __name__ == '__main__':
manager.run()
Change system return exit on failed tests | import os
from app import create_app, db
from flask.ext.script import Manager, Shell, Server
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
manager.add_command("runserver", Server(host="0.0.0.0"))
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db)
manager.add_command('shell', Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
successful = unittest.TextTestRunner(verbosity=2).run(tests).wasSuccessful()
import sys
sys.exit(successful is False)
if __name__ == '__main__':
manager.run()
| <commit_before>import os
from app import create_app, db
from flask.ext.script import Manager, Shell, Server
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
manager.add_command("runserver", Server(host="0.0.0.0"))
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db)
manager.add_command('shell', Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if __name__ == '__main__':
manager.run()
<commit_msg>Change system return exit on failed tests<commit_after> | import os
from app import create_app, db
from flask.ext.script import Manager, Shell, Server
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
manager.add_command("runserver", Server(host="0.0.0.0"))
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db)
manager.add_command('shell', Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
successful = unittest.TextTestRunner(verbosity=2).run(tests).wasSuccessful()
import sys
sys.exit(successful is False)
if __name__ == '__main__':
manager.run()
| import os
from app import create_app, db
from flask.ext.script import Manager, Shell, Server
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
manager.add_command("runserver", Server(host="0.0.0.0"))
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db)
manager.add_command('shell', Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if __name__ == '__main__':
manager.run()
Change system return exit on failed testsimport os
from app import create_app, db
from flask.ext.script import Manager, Shell, Server
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
manager.add_command("runserver", Server(host="0.0.0.0"))
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db)
manager.add_command('shell', Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
successful = unittest.TextTestRunner(verbosity=2).run(tests).wasSuccessful()
import sys
sys.exit(successful is False)
if __name__ == '__main__':
manager.run()
| <commit_before>import os
from app import create_app, db
from flask.ext.script import Manager, Shell, Server
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
manager.add_command("runserver", Server(host="0.0.0.0"))
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db)
manager.add_command('shell', Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if __name__ == '__main__':
manager.run()
<commit_msg>Change system return exit on failed tests<commit_after>import os
from app import create_app, db
from flask.ext.script import Manager, Shell, Server
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
manager.add_command("runserver", Server(host="0.0.0.0"))
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db)
manager.add_command('shell', Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
successful = unittest.TextTestRunner(verbosity=2).run(tests).wasSuccessful()
import sys
sys.exit(successful is False)
if __name__ == '__main__':
manager.run()
|
b80b781b8f446b8149b948a6ec4aeff63fd728ce | Orange/widgets/utils/plot/__init__.py | Orange/widgets/utils/plot/__init__.py | """
*************************
Plot classes and tools for use in Orange widgets
*************************
The main class of this module is :obj:`.OWPlot`, from which all plots
in visualization widgets should inherit.
This module also contains plot elements, which are normally used by the :obj:`.OWPlot`,
but can also be used directly or subclassed
"""
from .owplotgui import *
from .owpalette import *
from .owconstants import *
try:
from .owcurve import *
from .owpoint import *
from .owlegend import *
from .owaxis import *
from .owplot import *
from .owtools import *
except ImportError:
pass
| """
*************************
Plot classes and tools for use in Orange widgets
*************************
The main class of this module is :obj:`.OWPlot`, from which all plots
in visualization widgets should inherit.
This module also contains plot elements, which are normally used by the :obj:`.OWPlot`,
but can also be used directly or subclassed
"""
from .owplotgui import *
from .owpalette import *
from .owconstants import *
try:
from .owcurve import *
from .owpoint import *
from .owlegend import *
from .owaxis import *
from .owplot import *
from .owtools import *
except (ImportError, RuntimeError):
pass
| Handle PyQt 5.3 raising RuntimeError on incompatible orangeqt import | Handle PyQt 5.3 raising RuntimeError on incompatible orangeqt import
| Python | bsd-2-clause | cheral/orange3,cheral/orange3,cheral/orange3,cheral/orange3,cheral/orange3,cheral/orange3 | """
*************************
Plot classes and tools for use in Orange widgets
*************************
The main class of this module is :obj:`.OWPlot`, from which all plots
in visualization widgets should inherit.
This module also contains plot elements, which are normally used by the :obj:`.OWPlot`,
but can also be used directly or subclassed
"""
from .owplotgui import *
from .owpalette import *
from .owconstants import *
try:
from .owcurve import *
from .owpoint import *
from .owlegend import *
from .owaxis import *
from .owplot import *
from .owtools import *
except ImportError:
pass
Handle PyQt 5.3 raising RuntimeError on incompatible orangeqt import | """
*************************
Plot classes and tools for use in Orange widgets
*************************
The main class of this module is :obj:`.OWPlot`, from which all plots
in visualization widgets should inherit.
This module also contains plot elements, which are normally used by the :obj:`.OWPlot`,
but can also be used directly or subclassed
"""
from .owplotgui import *
from .owpalette import *
from .owconstants import *
try:
from .owcurve import *
from .owpoint import *
from .owlegend import *
from .owaxis import *
from .owplot import *
from .owtools import *
except (ImportError, RuntimeError):
pass
| <commit_before>"""
*************************
Plot classes and tools for use in Orange widgets
*************************
The main class of this module is :obj:`.OWPlot`, from which all plots
in visualization widgets should inherit.
This module also contains plot elements, which are normally used by the :obj:`.OWPlot`,
but can also be used directly or subclassed
"""
from .owplotgui import *
from .owpalette import *
from .owconstants import *
try:
from .owcurve import *
from .owpoint import *
from .owlegend import *
from .owaxis import *
from .owplot import *
from .owtools import *
except ImportError:
pass
<commit_msg>Handle PyQt 5.3 raising RuntimeError on incompatible orangeqt import<commit_after> | """
*************************
Plot classes and tools for use in Orange widgets
*************************
The main class of this module is :obj:`.OWPlot`, from which all plots
in visualization widgets should inherit.
This module also contains plot elements, which are normally used by the :obj:`.OWPlot`,
but can also be used directly or subclassed
"""
from .owplotgui import *
from .owpalette import *
from .owconstants import *
try:
from .owcurve import *
from .owpoint import *
from .owlegend import *
from .owaxis import *
from .owplot import *
from .owtools import *
except (ImportError, RuntimeError):
pass
| """
*************************
Plot classes and tools for use in Orange widgets
*************************
The main class of this module is :obj:`.OWPlot`, from which all plots
in visualization widgets should inherit.
This module also contains plot elements, which are normally used by the :obj:`.OWPlot`,
but can also be used directly or subclassed
"""
from .owplotgui import *
from .owpalette import *
from .owconstants import *
try:
from .owcurve import *
from .owpoint import *
from .owlegend import *
from .owaxis import *
from .owplot import *
from .owtools import *
except ImportError:
pass
Handle PyQt 5.3 raising RuntimeError on incompatible orangeqt import"""
*************************
Plot classes and tools for use in Orange widgets
*************************
The main class of this module is :obj:`.OWPlot`, from which all plots
in visualization widgets should inherit.
This module also contains plot elements, which are normally used by the :obj:`.OWPlot`,
but can also be used directly or subclassed
"""
from .owplotgui import *
from .owpalette import *
from .owconstants import *
try:
from .owcurve import *
from .owpoint import *
from .owlegend import *
from .owaxis import *
from .owplot import *
from .owtools import *
except (ImportError, RuntimeError):
pass
| <commit_before>"""
*************************
Plot classes and tools for use in Orange widgets
*************************
The main class of this module is :obj:`.OWPlot`, from which all plots
in visualization widgets should inherit.
This module also contains plot elements, which are normally used by the :obj:`.OWPlot`,
but can also be used directly or subclassed
"""
from .owplotgui import *
from .owpalette import *
from .owconstants import *
try:
from .owcurve import *
from .owpoint import *
from .owlegend import *
from .owaxis import *
from .owplot import *
from .owtools import *
except ImportError:
pass
<commit_msg>Handle PyQt 5.3 raising RuntimeError on incompatible orangeqt import<commit_after>"""
*************************
Plot classes and tools for use in Orange widgets
*************************
The main class of this module is :obj:`.OWPlot`, from which all plots
in visualization widgets should inherit.
This module also contains plot elements, which are normally used by the :obj:`.OWPlot`,
but can also be used directly or subclassed
"""
from .owplotgui import *
from .owpalette import *
from .owconstants import *
try:
from .owcurve import *
from .owpoint import *
from .owlegend import *
from .owaxis import *
from .owplot import *
from .owtools import *
except (ImportError, RuntimeError):
pass
|
59651470489a4479db6d9a79de3aacee6b9d7cd8 | travis/wait-until-cluster-initialised.py | travis/wait-until-cluster-initialised.py | #!/usr/bin/env python3
import json
import sys
import time
from urllib.request import urlopen
STATS_URL = "http://localhost:18001/stats"
MAXIMUM_TIME_SECONDS = 2 * 60
SLEEPING_INTERVAL_SECONDS = 1
STATUS_CODE_OK = 200
def is_initialised():
try:
response = urlopen(STATS_URL)
if (response.getcode() == STATUS_CODE_OK):
encoding = response.info().get_content_charset('utf-8')
content = response.read().decode(encoding)
return json.loads(content)['initialised']
else:
return False
except Exception as e:
return False
def wait_until_cluster_initialised():
start = time.time()
elapsed = 0.0
while elapsed < MAXIMUM_TIME_SECONDS:
if is_initialised():
print("Cluster initialised!")
break
elapsed = time.time() - start
print("Cluster not initialised... keep waiting... elapsed time: {1:.2f} seconds.".format(
SLEEPING_INTERVAL_SECONDS, elapsed))
time.sleep(SLEEPING_INTERVAL_SECONDS)
elapsed = time.time() - start
else:
sys.exit("Cluster not initialised after {} seconds. I give up!".format(MAXIMUM_TIME_SECONDS))
if __name__ == "__main__":
wait_until_cluster_initialised() | #!/usr/bin/env python3
import json
import sys
import time
from urllib.request import urlopen
STATS_URL = "http://localhost:18001/stats"
MAXIMUM_TIME_SECONDS = 2 * 60
SLEEPING_INTERVAL_SECONDS = 1
STATUS_CODE_OK = 200
def is_initialised():
try:
response = urlopen(STATS_URL)
if (response.getcode() == STATUS_CODE_OK):
encoding = response.info().get_content_charset('utf-8')
content = response.read().decode(encoding)
return json.loads(content)['initialised']
else:
return False
except Exception as e:
return False
def wait_until_cluster_initialised():
start = time.time()
elapsed = 0.0
while elapsed < MAXIMUM_TIME_SECONDS:
if is_initialised():
print("Cluster initialised!")
break
elapsed = time.time() - start
print("Cluster not initialised... keep waiting... elapsed time: {0:.2f} seconds.".format(elapsed))
time.sleep(SLEEPING_INTERVAL_SECONDS)
elapsed = time.time() - start
else:
sys.exit("Cluster not initialised after {} seconds. I give up!".format(MAXIMUM_TIME_SECONDS))
if __name__ == "__main__":
wait_until_cluster_initialised()
| Remove unused left over parameter | Remove unused left over parameter
| Python | apache-2.0 | codiply/barrio,codiply/barrio | #!/usr/bin/env python3
import json
import sys
import time
from urllib.request import urlopen
STATS_URL = "http://localhost:18001/stats"
MAXIMUM_TIME_SECONDS = 2 * 60
SLEEPING_INTERVAL_SECONDS = 1
STATUS_CODE_OK = 200
def is_initialised():
try:
response = urlopen(STATS_URL)
if (response.getcode() == STATUS_CODE_OK):
encoding = response.info().get_content_charset('utf-8')
content = response.read().decode(encoding)
return json.loads(content)['initialised']
else:
return False
except Exception as e:
return False
def wait_until_cluster_initialised():
start = time.time()
elapsed = 0.0
while elapsed < MAXIMUM_TIME_SECONDS:
if is_initialised():
print("Cluster initialised!")
break
elapsed = time.time() - start
print("Cluster not initialised... keep waiting... elapsed time: {1:.2f} seconds.".format(
SLEEPING_INTERVAL_SECONDS, elapsed))
time.sleep(SLEEPING_INTERVAL_SECONDS)
elapsed = time.time() - start
else:
sys.exit("Cluster not initialised after {} seconds. I give up!".format(MAXIMUM_TIME_SECONDS))
if __name__ == "__main__":
wait_until_cluster_initialised()Remove unused left over parameter | #!/usr/bin/env python3
import json
import sys
import time
from urllib.request import urlopen
STATS_URL = "http://localhost:18001/stats"
MAXIMUM_TIME_SECONDS = 2 * 60
SLEEPING_INTERVAL_SECONDS = 1
STATUS_CODE_OK = 200
def is_initialised():
try:
response = urlopen(STATS_URL)
if (response.getcode() == STATUS_CODE_OK):
encoding = response.info().get_content_charset('utf-8')
content = response.read().decode(encoding)
return json.loads(content)['initialised']
else:
return False
except Exception as e:
return False
def wait_until_cluster_initialised():
start = time.time()
elapsed = 0.0
while elapsed < MAXIMUM_TIME_SECONDS:
if is_initialised():
print("Cluster initialised!")
break
elapsed = time.time() - start
print("Cluster not initialised... keep waiting... elapsed time: {0:.2f} seconds.".format(elapsed))
time.sleep(SLEEPING_INTERVAL_SECONDS)
elapsed = time.time() - start
else:
sys.exit("Cluster not initialised after {} seconds. I give up!".format(MAXIMUM_TIME_SECONDS))
if __name__ == "__main__":
wait_until_cluster_initialised()
| <commit_before>#!/usr/bin/env python3
import json
import sys
import time
from urllib.request import urlopen
STATS_URL = "http://localhost:18001/stats"
MAXIMUM_TIME_SECONDS = 2 * 60
SLEEPING_INTERVAL_SECONDS = 1
STATUS_CODE_OK = 200
def is_initialised():
try:
response = urlopen(STATS_URL)
if (response.getcode() == STATUS_CODE_OK):
encoding = response.info().get_content_charset('utf-8')
content = response.read().decode(encoding)
return json.loads(content)['initialised']
else:
return False
except Exception as e:
return False
def wait_until_cluster_initialised():
start = time.time()
elapsed = 0.0
while elapsed < MAXIMUM_TIME_SECONDS:
if is_initialised():
print("Cluster initialised!")
break
elapsed = time.time() - start
print("Cluster not initialised... keep waiting... elapsed time: {1:.2f} seconds.".format(
SLEEPING_INTERVAL_SECONDS, elapsed))
time.sleep(SLEEPING_INTERVAL_SECONDS)
elapsed = time.time() - start
else:
sys.exit("Cluster not initialised after {} seconds. I give up!".format(MAXIMUM_TIME_SECONDS))
if __name__ == "__main__":
wait_until_cluster_initialised()<commit_msg>Remove unused left over parameter<commit_after> | #!/usr/bin/env python3
import json
import sys
import time
from urllib.request import urlopen
STATS_URL = "http://localhost:18001/stats"
MAXIMUM_TIME_SECONDS = 2 * 60
SLEEPING_INTERVAL_SECONDS = 1
STATUS_CODE_OK = 200
def is_initialised():
try:
response = urlopen(STATS_URL)
if (response.getcode() == STATUS_CODE_OK):
encoding = response.info().get_content_charset('utf-8')
content = response.read().decode(encoding)
return json.loads(content)['initialised']
else:
return False
except Exception as e:
return False
def wait_until_cluster_initialised():
start = time.time()
elapsed = 0.0
while elapsed < MAXIMUM_TIME_SECONDS:
if is_initialised():
print("Cluster initialised!")
break
elapsed = time.time() - start
print("Cluster not initialised... keep waiting... elapsed time: {0:.2f} seconds.".format(elapsed))
time.sleep(SLEEPING_INTERVAL_SECONDS)
elapsed = time.time() - start
else:
sys.exit("Cluster not initialised after {} seconds. I give up!".format(MAXIMUM_TIME_SECONDS))
if __name__ == "__main__":
wait_until_cluster_initialised()
| #!/usr/bin/env python3
import json
import sys
import time
from urllib.request import urlopen
STATS_URL = "http://localhost:18001/stats"
MAXIMUM_TIME_SECONDS = 2 * 60
SLEEPING_INTERVAL_SECONDS = 1
STATUS_CODE_OK = 200
def is_initialised():
try:
response = urlopen(STATS_URL)
if (response.getcode() == STATUS_CODE_OK):
encoding = response.info().get_content_charset('utf-8')
content = response.read().decode(encoding)
return json.loads(content)['initialised']
else:
return False
except Exception as e:
return False
def wait_until_cluster_initialised():
start = time.time()
elapsed = 0.0
while elapsed < MAXIMUM_TIME_SECONDS:
if is_initialised():
print("Cluster initialised!")
break
elapsed = time.time() - start
print("Cluster not initialised... keep waiting... elapsed time: {1:.2f} seconds.".format(
SLEEPING_INTERVAL_SECONDS, elapsed))
time.sleep(SLEEPING_INTERVAL_SECONDS)
elapsed = time.time() - start
else:
sys.exit("Cluster not initialised after {} seconds. I give up!".format(MAXIMUM_TIME_SECONDS))
if __name__ == "__main__":
wait_until_cluster_initialised()Remove unused left over parameter#!/usr/bin/env python3
import json
import sys
import time
from urllib.request import urlopen
STATS_URL = "http://localhost:18001/stats"
MAXIMUM_TIME_SECONDS = 2 * 60
SLEEPING_INTERVAL_SECONDS = 1
STATUS_CODE_OK = 200
def is_initialised():
try:
response = urlopen(STATS_URL)
if (response.getcode() == STATUS_CODE_OK):
encoding = response.info().get_content_charset('utf-8')
content = response.read().decode(encoding)
return json.loads(content)['initialised']
else:
return False
except Exception as e:
return False
def wait_until_cluster_initialised():
start = time.time()
elapsed = 0.0
while elapsed < MAXIMUM_TIME_SECONDS:
if is_initialised():
print("Cluster initialised!")
break
elapsed = time.time() - start
print("Cluster not initialised... keep waiting... elapsed time: {0:.2f} seconds.".format(elapsed))
time.sleep(SLEEPING_INTERVAL_SECONDS)
elapsed = time.time() - start
else:
sys.exit("Cluster not initialised after {} seconds. I give up!".format(MAXIMUM_TIME_SECONDS))
if __name__ == "__main__":
wait_until_cluster_initialised()
| <commit_before>#!/usr/bin/env python3
import json
import sys
import time
from urllib.request import urlopen
STATS_URL = "http://localhost:18001/stats"
MAXIMUM_TIME_SECONDS = 2 * 60
SLEEPING_INTERVAL_SECONDS = 1
STATUS_CODE_OK = 200
def is_initialised():
try:
response = urlopen(STATS_URL)
if (response.getcode() == STATUS_CODE_OK):
encoding = response.info().get_content_charset('utf-8')
content = response.read().decode(encoding)
return json.loads(content)['initialised']
else:
return False
except Exception as e:
return False
def wait_until_cluster_initialised():
start = time.time()
elapsed = 0.0
while elapsed < MAXIMUM_TIME_SECONDS:
if is_initialised():
print("Cluster initialised!")
break
elapsed = time.time() - start
print("Cluster not initialised... keep waiting... elapsed time: {1:.2f} seconds.".format(
SLEEPING_INTERVAL_SECONDS, elapsed))
time.sleep(SLEEPING_INTERVAL_SECONDS)
elapsed = time.time() - start
else:
sys.exit("Cluster not initialised after {} seconds. I give up!".format(MAXIMUM_TIME_SECONDS))
if __name__ == "__main__":
wait_until_cluster_initialised()<commit_msg>Remove unused left over parameter<commit_after>#!/usr/bin/env python3
import json
import sys
import time
from urllib.request import urlopen
STATS_URL = "http://localhost:18001/stats"
MAXIMUM_TIME_SECONDS = 2 * 60
SLEEPING_INTERVAL_SECONDS = 1
STATUS_CODE_OK = 200
def is_initialised():
try:
response = urlopen(STATS_URL)
if (response.getcode() == STATUS_CODE_OK):
encoding = response.info().get_content_charset('utf-8')
content = response.read().decode(encoding)
return json.loads(content)['initialised']
else:
return False
except Exception as e:
return False
def wait_until_cluster_initialised():
start = time.time()
elapsed = 0.0
while elapsed < MAXIMUM_TIME_SECONDS:
if is_initialised():
print("Cluster initialised!")
break
elapsed = time.time() - start
print("Cluster not initialised... keep waiting... elapsed time: {0:.2f} seconds.".format(elapsed))
time.sleep(SLEEPING_INTERVAL_SECONDS)
elapsed = time.time() - start
else:
sys.exit("Cluster not initialised after {} seconds. I give up!".format(MAXIMUM_TIME_SECONDS))
if __name__ == "__main__":
wait_until_cluster_initialised()
|
c1af56026da9669ff76908e4d89982b1c88fd30d | examples/custom_xmlrpc_client/server.py | examples/custom_xmlrpc_client/server.py | import random
import time
from SimpleXMLRPCServer import SimpleXMLRPCServer
def get_time():
time.sleep(random.random())
return time.time()
def get_random_number(low, high):
time.sleep(random.random())
return random.randint(low, high)
server = SimpleXMLRPCServer(("localhost", 8877))
print "Listening on port 8877..."
server.register_function(get_time, "get_time")
server.register_function(get_random_number, "get_random_number")
server.serve_forever()
| import random
import time
from SimpleXMLRPCServer import SimpleXMLRPCServer
def get_time():
time.sleep(random.random())
return time.time()
def get_random_number(low, high):
time.sleep(random.random())
return random.randint(low, high)
server = SimpleXMLRPCServer(("localhost", 8877))
print("Listening on port 8877...")
server.register_function(get_time, "get_time")
server.register_function(get_random_number, "get_random_number")
server.serve_forever()
| Use print() function in both Python 2 and Python 3 | Use print() function in both Python 2 and Python 3
Discovered via: __flake8 . --count --select=E9,F63,F72,F82 --show-source --statistics__
Legacy __print__ statements are syntax errors in Python 3 but __print()__ function works as expected in both Python 2 and Python 3. | Python | mit | locustio/locust,mbeacom/locust,mbeacom/locust,mbeacom/locust,locustio/locust,locustio/locust,heyman/locust,mbeacom/locust,locustio/locust | import random
import time
from SimpleXMLRPCServer import SimpleXMLRPCServer
def get_time():
time.sleep(random.random())
return time.time()
def get_random_number(low, high):
time.sleep(random.random())
return random.randint(low, high)
server = SimpleXMLRPCServer(("localhost", 8877))
print "Listening on port 8877..."
server.register_function(get_time, "get_time")
server.register_function(get_random_number, "get_random_number")
server.serve_forever()
Use print() function in both Python 2 and Python 3
Discovered via: __flake8 . --count --select=E9,F63,F72,F82 --show-source --statistics__
Legacy __print__ statements are syntax errors in Python 3 but __print()__ function works as expected in both Python 2 and Python 3. | import random
import time
from SimpleXMLRPCServer import SimpleXMLRPCServer
def get_time():
time.sleep(random.random())
return time.time()
def get_random_number(low, high):
time.sleep(random.random())
return random.randint(low, high)
server = SimpleXMLRPCServer(("localhost", 8877))
print("Listening on port 8877...")
server.register_function(get_time, "get_time")
server.register_function(get_random_number, "get_random_number")
server.serve_forever()
| <commit_before>import random
import time
from SimpleXMLRPCServer import SimpleXMLRPCServer
def get_time():
time.sleep(random.random())
return time.time()
def get_random_number(low, high):
time.sleep(random.random())
return random.randint(low, high)
server = SimpleXMLRPCServer(("localhost", 8877))
print "Listening on port 8877..."
server.register_function(get_time, "get_time")
server.register_function(get_random_number, "get_random_number")
server.serve_forever()
<commit_msg>Use print() function in both Python 2 and Python 3
Discovered via: __flake8 . --count --select=E9,F63,F72,F82 --show-source --statistics__
Legacy __print__ statements are syntax errors in Python 3 but __print()__ function works as expected in both Python 2 and Python 3.<commit_after> | import random
import time
from SimpleXMLRPCServer import SimpleXMLRPCServer
def get_time():
time.sleep(random.random())
return time.time()
def get_random_number(low, high):
time.sleep(random.random())
return random.randint(low, high)
server = SimpleXMLRPCServer(("localhost", 8877))
print("Listening on port 8877...")
server.register_function(get_time, "get_time")
server.register_function(get_random_number, "get_random_number")
server.serve_forever()
| import random
import time
from SimpleXMLRPCServer import SimpleXMLRPCServer
def get_time():
time.sleep(random.random())
return time.time()
def get_random_number(low, high):
time.sleep(random.random())
return random.randint(low, high)
server = SimpleXMLRPCServer(("localhost", 8877))
print "Listening on port 8877..."
server.register_function(get_time, "get_time")
server.register_function(get_random_number, "get_random_number")
server.serve_forever()
Use print() function in both Python 2 and Python 3
Discovered via: __flake8 . --count --select=E9,F63,F72,F82 --show-source --statistics__
Legacy __print__ statements are syntax errors in Python 3 but __print()__ function works as expected in both Python 2 and Python 3.import random
import time
from SimpleXMLRPCServer import SimpleXMLRPCServer
def get_time():
time.sleep(random.random())
return time.time()
def get_random_number(low, high):
time.sleep(random.random())
return random.randint(low, high)
server = SimpleXMLRPCServer(("localhost", 8877))
print("Listening on port 8877...")
server.register_function(get_time, "get_time")
server.register_function(get_random_number, "get_random_number")
server.serve_forever()
| <commit_before>import random
import time
from SimpleXMLRPCServer import SimpleXMLRPCServer
def get_time():
time.sleep(random.random())
return time.time()
def get_random_number(low, high):
time.sleep(random.random())
return random.randint(low, high)
server = SimpleXMLRPCServer(("localhost", 8877))
print "Listening on port 8877..."
server.register_function(get_time, "get_time")
server.register_function(get_random_number, "get_random_number")
server.serve_forever()
<commit_msg>Use print() function in both Python 2 and Python 3
Discovered via: __flake8 . --count --select=E9,F63,F72,F82 --show-source --statistics__
Legacy __print__ statements are syntax errors in Python 3 but __print()__ function works as expected in both Python 2 and Python 3.<commit_after>import random
import time
from SimpleXMLRPCServer import SimpleXMLRPCServer
def get_time():
time.sleep(random.random())
return time.time()
def get_random_number(low, high):
time.sleep(random.random())
return random.randint(low, high)
server = SimpleXMLRPCServer(("localhost", 8877))
print("Listening on port 8877...")
server.register_function(get_time, "get_time")
server.register_function(get_random_number, "get_random_number")
server.serve_forever()
|
d81dbd7b25cd44f730e979efe03eb6e5e1d87f1b | admin/commandRunner.py | admin/commandRunner.py | import configparser
import sys
import os
parser = configparser.ConfigParser()
parser.read("../halite.ini")
WORKERS = dict(parser.items("workerIPs"))
command = sys.argv[1]
print(command)
for name in WORKERS:
print("########"+name+"########")
print(WORKERS[name])
os.system("ssh root@"+WORKERS[name]+" '"+command+"'")
print("################\n")
| import pymysql
import configparser
import sys
import os
import os.path
parser = configparser.ConfigParser()
parser.read("../halite.ini")
DB_CONFIG = parser["database"]
keyPath = os.path.join("../", parser["aws"]["keyfilepath"])
db = pymysql.connect(host=DB_CONFIG["hostname"], user=DB_CONFIG['username'], passwd=DB_CONFIG['password'], db=DB_CONFIG['name'], cursorclass=pymysql.cursors.DictCursor)
cursor = db.cursor()
cursor.execute("select * from Worker")
workers = cursor.fetchall()
command = sys.argv[1]
for worker in workers:
print("########"+worker['ipAddress']+"########")
os.system("ssh -i \""+keyPath+"\" ubuntu@"+worker['ipAddress']+" '"+command+"'")
print("################\n")
| Switch command runner to using db | Switch command runner to using db
| Python | mit | HaliteChallenge/Halite,HaliteChallenge/Halite,yangle/HaliteIO,HaliteChallenge/Halite,lanyudhy/Halite-II,lanyudhy/Halite-II,yangle/HaliteIO,yangle/HaliteIO,lanyudhy/Halite-II,yangle/HaliteIO,yangle/HaliteIO,HaliteChallenge/Halite,HaliteChallenge/Halite-II,lanyudhy/Halite-II,HaliteChallenge/Halite,lanyudhy/Halite-II,lanyudhy/Halite-II,HaliteChallenge/Halite,HaliteChallenge/Halite,HaliteChallenge/Halite,HaliteChallenge/Halite,yangle/HaliteIO,yangle/HaliteIO,HaliteChallenge/Halite-II,yangle/HaliteIO,HaliteChallenge/Halite,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II,lanyudhy/Halite-II,HaliteChallenge/Halite-II,lanyudhy/Halite-II,yangle/HaliteIO,HaliteChallenge/Halite,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II,yangle/HaliteIO,yangle/HaliteIO,HaliteChallenge/Halite-II,HaliteChallenge/Halite,lanyudhy/Halite-II,yangle/HaliteIO,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II,lanyudhy/Halite-II,HaliteChallenge/Halite-II,lanyudhy/Halite-II,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II | import configparser
import sys
import os
parser = configparser.ConfigParser()
parser.read("../halite.ini")
WORKERS = dict(parser.items("workerIPs"))
command = sys.argv[1]
print(command)
for name in WORKERS:
print("########"+name+"########")
print(WORKERS[name])
os.system("ssh root@"+WORKERS[name]+" '"+command+"'")
print("################\n")
Switch command runner to using db | import pymysql
import configparser
import sys
import os
import os.path
parser = configparser.ConfigParser()
parser.read("../halite.ini")
DB_CONFIG = parser["database"]
keyPath = os.path.join("../", parser["aws"]["keyfilepath"])
db = pymysql.connect(host=DB_CONFIG["hostname"], user=DB_CONFIG['username'], passwd=DB_CONFIG['password'], db=DB_CONFIG['name'], cursorclass=pymysql.cursors.DictCursor)
cursor = db.cursor()
cursor.execute("select * from Worker")
workers = cursor.fetchall()
command = sys.argv[1]
for worker in workers:
print("########"+worker['ipAddress']+"########")
os.system("ssh -i \""+keyPath+"\" ubuntu@"+worker['ipAddress']+" '"+command+"'")
print("################\n")
| <commit_before>import configparser
import sys
import os
parser = configparser.ConfigParser()
parser.read("../halite.ini")
WORKERS = dict(parser.items("workerIPs"))
command = sys.argv[1]
print(command)
for name in WORKERS:
print("########"+name+"########")
print(WORKERS[name])
os.system("ssh root@"+WORKERS[name]+" '"+command+"'")
print("################\n")
<commit_msg>Switch command runner to using db<commit_after> | import pymysql
import configparser
import sys
import os
import os.path
parser = configparser.ConfigParser()
parser.read("../halite.ini")
DB_CONFIG = parser["database"]
keyPath = os.path.join("../", parser["aws"]["keyfilepath"])
db = pymysql.connect(host=DB_CONFIG["hostname"], user=DB_CONFIG['username'], passwd=DB_CONFIG['password'], db=DB_CONFIG['name'], cursorclass=pymysql.cursors.DictCursor)
cursor = db.cursor()
cursor.execute("select * from Worker")
workers = cursor.fetchall()
command = sys.argv[1]
for worker in workers:
print("########"+worker['ipAddress']+"########")
os.system("ssh -i \""+keyPath+"\" ubuntu@"+worker['ipAddress']+" '"+command+"'")
print("################\n")
| import configparser
import sys
import os
parser = configparser.ConfigParser()
parser.read("../halite.ini")
WORKERS = dict(parser.items("workerIPs"))
command = sys.argv[1]
print(command)
for name in WORKERS:
print("########"+name+"########")
print(WORKERS[name])
os.system("ssh root@"+WORKERS[name]+" '"+command+"'")
print("################\n")
Switch command runner to using dbimport pymysql
import configparser
import sys
import os
import os.path
parser = configparser.ConfigParser()
parser.read("../halite.ini")
DB_CONFIG = parser["database"]
keyPath = os.path.join("../", parser["aws"]["keyfilepath"])
db = pymysql.connect(host=DB_CONFIG["hostname"], user=DB_CONFIG['username'], passwd=DB_CONFIG['password'], db=DB_CONFIG['name'], cursorclass=pymysql.cursors.DictCursor)
cursor = db.cursor()
cursor.execute("select * from Worker")
workers = cursor.fetchall()
command = sys.argv[1]
for worker in workers:
print("########"+worker['ipAddress']+"########")
os.system("ssh -i \""+keyPath+"\" ubuntu@"+worker['ipAddress']+" '"+command+"'")
print("################\n")
| <commit_before>import configparser
import sys
import os
parser = configparser.ConfigParser()
parser.read("../halite.ini")
WORKERS = dict(parser.items("workerIPs"))
command = sys.argv[1]
print(command)
for name in WORKERS:
print("########"+name+"########")
print(WORKERS[name])
os.system("ssh root@"+WORKERS[name]+" '"+command+"'")
print("################\n")
<commit_msg>Switch command runner to using db<commit_after>import pymysql
import configparser
import sys
import os
import os.path
parser = configparser.ConfigParser()
parser.read("../halite.ini")
DB_CONFIG = parser["database"]
keyPath = os.path.join("../", parser["aws"]["keyfilepath"])
db = pymysql.connect(host=DB_CONFIG["hostname"], user=DB_CONFIG['username'], passwd=DB_CONFIG['password'], db=DB_CONFIG['name'], cursorclass=pymysql.cursors.DictCursor)
cursor = db.cursor()
cursor.execute("select * from Worker")
workers = cursor.fetchall()
command = sys.argv[1]
for worker in workers:
print("########"+worker['ipAddress']+"########")
os.system("ssh -i \""+keyPath+"\" ubuntu@"+worker['ipAddress']+" '"+command+"'")
print("################\n")
|
cba8bd7d3440cc643823e93036bc3b9ac938a412 | pinboard_linkrot.py | pinboard_linkrot.py | #!/usr/bin/env python
from __future__ import division
import requests
import json
import sys
from requests.exceptions import SSLError, InvalidSchema, ConnectionError
def get_link_status_code(link):
headers = {'User-agent':'Mozilla/5.0'}
try:
r = requests.get(link, headers = headers)
return r.status_code
except (SSLError, InvalidSchema, ConnectionError):
return 409
def is_valid_link(status_code):
if status_code == 200:
return True
else:
return False
def process_links(links):
bad_links = 0
try:
for link in links:
status_code = get_link_status_code(link['href'])
if not is_valid_link(status_code):
print 'Invalid link (%s): %s [%s]' % (status_code, link['description'], link['href'])
bad_links += 1
except KeyboardInterrupt:
pass
linkrot = int(bad_links/len(links)*100)
print '\n%s%% linkrot\n' % linkrot
def process_bookmarks_file(filename):
with open(filename) as f:
bookmarks = json.load(f)
process_links(bookmarks)
if __name__ == '__main__':
if len(sys.argv) != 2:
print 'Usage: pinboard_linkrot.py <bookmarks.json>'
exit(1)
process_bookmarks_file(sys.argv[1])
| #!/usr/bin/env python
from __future__ import division
import requests
import json
import sys
from requests.exceptions import SSLError, InvalidSchema, ConnectionError
def get_link_status_code(link):
headers = {'User-agent':'Mozilla/5.0'}
try:
r = requests.head(link, headers=headers, allow_redirects=True)
return r.status_code
except (SSLError, InvalidSchema, ConnectionError):
return 409
def is_valid_link(status_code):
if status_code == 200:
return True
else:
return False
def process_links(links):
bad_links = 0
try:
for link in links:
status_code = get_link_status_code(link['href'])
if not is_valid_link(status_code):
print 'Invalid link (%s): %s [%s]' % (status_code, link['description'], link['href'])
bad_links += 1
except KeyboardInterrupt:
pass
linkrot = int(bad_links/len(links)*100)
print '\n%s%% linkrot\n' % linkrot
def process_bookmarks_file(filename):
with open(filename) as f:
bookmarks = json.load(f)
process_links(bookmarks)
if __name__ == '__main__':
if len(sys.argv) != 2:
print 'Usage: pinboard_linkrot.py <bookmarks.json>'
exit(1)
process_bookmarks_file(sys.argv[1])
| Switch to head requests rather than get requests. | Switch to head requests rather than get requests.
| Python | mit | edgauthier/pinboard_linkrot | #!/usr/bin/env python
from __future__ import division
import requests
import json
import sys
from requests.exceptions import SSLError, InvalidSchema, ConnectionError
def get_link_status_code(link):
headers = {'User-agent':'Mozilla/5.0'}
try:
r = requests.get(link, headers = headers)
return r.status_code
except (SSLError, InvalidSchema, ConnectionError):
return 409
def is_valid_link(status_code):
if status_code == 200:
return True
else:
return False
def process_links(links):
bad_links = 0
try:
for link in links:
status_code = get_link_status_code(link['href'])
if not is_valid_link(status_code):
print 'Invalid link (%s): %s [%s]' % (status_code, link['description'], link['href'])
bad_links += 1
except KeyboardInterrupt:
pass
linkrot = int(bad_links/len(links)*100)
print '\n%s%% linkrot\n' % linkrot
def process_bookmarks_file(filename):
with open(filename) as f:
bookmarks = json.load(f)
process_links(bookmarks)
if __name__ == '__main__':
if len(sys.argv) != 2:
print 'Usage: pinboard_linkrot.py <bookmarks.json>'
exit(1)
process_bookmarks_file(sys.argv[1])
Switch to head requests rather than get requests. | #!/usr/bin/env python
from __future__ import division
import requests
import json
import sys
from requests.exceptions import SSLError, InvalidSchema, ConnectionError
def get_link_status_code(link):
headers = {'User-agent':'Mozilla/5.0'}
try:
r = requests.head(link, headers=headers, allow_redirects=True)
return r.status_code
except (SSLError, InvalidSchema, ConnectionError):
return 409
def is_valid_link(status_code):
if status_code == 200:
return True
else:
return False
def process_links(links):
bad_links = 0
try:
for link in links:
status_code = get_link_status_code(link['href'])
if not is_valid_link(status_code):
print 'Invalid link (%s): %s [%s]' % (status_code, link['description'], link['href'])
bad_links += 1
except KeyboardInterrupt:
pass
linkrot = int(bad_links/len(links)*100)
print '\n%s%% linkrot\n' % linkrot
def process_bookmarks_file(filename):
with open(filename) as f:
bookmarks = json.load(f)
process_links(bookmarks)
if __name__ == '__main__':
if len(sys.argv) != 2:
print 'Usage: pinboard_linkrot.py <bookmarks.json>'
exit(1)
process_bookmarks_file(sys.argv[1])
| <commit_before>#!/usr/bin/env python
from __future__ import division
import requests
import json
import sys
from requests.exceptions import SSLError, InvalidSchema, ConnectionError
def get_link_status_code(link):
headers = {'User-agent':'Mozilla/5.0'}
try:
r = requests.get(link, headers = headers)
return r.status_code
except (SSLError, InvalidSchema, ConnectionError):
return 409
def is_valid_link(status_code):
if status_code == 200:
return True
else:
return False
def process_links(links):
bad_links = 0
try:
for link in links:
status_code = get_link_status_code(link['href'])
if not is_valid_link(status_code):
print 'Invalid link (%s): %s [%s]' % (status_code, link['description'], link['href'])
bad_links += 1
except KeyboardInterrupt:
pass
linkrot = int(bad_links/len(links)*100)
print '\n%s%% linkrot\n' % linkrot
def process_bookmarks_file(filename):
with open(filename) as f:
bookmarks = json.load(f)
process_links(bookmarks)
if __name__ == '__main__':
if len(sys.argv) != 2:
print 'Usage: pinboard_linkrot.py <bookmarks.json>'
exit(1)
process_bookmarks_file(sys.argv[1])
<commit_msg>Switch to head requests rather than get requests.<commit_after> | #!/usr/bin/env python
from __future__ import division
import requests
import json
import sys
from requests.exceptions import SSLError, InvalidSchema, ConnectionError
def get_link_status_code(link):
headers = {'User-agent':'Mozilla/5.0'}
try:
r = requests.head(link, headers=headers, allow_redirects=True)
return r.status_code
except (SSLError, InvalidSchema, ConnectionError):
return 409
def is_valid_link(status_code):
if status_code == 200:
return True
else:
return False
def process_links(links):
bad_links = 0
try:
for link in links:
status_code = get_link_status_code(link['href'])
if not is_valid_link(status_code):
print 'Invalid link (%s): %s [%s]' % (status_code, link['description'], link['href'])
bad_links += 1
except KeyboardInterrupt:
pass
linkrot = int(bad_links/len(links)*100)
print '\n%s%% linkrot\n' % linkrot
def process_bookmarks_file(filename):
with open(filename) as f:
bookmarks = json.load(f)
process_links(bookmarks)
if __name__ == '__main__':
if len(sys.argv) != 2:
print 'Usage: pinboard_linkrot.py <bookmarks.json>'
exit(1)
process_bookmarks_file(sys.argv[1])
| #!/usr/bin/env python
from __future__ import division
import requests
import json
import sys
from requests.exceptions import SSLError, InvalidSchema, ConnectionError
def get_link_status_code(link):
headers = {'User-agent':'Mozilla/5.0'}
try:
r = requests.get(link, headers = headers)
return r.status_code
except (SSLError, InvalidSchema, ConnectionError):
return 409
def is_valid_link(status_code):
if status_code == 200:
return True
else:
return False
def process_links(links):
bad_links = 0
try:
for link in links:
status_code = get_link_status_code(link['href'])
if not is_valid_link(status_code):
print 'Invalid link (%s): %s [%s]' % (status_code, link['description'], link['href'])
bad_links += 1
except KeyboardInterrupt:
pass
linkrot = int(bad_links/len(links)*100)
print '\n%s%% linkrot\n' % linkrot
def process_bookmarks_file(filename):
with open(filename) as f:
bookmarks = json.load(f)
process_links(bookmarks)
if __name__ == '__main__':
if len(sys.argv) != 2:
print 'Usage: pinboard_linkrot.py <bookmarks.json>'
exit(1)
process_bookmarks_file(sys.argv[1])
Switch to head requests rather than get requests.#!/usr/bin/env python
from __future__ import division
import requests
import json
import sys
from requests.exceptions import SSLError, InvalidSchema, ConnectionError
def get_link_status_code(link):
headers = {'User-agent':'Mozilla/5.0'}
try:
r = requests.head(link, headers=headers, allow_redirects=True)
return r.status_code
except (SSLError, InvalidSchema, ConnectionError):
return 409
def is_valid_link(status_code):
if status_code == 200:
return True
else:
return False
def process_links(links):
bad_links = 0
try:
for link in links:
status_code = get_link_status_code(link['href'])
if not is_valid_link(status_code):
print 'Invalid link (%s): %s [%s]' % (status_code, link['description'], link['href'])
bad_links += 1
except KeyboardInterrupt:
pass
linkrot = int(bad_links/len(links)*100)
print '\n%s%% linkrot\n' % linkrot
def process_bookmarks_file(filename):
with open(filename) as f:
bookmarks = json.load(f)
process_links(bookmarks)
if __name__ == '__main__':
if len(sys.argv) != 2:
print 'Usage: pinboard_linkrot.py <bookmarks.json>'
exit(1)
process_bookmarks_file(sys.argv[1])
| <commit_before>#!/usr/bin/env python
from __future__ import division
import requests
import json
import sys
from requests.exceptions import SSLError, InvalidSchema, ConnectionError
def get_link_status_code(link):
headers = {'User-agent':'Mozilla/5.0'}
try:
r = requests.get(link, headers = headers)
return r.status_code
except (SSLError, InvalidSchema, ConnectionError):
return 409
def is_valid_link(status_code):
if status_code == 200:
return True
else:
return False
def process_links(links):
bad_links = 0
try:
for link in links:
status_code = get_link_status_code(link['href'])
if not is_valid_link(status_code):
print 'Invalid link (%s): %s [%s]' % (status_code, link['description'], link['href'])
bad_links += 1
except KeyboardInterrupt:
pass
linkrot = int(bad_links/len(links)*100)
print '\n%s%% linkrot\n' % linkrot
def process_bookmarks_file(filename):
with open(filename) as f:
bookmarks = json.load(f)
process_links(bookmarks)
if __name__ == '__main__':
if len(sys.argv) != 2:
print 'Usage: pinboard_linkrot.py <bookmarks.json>'
exit(1)
process_bookmarks_file(sys.argv[1])
<commit_msg>Switch to head requests rather than get requests.<commit_after>#!/usr/bin/env python
from __future__ import division
import requests
import json
import sys
from requests.exceptions import SSLError, InvalidSchema, ConnectionError
def get_link_status_code(link):
headers = {'User-agent':'Mozilla/5.0'}
try:
r = requests.head(link, headers=headers, allow_redirects=True)
return r.status_code
except (SSLError, InvalidSchema, ConnectionError):
return 409
def is_valid_link(status_code):
if status_code == 200:
return True
else:
return False
def process_links(links):
bad_links = 0
try:
for link in links:
status_code = get_link_status_code(link['href'])
if not is_valid_link(status_code):
print 'Invalid link (%s): %s [%s]' % (status_code, link['description'], link['href'])
bad_links += 1
except KeyboardInterrupt:
pass
linkrot = int(bad_links/len(links)*100)
print '\n%s%% linkrot\n' % linkrot
def process_bookmarks_file(filename):
with open(filename) as f:
bookmarks = json.load(f)
process_links(bookmarks)
if __name__ == '__main__':
if len(sys.argv) != 2:
print 'Usage: pinboard_linkrot.py <bookmarks.json>'
exit(1)
process_bookmarks_file(sys.argv[1])
|
4efbc87a912b62db062da0c277baf2ea007e29e2 | feedthefox/users/models.py | feedthefox/users/models.py | from django.db import models
from django.contrib.auth.models import AbstractUser
from django.utils.encoding import python_2_unicode_compatible
@python_2_unicode_compatible
class User(AbstractUser):
"""Basic Mozillian user profile."""
ircname = models.CharField(max_length=50, default='', blank=True)
avatar_url = models.URLField(max_length=400, default='', blank=True)
city = models.CharField(max_length=50, default='', blank=True)
country = models.CharField(max_length=50, default='', blank=True)
mozillians_url = models.URLField()
def __str__(self):
return self.user.get_full_name()
| from django.db import models
from django.contrib.auth.models import AbstractUser
from django.utils.encoding import python_2_unicode_compatible
@python_2_unicode_compatible
class User(AbstractUser):
"""Basic Mozillian user profile."""
ircname = models.CharField(max_length=50, default='', blank=True)
avatar_url = models.URLField(max_length=400, default='', blank=True)
city = models.CharField(max_length=50, default='', blank=True)
country = models.CharField(max_length=50, default='', blank=True)
mozillians_url = models.URLField()
def __str__(self):
return self.get_full_name()
| Fix str method for custom User model. | Fix str method for custom User model.
| Python | mpl-2.0 | mozilla/feedthefox,akatsoulas/feedthefox,mozilla/feedthefox,akatsoulas/feedthefox,akatsoulas/feedthefox,akatsoulas/feedthefox,mozilla/feedthefox,mozilla/feedthefox | from django.db import models
from django.contrib.auth.models import AbstractUser
from django.utils.encoding import python_2_unicode_compatible
@python_2_unicode_compatible
class User(AbstractUser):
"""Basic Mozillian user profile."""
ircname = models.CharField(max_length=50, default='', blank=True)
avatar_url = models.URLField(max_length=400, default='', blank=True)
city = models.CharField(max_length=50, default='', blank=True)
country = models.CharField(max_length=50, default='', blank=True)
mozillians_url = models.URLField()
def __str__(self):
return self.user.get_full_name()
Fix str method for custom User model. | from django.db import models
from django.contrib.auth.models import AbstractUser
from django.utils.encoding import python_2_unicode_compatible
@python_2_unicode_compatible
class User(AbstractUser):
"""Basic Mozillian user profile."""
ircname = models.CharField(max_length=50, default='', blank=True)
avatar_url = models.URLField(max_length=400, default='', blank=True)
city = models.CharField(max_length=50, default='', blank=True)
country = models.CharField(max_length=50, default='', blank=True)
mozillians_url = models.URLField()
def __str__(self):
return self.get_full_name()
| <commit_before>from django.db import models
from django.contrib.auth.models import AbstractUser
from django.utils.encoding import python_2_unicode_compatible
@python_2_unicode_compatible
class User(AbstractUser):
"""Basic Mozillian user profile."""
ircname = models.CharField(max_length=50, default='', blank=True)
avatar_url = models.URLField(max_length=400, default='', blank=True)
city = models.CharField(max_length=50, default='', blank=True)
country = models.CharField(max_length=50, default='', blank=True)
mozillians_url = models.URLField()
def __str__(self):
return self.user.get_full_name()
<commit_msg>Fix str method for custom User model.<commit_after> | from django.db import models
from django.contrib.auth.models import AbstractUser
from django.utils.encoding import python_2_unicode_compatible
@python_2_unicode_compatible
class User(AbstractUser):
"""Basic Mozillian user profile."""
ircname = models.CharField(max_length=50, default='', blank=True)
avatar_url = models.URLField(max_length=400, default='', blank=True)
city = models.CharField(max_length=50, default='', blank=True)
country = models.CharField(max_length=50, default='', blank=True)
mozillians_url = models.URLField()
def __str__(self):
return self.get_full_name()
| from django.db import models
from django.contrib.auth.models import AbstractUser
from django.utils.encoding import python_2_unicode_compatible
@python_2_unicode_compatible
class User(AbstractUser):
"""Basic Mozillian user profile."""
ircname = models.CharField(max_length=50, default='', blank=True)
avatar_url = models.URLField(max_length=400, default='', blank=True)
city = models.CharField(max_length=50, default='', blank=True)
country = models.CharField(max_length=50, default='', blank=True)
mozillians_url = models.URLField()
def __str__(self):
return self.user.get_full_name()
Fix str method for custom User model.from django.db import models
from django.contrib.auth.models import AbstractUser
from django.utils.encoding import python_2_unicode_compatible
@python_2_unicode_compatible
class User(AbstractUser):
"""Basic Mozillian user profile."""
ircname = models.CharField(max_length=50, default='', blank=True)
avatar_url = models.URLField(max_length=400, default='', blank=True)
city = models.CharField(max_length=50, default='', blank=True)
country = models.CharField(max_length=50, default='', blank=True)
mozillians_url = models.URLField()
def __str__(self):
return self.get_full_name()
| <commit_before>from django.db import models
from django.contrib.auth.models import AbstractUser
from django.utils.encoding import python_2_unicode_compatible
@python_2_unicode_compatible
class User(AbstractUser):
"""Basic Mozillian user profile."""
ircname = models.CharField(max_length=50, default='', blank=True)
avatar_url = models.URLField(max_length=400, default='', blank=True)
city = models.CharField(max_length=50, default='', blank=True)
country = models.CharField(max_length=50, default='', blank=True)
mozillians_url = models.URLField()
def __str__(self):
return self.user.get_full_name()
<commit_msg>Fix str method for custom User model.<commit_after>from django.db import models
from django.contrib.auth.models import AbstractUser
from django.utils.encoding import python_2_unicode_compatible
@python_2_unicode_compatible
class User(AbstractUser):
"""Basic Mozillian user profile."""
ircname = models.CharField(max_length=50, default='', blank=True)
avatar_url = models.URLField(max_length=400, default='', blank=True)
city = models.CharField(max_length=50, default='', blank=True)
country = models.CharField(max_length=50, default='', blank=True)
mozillians_url = models.URLField()
def __str__(self):
return self.get_full_name()
|
91f107ef2ebdaf7ff210b9f36e2c810441f389e7 | services/rdio.py | services/rdio.py | from werkzeug.urls import url_decode
from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
signature_type = SIGNATURE_TYPE_BODY
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
| from werkzeug.urls import url_decode
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
| Allow Rdio to use default signature handling | Allow Rdio to use default signature handling
| Python | bsd-3-clause | foauth/oauth-proxy,foauth/foauth.org,foauth/foauth.org,foauth/foauth.org | from werkzeug.urls import url_decode
from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
signature_type = SIGNATURE_TYPE_BODY
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
Allow Rdio to use default signature handling | from werkzeug.urls import url_decode
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
| <commit_before>from werkzeug.urls import url_decode
from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
signature_type = SIGNATURE_TYPE_BODY
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
<commit_msg>Allow Rdio to use default signature handling<commit_after> | from werkzeug.urls import url_decode
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
| from werkzeug.urls import url_decode
from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
signature_type = SIGNATURE_TYPE_BODY
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
Allow Rdio to use default signature handlingfrom werkzeug.urls import url_decode
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
| <commit_before>from werkzeug.urls import url_decode
from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
signature_type = SIGNATURE_TYPE_BODY
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
<commit_msg>Allow Rdio to use default signature handling<commit_after>from werkzeug.urls import url_decode
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
|
7dcda004fb2cc61b075e7ef67c8c33ebfd70786c | bashhub/view/status.py | bashhub/view/status.py | import dateutil.parser
import datetime
import humanize
status_view ="""\
=== Bashhub Status
http://bashhub.com/u/{0}
Total Commands: {1}
Total Sessions: {2}
Total Systems: {3}
===
Session PID {4} Started {5}
Commands In Session: {6}
Commands Today: {7}
"""
def build_status_view(model):
date = datetime.datetime.fromtimestamp(model.session_start_time/1000.0)
date_str = humanize.naturaltime(date)
return status_view.format(model.username, model.total_commands,
model.total_sessions, model.total_systems, model.session_name,
date_str, model.session_total_commands, model.total_commands_today)
| import dateutil.parser
import datetime
import humanize
status_view ="""\
=== Bashhub Status
https://bashhub.com/u/{0}
Total Commands: {1}
Total Sessions: {2}
Total Systems: {3}
===
Session PID {4} Started {5}
Commands In Session: {6}
Commands Today: {7}
"""
def build_status_view(model):
date = datetime.datetime.fromtimestamp(model.session_start_time/1000.0)
date_str = humanize.naturaltime(date)
return status_view.format(model.username, model.total_commands,
model.total_sessions, model.total_systems, model.session_name,
date_str, model.session_total_commands, model.total_commands_today)
| Change http url to https | Change http url to https
| Python | apache-2.0 | rcaloras/bashhub-client,rcaloras/bashhub-client | import dateutil.parser
import datetime
import humanize
status_view ="""\
=== Bashhub Status
http://bashhub.com/u/{0}
Total Commands: {1}
Total Sessions: {2}
Total Systems: {3}
===
Session PID {4} Started {5}
Commands In Session: {6}
Commands Today: {7}
"""
def build_status_view(model):
date = datetime.datetime.fromtimestamp(model.session_start_time/1000.0)
date_str = humanize.naturaltime(date)
return status_view.format(model.username, model.total_commands,
model.total_sessions, model.total_systems, model.session_name,
date_str, model.session_total_commands, model.total_commands_today)
Change http url to https | import dateutil.parser
import datetime
import humanize
status_view ="""\
=== Bashhub Status
https://bashhub.com/u/{0}
Total Commands: {1}
Total Sessions: {2}
Total Systems: {3}
===
Session PID {4} Started {5}
Commands In Session: {6}
Commands Today: {7}
"""
def build_status_view(model):
date = datetime.datetime.fromtimestamp(model.session_start_time/1000.0)
date_str = humanize.naturaltime(date)
return status_view.format(model.username, model.total_commands,
model.total_sessions, model.total_systems, model.session_name,
date_str, model.session_total_commands, model.total_commands_today)
| <commit_before>import dateutil.parser
import datetime
import humanize
status_view ="""\
=== Bashhub Status
http://bashhub.com/u/{0}
Total Commands: {1}
Total Sessions: {2}
Total Systems: {3}
===
Session PID {4} Started {5}
Commands In Session: {6}
Commands Today: {7}
"""
def build_status_view(model):
date = datetime.datetime.fromtimestamp(model.session_start_time/1000.0)
date_str = humanize.naturaltime(date)
return status_view.format(model.username, model.total_commands,
model.total_sessions, model.total_systems, model.session_name,
date_str, model.session_total_commands, model.total_commands_today)
<commit_msg>Change http url to https<commit_after> | import dateutil.parser
import datetime
import humanize
status_view ="""\
=== Bashhub Status
https://bashhub.com/u/{0}
Total Commands: {1}
Total Sessions: {2}
Total Systems: {3}
===
Session PID {4} Started {5}
Commands In Session: {6}
Commands Today: {7}
"""
def build_status_view(model):
date = datetime.datetime.fromtimestamp(model.session_start_time/1000.0)
date_str = humanize.naturaltime(date)
return status_view.format(model.username, model.total_commands,
model.total_sessions, model.total_systems, model.session_name,
date_str, model.session_total_commands, model.total_commands_today)
| import dateutil.parser
import datetime
import humanize
status_view ="""\
=== Bashhub Status
http://bashhub.com/u/{0}
Total Commands: {1}
Total Sessions: {2}
Total Systems: {3}
===
Session PID {4} Started {5}
Commands In Session: {6}
Commands Today: {7}
"""
def build_status_view(model):
date = datetime.datetime.fromtimestamp(model.session_start_time/1000.0)
date_str = humanize.naturaltime(date)
return status_view.format(model.username, model.total_commands,
model.total_sessions, model.total_systems, model.session_name,
date_str, model.session_total_commands, model.total_commands_today)
Change http url to httpsimport dateutil.parser
import datetime
import humanize
status_view ="""\
=== Bashhub Status
https://bashhub.com/u/{0}
Total Commands: {1}
Total Sessions: {2}
Total Systems: {3}
===
Session PID {4} Started {5}
Commands In Session: {6}
Commands Today: {7}
"""
def build_status_view(model):
date = datetime.datetime.fromtimestamp(model.session_start_time/1000.0)
date_str = humanize.naturaltime(date)
return status_view.format(model.username, model.total_commands,
model.total_sessions, model.total_systems, model.session_name,
date_str, model.session_total_commands, model.total_commands_today)
| <commit_before>import dateutil.parser
import datetime
import humanize
status_view ="""\
=== Bashhub Status
http://bashhub.com/u/{0}
Total Commands: {1}
Total Sessions: {2}
Total Systems: {3}
===
Session PID {4} Started {5}
Commands In Session: {6}
Commands Today: {7}
"""
def build_status_view(model):
date = datetime.datetime.fromtimestamp(model.session_start_time/1000.0)
date_str = humanize.naturaltime(date)
return status_view.format(model.username, model.total_commands,
model.total_sessions, model.total_systems, model.session_name,
date_str, model.session_total_commands, model.total_commands_today)
<commit_msg>Change http url to https<commit_after>import dateutil.parser
import datetime
import humanize
status_view ="""\
=== Bashhub Status
https://bashhub.com/u/{0}
Total Commands: {1}
Total Sessions: {2}
Total Systems: {3}
===
Session PID {4} Started {5}
Commands In Session: {6}
Commands Today: {7}
"""
def build_status_view(model):
date = datetime.datetime.fromtimestamp(model.session_start_time/1000.0)
date_str = humanize.naturaltime(date)
return status_view.format(model.username, model.total_commands,
model.total_sessions, model.total_systems, model.session_name,
date_str, model.session_total_commands, model.total_commands_today)
|
7959c38d82090db6a66c7d81a4adba089c9a884f | brains/orders/views.py | brains/orders/views.py | import math
from django.shortcuts import render
from django.template import RequestContext
from django.http import HttpResponseRedirect
from orders.models import Order
def index(request, x, y):
if request.META['HTTP_REFERER'] not in ('http://www.urbandead.com/map.cgi', 'http://urbandead.com/map.cgi'):
return HttpResponseRedirect('http://www.youtube.com/watch?v=lWKQiZVBtu4')
x = int(x)
y = int(y)
# Fetch the orders
fresh_order = list(Order.objects.all().order_by('-date'))
fresh_order.sort(key=lambda o: user_coord_dist(x, y, o.x, o.y))
# Return orders sorted by distance
return render(request, 'orders/orders.html', dict(orders=fresh_order))
def user_coord_dist(x1, y1, x2, y2):
# Pythagoras
return math.sqrt(math.pow(x2-x1,2)+math.pow(y2-y1,2))
| import math
from django.shortcuts import render
from django.template import RequestContext
from django.http import HttpResponseRedirect
from orders.models import Order
def index(request, x, y):
if request.META.get('HTTP_REFERER', None) not in ('http://www.urbandead.com/map.cgi', 'http://urbandead.com/map.cgi'):
return HttpResponseRedirect('http://www.youtube.com/watch?v=lWKQiZVBtu4')
x = int(x)
y = int(y)
# Fetch the orders
fresh_order = list(Order.objects.all().order_by('-date'))
fresh_order.sort(key=lambda o: user_coord_dist(x, y, o.x, o.y))
# Return orders sorted by distance
return render(request, 'orders/orders.html', dict(orders=fresh_order))
def user_coord_dist(x1, y1, x2, y2):
# Pythagoras
return math.sqrt(math.pow(x2-x1,2)+math.pow(y2-y1,2))
| Test things first you big dummy | Test things first you big dummy
| Python | bsd-3-clause | crisisking/udbraaains,crisisking/udbraaains,crisisking/udbraaains,crisisking/udbraaains | import math
from django.shortcuts import render
from django.template import RequestContext
from django.http import HttpResponseRedirect
from orders.models import Order
def index(request, x, y):
if request.META['HTTP_REFERER'] not in ('http://www.urbandead.com/map.cgi', 'http://urbandead.com/map.cgi'):
return HttpResponseRedirect('http://www.youtube.com/watch?v=lWKQiZVBtu4')
x = int(x)
y = int(y)
# Fetch the orders
fresh_order = list(Order.objects.all().order_by('-date'))
fresh_order.sort(key=lambda o: user_coord_dist(x, y, o.x, o.y))
# Return orders sorted by distance
return render(request, 'orders/orders.html', dict(orders=fresh_order))
def user_coord_dist(x1, y1, x2, y2):
# Pythagoras
return math.sqrt(math.pow(x2-x1,2)+math.pow(y2-y1,2))
Test things first you big dummy | import math
from django.shortcuts import render
from django.template import RequestContext
from django.http import HttpResponseRedirect
from orders.models import Order
def index(request, x, y):
if request.META.get('HTTP_REFERER', None) not in ('http://www.urbandead.com/map.cgi', 'http://urbandead.com/map.cgi'):
return HttpResponseRedirect('http://www.youtube.com/watch?v=lWKQiZVBtu4')
x = int(x)
y = int(y)
# Fetch the orders
fresh_order = list(Order.objects.all().order_by('-date'))
fresh_order.sort(key=lambda o: user_coord_dist(x, y, o.x, o.y))
# Return orders sorted by distance
return render(request, 'orders/orders.html', dict(orders=fresh_order))
def user_coord_dist(x1, y1, x2, y2):
# Pythagoras
return math.sqrt(math.pow(x2-x1,2)+math.pow(y2-y1,2))
| <commit_before>import math
from django.shortcuts import render
from django.template import RequestContext
from django.http import HttpResponseRedirect
from orders.models import Order
def index(request, x, y):
if request.META['HTTP_REFERER'] not in ('http://www.urbandead.com/map.cgi', 'http://urbandead.com/map.cgi'):
return HttpResponseRedirect('http://www.youtube.com/watch?v=lWKQiZVBtu4')
x = int(x)
y = int(y)
# Fetch the orders
fresh_order = list(Order.objects.all().order_by('-date'))
fresh_order.sort(key=lambda o: user_coord_dist(x, y, o.x, o.y))
# Return orders sorted by distance
return render(request, 'orders/orders.html', dict(orders=fresh_order))
def user_coord_dist(x1, y1, x2, y2):
# Pythagoras
return math.sqrt(math.pow(x2-x1,2)+math.pow(y2-y1,2))
<commit_msg>Test things first you big dummy<commit_after> | import math
from django.shortcuts import render
from django.template import RequestContext
from django.http import HttpResponseRedirect
from orders.models import Order
def index(request, x, y):
if request.META.get('HTTP_REFERER', None) not in ('http://www.urbandead.com/map.cgi', 'http://urbandead.com/map.cgi'):
return HttpResponseRedirect('http://www.youtube.com/watch?v=lWKQiZVBtu4')
x = int(x)
y = int(y)
# Fetch the orders
fresh_order = list(Order.objects.all().order_by('-date'))
fresh_order.sort(key=lambda o: user_coord_dist(x, y, o.x, o.y))
# Return orders sorted by distance
return render(request, 'orders/orders.html', dict(orders=fresh_order))
def user_coord_dist(x1, y1, x2, y2):
# Pythagoras
return math.sqrt(math.pow(x2-x1,2)+math.pow(y2-y1,2))
| import math
from django.shortcuts import render
from django.template import RequestContext
from django.http import HttpResponseRedirect
from orders.models import Order
def index(request, x, y):
if request.META['HTTP_REFERER'] not in ('http://www.urbandead.com/map.cgi', 'http://urbandead.com/map.cgi'):
return HttpResponseRedirect('http://www.youtube.com/watch?v=lWKQiZVBtu4')
x = int(x)
y = int(y)
# Fetch the orders
fresh_order = list(Order.objects.all().order_by('-date'))
fresh_order.sort(key=lambda o: user_coord_dist(x, y, o.x, o.y))
# Return orders sorted by distance
return render(request, 'orders/orders.html', dict(orders=fresh_order))
def user_coord_dist(x1, y1, x2, y2):
# Pythagoras
return math.sqrt(math.pow(x2-x1,2)+math.pow(y2-y1,2))
Test things first you big dummyimport math
from django.shortcuts import render
from django.template import RequestContext
from django.http import HttpResponseRedirect
from orders.models import Order
def index(request, x, y):
if request.META.get('HTTP_REFERER', None) not in ('http://www.urbandead.com/map.cgi', 'http://urbandead.com/map.cgi'):
return HttpResponseRedirect('http://www.youtube.com/watch?v=lWKQiZVBtu4')
x = int(x)
y = int(y)
# Fetch the orders
fresh_order = list(Order.objects.all().order_by('-date'))
fresh_order.sort(key=lambda o: user_coord_dist(x, y, o.x, o.y))
# Return orders sorted by distance
return render(request, 'orders/orders.html', dict(orders=fresh_order))
def user_coord_dist(x1, y1, x2, y2):
# Pythagoras
return math.sqrt(math.pow(x2-x1,2)+math.pow(y2-y1,2))
| <commit_before>import math
from django.shortcuts import render
from django.template import RequestContext
from django.http import HttpResponseRedirect
from orders.models import Order
def index(request, x, y):
if request.META['HTTP_REFERER'] not in ('http://www.urbandead.com/map.cgi', 'http://urbandead.com/map.cgi'):
return HttpResponseRedirect('http://www.youtube.com/watch?v=lWKQiZVBtu4')
x = int(x)
y = int(y)
# Fetch the orders
fresh_order = list(Order.objects.all().order_by('-date'))
fresh_order.sort(key=lambda o: user_coord_dist(x, y, o.x, o.y))
# Return orders sorted by distance
return render(request, 'orders/orders.html', dict(orders=fresh_order))
def user_coord_dist(x1, y1, x2, y2):
# Pythagoras
return math.sqrt(math.pow(x2-x1,2)+math.pow(y2-y1,2))
<commit_msg>Test things first you big dummy<commit_after>import math
from django.shortcuts import render
from django.template import RequestContext
from django.http import HttpResponseRedirect
from orders.models import Order
def index(request, x, y):
if request.META.get('HTTP_REFERER', None) not in ('http://www.urbandead.com/map.cgi', 'http://urbandead.com/map.cgi'):
return HttpResponseRedirect('http://www.youtube.com/watch?v=lWKQiZVBtu4')
x = int(x)
y = int(y)
# Fetch the orders
fresh_order = list(Order.objects.all().order_by('-date'))
fresh_order.sort(key=lambda o: user_coord_dist(x, y, o.x, o.y))
# Return orders sorted by distance
return render(request, 'orders/orders.html', dict(orders=fresh_order))
def user_coord_dist(x1, y1, x2, y2):
# Pythagoras
return math.sqrt(math.pow(x2-x1,2)+math.pow(y2-y1,2))
|
81489c115704c5df83ef7607121c8c20ab2ab2b0 | packages/mono-llvm.py | packages/mono-llvm.py | GitHubTarballPackage ('mono', 'llvm', '3.0', '292aa8712c3120b03f9aa1d201b2e7949adf35c3',
configure = './configure --prefix="%{prefix}" --enable-optimized --enable-targets="x86 x86_64" --target=i386-apple-darwin10.8.0',
override_properties = { 'make': 'make' }
)
| GitHubTarballPackage ('mono', 'llvm', '3.0', '292aa8712c3120b03f9aa1d201b2e7949adf35c3',
configure = './configure --prefix="%{prefix}" --enable-optimized --enable-targets="x86 x86_64" --build=i386-apple-darwin10.8.0',
override_properties = { 'make': 'make' }
)
| Set --build instead of --target. | Set --build instead of --target.
| Python | mit | mono/bockbuild,BansheeMediaPlayer/bockbuild,BansheeMediaPlayer/bockbuild,mono/bockbuild,BansheeMediaPlayer/bockbuild | GitHubTarballPackage ('mono', 'llvm', '3.0', '292aa8712c3120b03f9aa1d201b2e7949adf35c3',
configure = './configure --prefix="%{prefix}" --enable-optimized --enable-targets="x86 x86_64" --target=i386-apple-darwin10.8.0',
override_properties = { 'make': 'make' }
)
Set --build instead of --target. | GitHubTarballPackage ('mono', 'llvm', '3.0', '292aa8712c3120b03f9aa1d201b2e7949adf35c3',
configure = './configure --prefix="%{prefix}" --enable-optimized --enable-targets="x86 x86_64" --build=i386-apple-darwin10.8.0',
override_properties = { 'make': 'make' }
)
| <commit_before>GitHubTarballPackage ('mono', 'llvm', '3.0', '292aa8712c3120b03f9aa1d201b2e7949adf35c3',
configure = './configure --prefix="%{prefix}" --enable-optimized --enable-targets="x86 x86_64" --target=i386-apple-darwin10.8.0',
override_properties = { 'make': 'make' }
)
<commit_msg>Set --build instead of --target.<commit_after> | GitHubTarballPackage ('mono', 'llvm', '3.0', '292aa8712c3120b03f9aa1d201b2e7949adf35c3',
configure = './configure --prefix="%{prefix}" --enable-optimized --enable-targets="x86 x86_64" --build=i386-apple-darwin10.8.0',
override_properties = { 'make': 'make' }
)
| GitHubTarballPackage ('mono', 'llvm', '3.0', '292aa8712c3120b03f9aa1d201b2e7949adf35c3',
configure = './configure --prefix="%{prefix}" --enable-optimized --enable-targets="x86 x86_64" --target=i386-apple-darwin10.8.0',
override_properties = { 'make': 'make' }
)
Set --build instead of --target.GitHubTarballPackage ('mono', 'llvm', '3.0', '292aa8712c3120b03f9aa1d201b2e7949adf35c3',
configure = './configure --prefix="%{prefix}" --enable-optimized --enable-targets="x86 x86_64" --build=i386-apple-darwin10.8.0',
override_properties = { 'make': 'make' }
)
| <commit_before>GitHubTarballPackage ('mono', 'llvm', '3.0', '292aa8712c3120b03f9aa1d201b2e7949adf35c3',
configure = './configure --prefix="%{prefix}" --enable-optimized --enable-targets="x86 x86_64" --target=i386-apple-darwin10.8.0',
override_properties = { 'make': 'make' }
)
<commit_msg>Set --build instead of --target.<commit_after>GitHubTarballPackage ('mono', 'llvm', '3.0', '292aa8712c3120b03f9aa1d201b2e7949adf35c3',
configure = './configure --prefix="%{prefix}" --enable-optimized --enable-targets="x86 x86_64" --build=i386-apple-darwin10.8.0',
override_properties = { 'make': 'make' }
)
|
9a7c84cab0931f2998af990200c4412f23cc2034 | scripts/run_unit_test.py | scripts/run_unit_test.py | #!/usr/bin/env python
import serial
import os
# Make and flash the unit test
FILE_LOCATION = os.path.dirname(os.path.abspath(__file__))
os.system("cd " + FILE_LOCATION + " ../")
print os.system("make flash_unit_test")
# Ask the user to reset the board
raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...")
# Open a serial port
ser = serial.Serial("/dev/serial/by-id/usb-PopaGator_Toad-if00", 115200)
# Send data to start USB OTG
ser.write("start")
# Read until we see the finished text
result = ''
try:
while True:
num_chars = ser.inWaiting()
if num_chars:
result += ser.read(num_chars)
if result.find("Finished") != -1:
break
finally:
# Print the result so the user can see and close the serial port
print result
ser.close() | #!/usr/bin/env python
import serial
import os
# Make and flash the unit test
FILE_LOCATION = os.path.dirname(os.path.abspath(__file__))
os.chdir(FILE_LOCATION + "/../")
print os.system("make flash_unit_test")
# Ask the user to reset the board
raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...")
# Open a serial port
ser = serial.Serial("/dev/serial/by-id/usb-PopaGator_Toad-if00", 115200)
# Send data to start USB OTG
ser.write("start")
# Read until we see the finished text
result = ''
try:
while True:
num_chars = ser.inWaiting()
if num_chars:
result += ser.read(num_chars)
if result.find("Finished") != -1:
break
finally:
# Print the result so the user can see and close the serial port
print result
ser.close() | Add ability to run unit test script from anywhere | UNIT_TEST: Add ability to run unit test script from anywhere
| Python | mit | fnivek/Pop-a-Gator,fnivek/Pop-a-Gator,fnivek/Pop-a-Gator | #!/usr/bin/env python
import serial
import os
# Make and flash the unit test
FILE_LOCATION = os.path.dirname(os.path.abspath(__file__))
os.system("cd " + FILE_LOCATION + " ../")
print os.system("make flash_unit_test")
# Ask the user to reset the board
raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...")
# Open a serial port
ser = serial.Serial("/dev/serial/by-id/usb-PopaGator_Toad-if00", 115200)
# Send data to start USB OTG
ser.write("start")
# Read until we see the finished text
result = ''
try:
while True:
num_chars = ser.inWaiting()
if num_chars:
result += ser.read(num_chars)
if result.find("Finished") != -1:
break
finally:
# Print the result so the user can see and close the serial port
print result
ser.close()UNIT_TEST: Add ability to run unit test script from anywhere | #!/usr/bin/env python
import serial
import os
# Make and flash the unit test
FILE_LOCATION = os.path.dirname(os.path.abspath(__file__))
os.chdir(FILE_LOCATION + "/../")
print os.system("make flash_unit_test")
# Ask the user to reset the board
raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...")
# Open a serial port
ser = serial.Serial("/dev/serial/by-id/usb-PopaGator_Toad-if00", 115200)
# Send data to start USB OTG
ser.write("start")
# Read until we see the finished text
result = ''
try:
while True:
num_chars = ser.inWaiting()
if num_chars:
result += ser.read(num_chars)
if result.find("Finished") != -1:
break
finally:
# Print the result so the user can see and close the serial port
print result
ser.close() | <commit_before>#!/usr/bin/env python
import serial
import os
# Make and flash the unit test
FILE_LOCATION = os.path.dirname(os.path.abspath(__file__))
os.system("cd " + FILE_LOCATION + " ../")
print os.system("make flash_unit_test")
# Ask the user to reset the board
raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...")
# Open a serial port
ser = serial.Serial("/dev/serial/by-id/usb-PopaGator_Toad-if00", 115200)
# Send data to start USB OTG
ser.write("start")
# Read until we see the finished text
result = ''
try:
while True:
num_chars = ser.inWaiting()
if num_chars:
result += ser.read(num_chars)
if result.find("Finished") != -1:
break
finally:
# Print the result so the user can see and close the serial port
print result
ser.close()<commit_msg>UNIT_TEST: Add ability to run unit test script from anywhere<commit_after> | #!/usr/bin/env python
import serial
import os
# Make and flash the unit test
FILE_LOCATION = os.path.dirname(os.path.abspath(__file__))
os.chdir(FILE_LOCATION + "/../")
print os.system("make flash_unit_test")
# Ask the user to reset the board
raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...")
# Open a serial port
ser = serial.Serial("/dev/serial/by-id/usb-PopaGator_Toad-if00", 115200)
# Send data to start USB OTG
ser.write("start")
# Read until we see the finished text
result = ''
try:
while True:
num_chars = ser.inWaiting()
if num_chars:
result += ser.read(num_chars)
if result.find("Finished") != -1:
break
finally:
# Print the result so the user can see and close the serial port
print result
ser.close() | #!/usr/bin/env python
import serial
import os
# Make and flash the unit test
FILE_LOCATION = os.path.dirname(os.path.abspath(__file__))
os.system("cd " + FILE_LOCATION + " ../")
print os.system("make flash_unit_test")
# Ask the user to reset the board
raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...")
# Open a serial port
ser = serial.Serial("/dev/serial/by-id/usb-PopaGator_Toad-if00", 115200)
# Send data to start USB OTG
ser.write("start")
# Read until we see the finished text
result = ''
try:
while True:
num_chars = ser.inWaiting()
if num_chars:
result += ser.read(num_chars)
if result.find("Finished") != -1:
break
finally:
# Print the result so the user can see and close the serial port
print result
ser.close()UNIT_TEST: Add ability to run unit test script from anywhere#!/usr/bin/env python
import serial
import os
# Make and flash the unit test
FILE_LOCATION = os.path.dirname(os.path.abspath(__file__))
os.chdir(FILE_LOCATION + "/../")
print os.system("make flash_unit_test")
# Ask the user to reset the board
raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...")
# Open a serial port
ser = serial.Serial("/dev/serial/by-id/usb-PopaGator_Toad-if00", 115200)
# Send data to start USB OTG
ser.write("start")
# Read until we see the finished text
result = ''
try:
while True:
num_chars = ser.inWaiting()
if num_chars:
result += ser.read(num_chars)
if result.find("Finished") != -1:
break
finally:
# Print the result so the user can see and close the serial port
print result
ser.close() | <commit_before>#!/usr/bin/env python
import serial
import os
# Make and flash the unit test
FILE_LOCATION = os.path.dirname(os.path.abspath(__file__))
os.system("cd " + FILE_LOCATION + " ../")
print os.system("make flash_unit_test")
# Ask the user to reset the board
raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...")
# Open a serial port
ser = serial.Serial("/dev/serial/by-id/usb-PopaGator_Toad-if00", 115200)
# Send data to start USB OTG
ser.write("start")
# Read until we see the finished text
result = ''
try:
while True:
num_chars = ser.inWaiting()
if num_chars:
result += ser.read(num_chars)
if result.find("Finished") != -1:
break
finally:
# Print the result so the user can see and close the serial port
print result
ser.close()<commit_msg>UNIT_TEST: Add ability to run unit test script from anywhere<commit_after>#!/usr/bin/env python
import serial
import os
# Make and flash the unit test
FILE_LOCATION = os.path.dirname(os.path.abspath(__file__))
os.chdir(FILE_LOCATION + "/../")
print os.system("make flash_unit_test")
# Ask the user to reset the board
raw_input("\nPlease press the phsyical reset button on the STM32F4Discovery board and then press enter to continue...")
# Open a serial port
ser = serial.Serial("/dev/serial/by-id/usb-PopaGator_Toad-if00", 115200)
# Send data to start USB OTG
ser.write("start")
# Read until we see the finished text
result = ''
try:
while True:
num_chars = ser.inWaiting()
if num_chars:
result += ser.read(num_chars)
if result.find("Finished") != -1:
break
finally:
# Print the result so the user can see and close the serial port
print result
ser.close() |
e2c3c9f50f3bdb537ef863d7cff80d4fd5e27911 | test/test_api.py | test/test_api.py | import unittest
import sys
import appdirs
if sys.version_info[0] < 3:
STRING_TYPE = basestring
else:
STRING_TYPE = str
class Test_AppDir(unittest.TestCase):
def test_metadata(self):
self.assertTrue(hasattr(appdirs, "__version__"))
self.assertTrue(hasattr(appdirs, "__version_info__"))
def test_helpers(self):
self.assertIsInstance(
appdirs.user_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.site_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_cache_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_log_dir('MyApp', 'MyCompany'), STRING_TYPE)
def test_dirs(self):
dirs = appdirs.AppDirs('MyApp', 'MyCompany', version='1.0')
self.assertIsInstance(dirs.user_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.site_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_cache_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_log_dir, STRING_TYPE)
if __name__ == "__main__":
unittest.main()
| import sys
import appdirs
if sys.version_info < (2, 7):
import unittest2 as unittest
else:
import unittest
if sys.version_info[0] < 3:
STRING_TYPE = basestring
else:
STRING_TYPE = str
class Test_AppDir(unittest.TestCase):
def test_metadata(self):
self.assertTrue(hasattr(appdirs, "__version__"))
self.assertTrue(hasattr(appdirs, "__version_info__"))
def test_helpers(self):
self.assertIsInstance(
appdirs.user_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.site_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_cache_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_log_dir('MyApp', 'MyCompany'), STRING_TYPE)
def test_dirs(self):
dirs = appdirs.AppDirs('MyApp', 'MyCompany', version='1.0')
self.assertIsInstance(dirs.user_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.site_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_cache_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_log_dir, STRING_TYPE)
if __name__ == "__main__":
unittest.main()
| Use unittest2 for Python < 2.7. | Use unittest2 for Python < 2.7.
| Python | mit | platformdirs/platformdirs | import unittest
import sys
import appdirs
if sys.version_info[0] < 3:
STRING_TYPE = basestring
else:
STRING_TYPE = str
class Test_AppDir(unittest.TestCase):
def test_metadata(self):
self.assertTrue(hasattr(appdirs, "__version__"))
self.assertTrue(hasattr(appdirs, "__version_info__"))
def test_helpers(self):
self.assertIsInstance(
appdirs.user_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.site_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_cache_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_log_dir('MyApp', 'MyCompany'), STRING_TYPE)
def test_dirs(self):
dirs = appdirs.AppDirs('MyApp', 'MyCompany', version='1.0')
self.assertIsInstance(dirs.user_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.site_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_cache_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_log_dir, STRING_TYPE)
if __name__ == "__main__":
unittest.main()
Use unittest2 for Python < 2.7. | import sys
import appdirs
if sys.version_info < (2, 7):
import unittest2 as unittest
else:
import unittest
if sys.version_info[0] < 3:
STRING_TYPE = basestring
else:
STRING_TYPE = str
class Test_AppDir(unittest.TestCase):
def test_metadata(self):
self.assertTrue(hasattr(appdirs, "__version__"))
self.assertTrue(hasattr(appdirs, "__version_info__"))
def test_helpers(self):
self.assertIsInstance(
appdirs.user_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.site_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_cache_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_log_dir('MyApp', 'MyCompany'), STRING_TYPE)
def test_dirs(self):
dirs = appdirs.AppDirs('MyApp', 'MyCompany', version='1.0')
self.assertIsInstance(dirs.user_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.site_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_cache_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_log_dir, STRING_TYPE)
if __name__ == "__main__":
unittest.main()
| <commit_before>import unittest
import sys
import appdirs
if sys.version_info[0] < 3:
STRING_TYPE = basestring
else:
STRING_TYPE = str
class Test_AppDir(unittest.TestCase):
def test_metadata(self):
self.assertTrue(hasattr(appdirs, "__version__"))
self.assertTrue(hasattr(appdirs, "__version_info__"))
def test_helpers(self):
self.assertIsInstance(
appdirs.user_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.site_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_cache_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_log_dir('MyApp', 'MyCompany'), STRING_TYPE)
def test_dirs(self):
dirs = appdirs.AppDirs('MyApp', 'MyCompany', version='1.0')
self.assertIsInstance(dirs.user_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.site_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_cache_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_log_dir, STRING_TYPE)
if __name__ == "__main__":
unittest.main()
<commit_msg>Use unittest2 for Python < 2.7.<commit_after> | import sys
import appdirs
if sys.version_info < (2, 7):
import unittest2 as unittest
else:
import unittest
if sys.version_info[0] < 3:
STRING_TYPE = basestring
else:
STRING_TYPE = str
class Test_AppDir(unittest.TestCase):
def test_metadata(self):
self.assertTrue(hasattr(appdirs, "__version__"))
self.assertTrue(hasattr(appdirs, "__version_info__"))
def test_helpers(self):
self.assertIsInstance(
appdirs.user_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.site_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_cache_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_log_dir('MyApp', 'MyCompany'), STRING_TYPE)
def test_dirs(self):
dirs = appdirs.AppDirs('MyApp', 'MyCompany', version='1.0')
self.assertIsInstance(dirs.user_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.site_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_cache_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_log_dir, STRING_TYPE)
if __name__ == "__main__":
unittest.main()
| import unittest
import sys
import appdirs
if sys.version_info[0] < 3:
STRING_TYPE = basestring
else:
STRING_TYPE = str
class Test_AppDir(unittest.TestCase):
def test_metadata(self):
self.assertTrue(hasattr(appdirs, "__version__"))
self.assertTrue(hasattr(appdirs, "__version_info__"))
def test_helpers(self):
self.assertIsInstance(
appdirs.user_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.site_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_cache_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_log_dir('MyApp', 'MyCompany'), STRING_TYPE)
def test_dirs(self):
dirs = appdirs.AppDirs('MyApp', 'MyCompany', version='1.0')
self.assertIsInstance(dirs.user_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.site_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_cache_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_log_dir, STRING_TYPE)
if __name__ == "__main__":
unittest.main()
Use unittest2 for Python < 2.7.import sys
import appdirs
if sys.version_info < (2, 7):
import unittest2 as unittest
else:
import unittest
if sys.version_info[0] < 3:
STRING_TYPE = basestring
else:
STRING_TYPE = str
class Test_AppDir(unittest.TestCase):
def test_metadata(self):
self.assertTrue(hasattr(appdirs, "__version__"))
self.assertTrue(hasattr(appdirs, "__version_info__"))
def test_helpers(self):
self.assertIsInstance(
appdirs.user_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.site_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_cache_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_log_dir('MyApp', 'MyCompany'), STRING_TYPE)
def test_dirs(self):
dirs = appdirs.AppDirs('MyApp', 'MyCompany', version='1.0')
self.assertIsInstance(dirs.user_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.site_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_cache_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_log_dir, STRING_TYPE)
if __name__ == "__main__":
unittest.main()
| <commit_before>import unittest
import sys
import appdirs
if sys.version_info[0] < 3:
STRING_TYPE = basestring
else:
STRING_TYPE = str
class Test_AppDir(unittest.TestCase):
def test_metadata(self):
self.assertTrue(hasattr(appdirs, "__version__"))
self.assertTrue(hasattr(appdirs, "__version_info__"))
def test_helpers(self):
self.assertIsInstance(
appdirs.user_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.site_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_cache_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_log_dir('MyApp', 'MyCompany'), STRING_TYPE)
def test_dirs(self):
dirs = appdirs.AppDirs('MyApp', 'MyCompany', version='1.0')
self.assertIsInstance(dirs.user_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.site_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_cache_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_log_dir, STRING_TYPE)
if __name__ == "__main__":
unittest.main()
<commit_msg>Use unittest2 for Python < 2.7.<commit_after>import sys
import appdirs
if sys.version_info < (2, 7):
import unittest2 as unittest
else:
import unittest
if sys.version_info[0] < 3:
STRING_TYPE = basestring
else:
STRING_TYPE = str
class Test_AppDir(unittest.TestCase):
def test_metadata(self):
self.assertTrue(hasattr(appdirs, "__version__"))
self.assertTrue(hasattr(appdirs, "__version_info__"))
def test_helpers(self):
self.assertIsInstance(
appdirs.user_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.site_data_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_cache_dir('MyApp', 'MyCompany'), STRING_TYPE)
self.assertIsInstance(
appdirs.user_log_dir('MyApp', 'MyCompany'), STRING_TYPE)
def test_dirs(self):
dirs = appdirs.AppDirs('MyApp', 'MyCompany', version='1.0')
self.assertIsInstance(dirs.user_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.site_data_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_cache_dir, STRING_TYPE)
self.assertIsInstance(dirs.user_log_dir, STRING_TYPE)
if __name__ == "__main__":
unittest.main()
|
53fa37b1e8a97c214a0a3c1f95be53dbe4d3d442 | comics/comics/wumovg.py | comics/comics/wumovg.py | from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = 'Wulffmorgenthaler (vg.no)'
language = 'no'
url = 'http://heltnormalt.no/wumo'
rights = 'Mikael Wulff & Anders Morgenthaler'
class Crawler(CrawlerBase):
history_capable_date = '2013-01-26'
schedule = 'Mo,Tu,We,Th,Fr,Sa,Su'
time_zone = 'Europe/Oslo'
def crawl(self, pub_date):
url = 'http://heltnormalt.no/img/wumo/%s.jpg' % (
pub_date.strftime('%Y/%m/%d'))
return CrawlerImage(url)
| from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = 'Wumo (vg.no)'
language = 'no'
url = 'http://heltnormalt.no/wumo'
rights = 'Mikael Wulff & Anders Morgenthaler'
class Crawler(CrawlerBase):
history_capable_date = '2013-01-26'
schedule = 'Mo,Tu,We,Th,Fr,Sa,Su'
time_zone = 'Europe/Oslo'
def crawl(self, pub_date):
url = 'http://heltnormalt.no/img/wumo/%s.jpg' % (
pub_date.strftime('%Y/%m/%d'))
return CrawlerImage(url)
| Update title of 'Wumo' crawlers, part two | Update title of 'Wumo' crawlers, part two
| Python | agpl-3.0 | jodal/comics,datagutten/comics,jodal/comics,jodal/comics,datagutten/comics,datagutten/comics,jodal/comics,datagutten/comics | from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = 'Wulffmorgenthaler (vg.no)'
language = 'no'
url = 'http://heltnormalt.no/wumo'
rights = 'Mikael Wulff & Anders Morgenthaler'
class Crawler(CrawlerBase):
history_capable_date = '2013-01-26'
schedule = 'Mo,Tu,We,Th,Fr,Sa,Su'
time_zone = 'Europe/Oslo'
def crawl(self, pub_date):
url = 'http://heltnormalt.no/img/wumo/%s.jpg' % (
pub_date.strftime('%Y/%m/%d'))
return CrawlerImage(url)
Update title of 'Wumo' crawlers, part two | from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = 'Wumo (vg.no)'
language = 'no'
url = 'http://heltnormalt.no/wumo'
rights = 'Mikael Wulff & Anders Morgenthaler'
class Crawler(CrawlerBase):
history_capable_date = '2013-01-26'
schedule = 'Mo,Tu,We,Th,Fr,Sa,Su'
time_zone = 'Europe/Oslo'
def crawl(self, pub_date):
url = 'http://heltnormalt.no/img/wumo/%s.jpg' % (
pub_date.strftime('%Y/%m/%d'))
return CrawlerImage(url)
| <commit_before>from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = 'Wulffmorgenthaler (vg.no)'
language = 'no'
url = 'http://heltnormalt.no/wumo'
rights = 'Mikael Wulff & Anders Morgenthaler'
class Crawler(CrawlerBase):
history_capable_date = '2013-01-26'
schedule = 'Mo,Tu,We,Th,Fr,Sa,Su'
time_zone = 'Europe/Oslo'
def crawl(self, pub_date):
url = 'http://heltnormalt.no/img/wumo/%s.jpg' % (
pub_date.strftime('%Y/%m/%d'))
return CrawlerImage(url)
<commit_msg>Update title of 'Wumo' crawlers, part two<commit_after> | from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = 'Wumo (vg.no)'
language = 'no'
url = 'http://heltnormalt.no/wumo'
rights = 'Mikael Wulff & Anders Morgenthaler'
class Crawler(CrawlerBase):
history_capable_date = '2013-01-26'
schedule = 'Mo,Tu,We,Th,Fr,Sa,Su'
time_zone = 'Europe/Oslo'
def crawl(self, pub_date):
url = 'http://heltnormalt.no/img/wumo/%s.jpg' % (
pub_date.strftime('%Y/%m/%d'))
return CrawlerImage(url)
| from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = 'Wulffmorgenthaler (vg.no)'
language = 'no'
url = 'http://heltnormalt.no/wumo'
rights = 'Mikael Wulff & Anders Morgenthaler'
class Crawler(CrawlerBase):
history_capable_date = '2013-01-26'
schedule = 'Mo,Tu,We,Th,Fr,Sa,Su'
time_zone = 'Europe/Oslo'
def crawl(self, pub_date):
url = 'http://heltnormalt.no/img/wumo/%s.jpg' % (
pub_date.strftime('%Y/%m/%d'))
return CrawlerImage(url)
Update title of 'Wumo' crawlers, part twofrom comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = 'Wumo (vg.no)'
language = 'no'
url = 'http://heltnormalt.no/wumo'
rights = 'Mikael Wulff & Anders Morgenthaler'
class Crawler(CrawlerBase):
history_capable_date = '2013-01-26'
schedule = 'Mo,Tu,We,Th,Fr,Sa,Su'
time_zone = 'Europe/Oslo'
def crawl(self, pub_date):
url = 'http://heltnormalt.no/img/wumo/%s.jpg' % (
pub_date.strftime('%Y/%m/%d'))
return CrawlerImage(url)
| <commit_before>from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = 'Wulffmorgenthaler (vg.no)'
language = 'no'
url = 'http://heltnormalt.no/wumo'
rights = 'Mikael Wulff & Anders Morgenthaler'
class Crawler(CrawlerBase):
history_capable_date = '2013-01-26'
schedule = 'Mo,Tu,We,Th,Fr,Sa,Su'
time_zone = 'Europe/Oslo'
def crawl(self, pub_date):
url = 'http://heltnormalt.no/img/wumo/%s.jpg' % (
pub_date.strftime('%Y/%m/%d'))
return CrawlerImage(url)
<commit_msg>Update title of 'Wumo' crawlers, part two<commit_after>from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = 'Wumo (vg.no)'
language = 'no'
url = 'http://heltnormalt.no/wumo'
rights = 'Mikael Wulff & Anders Morgenthaler'
class Crawler(CrawlerBase):
history_capable_date = '2013-01-26'
schedule = 'Mo,Tu,We,Th,Fr,Sa,Su'
time_zone = 'Europe/Oslo'
def crawl(self, pub_date):
url = 'http://heltnormalt.no/img/wumo/%s.jpg' % (
pub_date.strftime('%Y/%m/%d'))
return CrawlerImage(url)
|
534066b1228bb0070c1d62445155afa696a37921 | contrail_provisioning/config/templates/contrail_plugin_ini.py | contrail_provisioning/config/templates/contrail_plugin_ini.py | import string
template = string.Template("""
[APISERVER]
api_server_ip = $__contrail_api_server_ip__
api_server_port = $__contrail_api_server_port__
multi_tenancy = $__contrail_multi_tenancy__
#use_ssl = False
#insecure = False
#certfile=$__contrail_api_server_cert_file__
#keyfile=$__contrail_api_server_key_file__
#cafile=$__contrail_api_server_ca_file__
contrail_extensions = ipam:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_ipam.NeutronPluginContrailIpam,policy:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_policy.NeutronPluginContrailPolicy,route-table:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_vpc.NeutronPluginContrailVpc,contrail:None
[COLLECTOR]
analytics_api_ip = $__contrail_analytics_server_ip__
analytics_api_port = $__contrail_analytics_server_port__
[KEYSTONE]
auth_url = $__contrail_ks_auth_protocol__://$__contrail_keystone_ip__:$__contrail_ks_auth_port__/v2.0
admin_user=$__contrail_admin_user__
admin_password=$__contrail_admin_password__
admin_tenant_name=$__contrail_admin_tenant_name__
""")
| import string
template = string.Template("""
[APISERVER]
api_server_ip = $__contrail_api_server_ip__
api_server_port = $__contrail_api_server_port__
multi_tenancy = $__contrail_multi_tenancy__
#use_ssl = False
#insecure = False
#certfile=$__contrail_api_server_cert_file__
#keyfile=$__contrail_api_server_key_file__
#cafile=$__contrail_api_server_ca_file__
contrail_extensions = ipam:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_ipam.NeutronPluginContrailIpam,policy:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_policy.NeutronPluginContrailPolicy,route-table:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_vpc.NeutronPluginContrailVpc,contrail:None,service-interface:None,vf-binding:None
[COLLECTOR]
analytics_api_ip = $__contrail_analytics_server_ip__
analytics_api_port = $__contrail_analytics_server_port__
[KEYSTONE]
auth_url = $__contrail_ks_auth_protocol__://$__contrail_keystone_ip__:$__contrail_ks_auth_port__/v2.0
admin_user=$__contrail_admin_user__
admin_password=$__contrail_admin_password__
admin_tenant_name=$__contrail_admin_tenant_name__
""")
| Enable service-interface and vf-binding extensions by default in contrail based provisioning. | Enable service-interface and vf-binding extensions by default in
contrail based provisioning.
Change-Id: I5916f41cdf12ad54e74c0f76de244ed60f57aea5
Partial-Bug: 1556336
| Python | apache-2.0 | Juniper/contrail-provisioning,Juniper/contrail-provisioning | import string
template = string.Template("""
[APISERVER]
api_server_ip = $__contrail_api_server_ip__
api_server_port = $__contrail_api_server_port__
multi_tenancy = $__contrail_multi_tenancy__
#use_ssl = False
#insecure = False
#certfile=$__contrail_api_server_cert_file__
#keyfile=$__contrail_api_server_key_file__
#cafile=$__contrail_api_server_ca_file__
contrail_extensions = ipam:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_ipam.NeutronPluginContrailIpam,policy:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_policy.NeutronPluginContrailPolicy,route-table:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_vpc.NeutronPluginContrailVpc,contrail:None
[COLLECTOR]
analytics_api_ip = $__contrail_analytics_server_ip__
analytics_api_port = $__contrail_analytics_server_port__
[KEYSTONE]
auth_url = $__contrail_ks_auth_protocol__://$__contrail_keystone_ip__:$__contrail_ks_auth_port__/v2.0
admin_user=$__contrail_admin_user__
admin_password=$__contrail_admin_password__
admin_tenant_name=$__contrail_admin_tenant_name__
""")
Enable service-interface and vf-binding extensions by default in
contrail based provisioning.
Change-Id: I5916f41cdf12ad54e74c0f76de244ed60f57aea5
Partial-Bug: 1556336 | import string
template = string.Template("""
[APISERVER]
api_server_ip = $__contrail_api_server_ip__
api_server_port = $__contrail_api_server_port__
multi_tenancy = $__contrail_multi_tenancy__
#use_ssl = False
#insecure = False
#certfile=$__contrail_api_server_cert_file__
#keyfile=$__contrail_api_server_key_file__
#cafile=$__contrail_api_server_ca_file__
contrail_extensions = ipam:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_ipam.NeutronPluginContrailIpam,policy:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_policy.NeutronPluginContrailPolicy,route-table:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_vpc.NeutronPluginContrailVpc,contrail:None,service-interface:None,vf-binding:None
[COLLECTOR]
analytics_api_ip = $__contrail_analytics_server_ip__
analytics_api_port = $__contrail_analytics_server_port__
[KEYSTONE]
auth_url = $__contrail_ks_auth_protocol__://$__contrail_keystone_ip__:$__contrail_ks_auth_port__/v2.0
admin_user=$__contrail_admin_user__
admin_password=$__contrail_admin_password__
admin_tenant_name=$__contrail_admin_tenant_name__
""")
| <commit_before>import string
template = string.Template("""
[APISERVER]
api_server_ip = $__contrail_api_server_ip__
api_server_port = $__contrail_api_server_port__
multi_tenancy = $__contrail_multi_tenancy__
#use_ssl = False
#insecure = False
#certfile=$__contrail_api_server_cert_file__
#keyfile=$__contrail_api_server_key_file__
#cafile=$__contrail_api_server_ca_file__
contrail_extensions = ipam:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_ipam.NeutronPluginContrailIpam,policy:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_policy.NeutronPluginContrailPolicy,route-table:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_vpc.NeutronPluginContrailVpc,contrail:None
[COLLECTOR]
analytics_api_ip = $__contrail_analytics_server_ip__
analytics_api_port = $__contrail_analytics_server_port__
[KEYSTONE]
auth_url = $__contrail_ks_auth_protocol__://$__contrail_keystone_ip__:$__contrail_ks_auth_port__/v2.0
admin_user=$__contrail_admin_user__
admin_password=$__contrail_admin_password__
admin_tenant_name=$__contrail_admin_tenant_name__
""")
<commit_msg>Enable service-interface and vf-binding extensions by default in
contrail based provisioning.
Change-Id: I5916f41cdf12ad54e74c0f76de244ed60f57aea5
Partial-Bug: 1556336<commit_after> | import string
template = string.Template("""
[APISERVER]
api_server_ip = $__contrail_api_server_ip__
api_server_port = $__contrail_api_server_port__
multi_tenancy = $__contrail_multi_tenancy__
#use_ssl = False
#insecure = False
#certfile=$__contrail_api_server_cert_file__
#keyfile=$__contrail_api_server_key_file__
#cafile=$__contrail_api_server_ca_file__
contrail_extensions = ipam:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_ipam.NeutronPluginContrailIpam,policy:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_policy.NeutronPluginContrailPolicy,route-table:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_vpc.NeutronPluginContrailVpc,contrail:None,service-interface:None,vf-binding:None
[COLLECTOR]
analytics_api_ip = $__contrail_analytics_server_ip__
analytics_api_port = $__contrail_analytics_server_port__
[KEYSTONE]
auth_url = $__contrail_ks_auth_protocol__://$__contrail_keystone_ip__:$__contrail_ks_auth_port__/v2.0
admin_user=$__contrail_admin_user__
admin_password=$__contrail_admin_password__
admin_tenant_name=$__contrail_admin_tenant_name__
""")
| import string
template = string.Template("""
[APISERVER]
api_server_ip = $__contrail_api_server_ip__
api_server_port = $__contrail_api_server_port__
multi_tenancy = $__contrail_multi_tenancy__
#use_ssl = False
#insecure = False
#certfile=$__contrail_api_server_cert_file__
#keyfile=$__contrail_api_server_key_file__
#cafile=$__contrail_api_server_ca_file__
contrail_extensions = ipam:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_ipam.NeutronPluginContrailIpam,policy:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_policy.NeutronPluginContrailPolicy,route-table:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_vpc.NeutronPluginContrailVpc,contrail:None
[COLLECTOR]
analytics_api_ip = $__contrail_analytics_server_ip__
analytics_api_port = $__contrail_analytics_server_port__
[KEYSTONE]
auth_url = $__contrail_ks_auth_protocol__://$__contrail_keystone_ip__:$__contrail_ks_auth_port__/v2.0
admin_user=$__contrail_admin_user__
admin_password=$__contrail_admin_password__
admin_tenant_name=$__contrail_admin_tenant_name__
""")
Enable service-interface and vf-binding extensions by default in
contrail based provisioning.
Change-Id: I5916f41cdf12ad54e74c0f76de244ed60f57aea5
Partial-Bug: 1556336import string
template = string.Template("""
[APISERVER]
api_server_ip = $__contrail_api_server_ip__
api_server_port = $__contrail_api_server_port__
multi_tenancy = $__contrail_multi_tenancy__
#use_ssl = False
#insecure = False
#certfile=$__contrail_api_server_cert_file__
#keyfile=$__contrail_api_server_key_file__
#cafile=$__contrail_api_server_ca_file__
contrail_extensions = ipam:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_ipam.NeutronPluginContrailIpam,policy:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_policy.NeutronPluginContrailPolicy,route-table:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_vpc.NeutronPluginContrailVpc,contrail:None,service-interface:None,vf-binding:None
[COLLECTOR]
analytics_api_ip = $__contrail_analytics_server_ip__
analytics_api_port = $__contrail_analytics_server_port__
[KEYSTONE]
auth_url = $__contrail_ks_auth_protocol__://$__contrail_keystone_ip__:$__contrail_ks_auth_port__/v2.0
admin_user=$__contrail_admin_user__
admin_password=$__contrail_admin_password__
admin_tenant_name=$__contrail_admin_tenant_name__
""")
| <commit_before>import string
template = string.Template("""
[APISERVER]
api_server_ip = $__contrail_api_server_ip__
api_server_port = $__contrail_api_server_port__
multi_tenancy = $__contrail_multi_tenancy__
#use_ssl = False
#insecure = False
#certfile=$__contrail_api_server_cert_file__
#keyfile=$__contrail_api_server_key_file__
#cafile=$__contrail_api_server_ca_file__
contrail_extensions = ipam:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_ipam.NeutronPluginContrailIpam,policy:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_policy.NeutronPluginContrailPolicy,route-table:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_vpc.NeutronPluginContrailVpc,contrail:None
[COLLECTOR]
analytics_api_ip = $__contrail_analytics_server_ip__
analytics_api_port = $__contrail_analytics_server_port__
[KEYSTONE]
auth_url = $__contrail_ks_auth_protocol__://$__contrail_keystone_ip__:$__contrail_ks_auth_port__/v2.0
admin_user=$__contrail_admin_user__
admin_password=$__contrail_admin_password__
admin_tenant_name=$__contrail_admin_tenant_name__
""")
<commit_msg>Enable service-interface and vf-binding extensions by default in
contrail based provisioning.
Change-Id: I5916f41cdf12ad54e74c0f76de244ed60f57aea5
Partial-Bug: 1556336<commit_after>import string
template = string.Template("""
[APISERVER]
api_server_ip = $__contrail_api_server_ip__
api_server_port = $__contrail_api_server_port__
multi_tenancy = $__contrail_multi_tenancy__
#use_ssl = False
#insecure = False
#certfile=$__contrail_api_server_cert_file__
#keyfile=$__contrail_api_server_key_file__
#cafile=$__contrail_api_server_ca_file__
contrail_extensions = ipam:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_ipam.NeutronPluginContrailIpam,policy:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_policy.NeutronPluginContrailPolicy,route-table:neutron_plugin_contrail.plugins.opencontrail.contrail_plugin_vpc.NeutronPluginContrailVpc,contrail:None,service-interface:None,vf-binding:None
[COLLECTOR]
analytics_api_ip = $__contrail_analytics_server_ip__
analytics_api_port = $__contrail_analytics_server_port__
[KEYSTONE]
auth_url = $__contrail_ks_auth_protocol__://$__contrail_keystone_ip__:$__contrail_ks_auth_port__/v2.0
admin_user=$__contrail_admin_user__
admin_password=$__contrail_admin_password__
admin_tenant_name=$__contrail_admin_tenant_name__
""")
|
eb987c4ca71ec53db46c0a8afa4265f70671330d | geotrek/settings/env_tests.py | geotrek/settings/env_tests.py | #
# Django Tests
# ..........................
TEST = True
CELERY_ALWAYS_EAGER = True
TEST_EXCLUDE = ('django',)
INSTALLED_APPS += (
'geotrek.diving',
'geotrek.sensitivity',
'geotrek.outdoor',
)
LOGGING['handlers']['console']['level'] = 'CRITICAL'
LANGUAGE_CODE = 'en'
MODELTRANSLATION_DEFAULT_LANGUAGE = 'en'
MODELTRANSLATION_LANGUAGES = ('en', 'es', 'fr', 'it')
LAND_BBOX_AREAS_ENABLED = True
class DisableMigrations():
def __contains__(self, item):
return True
def __getitem__(self, item):
return None
MIGRATION_MODULES = DisableMigrations()
ADMINS = (
('test', 'test@test.com'),
)
MANAGERS = ADMINS
TEST_RUNNER = 'geotrek.test_runner.TestRunner'
| #
# Django Tests
# ..........................
TEST = True
CELERY_ALWAYS_EAGER = True
TEST_EXCLUDE = ('django',)
INSTALLED_APPS += (
'geotrek.diving',
'geotrek.sensitivity',
'geotrek.outdoor',
'drf_yasg',
)
LOGGING['handlers']['console']['level'] = 'CRITICAL'
LANGUAGE_CODE = 'en'
MODELTRANSLATION_DEFAULT_LANGUAGE = 'en'
MODELTRANSLATION_LANGUAGES = ('en', 'es', 'fr', 'it')
LAND_BBOX_AREAS_ENABLED = True
class DisableMigrations():
def __contains__(self, item):
return True
def __getitem__(self, item):
return None
MIGRATION_MODULES = DisableMigrations()
ADMINS = (
('test', 'test@test.com'),
)
MANAGERS = ADMINS
TEST_RUNNER = 'geotrek.test_runner.TestRunner'
| Enable drf_yasg in test settings | Enable drf_yasg in test settings
| Python | bsd-2-clause | GeotrekCE/Geotrek-admin,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek,makinacorpus/Geotrek,GeotrekCE/Geotrek-admin,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek,makinacorpus/Geotrek | #
# Django Tests
# ..........................
TEST = True
CELERY_ALWAYS_EAGER = True
TEST_EXCLUDE = ('django',)
INSTALLED_APPS += (
'geotrek.diving',
'geotrek.sensitivity',
'geotrek.outdoor',
)
LOGGING['handlers']['console']['level'] = 'CRITICAL'
LANGUAGE_CODE = 'en'
MODELTRANSLATION_DEFAULT_LANGUAGE = 'en'
MODELTRANSLATION_LANGUAGES = ('en', 'es', 'fr', 'it')
LAND_BBOX_AREAS_ENABLED = True
class DisableMigrations():
def __contains__(self, item):
return True
def __getitem__(self, item):
return None
MIGRATION_MODULES = DisableMigrations()
ADMINS = (
('test', 'test@test.com'),
)
MANAGERS = ADMINS
TEST_RUNNER = 'geotrek.test_runner.TestRunner'
Enable drf_yasg in test settings | #
# Django Tests
# ..........................
TEST = True
CELERY_ALWAYS_EAGER = True
TEST_EXCLUDE = ('django',)
INSTALLED_APPS += (
'geotrek.diving',
'geotrek.sensitivity',
'geotrek.outdoor',
'drf_yasg',
)
LOGGING['handlers']['console']['level'] = 'CRITICAL'
LANGUAGE_CODE = 'en'
MODELTRANSLATION_DEFAULT_LANGUAGE = 'en'
MODELTRANSLATION_LANGUAGES = ('en', 'es', 'fr', 'it')
LAND_BBOX_AREAS_ENABLED = True
class DisableMigrations():
def __contains__(self, item):
return True
def __getitem__(self, item):
return None
MIGRATION_MODULES = DisableMigrations()
ADMINS = (
('test', 'test@test.com'),
)
MANAGERS = ADMINS
TEST_RUNNER = 'geotrek.test_runner.TestRunner'
| <commit_before>#
# Django Tests
# ..........................
TEST = True
CELERY_ALWAYS_EAGER = True
TEST_EXCLUDE = ('django',)
INSTALLED_APPS += (
'geotrek.diving',
'geotrek.sensitivity',
'geotrek.outdoor',
)
LOGGING['handlers']['console']['level'] = 'CRITICAL'
LANGUAGE_CODE = 'en'
MODELTRANSLATION_DEFAULT_LANGUAGE = 'en'
MODELTRANSLATION_LANGUAGES = ('en', 'es', 'fr', 'it')
LAND_BBOX_AREAS_ENABLED = True
class DisableMigrations():
def __contains__(self, item):
return True
def __getitem__(self, item):
return None
MIGRATION_MODULES = DisableMigrations()
ADMINS = (
('test', 'test@test.com'),
)
MANAGERS = ADMINS
TEST_RUNNER = 'geotrek.test_runner.TestRunner'
<commit_msg>Enable drf_yasg in test settings<commit_after> | #
# Django Tests
# ..........................
TEST = True
CELERY_ALWAYS_EAGER = True
TEST_EXCLUDE = ('django',)
INSTALLED_APPS += (
'geotrek.diving',
'geotrek.sensitivity',
'geotrek.outdoor',
'drf_yasg',
)
LOGGING['handlers']['console']['level'] = 'CRITICAL'
LANGUAGE_CODE = 'en'
MODELTRANSLATION_DEFAULT_LANGUAGE = 'en'
MODELTRANSLATION_LANGUAGES = ('en', 'es', 'fr', 'it')
LAND_BBOX_AREAS_ENABLED = True
class DisableMigrations():
def __contains__(self, item):
return True
def __getitem__(self, item):
return None
MIGRATION_MODULES = DisableMigrations()
ADMINS = (
('test', 'test@test.com'),
)
MANAGERS = ADMINS
TEST_RUNNER = 'geotrek.test_runner.TestRunner'
| #
# Django Tests
# ..........................
TEST = True
CELERY_ALWAYS_EAGER = True
TEST_EXCLUDE = ('django',)
INSTALLED_APPS += (
'geotrek.diving',
'geotrek.sensitivity',
'geotrek.outdoor',
)
LOGGING['handlers']['console']['level'] = 'CRITICAL'
LANGUAGE_CODE = 'en'
MODELTRANSLATION_DEFAULT_LANGUAGE = 'en'
MODELTRANSLATION_LANGUAGES = ('en', 'es', 'fr', 'it')
LAND_BBOX_AREAS_ENABLED = True
class DisableMigrations():
def __contains__(self, item):
return True
def __getitem__(self, item):
return None
MIGRATION_MODULES = DisableMigrations()
ADMINS = (
('test', 'test@test.com'),
)
MANAGERS = ADMINS
TEST_RUNNER = 'geotrek.test_runner.TestRunner'
Enable drf_yasg in test settings#
# Django Tests
# ..........................
TEST = True
CELERY_ALWAYS_EAGER = True
TEST_EXCLUDE = ('django',)
INSTALLED_APPS += (
'geotrek.diving',
'geotrek.sensitivity',
'geotrek.outdoor',
'drf_yasg',
)
LOGGING['handlers']['console']['level'] = 'CRITICAL'
LANGUAGE_CODE = 'en'
MODELTRANSLATION_DEFAULT_LANGUAGE = 'en'
MODELTRANSLATION_LANGUAGES = ('en', 'es', 'fr', 'it')
LAND_BBOX_AREAS_ENABLED = True
class DisableMigrations():
def __contains__(self, item):
return True
def __getitem__(self, item):
return None
MIGRATION_MODULES = DisableMigrations()
ADMINS = (
('test', 'test@test.com'),
)
MANAGERS = ADMINS
TEST_RUNNER = 'geotrek.test_runner.TestRunner'
| <commit_before>#
# Django Tests
# ..........................
TEST = True
CELERY_ALWAYS_EAGER = True
TEST_EXCLUDE = ('django',)
INSTALLED_APPS += (
'geotrek.diving',
'geotrek.sensitivity',
'geotrek.outdoor',
)
LOGGING['handlers']['console']['level'] = 'CRITICAL'
LANGUAGE_CODE = 'en'
MODELTRANSLATION_DEFAULT_LANGUAGE = 'en'
MODELTRANSLATION_LANGUAGES = ('en', 'es', 'fr', 'it')
LAND_BBOX_AREAS_ENABLED = True
class DisableMigrations():
def __contains__(self, item):
return True
def __getitem__(self, item):
return None
MIGRATION_MODULES = DisableMigrations()
ADMINS = (
('test', 'test@test.com'),
)
MANAGERS = ADMINS
TEST_RUNNER = 'geotrek.test_runner.TestRunner'
<commit_msg>Enable drf_yasg in test settings<commit_after>#
# Django Tests
# ..........................
TEST = True
CELERY_ALWAYS_EAGER = True
TEST_EXCLUDE = ('django',)
INSTALLED_APPS += (
'geotrek.diving',
'geotrek.sensitivity',
'geotrek.outdoor',
'drf_yasg',
)
LOGGING['handlers']['console']['level'] = 'CRITICAL'
LANGUAGE_CODE = 'en'
MODELTRANSLATION_DEFAULT_LANGUAGE = 'en'
MODELTRANSLATION_LANGUAGES = ('en', 'es', 'fr', 'it')
LAND_BBOX_AREAS_ENABLED = True
class DisableMigrations():
def __contains__(self, item):
return True
def __getitem__(self, item):
return None
MIGRATION_MODULES = DisableMigrations()
ADMINS = (
('test', 'test@test.com'),
)
MANAGERS = ADMINS
TEST_RUNNER = 'geotrek.test_runner.TestRunner'
|
456de4b1184780b9179ee9e6572a3f62cf22550a | tests/test_tools/simple_project.py | tests/test_tools/simple_project.py | project_1_yaml = {
'common': {
'sources': ['sources/main.cpp'],
'includes': ['includes/header1.h'],
'target': ['mbed-lpc1768']
}
}
projects_1_yaml = {
'projects': {
'project_1' : ['test_workspace/project_1.yaml']
},
}
| project_1_yaml = {
'common': {
'sources': ['sources/main.cpp'],
'includes': ['includes/header1.h'],
'target': ['mbed-lpc1768'],
'linker_file': ['linker_script'],
}
}
projects_1_yaml = {
'projects': {
'project_1' : ['test_workspace/project_1.yaml']
},
}
| Test - add linker script for tools project | Test - add linker script for tools project
| Python | apache-2.0 | molejar/project_generator,hwfwgrp/project_generator,0xc0170/project_generator,sarahmarshy/project_generator,ohagendorf/project_generator,project-generator/project_generator | project_1_yaml = {
'common': {
'sources': ['sources/main.cpp'],
'includes': ['includes/header1.h'],
'target': ['mbed-lpc1768']
}
}
projects_1_yaml = {
'projects': {
'project_1' : ['test_workspace/project_1.yaml']
},
}
Test - add linker script for tools project | project_1_yaml = {
'common': {
'sources': ['sources/main.cpp'],
'includes': ['includes/header1.h'],
'target': ['mbed-lpc1768'],
'linker_file': ['linker_script'],
}
}
projects_1_yaml = {
'projects': {
'project_1' : ['test_workspace/project_1.yaml']
},
}
| <commit_before>project_1_yaml = {
'common': {
'sources': ['sources/main.cpp'],
'includes': ['includes/header1.h'],
'target': ['mbed-lpc1768']
}
}
projects_1_yaml = {
'projects': {
'project_1' : ['test_workspace/project_1.yaml']
},
}
<commit_msg>Test - add linker script for tools project<commit_after> | project_1_yaml = {
'common': {
'sources': ['sources/main.cpp'],
'includes': ['includes/header1.h'],
'target': ['mbed-lpc1768'],
'linker_file': ['linker_script'],
}
}
projects_1_yaml = {
'projects': {
'project_1' : ['test_workspace/project_1.yaml']
},
}
| project_1_yaml = {
'common': {
'sources': ['sources/main.cpp'],
'includes': ['includes/header1.h'],
'target': ['mbed-lpc1768']
}
}
projects_1_yaml = {
'projects': {
'project_1' : ['test_workspace/project_1.yaml']
},
}
Test - add linker script for tools projectproject_1_yaml = {
'common': {
'sources': ['sources/main.cpp'],
'includes': ['includes/header1.h'],
'target': ['mbed-lpc1768'],
'linker_file': ['linker_script'],
}
}
projects_1_yaml = {
'projects': {
'project_1' : ['test_workspace/project_1.yaml']
},
}
| <commit_before>project_1_yaml = {
'common': {
'sources': ['sources/main.cpp'],
'includes': ['includes/header1.h'],
'target': ['mbed-lpc1768']
}
}
projects_1_yaml = {
'projects': {
'project_1' : ['test_workspace/project_1.yaml']
},
}
<commit_msg>Test - add linker script for tools project<commit_after>project_1_yaml = {
'common': {
'sources': ['sources/main.cpp'],
'includes': ['includes/header1.h'],
'target': ['mbed-lpc1768'],
'linker_file': ['linker_script'],
}
}
projects_1_yaml = {
'projects': {
'project_1' : ['test_workspace/project_1.yaml']
},
}
|
eca27464cc2c23a84e56e1d432a080ca663d04fb | src/dicomweb_client/__init__.py | src/dicomweb_client/__init__.py | __version__ = '0.9.1'
from dicomweb_client.api import DICOMwebClient
| __version__ = '0.9.2'
from dicomweb_client.api import DICOMwebClient
| Increase version to 0.9.2 for release | Increase version to 0.9.2 for release
| Python | mit | MGHComputationalPathology/dicomweb-client | __version__ = '0.9.1'
from dicomweb_client.api import DICOMwebClient
Increase version to 0.9.2 for release | __version__ = '0.9.2'
from dicomweb_client.api import DICOMwebClient
| <commit_before>__version__ = '0.9.1'
from dicomweb_client.api import DICOMwebClient
<commit_msg>Increase version to 0.9.2 for release<commit_after> | __version__ = '0.9.2'
from dicomweb_client.api import DICOMwebClient
| __version__ = '0.9.1'
from dicomweb_client.api import DICOMwebClient
Increase version to 0.9.2 for release__version__ = '0.9.2'
from dicomweb_client.api import DICOMwebClient
| <commit_before>__version__ = '0.9.1'
from dicomweb_client.api import DICOMwebClient
<commit_msg>Increase version to 0.9.2 for release<commit_after>__version__ = '0.9.2'
from dicomweb_client.api import DICOMwebClient
|
47aeeaad68ea0c9246ec68b7a49f385a4b7fe9cf | socketio/policyserver.py | socketio/policyserver.py | from gevent.server import StreamServer
__all__ = ['FlashPolicyServer']
class FlashPolicyServer(StreamServer):
policy = """<?xml version="1.0"?><!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
<cross-domain-policy><allow-access-from domain="*" to-ports="*"/></cross-domain-policy>"""
def __init__(self, listener=None, backlog=None):
if listener is None:
listener = ('0.0.0.0', 10843)
StreamServer.__init__(self, listener=listener, backlog=backlog)
def handle(self, socket, address):
socket.sendall(self.policy)
| from gevent.server import StreamServer
__all__ = ['FlashPolicyServer']
class FlashPolicyServer(StreamServer):
policy = """<?xml version="1.0"?><!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
<cross-domain-policy><allow-access-from domain="*" to-ports="*"/></cross-domain-policy>"""
def __init__(self, listener=None, backlog=None):
if listener is None:
listener = ('0.0.0.0', 10843)
StreamServer.__init__(self, listener=listener, backlog=backlog)
def handle(self, sock, address):
# send and read functions should not wait longer than three seconds
sock.settimeout(3)
try:
# try to receive at most 128 bytes (`POLICYREQUEST` is shorter)
# Interestingly if we dont do this and we write to the spcket directly
# I am getting strange errors.
input = sock.recv(128)
if input.startswith(FlashPolicyServer.policyrequest):
sock.sendall(FlashPolicyServer.policy)
except socket.timeout:
pass
sock.close()
| Fix to make sure we dont get errors in gevent socket write call when we are writing the policy file back | Fix to make sure we dont get errors in gevent socket write call when we are writing the policy file back
Conflicts:
socketio/policyserver.py
| Python | bsd-3-clause | abourget/gevent-socketio,arnuschky/gevent-socketio,bobvandevijver/gevent-socketio,Eugeny/gevent-socketio,hzruandd/gevent-socketio,gutomaia/gevent-socketio,gutomaia/gevent-socketio,yacneyac/gevent-socketio,smurfix/gevent-socketio,gutomaia/gevent-socketio,smurfix/gevent-socketio,Eugeny/gevent-socketio,kazmiruk/gevent-socketio,arnuschky/gevent-socketio,kazmiruk/gevent-socketio,abourget/gevent-socketio,smurfix/gevent-socketio,theskumar-archive/gevent-socketio,theskumar-archive/gevent-socketio,bobvandevijver/gevent-socketio,hzruandd/gevent-socketio,yacneyac/gevent-socketio | from gevent.server import StreamServer
__all__ = ['FlashPolicyServer']
class FlashPolicyServer(StreamServer):
policy = """<?xml version="1.0"?><!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
<cross-domain-policy><allow-access-from domain="*" to-ports="*"/></cross-domain-policy>"""
def __init__(self, listener=None, backlog=None):
if listener is None:
listener = ('0.0.0.0', 10843)
StreamServer.__init__(self, listener=listener, backlog=backlog)
def handle(self, socket, address):
socket.sendall(self.policy)
Fix to make sure we dont get errors in gevent socket write call when we are writing the policy file back
Conflicts:
socketio/policyserver.py | from gevent.server import StreamServer
__all__ = ['FlashPolicyServer']
class FlashPolicyServer(StreamServer):
policy = """<?xml version="1.0"?><!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
<cross-domain-policy><allow-access-from domain="*" to-ports="*"/></cross-domain-policy>"""
def __init__(self, listener=None, backlog=None):
if listener is None:
listener = ('0.0.0.0', 10843)
StreamServer.__init__(self, listener=listener, backlog=backlog)
def handle(self, sock, address):
# send and read functions should not wait longer than three seconds
sock.settimeout(3)
try:
# try to receive at most 128 bytes (`POLICYREQUEST` is shorter)
# Interestingly if we dont do this and we write to the spcket directly
# I am getting strange errors.
input = sock.recv(128)
if input.startswith(FlashPolicyServer.policyrequest):
sock.sendall(FlashPolicyServer.policy)
except socket.timeout:
pass
sock.close()
| <commit_before>from gevent.server import StreamServer
__all__ = ['FlashPolicyServer']
class FlashPolicyServer(StreamServer):
policy = """<?xml version="1.0"?><!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
<cross-domain-policy><allow-access-from domain="*" to-ports="*"/></cross-domain-policy>"""
def __init__(self, listener=None, backlog=None):
if listener is None:
listener = ('0.0.0.0', 10843)
StreamServer.__init__(self, listener=listener, backlog=backlog)
def handle(self, socket, address):
socket.sendall(self.policy)
<commit_msg>Fix to make sure we dont get errors in gevent socket write call when we are writing the policy file back
Conflicts:
socketio/policyserver.py<commit_after> | from gevent.server import StreamServer
__all__ = ['FlashPolicyServer']
class FlashPolicyServer(StreamServer):
policy = """<?xml version="1.0"?><!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
<cross-domain-policy><allow-access-from domain="*" to-ports="*"/></cross-domain-policy>"""
def __init__(self, listener=None, backlog=None):
if listener is None:
listener = ('0.0.0.0', 10843)
StreamServer.__init__(self, listener=listener, backlog=backlog)
def handle(self, sock, address):
# send and read functions should not wait longer than three seconds
sock.settimeout(3)
try:
# try to receive at most 128 bytes (`POLICYREQUEST` is shorter)
# Interestingly if we dont do this and we write to the spcket directly
# I am getting strange errors.
input = sock.recv(128)
if input.startswith(FlashPolicyServer.policyrequest):
sock.sendall(FlashPolicyServer.policy)
except socket.timeout:
pass
sock.close()
| from gevent.server import StreamServer
__all__ = ['FlashPolicyServer']
class FlashPolicyServer(StreamServer):
policy = """<?xml version="1.0"?><!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
<cross-domain-policy><allow-access-from domain="*" to-ports="*"/></cross-domain-policy>"""
def __init__(self, listener=None, backlog=None):
if listener is None:
listener = ('0.0.0.0', 10843)
StreamServer.__init__(self, listener=listener, backlog=backlog)
def handle(self, socket, address):
socket.sendall(self.policy)
Fix to make sure we dont get errors in gevent socket write call when we are writing the policy file back
Conflicts:
socketio/policyserver.pyfrom gevent.server import StreamServer
__all__ = ['FlashPolicyServer']
class FlashPolicyServer(StreamServer):
policy = """<?xml version="1.0"?><!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
<cross-domain-policy><allow-access-from domain="*" to-ports="*"/></cross-domain-policy>"""
def __init__(self, listener=None, backlog=None):
if listener is None:
listener = ('0.0.0.0', 10843)
StreamServer.__init__(self, listener=listener, backlog=backlog)
def handle(self, sock, address):
# send and read functions should not wait longer than three seconds
sock.settimeout(3)
try:
# try to receive at most 128 bytes (`POLICYREQUEST` is shorter)
# Interestingly if we dont do this and we write to the spcket directly
# I am getting strange errors.
input = sock.recv(128)
if input.startswith(FlashPolicyServer.policyrequest):
sock.sendall(FlashPolicyServer.policy)
except socket.timeout:
pass
sock.close()
| <commit_before>from gevent.server import StreamServer
__all__ = ['FlashPolicyServer']
class FlashPolicyServer(StreamServer):
policy = """<?xml version="1.0"?><!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
<cross-domain-policy><allow-access-from domain="*" to-ports="*"/></cross-domain-policy>"""
def __init__(self, listener=None, backlog=None):
if listener is None:
listener = ('0.0.0.0', 10843)
StreamServer.__init__(self, listener=listener, backlog=backlog)
def handle(self, socket, address):
socket.sendall(self.policy)
<commit_msg>Fix to make sure we dont get errors in gevent socket write call when we are writing the policy file back
Conflicts:
socketio/policyserver.py<commit_after>from gevent.server import StreamServer
__all__ = ['FlashPolicyServer']
class FlashPolicyServer(StreamServer):
policy = """<?xml version="1.0"?><!DOCTYPE cross-domain-policy SYSTEM "http://www.macromedia.com/xml/dtds/cross-domain-policy.dtd">
<cross-domain-policy><allow-access-from domain="*" to-ports="*"/></cross-domain-policy>"""
def __init__(self, listener=None, backlog=None):
if listener is None:
listener = ('0.0.0.0', 10843)
StreamServer.__init__(self, listener=listener, backlog=backlog)
def handle(self, sock, address):
# send and read functions should not wait longer than three seconds
sock.settimeout(3)
try:
# try to receive at most 128 bytes (`POLICYREQUEST` is shorter)
# Interestingly if we dont do this and we write to the spcket directly
# I am getting strange errors.
input = sock.recv(128)
if input.startswith(FlashPolicyServer.policyrequest):
sock.sendall(FlashPolicyServer.policy)
except socket.timeout:
pass
sock.close()
|
fc04d8f2629e5fef10cf62749e7c91e6b7d2d557 | cms/djangoapps/contentstore/views/session_kv_store.py | cms/djangoapps/contentstore/views/session_kv_store.py | """
An :class:`~xblock.runtime.KeyValueStore` that stores data in the django session
"""
from __future__ import absolute_import
from xblock.runtime import KeyValueStore
class SessionKeyValueStore(KeyValueStore):
def __init__(self, request):
self._session = request.session
def get(self, key):
return self._session[tuple(key)]
def set(self, key, value):
self._session[tuple(key)] = value
def delete(self, key):
del self._session[tuple(key)]
def has(self, key):
return tuple(key) in self._session
| """
An :class:`~xblock.runtime.KeyValueStore` that stores data in the django session
"""
from __future__ import absolute_import
from xblock.runtime import KeyValueStore
def stringify(key):
return repr(tuple(key))
class SessionKeyValueStore(KeyValueStore):
def __init__(self, request):
self._session = request.session
def get(self, key):
return self._session[stringify(key)]
def set(self, key, value):
self._session[stringify(key)] = value
def delete(self, key):
del self._session[stringify(key)]
def has(self, key):
return stringify(key) in self._session
| Use strings instead of tuples as keys in SessionKeyValueStore | Use strings instead of tuples as keys in SessionKeyValueStore
Some Django packages expect only strings as keys in the user session,
and it is also a recommended practice in the Django manual.
| Python | agpl-3.0 | ESOedX/edx-platform,B-MOOC/edx-platform,jswope00/griffinx,openfun/edx-platform,MakeHer/edx-platform,benpatterson/edx-platform,stvstnfrd/edx-platform,AkA84/edx-platform,Softmotions/edx-platform,IONISx/edx-platform,kmoocdev2/edx-platform,nanolearningllc/edx-platform-cypress,unicri/edx-platform,romain-li/edx-platform,xinjiguaike/edx-platform,mbareta/edx-platform-ft,RPI-OPENEDX/edx-platform,simbs/edx-platform,msegado/edx-platform,4eek/edx-platform,longmen21/edx-platform,cognitiveclass/edx-platform,eduNEXT/edx-platform,hastexo/edx-platform,openfun/edx-platform,eestay/edx-platform,shurihell/testasia,DNFcode/edx-platform,hamzehd/edx-platform,iivic/BoiseStateX,jruiperezv/ANALYSE,atsolakid/edx-platform,arifsetiawan/edx-platform,leansoft/edx-platform,chauhanhardik/populo,B-MOOC/edx-platform,doismellburning/edx-platform,Edraak/edraak-platform,nanolearningllc/edx-platform-cypress-2,simbs/edx-platform,tanmaykm/edx-platform,Edraak/circleci-edx-platform,Shrhawk/edx-platform,kmoocdev/edx-platform,devs1991/test_edx_docmode,martynovp/edx-platform,shubhdev/edx-platform,chauhanhardik/populo_2,kmoocdev2/edx-platform,inares/edx-platform,AkA84/edx-platform,jazkarta/edx-platform-for-isc,chand3040/cloud_that,romain-li/edx-platform,zhenzhai/edx-platform,nttks/edx-platform,zerobatu/edx-platform,gymnasium/edx-platform,cyanna/edx-platform,beni55/edx-platform,chudaol/edx-platform,shubhdev/edxOnBaadal,sameetb-cuelogic/edx-platform-test,atsolakid/edx-platform,cyanna/edx-platform,CourseTalk/edx-platform,rue89-tech/edx-platform,DNFcode/edx-platform,wwj718/edx-platform,ahmadio/edx-platform,mitocw/edx-platform,unicri/edx-platform,cognitiveclass/edx-platform,Edraak/edx-platform,andyzsf/edx,shashank971/edx-platform,nikolas/edx-platform,antoviaque/edx-platform,pabloborrego93/edx-platform,Endika/edx-platform,leansoft/edx-platform,devs1991/test_edx_docmode,shashank971/edx-platform,waheedahmed/edx-platform,xuxiao19910803/edx,halvertoluke/edx-platform,unicri/edx-platform,EDUlib/edx-platform,franosincic/edx-platform,appliedx/edx-platform,Stanford-Online/edx-platform,jazkarta/edx-platform-for-isc,DNFcode/edx-platform,wwj718/ANALYSE,wwj718/edx-platform,kmoocdev/edx-platform,jelugbo/tundex,kxliugang/edx-platform,gymnasium/edx-platform,jazkarta/edx-platform,EDUlib/edx-platform,teltek/edx-platform,MSOpenTech/edx-platform,angelapper/edx-platform,jazkarta/edx-platform,xuxiao19910803/edx,jamesblunt/edx-platform,TeachAtTUM/edx-platform,chudaol/edx-platform,halvertoluke/edx-platform,kamalx/edx-platform,cselis86/edx-platform,jelugbo/tundex,J861449197/edx-platform,etzhou/edx-platform,zerobatu/edx-platform,chauhanhardik/populo,AkA84/edx-platform,nttks/jenkins-test,proversity-org/edx-platform,jelugbo/tundex,chand3040/cloud_that,ubc/edx-platform,JCBarahona/edX,franosincic/edx-platform,OmarIthawi/edx-platform,philanthropy-u/edx-platform,cpennington/edx-platform,nanolearningllc/edx-platform-cypress,deepsrijit1105/edx-platform,sameetb-cuelogic/edx-platform-test,tiagochiavericosta/edx-platform,atsolakid/edx-platform,etzhou/edx-platform,fintech-circle/edx-platform,kursitet/edx-platform,devs1991/test_edx_docmode,rue89-tech/edx-platform,zerobatu/edx-platform,valtech-mooc/edx-platform,Endika/edx-platform,doismellburning/edx-platform,a-parhom/edx-platform,edx/edx-platform,mushtaqak/edx-platform,pomegranited/edx-platform,ferabra/edx-platform,cecep-edu/edx-platform,iivic/BoiseStateX,Shrhawk/edx-platform,jazkarta/edx-platform-for-isc,jbassen/edx-platform,dsajkl/123,arifsetiawan/edx-platform,xuxiao19910803/edx-platform,openfun/edx-platform,JCBarahona/edX,beni55/edx-platform,EDUlib/edx-platform,cyanna/edx-platform,cyanna/edx-platform,vismartltd/edx-platform,kxliugang/edx-platform,louyihua/edx-platform,jbassen/edx-platform,motion2015/a3,zofuthan/edx-platform,raccoongang/edx-platform,mtlchun/edx,analyseuc3m/ANALYSE-v1,eemirtekin/edx-platform,devs1991/test_edx_docmode,CredoReference/edx-platform,antonve/s4-project-mooc,ak2703/edx-platform,appsembler/edx-platform,beacloudgenius/edx-platform,cpennington/edx-platform,itsjeyd/edx-platform,ubc/edx-platform,hamzehd/edx-platform,UXE/local-edx,Kalyzee/edx-platform,alu042/edx-platform,jzoldak/edx-platform,naresh21/synergetics-edx-platform,edx-solutions/edx-platform,benpatterson/edx-platform,vasyarv/edx-platform,vikas1885/test1,Edraak/edx-platform,SravanthiSinha/edx-platform,bitifirefly/edx-platform,tiagochiavericosta/edx-platform,eduNEXT/edx-platform,y12uc231/edx-platform,alu042/edx-platform,sameetb-cuelogic/edx-platform-test,caesar2164/edx-platform,benpatterson/edx-platform,Ayub-Khan/edx-platform,leansoft/edx-platform,edry/edx-platform,DNFcode/edx-platform,jazkarta/edx-platform,valtech-mooc/edx-platform,Lektorium-LLC/edx-platform,dsajkl/reqiop,nagyistoce/edx-platform,ahmadio/edx-platform,rismalrv/edx-platform,ovnicraft/edx-platform,ahmadio/edx-platform,Livit/Livit.Learn.EdX,ahmadiga/min_edx,ampax/edx-platform-backup,beni55/edx-platform,TeachAtTUM/edx-platform,ak2703/edx-platform,kamalx/edx-platform,mitocw/edx-platform,AkA84/edx-platform,louyihua/edx-platform,JioEducation/edx-platform,chand3040/cloud_that,mjirayu/sit_academy,peterm-itr/edx-platform,jjmiranda/edx-platform,Kalyzee/edx-platform,dsajkl/reqiop,SravanthiSinha/edx-platform,edry/edx-platform,teltek/edx-platform,doismellburning/edx-platform,appliedx/edx-platform,pomegranited/edx-platform,Endika/edx-platform,nagyistoce/edx-platform,dkarakats/edx-platform,wwj718/ANALYSE,Livit/Livit.Learn.EdX,cognitiveclass/edx-platform,jolyonb/edx-platform,eduNEXT/edunext-platform,angelapper/edx-platform,vikas1885/test1,bitifirefly/edx-platform,TeachAtTUM/edx-platform,marcore/edx-platform,waheedahmed/edx-platform,bigdatauniversity/edx-platform,playm2mboy/edx-platform,jonathan-beard/edx-platform,pabloborrego93/edx-platform,dcosentino/edx-platform,jazkarta/edx-platform,RPI-OPENEDX/edx-platform,ovnicraft/edx-platform,kamalx/edx-platform,chand3040/cloud_that,ampax/edx-platform-backup,waheedahmed/edx-platform,IndonesiaX/edx-platform,kmoocdev/edx-platform,RPI-OPENEDX/edx-platform,defance/edx-platform,nttks/edx-platform,shabab12/edx-platform,longmen21/edx-platform,gsehub/edx-platform,Semi-global/edx-platform,marcore/edx-platform,dsajkl/123,appsembler/edx-platform,jzoldak/edx-platform,shashank971/edx-platform,etzhou/edx-platform,nagyistoce/edx-platform,MSOpenTech/edx-platform,CourseTalk/edx-platform,mahendra-r/edx-platform,rue89-tech/edx-platform,ampax/edx-platform-backup,beni55/edx-platform,chrisndodge/edx-platform,motion2015/a3,IONISx/edx-platform,SravanthiSinha/edx-platform,Shrhawk/edx-platform,alexthered/kienhoc-platform,mitocw/edx-platform,proversity-org/edx-platform,vikas1885/test1,mcgachey/edx-platform,jbzdak/edx-platform,BehavioralInsightsTeam/edx-platform,devs1991/test_edx_docmode,ahmadiga/min_edx,motion2015/a3,appliedx/edx-platform,bigdatauniversity/edx-platform,solashirai/edx-platform,UXE/local-edx,jonathan-beard/edx-platform,iivic/BoiseStateX,OmarIthawi/edx-platform,4eek/edx-platform,Kalyzee/edx-platform,jazkarta/edx-platform-for-isc,Edraak/edx-platform,nanolearningllc/edx-platform-cypress,vismartltd/edx-platform,angelapper/edx-platform,playm2mboy/edx-platform,fintech-circle/edx-platform,bigdatauniversity/edx-platform,nagyistoce/edx-platform,Lektorium-LLC/edx-platform,bigdatauniversity/edx-platform,don-github/edx-platform,zubair-arbi/edx-platform,ubc/edx-platform,prarthitm/edxplatform,SravanthiSinha/edx-platform,fintech-circle/edx-platform,ferabra/edx-platform,antoviaque/edx-platform,marcore/edx-platform,proversity-org/edx-platform,chudaol/edx-platform,tiagochiavericosta/edx-platform,jbassen/edx-platform,rismalrv/edx-platform,cselis86/edx-platform,shubhdev/edxOnBaadal,ferabra/edx-platform,ubc/edx-platform,rismalrv/edx-platform,doismellburning/edx-platform,cognitiveclass/edx-platform,rhndg/openedx,utecuy/edx-platform,ahmadiga/min_edx,doganov/edx-platform,nttks/jenkins-test,andyzsf/edx,defance/edx-platform,chudaol/edx-platform,mjirayu/sit_academy,halvertoluke/edx-platform,CourseTalk/edx-platform,MakeHer/edx-platform,Softmotions/edx-platform,DefyVentures/edx-platform,Softmotions/edx-platform,arifsetiawan/edx-platform,xingyepei/edx-platform,zadgroup/edx-platform,jruiperezv/ANALYSE,proversity-org/edx-platform,JioEducation/edx-platform,pabloborrego93/edx-platform,don-github/edx-platform,edry/edx-platform,pepeportela/edx-platform,utecuy/edx-platform,nagyistoce/edx-platform,MSOpenTech/edx-platform,arbrandes/edx-platform,utecuy/edx-platform,Kalyzee/edx-platform,J861449197/edx-platform,synergeticsedx/deployment-wipro,eestay/edx-platform,ovnicraft/edx-platform,SivilTaram/edx-platform,eestay/edx-platform,Ayub-Khan/edx-platform,SivilTaram/edx-platform,BehavioralInsightsTeam/edx-platform,UXE/local-edx,antonve/s4-project-mooc,zadgroup/edx-platform,fly19890211/edx-platform,longmen21/edx-platform,chauhanhardik/populo_2,mbareta/edx-platform-ft,philanthropy-u/edx-platform,hamzehd/edx-platform,nanolearningllc/edx-platform-cypress-2,ampax/edx-platform,UOMx/edx-platform,cecep-edu/edx-platform,dsajkl/reqiop,jswope00/griffinx,ahmadiga/min_edx,don-github/edx-platform,shashank971/edx-platform,10clouds/edx-platform,IndonesiaX/edx-platform,mjirayu/sit_academy,unicri/edx-platform,shabab12/edx-platform,valtech-mooc/edx-platform,miptliot/edx-platform,UOMx/edx-platform,zofuthan/edx-platform,alu042/edx-platform,xinjiguaike/edx-platform,xuxiao19910803/edx,edry/edx-platform,ak2703/edx-platform,doganov/edx-platform,mbareta/edx-platform-ft,shubhdev/edx-platform,chauhanhardik/populo,cecep-edu/edx-platform,louyihua/edx-platform,Semi-global/edx-platform,gsehub/edx-platform,Endika/edx-platform,shubhdev/openedx,ESOedX/edx-platform,stvstnfrd/edx-platform,Edraak/edx-platform,appliedx/edx-platform,mtlchun/edx,jelugbo/tundex,arifsetiawan/edx-platform,nikolas/edx-platform,amir-qayyum-khan/edx-platform,dcosentino/edx-platform,cecep-edu/edx-platform,adoosii/edx-platform,arbrandes/edx-platform,xingyepei/edx-platform,Lektorium-LLC/edx-platform,naresh21/synergetics-edx-platform,procangroup/edx-platform,mahendra-r/edx-platform,fly19890211/edx-platform,msegado/edx-platform,CredoReference/edx-platform,kxliugang/edx-platform,rue89-tech/edx-platform,dkarakats/edx-platform,benpatterson/edx-platform,stvstnfrd/edx-platform,romain-li/edx-platform,alu042/edx-platform,ampax/edx-platform,ampax/edx-platform-backup,valtech-mooc/edx-platform,motion2015/a3,caesar2164/edx-platform,tiagochiavericosta/edx-platform,bigdatauniversity/edx-platform,JCBarahona/edX,analyseuc3m/ANALYSE-v1,IONISx/edx-platform,xuxiao19910803/edx-platform,jazztpt/edx-platform,mushtaqak/edx-platform,atsolakid/edx-platform,eduNEXT/edunext-platform,mjirayu/sit_academy,xingyepei/edx-platform,itsjeyd/edx-platform,DefyVentures/edx-platform,antoviaque/edx-platform,10clouds/edx-platform,eemirtekin/edx-platform,shabab12/edx-platform,synergeticsedx/deployment-wipro,nikolas/edx-platform,edx/edx-platform,antoviaque/edx-platform,wwj718/edx-platform,bitifirefly/edx-platform,mcgachey/edx-platform,defance/edx-platform,alexthered/kienhoc-platform,vasyarv/edx-platform,lduarte1991/edx-platform,ZLLab-Mooc/edx-platform,jruiperezv/ANALYSE,chrisndodge/edx-platform,sameetb-cuelogic/edx-platform-test,iivic/BoiseStateX,sameetb-cuelogic/edx-platform-test,zerobatu/edx-platform,xuxiao19910803/edx,atsolakid/edx-platform,jbassen/edx-platform,cselis86/edx-platform,ZLLab-Mooc/edx-platform,wwj718/edx-platform,nanolearningllc/edx-platform-cypress-2,msegado/edx-platform,JCBarahona/edX,miptliot/edx-platform,chrisndodge/edx-platform,shubhdev/edxOnBaadal,knehez/edx-platform,Ayub-Khan/edx-platform,dkarakats/edx-platform,xuxiao19910803/edx-platform,SivilTaram/edx-platform,B-MOOC/edx-platform,adoosii/edx-platform,Edraak/edraak-platform,kursitet/edx-platform,beni55/edx-platform,Livit/Livit.Learn.EdX,kursitet/edx-platform,ahmadio/edx-platform,jbzdak/edx-platform,caesar2164/edx-platform,marcore/edx-platform,jruiperezv/ANALYSE,jamesblunt/edx-platform,kxliugang/edx-platform,mtlchun/edx,alexthered/kienhoc-platform,deepsrijit1105/edx-platform,fly19890211/edx-platform,edry/edx-platform,jamiefolsom/edx-platform,xingyepei/edx-platform,Shrhawk/edx-platform,beacloudgenius/edx-platform,appsembler/edx-platform,shubhdev/edxOnBaadal,motion2015/edx-platform,beacloudgenius/edx-platform,gymnasium/edx-platform,ESOedX/edx-platform,cyanna/edx-platform,jamesblunt/edx-platform,ampax/edx-platform-backup,beacloudgenius/edx-platform,devs1991/test_edx_docmode,vasyarv/edx-platform,IndonesiaX/edx-platform,olexiim/edx-platform,knehez/edx-platform,cecep-edu/edx-platform,arbrandes/edx-platform,dcosentino/edx-platform,arifsetiawan/edx-platform,Stanford-Online/edx-platform,cognitiveclass/edx-platform,IONISx/edx-platform,Edraak/edx-platform,vasyarv/edx-platform,iivic/BoiseStateX,DNFcode/edx-platform,chauhanhardik/populo_2,zadgroup/edx-platform,zubair-arbi/edx-platform,zofuthan/edx-platform,jjmiranda/edx-platform,procangroup/edx-platform,rismalrv/edx-platform,CourseTalk/edx-platform,olexiim/edx-platform,ak2703/edx-platform,mahendra-r/edx-platform,nttks/edx-platform,bitifirefly/edx-platform,shubhdev/openedx,franosincic/edx-platform,kmoocdev2/edx-platform,jswope00/griffinx,mushtaqak/edx-platform,synergeticsedx/deployment-wipro,solashirai/edx-platform,MakeHer/edx-platform,xuxiao19910803/edx-platform,antonve/s4-project-mooc,stvstnfrd/edx-platform,Edraak/circleci-edx-platform,polimediaupv/edx-platform,vasyarv/edx-platform,antonve/s4-project-mooc,jolyonb/edx-platform,ahmadiga/min_edx,arbrandes/edx-platform,procangroup/edx-platform,shubhdev/edx-platform,utecuy/edx-platform,chauhanhardik/populo,inares/edx-platform,devs1991/test_edx_docmode,BehavioralInsightsTeam/edx-platform,caesar2164/edx-platform,motion2015/edx-platform,ESOedX/edx-platform,nttks/jenkins-test,Lektorium-LLC/edx-platform,eduNEXT/edunext-platform,don-github/edx-platform,solashirai/edx-platform,martynovp/edx-platform,kmoocdev/edx-platform,CredoReference/edx-platform,msegado/edx-platform,IndonesiaX/edx-platform,polimediaupv/edx-platform,y12uc231/edx-platform,kursitet/edx-platform,jamesblunt/edx-platform,don-github/edx-platform,RPI-OPENEDX/edx-platform,jonathan-beard/edx-platform,jonathan-beard/edx-platform,angelapper/edx-platform,xinjiguaike/edx-platform,wwj718/ANALYSE,gsehub/edx-platform,antonve/s4-project-mooc,cpennington/edx-platform,adoosii/edx-platform,solashirai/edx-platform,EDUlib/edx-platform,longmen21/edx-platform,jazkarta/edx-platform-for-isc,mushtaqak/edx-platform,chudaol/edx-platform,MSOpenTech/edx-platform,pepeportela/edx-platform,mcgachey/edx-platform,philanthropy-u/edx-platform,jswope00/griffinx,lduarte1991/edx-platform,pomegranited/edx-platform,solashirai/edx-platform,rismalrv/edx-platform,tanmaykm/edx-platform,hastexo/edx-platform,eemirtekin/edx-platform,vismartltd/edx-platform,Edraak/circleci-edx-platform,zadgroup/edx-platform,jruiperezv/ANALYSE,eestay/edx-platform,edx/edx-platform,mitocw/edx-platform,Edraak/circleci-edx-platform,zhenzhai/edx-platform,J861449197/edx-platform,synergeticsedx/deployment-wipro,teltek/edx-platform,fintech-circle/edx-platform,beacloudgenius/edx-platform,romain-li/edx-platform,shubhdev/edxOnBaadal,raccoongang/edx-platform,a-parhom/edx-platform,4eek/edx-platform,JioEducation/edx-platform,rhndg/openedx,tiagochiavericosta/edx-platform,waheedahmed/edx-platform,nttks/edx-platform,philanthropy-u/edx-platform,JioEducation/edx-platform,pepeportela/edx-platform,doganov/edx-platform,ovnicraft/edx-platform,shubhdev/edx-platform,waheedahmed/edx-platform,chauhanhardik/populo,jswope00/griffinx,shubhdev/edx-platform,teltek/edx-platform,cselis86/edx-platform,lduarte1991/edx-platform,kamalx/edx-platform,Ayub-Khan/edx-platform,alexthered/kienhoc-platform,lduarte1991/edx-platform,ZLLab-Mooc/edx-platform,alexthered/kienhoc-platform,xinjiguaike/edx-platform,jamiefolsom/edx-platform,UXE/local-edx,Edraak/circleci-edx-platform,raccoongang/edx-platform,shurihell/testasia,jazztpt/edx-platform,prarthitm/edxplatform,10clouds/edx-platform,motion2015/edx-platform,andyzsf/edx,xingyepei/edx-platform,wwj718/edx-platform,edx-solutions/edx-platform,msegado/edx-platform,naresh21/synergetics-edx-platform,shubhdev/openedx,ubc/edx-platform,Kalyzee/edx-platform,kamalx/edx-platform,shurihell/testasia,nanolearningllc/edx-platform-cypress,AkA84/edx-platform,zadgroup/edx-platform,inares/edx-platform,CredoReference/edx-platform,fly19890211/edx-platform,Semi-global/edx-platform,OmarIthawi/edx-platform,Softmotions/edx-platform,shubhdev/openedx,deepsrijit1105/edx-platform,shurihell/testasia,Softmotions/edx-platform,miptliot/edx-platform,martynovp/edx-platform,mahendra-r/edx-platform,zhenzhai/edx-platform,ahmedaljazzar/edx-platform,ahmedaljazzar/edx-platform,knehez/edx-platform,Semi-global/edx-platform,hastexo/edx-platform,jjmiranda/edx-platform,mcgachey/edx-platform,hamzehd/edx-platform,nanolearningllc/edx-platform-cypress-2,4eek/edx-platform,jamesblunt/edx-platform,jbzdak/edx-platform,unicri/edx-platform,playm2mboy/edx-platform,cpennington/edx-platform,leansoft/edx-platform,naresh21/synergetics-edx-platform,SivilTaram/edx-platform,jonathan-beard/edx-platform,peterm-itr/edx-platform,amir-qayyum-khan/edx-platform,ovnicraft/edx-platform,mjirayu/sit_academy,UOMx/edx-platform,doganov/edx-platform,MSOpenTech/edx-platform,motion2015/a3,nanolearningllc/edx-platform-cypress,etzhou/edx-platform,ahmadio/edx-platform,jazztpt/edx-platform,jazztpt/edx-platform,playm2mboy/edx-platform,dsajkl/123,openfun/edx-platform,RPI-OPENEDX/edx-platform,raccoongang/edx-platform,motion2015/edx-platform,dsajkl/123,dsajkl/123,leansoft/edx-platform,louyihua/edx-platform,mahendra-r/edx-platform,chauhanhardik/populo_2,adoosii/edx-platform,y12uc231/edx-platform,valtech-mooc/edx-platform,dkarakats/edx-platform,rhndg/openedx,franosincic/edx-platform,Shrhawk/edx-platform,nttks/edx-platform,mtlchun/edx,kmoocdev/edx-platform,longmen21/edx-platform,edx-solutions/edx-platform,cselis86/edx-platform,zubair-arbi/edx-platform,eestay/edx-platform,gsehub/edx-platform,inares/edx-platform,jamiefolsom/edx-platform,4eek/edx-platform,analyseuc3m/ANALYSE-v1,amir-qayyum-khan/edx-platform,peterm-itr/edx-platform,rue89-tech/edx-platform,ak2703/edx-platform,jbzdak/edx-platform,ampax/edx-platform,DefyVentures/edx-platform,jzoldak/edx-platform,ferabra/edx-platform,adoosii/edx-platform,prarthitm/edxplatform,dcosentino/edx-platform,jjmiranda/edx-platform,deepsrijit1105/edx-platform,eduNEXT/edx-platform,jolyonb/edx-platform,polimediaupv/edx-platform,pomegranited/edx-platform,y12uc231/edx-platform,franosincic/edx-platform,Edraak/edraak-platform,vismartltd/edx-platform,xuxiao19910803/edx-platform,openfun/edx-platform,eemirtekin/edx-platform,doismellburning/edx-platform,miptliot/edx-platform,eduNEXT/edunext-platform,itsjeyd/edx-platform,zerobatu/edx-platform,simbs/edx-platform,a-parhom/edx-platform,andyzsf/edx,10clouds/edx-platform,fly19890211/edx-platform,zubair-arbi/edx-platform,jelugbo/tundex,etzhou/edx-platform,hastexo/edx-platform,rhndg/openedx,zhenzhai/edx-platform,eemirtekin/edx-platform,edx/edx-platform,jamiefolsom/edx-platform,procangroup/edx-platform,jbzdak/edx-platform,SivilTaram/edx-platform,polimediaupv/edx-platform,SravanthiSinha/edx-platform,itsjeyd/edx-platform,analyseuc3m/ANALYSE-v1,TeachAtTUM/edx-platform,bitifirefly/edx-platform,DefyVentures/edx-platform,mbareta/edx-platform-ft,wwj718/ANALYSE,Stanford-Online/edx-platform,shurihell/testasia,J861449197/edx-platform,Stanford-Online/edx-platform,ZLLab-Mooc/edx-platform,olexiim/edx-platform,jbassen/edx-platform,chauhanhardik/populo_2,dsajkl/reqiop,nikolas/edx-platform,appsembler/edx-platform,shashank971/edx-platform,prarthitm/edxplatform,vikas1885/test1,simbs/edx-platform,tanmaykm/edx-platform,jamiefolsom/edx-platform,ZLLab-Mooc/edx-platform,romain-li/edx-platform,nanolearningllc/edx-platform-cypress-2,IndonesiaX/edx-platform,peterm-itr/edx-platform,shubhdev/openedx,jazkarta/edx-platform,olexiim/edx-platform,knehez/edx-platform,dkarakats/edx-platform,zhenzhai/edx-platform,B-MOOC/edx-platform,BehavioralInsightsTeam/edx-platform,motion2015/edx-platform,devs1991/test_edx_docmode,martynovp/edx-platform,vikas1885/test1,chand3040/cloud_that,ahmedaljazzar/edx-platform,tanmaykm/edx-platform,UOMx/edx-platform,mcgachey/edx-platform,a-parhom/edx-platform,utecuy/edx-platform,kmoocdev2/edx-platform,playm2mboy/edx-platform,DefyVentures/edx-platform,shabab12/edx-platform,inares/edx-platform,edx-solutions/edx-platform,JCBarahona/edX,zubair-arbi/edx-platform,ahmedaljazzar/edx-platform,Ayub-Khan/edx-platform,y12uc231/edx-platform,simbs/edx-platform,kursitet/edx-platform,mushtaqak/edx-platform,appliedx/edx-platform,ferabra/edx-platform,eduNEXT/edx-platform,chrisndodge/edx-platform,martynovp/edx-platform,dcosentino/edx-platform,jazztpt/edx-platform,xinjiguaike/edx-platform,IONISx/edx-platform,pomegranited/edx-platform,knehez/edx-platform,hamzehd/edx-platform,halvertoluke/edx-platform,ampax/edx-platform,xuxiao19910803/edx,MakeHer/edx-platform,Semi-global/edx-platform,nttks/jenkins-test,MakeHer/edx-platform,Livit/Livit.Learn.EdX,OmarIthawi/edx-platform,jolyonb/edx-platform,rhndg/openedx,B-MOOC/edx-platform,kmoocdev2/edx-platform,halvertoluke/edx-platform,zofuthan/edx-platform,jzoldak/edx-platform,amir-qayyum-khan/edx-platform,benpatterson/edx-platform,defance/edx-platform,polimediaupv/edx-platform,kxliugang/edx-platform,pepeportela/edx-platform,gymnasium/edx-platform,Edraak/edraak-platform,olexiim/edx-platform,J861449197/edx-platform,nttks/jenkins-test,zofuthan/edx-platform,pabloborrego93/edx-platform,nikolas/edx-platform,wwj718/ANALYSE,doganov/edx-platform,mtlchun/edx,vismartltd/edx-platform | """
An :class:`~xblock.runtime.KeyValueStore` that stores data in the django session
"""
from __future__ import absolute_import
from xblock.runtime import KeyValueStore
class SessionKeyValueStore(KeyValueStore):
def __init__(self, request):
self._session = request.session
def get(self, key):
return self._session[tuple(key)]
def set(self, key, value):
self._session[tuple(key)] = value
def delete(self, key):
del self._session[tuple(key)]
def has(self, key):
return tuple(key) in self._session
Use strings instead of tuples as keys in SessionKeyValueStore
Some Django packages expect only strings as keys in the user session,
and it is also a recommended practice in the Django manual. | """
An :class:`~xblock.runtime.KeyValueStore` that stores data in the django session
"""
from __future__ import absolute_import
from xblock.runtime import KeyValueStore
def stringify(key):
return repr(tuple(key))
class SessionKeyValueStore(KeyValueStore):
def __init__(self, request):
self._session = request.session
def get(self, key):
return self._session[stringify(key)]
def set(self, key, value):
self._session[stringify(key)] = value
def delete(self, key):
del self._session[stringify(key)]
def has(self, key):
return stringify(key) in self._session
| <commit_before>"""
An :class:`~xblock.runtime.KeyValueStore` that stores data in the django session
"""
from __future__ import absolute_import
from xblock.runtime import KeyValueStore
class SessionKeyValueStore(KeyValueStore):
def __init__(self, request):
self._session = request.session
def get(self, key):
return self._session[tuple(key)]
def set(self, key, value):
self._session[tuple(key)] = value
def delete(self, key):
del self._session[tuple(key)]
def has(self, key):
return tuple(key) in self._session
<commit_msg>Use strings instead of tuples as keys in SessionKeyValueStore
Some Django packages expect only strings as keys in the user session,
and it is also a recommended practice in the Django manual.<commit_after> | """
An :class:`~xblock.runtime.KeyValueStore` that stores data in the django session
"""
from __future__ import absolute_import
from xblock.runtime import KeyValueStore
def stringify(key):
return repr(tuple(key))
class SessionKeyValueStore(KeyValueStore):
def __init__(self, request):
self._session = request.session
def get(self, key):
return self._session[stringify(key)]
def set(self, key, value):
self._session[stringify(key)] = value
def delete(self, key):
del self._session[stringify(key)]
def has(self, key):
return stringify(key) in self._session
| """
An :class:`~xblock.runtime.KeyValueStore` that stores data in the django session
"""
from __future__ import absolute_import
from xblock.runtime import KeyValueStore
class SessionKeyValueStore(KeyValueStore):
def __init__(self, request):
self._session = request.session
def get(self, key):
return self._session[tuple(key)]
def set(self, key, value):
self._session[tuple(key)] = value
def delete(self, key):
del self._session[tuple(key)]
def has(self, key):
return tuple(key) in self._session
Use strings instead of tuples as keys in SessionKeyValueStore
Some Django packages expect only strings as keys in the user session,
and it is also a recommended practice in the Django manual."""
An :class:`~xblock.runtime.KeyValueStore` that stores data in the django session
"""
from __future__ import absolute_import
from xblock.runtime import KeyValueStore
def stringify(key):
return repr(tuple(key))
class SessionKeyValueStore(KeyValueStore):
def __init__(self, request):
self._session = request.session
def get(self, key):
return self._session[stringify(key)]
def set(self, key, value):
self._session[stringify(key)] = value
def delete(self, key):
del self._session[stringify(key)]
def has(self, key):
return stringify(key) in self._session
| <commit_before>"""
An :class:`~xblock.runtime.KeyValueStore` that stores data in the django session
"""
from __future__ import absolute_import
from xblock.runtime import KeyValueStore
class SessionKeyValueStore(KeyValueStore):
def __init__(self, request):
self._session = request.session
def get(self, key):
return self._session[tuple(key)]
def set(self, key, value):
self._session[tuple(key)] = value
def delete(self, key):
del self._session[tuple(key)]
def has(self, key):
return tuple(key) in self._session
<commit_msg>Use strings instead of tuples as keys in SessionKeyValueStore
Some Django packages expect only strings as keys in the user session,
and it is also a recommended practice in the Django manual.<commit_after>"""
An :class:`~xblock.runtime.KeyValueStore` that stores data in the django session
"""
from __future__ import absolute_import
from xblock.runtime import KeyValueStore
def stringify(key):
return repr(tuple(key))
class SessionKeyValueStore(KeyValueStore):
def __init__(self, request):
self._session = request.session
def get(self, key):
return self._session[stringify(key)]
def set(self, key, value):
self._session[stringify(key)] = value
def delete(self, key):
del self._session[stringify(key)]
def has(self, key):
return stringify(key) in self._session
|
5eb3f2c61c2b61e1bad7faa006e5503bd9a20edf | uni_form/util.py | uni_form/util.py | from django import forms
from django.forms.widgets import Input
class SubmitButtonWidget(Input):
"""
A widget that handles a submit button.
"""
input_type = 'submit'
def render(self, name, value, attrs=None):
return super(SubmitButtonWidget, self).render(name,
self.attrs['value'], attrs)
class BaseInput(forms.Field):
"""
An base Input class to reduce the amount of code in the Input classes.
"""
widget = SubmitButtonWidget
def __init__(self, **kwargs):
if not 'label' in kwargs:
kwargs['label'] = ''
if not 'required' in kwargs:
kwargs['required'] = False
if 'value' in kwargs:
self._widget_attrs = {'value': kwargs['value']}
del kwargs['value']
else:
self._widget_attrs = {'value': 'Submit'}
super(BaseInput, self).__init__(**kwargs)
def widget_attrs(self, widget):
return self._widget_attrs
class Toggle(object):
"""
A container for holder toggled items such as fields and buttons.
"""
fields = [] | class BaseInput(object):
"""
An base Input class to reduce the amount of code in the Input classes.
"""
def __init__(self,name,value):
self.name = name
self.value = value
class Toggle(object):
"""
A container for holder toggled items such as fields and buttons.
"""
fields = [] | Revert "Made BaseInput inherit from forms.Field so inputs can be used in layouts. Added a SubmitButtonWidget." | Revert "Made BaseInput inherit from forms.Field so inputs can be used in layouts. Added a SubmitButtonWidget."
This reverts commit aa571b2e1fd177491895cc263b192467431b90c2.
| Python | mit | HungryCloud/django-crispy-forms,spectras/django-crispy-forms,iris-edu-int/django-crispy-forms,scuml/django-crispy-forms,ngenovictor/django-crispy-forms,CashStar/django-uni-form,PetrDlouhy/django-crispy-forms,RamezIssac/django-crispy-forms,jcomeauictx/django-crispy-forms,tarunlnmiit/django-crispy-forms,CashStar/django-uni-form,pydanny/django-uni-form,avsd/django-crispy-forms,agepoly/django-crispy-forms,django-crispy-forms/django-crispy-forms,carltongibson/django-crispy-forms,pjdelport/django-crispy-forms,iris-edu/django-crispy-forms,IanLee1521/django-crispy-forms,dzhuang/django-crispy-forms,IanLee1521/django-crispy-forms,zixan/django-crispy-forms,ionelmc/django-uni-form,spectras/django-crispy-forms,dessibelle/django-crispy-forms,bouttier/django-crispy-forms,HungryCloud/django-crispy-forms,jcomeauictx/django-crispy-forms,tarunlnmiit/django-crispy-forms,eykanal/django-crispy-forms,HungryCloud/django-crispy-forms,maraujop/django-crispy-forms,iedparis8/django-crispy-forms,VishvajitP/django-crispy-forms,damienjones/django-crispy-forms,saydulk/django-crispy-forms,impulse-cloud/django-crispy-forms,VishvajitP/django-crispy-forms,saydulk/django-crispy-forms,scuml/django-crispy-forms,carltongibson/django-crispy-forms,alanwj/django-crispy-forms,rfleschenberg/django-crispy-forms,eykanal/django-crispy-forms,alanwj/django-crispy-forms,uranusjr/django-crispy-forms-ng,iris-edu-int/django-crispy-forms,zixan/django-crispy-forms,treyhunner/django-crispy-forms,maraujop/django-crispy-forms,davidszotten/django-crispy-forms,iedparis8/django-crispy-forms,avsd/django-crispy-forms,ngenovictor/django-crispy-forms,jtyoung/django-crispy-forms,davidszotten/django-crispy-forms,Stranger6667/django-crispy-forms,dessibelle/django-crispy-forms,PetrDlouhy/django-crispy-forms,dzhuang/django-crispy-forms,treyhunner/django-crispy-forms,agepoly/django-crispy-forms,uranusjr/django-crispy-forms-ng,django-crispy-forms/django-crispy-forms,RamezIssac/django-crispy-forms,damienjones/django-crispy-forms,bouttier/django-crispy-forms,rfleschenberg/django-crispy-forms,smirolo/django-crispy-forms,jtyoung/django-crispy-forms,iris-edu/django-crispy-forms,schrd/django-crispy-forms,Stranger6667/django-crispy-forms,pydanny/django-uni-form,impulse-cloud/django-crispy-forms,smirolo/django-crispy-forms,schrd/django-crispy-forms | from django import forms
from django.forms.widgets import Input
class SubmitButtonWidget(Input):
"""
A widget that handles a submit button.
"""
input_type = 'submit'
def render(self, name, value, attrs=None):
return super(SubmitButtonWidget, self).render(name,
self.attrs['value'], attrs)
class BaseInput(forms.Field):
"""
An base Input class to reduce the amount of code in the Input classes.
"""
widget = SubmitButtonWidget
def __init__(self, **kwargs):
if not 'label' in kwargs:
kwargs['label'] = ''
if not 'required' in kwargs:
kwargs['required'] = False
if 'value' in kwargs:
self._widget_attrs = {'value': kwargs['value']}
del kwargs['value']
else:
self._widget_attrs = {'value': 'Submit'}
super(BaseInput, self).__init__(**kwargs)
def widget_attrs(self, widget):
return self._widget_attrs
class Toggle(object):
"""
A container for holder toggled items such as fields and buttons.
"""
fields = []Revert "Made BaseInput inherit from forms.Field so inputs can be used in layouts. Added a SubmitButtonWidget."
This reverts commit aa571b2e1fd177491895cc263b192467431b90c2. | class BaseInput(object):
"""
An base Input class to reduce the amount of code in the Input classes.
"""
def __init__(self,name,value):
self.name = name
self.value = value
class Toggle(object):
"""
A container for holder toggled items such as fields and buttons.
"""
fields = [] | <commit_before>from django import forms
from django.forms.widgets import Input
class SubmitButtonWidget(Input):
"""
A widget that handles a submit button.
"""
input_type = 'submit'
def render(self, name, value, attrs=None):
return super(SubmitButtonWidget, self).render(name,
self.attrs['value'], attrs)
class BaseInput(forms.Field):
"""
An base Input class to reduce the amount of code in the Input classes.
"""
widget = SubmitButtonWidget
def __init__(self, **kwargs):
if not 'label' in kwargs:
kwargs['label'] = ''
if not 'required' in kwargs:
kwargs['required'] = False
if 'value' in kwargs:
self._widget_attrs = {'value': kwargs['value']}
del kwargs['value']
else:
self._widget_attrs = {'value': 'Submit'}
super(BaseInput, self).__init__(**kwargs)
def widget_attrs(self, widget):
return self._widget_attrs
class Toggle(object):
"""
A container for holder toggled items such as fields and buttons.
"""
fields = []<commit_msg>Revert "Made BaseInput inherit from forms.Field so inputs can be used in layouts. Added a SubmitButtonWidget."
This reverts commit aa571b2e1fd177491895cc263b192467431b90c2.<commit_after> | class BaseInput(object):
"""
An base Input class to reduce the amount of code in the Input classes.
"""
def __init__(self,name,value):
self.name = name
self.value = value
class Toggle(object):
"""
A container for holder toggled items such as fields and buttons.
"""
fields = [] | from django import forms
from django.forms.widgets import Input
class SubmitButtonWidget(Input):
"""
A widget that handles a submit button.
"""
input_type = 'submit'
def render(self, name, value, attrs=None):
return super(SubmitButtonWidget, self).render(name,
self.attrs['value'], attrs)
class BaseInput(forms.Field):
"""
An base Input class to reduce the amount of code in the Input classes.
"""
widget = SubmitButtonWidget
def __init__(self, **kwargs):
if not 'label' in kwargs:
kwargs['label'] = ''
if not 'required' in kwargs:
kwargs['required'] = False
if 'value' in kwargs:
self._widget_attrs = {'value': kwargs['value']}
del kwargs['value']
else:
self._widget_attrs = {'value': 'Submit'}
super(BaseInput, self).__init__(**kwargs)
def widget_attrs(self, widget):
return self._widget_attrs
class Toggle(object):
"""
A container for holder toggled items such as fields and buttons.
"""
fields = []Revert "Made BaseInput inherit from forms.Field so inputs can be used in layouts. Added a SubmitButtonWidget."
This reverts commit aa571b2e1fd177491895cc263b192467431b90c2.class BaseInput(object):
"""
An base Input class to reduce the amount of code in the Input classes.
"""
def __init__(self,name,value):
self.name = name
self.value = value
class Toggle(object):
"""
A container for holder toggled items such as fields and buttons.
"""
fields = [] | <commit_before>from django import forms
from django.forms.widgets import Input
class SubmitButtonWidget(Input):
"""
A widget that handles a submit button.
"""
input_type = 'submit'
def render(self, name, value, attrs=None):
return super(SubmitButtonWidget, self).render(name,
self.attrs['value'], attrs)
class BaseInput(forms.Field):
"""
An base Input class to reduce the amount of code in the Input classes.
"""
widget = SubmitButtonWidget
def __init__(self, **kwargs):
if not 'label' in kwargs:
kwargs['label'] = ''
if not 'required' in kwargs:
kwargs['required'] = False
if 'value' in kwargs:
self._widget_attrs = {'value': kwargs['value']}
del kwargs['value']
else:
self._widget_attrs = {'value': 'Submit'}
super(BaseInput, self).__init__(**kwargs)
def widget_attrs(self, widget):
return self._widget_attrs
class Toggle(object):
"""
A container for holder toggled items such as fields and buttons.
"""
fields = []<commit_msg>Revert "Made BaseInput inherit from forms.Field so inputs can be used in layouts. Added a SubmitButtonWidget."
This reverts commit aa571b2e1fd177491895cc263b192467431b90c2.<commit_after>class BaseInput(object):
"""
An base Input class to reduce the amount of code in the Input classes.
"""
def __init__(self,name,value):
self.name = name
self.value = value
class Toggle(object):
"""
A container for holder toggled items such as fields and buttons.
"""
fields = [] |
c306f6963e53b971674421eddca7f6b5c913281e | core/data/DataWriter.py | core/data/DataWriter.py | """
DataWriter.py
"""
from DataController import DataController
from DataReader import DataReader
from vtk import vtkMetaImageWriter
from vtk import vtkXMLImageDataWriter
class DataWriter(DataController):
"""
DataWriter writes an image data object to
disk using the provided format.
"""
def __init__(self):
super(DataWriter, self).__init__()
self.supportedExtensions = [DataReader.TypeMHD,
DataReader.TypeVTI]
def WriteToFile(self, imageData, exportFileName, fileType):
if fileType == DataReader.TypeMHA:
if not exportFileName.endswith(".mhd"):
exportFileName = exportFileName + ".mhd"
writer = vtkMetaImageWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
elif fileType == DataReader.TypeVTI:
writer = vtkXMLImageDataWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
else:
raise NotImplementedError("No writing support for type " + str(fileType))
| """
DataWriter.py
"""
from DataController import DataController
from DataReader import DataReader
from vtk import vtkMetaImageWriter
from vtk import vtkXMLImageDataWriter
class DataWriter(DataController):
"""
DataWriter writes an image data object to
disk using the provided format.
"""
def __init__(self):
super(DataWriter, self).__init__()
self.supportedExtensions = [DataReader.TypeMHD,
DataReader.TypeVTI]
def WriteToFile(self, imageData, exportFileName, fileType):
if fileType == DataReader.TypeMHD:
if not exportFileName.endswith(".mhd"):
exportFileName = exportFileName + ".mhd"
writer = vtkMetaImageWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
elif fileType == DataReader.TypeVTI:
writer = vtkXMLImageDataWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
else:
raise NotImplementedError("No writing support for type " + str(fileType))
| Fix for comparing with the wrong data type. | Fix for comparing with the wrong data type.
| Python | mit | berendkleinhaneveld/Registrationshop,berendkleinhaneveld/Registrationshop | """
DataWriter.py
"""
from DataController import DataController
from DataReader import DataReader
from vtk import vtkMetaImageWriter
from vtk import vtkXMLImageDataWriter
class DataWriter(DataController):
"""
DataWriter writes an image data object to
disk using the provided format.
"""
def __init__(self):
super(DataWriter, self).__init__()
self.supportedExtensions = [DataReader.TypeMHD,
DataReader.TypeVTI]
def WriteToFile(self, imageData, exportFileName, fileType):
if fileType == DataReader.TypeMHA:
if not exportFileName.endswith(".mhd"):
exportFileName = exportFileName + ".mhd"
writer = vtkMetaImageWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
elif fileType == DataReader.TypeVTI:
writer = vtkXMLImageDataWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
else:
raise NotImplementedError("No writing support for type " + str(fileType))
Fix for comparing with the wrong data type. | """
DataWriter.py
"""
from DataController import DataController
from DataReader import DataReader
from vtk import vtkMetaImageWriter
from vtk import vtkXMLImageDataWriter
class DataWriter(DataController):
"""
DataWriter writes an image data object to
disk using the provided format.
"""
def __init__(self):
super(DataWriter, self).__init__()
self.supportedExtensions = [DataReader.TypeMHD,
DataReader.TypeVTI]
def WriteToFile(self, imageData, exportFileName, fileType):
if fileType == DataReader.TypeMHD:
if not exportFileName.endswith(".mhd"):
exportFileName = exportFileName + ".mhd"
writer = vtkMetaImageWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
elif fileType == DataReader.TypeVTI:
writer = vtkXMLImageDataWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
else:
raise NotImplementedError("No writing support for type " + str(fileType))
| <commit_before>"""
DataWriter.py
"""
from DataController import DataController
from DataReader import DataReader
from vtk import vtkMetaImageWriter
from vtk import vtkXMLImageDataWriter
class DataWriter(DataController):
"""
DataWriter writes an image data object to
disk using the provided format.
"""
def __init__(self):
super(DataWriter, self).__init__()
self.supportedExtensions = [DataReader.TypeMHD,
DataReader.TypeVTI]
def WriteToFile(self, imageData, exportFileName, fileType):
if fileType == DataReader.TypeMHA:
if not exportFileName.endswith(".mhd"):
exportFileName = exportFileName + ".mhd"
writer = vtkMetaImageWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
elif fileType == DataReader.TypeVTI:
writer = vtkXMLImageDataWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
else:
raise NotImplementedError("No writing support for type " + str(fileType))
<commit_msg>Fix for comparing with the wrong data type.<commit_after> | """
DataWriter.py
"""
from DataController import DataController
from DataReader import DataReader
from vtk import vtkMetaImageWriter
from vtk import vtkXMLImageDataWriter
class DataWriter(DataController):
"""
DataWriter writes an image data object to
disk using the provided format.
"""
def __init__(self):
super(DataWriter, self).__init__()
self.supportedExtensions = [DataReader.TypeMHD,
DataReader.TypeVTI]
def WriteToFile(self, imageData, exportFileName, fileType):
if fileType == DataReader.TypeMHD:
if not exportFileName.endswith(".mhd"):
exportFileName = exportFileName + ".mhd"
writer = vtkMetaImageWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
elif fileType == DataReader.TypeVTI:
writer = vtkXMLImageDataWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
else:
raise NotImplementedError("No writing support for type " + str(fileType))
| """
DataWriter.py
"""
from DataController import DataController
from DataReader import DataReader
from vtk import vtkMetaImageWriter
from vtk import vtkXMLImageDataWriter
class DataWriter(DataController):
"""
DataWriter writes an image data object to
disk using the provided format.
"""
def __init__(self):
super(DataWriter, self).__init__()
self.supportedExtensions = [DataReader.TypeMHD,
DataReader.TypeVTI]
def WriteToFile(self, imageData, exportFileName, fileType):
if fileType == DataReader.TypeMHA:
if not exportFileName.endswith(".mhd"):
exportFileName = exportFileName + ".mhd"
writer = vtkMetaImageWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
elif fileType == DataReader.TypeVTI:
writer = vtkXMLImageDataWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
else:
raise NotImplementedError("No writing support for type " + str(fileType))
Fix for comparing with the wrong data type."""
DataWriter.py
"""
from DataController import DataController
from DataReader import DataReader
from vtk import vtkMetaImageWriter
from vtk import vtkXMLImageDataWriter
class DataWriter(DataController):
"""
DataWriter writes an image data object to
disk using the provided format.
"""
def __init__(self):
super(DataWriter, self).__init__()
self.supportedExtensions = [DataReader.TypeMHD,
DataReader.TypeVTI]
def WriteToFile(self, imageData, exportFileName, fileType):
if fileType == DataReader.TypeMHD:
if not exportFileName.endswith(".mhd"):
exportFileName = exportFileName + ".mhd"
writer = vtkMetaImageWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
elif fileType == DataReader.TypeVTI:
writer = vtkXMLImageDataWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
else:
raise NotImplementedError("No writing support for type " + str(fileType))
| <commit_before>"""
DataWriter.py
"""
from DataController import DataController
from DataReader import DataReader
from vtk import vtkMetaImageWriter
from vtk import vtkXMLImageDataWriter
class DataWriter(DataController):
"""
DataWriter writes an image data object to
disk using the provided format.
"""
def __init__(self):
super(DataWriter, self).__init__()
self.supportedExtensions = [DataReader.TypeMHD,
DataReader.TypeVTI]
def WriteToFile(self, imageData, exportFileName, fileType):
if fileType == DataReader.TypeMHA:
if not exportFileName.endswith(".mhd"):
exportFileName = exportFileName + ".mhd"
writer = vtkMetaImageWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
elif fileType == DataReader.TypeVTI:
writer = vtkXMLImageDataWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
else:
raise NotImplementedError("No writing support for type " + str(fileType))
<commit_msg>Fix for comparing with the wrong data type.<commit_after>"""
DataWriter.py
"""
from DataController import DataController
from DataReader import DataReader
from vtk import vtkMetaImageWriter
from vtk import vtkXMLImageDataWriter
class DataWriter(DataController):
"""
DataWriter writes an image data object to
disk using the provided format.
"""
def __init__(self):
super(DataWriter, self).__init__()
self.supportedExtensions = [DataReader.TypeMHD,
DataReader.TypeVTI]
def WriteToFile(self, imageData, exportFileName, fileType):
if fileType == DataReader.TypeMHD:
if not exportFileName.endswith(".mhd"):
exportFileName = exportFileName + ".mhd"
writer = vtkMetaImageWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
elif fileType == DataReader.TypeVTI:
writer = vtkXMLImageDataWriter()
writer.SetFileName(exportFileName)
writer.SetInputData(imageData)
writer.Write()
else:
raise NotImplementedError("No writing support for type " + str(fileType))
|
bfe884723d06252648cb95fdfc0f9dd0f804795f | proxyswitch/driver.py | proxyswitch/driver.py | from flask import Flask
class Driver:
'''
Holds the driver state so the flasked script can change behaviour based
on what the user injects via HTTP
'''
name = 'nobody'
def start(self, name):
self.name = name
return self.name
def stop(self):
self.name = 'nobody'
return self.name
app = Flask('proxapp')
driver = Driver()
def register(context):
context.app_registry.add(app, 'proxapp', 5000)
return context
@app.route('/')
def index():
return 'Try /start/:driver or /stop/:driver instead'
@app.route('/stop')
def stop_driver():
driver.stop()
return 'No drivers running\n'
@app.route('/<driver_name>/start')
def start_driver(driver_name):
print(driver.start(driver_name))
return '{} driver started\n'.format(driver_name)
| from flask import Flask
class Driver:
'''
Holds the driver state so the flasked script can change behaviour based
on what the user injects via HTTP
'''
name = 'nobody'
def start(self, name):
self.name = name
return self.name
def stop(self):
self.name = 'nobody'
return self.name
app = Flask('proxapp')
app.host = '0.0.0.0'
driver = Driver()
def register(context):
context.app_registry.add(app, 'proxapp', 5000)
return context
@app.route('/')
def index():
return 'Try /start/:driver or /stop/:driver instead'
@app.route('/stop')
def stop_driver():
driver.stop()
return 'No drivers running\n'
@app.route('/<driver_name>/start')
def start_driver(driver_name):
print(driver.start(driver_name))
return '{} driver started\n'.format(driver_name)
| Change Flask interface to 0.0.0.0 | Change Flask interface to 0.0.0.0
| Python | mit | ustwo/mastermind,ustwo/mastermind | from flask import Flask
class Driver:
'''
Holds the driver state so the flasked script can change behaviour based
on what the user injects via HTTP
'''
name = 'nobody'
def start(self, name):
self.name = name
return self.name
def stop(self):
self.name = 'nobody'
return self.name
app = Flask('proxapp')
driver = Driver()
def register(context):
context.app_registry.add(app, 'proxapp', 5000)
return context
@app.route('/')
def index():
return 'Try /start/:driver or /stop/:driver instead'
@app.route('/stop')
def stop_driver():
driver.stop()
return 'No drivers running\n'
@app.route('/<driver_name>/start')
def start_driver(driver_name):
print(driver.start(driver_name))
return '{} driver started\n'.format(driver_name)
Change Flask interface to 0.0.0.0 | from flask import Flask
class Driver:
'''
Holds the driver state so the flasked script can change behaviour based
on what the user injects via HTTP
'''
name = 'nobody'
def start(self, name):
self.name = name
return self.name
def stop(self):
self.name = 'nobody'
return self.name
app = Flask('proxapp')
app.host = '0.0.0.0'
driver = Driver()
def register(context):
context.app_registry.add(app, 'proxapp', 5000)
return context
@app.route('/')
def index():
return 'Try /start/:driver or /stop/:driver instead'
@app.route('/stop')
def stop_driver():
driver.stop()
return 'No drivers running\n'
@app.route('/<driver_name>/start')
def start_driver(driver_name):
print(driver.start(driver_name))
return '{} driver started\n'.format(driver_name)
| <commit_before>from flask import Flask
class Driver:
'''
Holds the driver state so the flasked script can change behaviour based
on what the user injects via HTTP
'''
name = 'nobody'
def start(self, name):
self.name = name
return self.name
def stop(self):
self.name = 'nobody'
return self.name
app = Flask('proxapp')
driver = Driver()
def register(context):
context.app_registry.add(app, 'proxapp', 5000)
return context
@app.route('/')
def index():
return 'Try /start/:driver or /stop/:driver instead'
@app.route('/stop')
def stop_driver():
driver.stop()
return 'No drivers running\n'
@app.route('/<driver_name>/start')
def start_driver(driver_name):
print(driver.start(driver_name))
return '{} driver started\n'.format(driver_name)
<commit_msg>Change Flask interface to 0.0.0.0<commit_after> | from flask import Flask
class Driver:
'''
Holds the driver state so the flasked script can change behaviour based
on what the user injects via HTTP
'''
name = 'nobody'
def start(self, name):
self.name = name
return self.name
def stop(self):
self.name = 'nobody'
return self.name
app = Flask('proxapp')
app.host = '0.0.0.0'
driver = Driver()
def register(context):
context.app_registry.add(app, 'proxapp', 5000)
return context
@app.route('/')
def index():
return 'Try /start/:driver or /stop/:driver instead'
@app.route('/stop')
def stop_driver():
driver.stop()
return 'No drivers running\n'
@app.route('/<driver_name>/start')
def start_driver(driver_name):
print(driver.start(driver_name))
return '{} driver started\n'.format(driver_name)
| from flask import Flask
class Driver:
'''
Holds the driver state so the flasked script can change behaviour based
on what the user injects via HTTP
'''
name = 'nobody'
def start(self, name):
self.name = name
return self.name
def stop(self):
self.name = 'nobody'
return self.name
app = Flask('proxapp')
driver = Driver()
def register(context):
context.app_registry.add(app, 'proxapp', 5000)
return context
@app.route('/')
def index():
return 'Try /start/:driver or /stop/:driver instead'
@app.route('/stop')
def stop_driver():
driver.stop()
return 'No drivers running\n'
@app.route('/<driver_name>/start')
def start_driver(driver_name):
print(driver.start(driver_name))
return '{} driver started\n'.format(driver_name)
Change Flask interface to 0.0.0.0from flask import Flask
class Driver:
'''
Holds the driver state so the flasked script can change behaviour based
on what the user injects via HTTP
'''
name = 'nobody'
def start(self, name):
self.name = name
return self.name
def stop(self):
self.name = 'nobody'
return self.name
app = Flask('proxapp')
app.host = '0.0.0.0'
driver = Driver()
def register(context):
context.app_registry.add(app, 'proxapp', 5000)
return context
@app.route('/')
def index():
return 'Try /start/:driver or /stop/:driver instead'
@app.route('/stop')
def stop_driver():
driver.stop()
return 'No drivers running\n'
@app.route('/<driver_name>/start')
def start_driver(driver_name):
print(driver.start(driver_name))
return '{} driver started\n'.format(driver_name)
| <commit_before>from flask import Flask
class Driver:
'''
Holds the driver state so the flasked script can change behaviour based
on what the user injects via HTTP
'''
name = 'nobody'
def start(self, name):
self.name = name
return self.name
def stop(self):
self.name = 'nobody'
return self.name
app = Flask('proxapp')
driver = Driver()
def register(context):
context.app_registry.add(app, 'proxapp', 5000)
return context
@app.route('/')
def index():
return 'Try /start/:driver or /stop/:driver instead'
@app.route('/stop')
def stop_driver():
driver.stop()
return 'No drivers running\n'
@app.route('/<driver_name>/start')
def start_driver(driver_name):
print(driver.start(driver_name))
return '{} driver started\n'.format(driver_name)
<commit_msg>Change Flask interface to 0.0.0.0<commit_after>from flask import Flask
class Driver:
'''
Holds the driver state so the flasked script can change behaviour based
on what the user injects via HTTP
'''
name = 'nobody'
def start(self, name):
self.name = name
return self.name
def stop(self):
self.name = 'nobody'
return self.name
app = Flask('proxapp')
app.host = '0.0.0.0'
driver = Driver()
def register(context):
context.app_registry.add(app, 'proxapp', 5000)
return context
@app.route('/')
def index():
return 'Try /start/:driver or /stop/:driver instead'
@app.route('/stop')
def stop_driver():
driver.stop()
return 'No drivers running\n'
@app.route('/<driver_name>/start')
def start_driver(driver_name):
print(driver.start(driver_name))
return '{} driver started\n'.format(driver_name)
|
323f897e3550f41edc139352a6ac9d95ddf7228d | seriesly/helper/context_processors.py | seriesly/helper/context_processors.py | from django.conf import settings
def site_info(request):
return {'DOMAIN_URL': settings.DOMAIN_URL,
'SECURE_DOMAIN_URL': settings.SECURE_DOMAIN_URL,
'DEBUG': settings.DEBUG} | from django.conf import settings
def site_info(request):
return {'DOMAIN_URL': settings.DOMAIN_URL,
'SECURE_DOMAIN_URL': settings.SECURE_DOMAIN_URL,
'DEFAULT_FROM_EMAIL': settings.DEFAULT_FROM_EMAIL,
'DEBUG': settings.DEBUG}
| Add DEFAULT_FROM_EMAIL to default template context | Add DEFAULT_FROM_EMAIL to default template context
| Python | agpl-3.0 | maxgraser/seriesly,maxgraser/seriesly,stefanw/seriesly,stefanw/seriesly,maxgraser/seriesly | from django.conf import settings
def site_info(request):
return {'DOMAIN_URL': settings.DOMAIN_URL,
'SECURE_DOMAIN_URL': settings.SECURE_DOMAIN_URL,
'DEBUG': settings.DEBUG}Add DEFAULT_FROM_EMAIL to default template context | from django.conf import settings
def site_info(request):
return {'DOMAIN_URL': settings.DOMAIN_URL,
'SECURE_DOMAIN_URL': settings.SECURE_DOMAIN_URL,
'DEFAULT_FROM_EMAIL': settings.DEFAULT_FROM_EMAIL,
'DEBUG': settings.DEBUG}
| <commit_before>from django.conf import settings
def site_info(request):
return {'DOMAIN_URL': settings.DOMAIN_URL,
'SECURE_DOMAIN_URL': settings.SECURE_DOMAIN_URL,
'DEBUG': settings.DEBUG}<commit_msg>Add DEFAULT_FROM_EMAIL to default template context<commit_after> | from django.conf import settings
def site_info(request):
return {'DOMAIN_URL': settings.DOMAIN_URL,
'SECURE_DOMAIN_URL': settings.SECURE_DOMAIN_URL,
'DEFAULT_FROM_EMAIL': settings.DEFAULT_FROM_EMAIL,
'DEBUG': settings.DEBUG}
| from django.conf import settings
def site_info(request):
return {'DOMAIN_URL': settings.DOMAIN_URL,
'SECURE_DOMAIN_URL': settings.SECURE_DOMAIN_URL,
'DEBUG': settings.DEBUG}Add DEFAULT_FROM_EMAIL to default template contextfrom django.conf import settings
def site_info(request):
return {'DOMAIN_URL': settings.DOMAIN_URL,
'SECURE_DOMAIN_URL': settings.SECURE_DOMAIN_URL,
'DEFAULT_FROM_EMAIL': settings.DEFAULT_FROM_EMAIL,
'DEBUG': settings.DEBUG}
| <commit_before>from django.conf import settings
def site_info(request):
return {'DOMAIN_URL': settings.DOMAIN_URL,
'SECURE_DOMAIN_URL': settings.SECURE_DOMAIN_URL,
'DEBUG': settings.DEBUG}<commit_msg>Add DEFAULT_FROM_EMAIL to default template context<commit_after>from django.conf import settings
def site_info(request):
return {'DOMAIN_URL': settings.DOMAIN_URL,
'SECURE_DOMAIN_URL': settings.SECURE_DOMAIN_URL,
'DEFAULT_FROM_EMAIL': settings.DEFAULT_FROM_EMAIL,
'DEBUG': settings.DEBUG}
|
fd054790ce32c3918f6edbe824540c09d7efce59 | stagehand/providers/__init__.py | stagehand/providers/__init__.py | import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
| import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
| Remove tvrage from active providers as site is shut down | Remove tvrage from active providers as site is shut down
| Python | mit | jtackaberry/stagehand,jtackaberry/stagehand | import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
Remove tvrage from active providers as site is shut down | import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
| <commit_before>import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
<commit_msg>Remove tvrage from active providers as site is shut down<commit_after> | import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
| import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
Remove tvrage from active providers as site is shut downimport asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
| <commit_before>import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb', 'tvrage'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
<commit_msg>Remove tvrage from active providers as site is shut down<commit_after>import asyncio
from ..utils import load_plugins, invoke_plugins
from .base import ProviderError
plugins, broken_plugins = load_plugins('providers', ['thetvdb'])
@asyncio.coroutine
def start(manager):
"""
Called when the manager is starting.
"""
yield from invoke_plugins(plugins, 'start', manager)
for name, error in broken_plugins.items():
log.warning('failed to load provider plugin %s: %s', name, error)
|
22b5f7ecc6057252ec77d037522b5783c5f86c1f | mcmodfixes.py | mcmodfixes.py | #!/usr/bin/python
# Fixes and mod-specific data for various mods' mcmod.info files
DEP_BLACKLIST = set((
"mod_MinecraftForge", # we always have Forge
"Forge", # typo for mod_MinecraftForge
"Industrialcraft", # typo for IC2
"GUI_Api", # typo for GuiAPI and not needed on server
))
DEP_ADDITIONS = {
"gregtech": ["IC2"],
"MineFactoryReloaded": ["PowerCrystalsCore"],
"NetherOres": ["PowerCrystalsCore"],
"PowerConverters": ["PowerCrystalsCore"],
"FlatBedrock": ["PowerCrystalsCore"],
"immibis-microblocks": ["ImmibisCore"],
}
def getExtraDeps(mod):
for k, v in DEP_ADDITIONS.iteritems():
if mod.startswith(k):
return set(v)
return set()
def fixDeps(mod, deps):
deps = set(deps)
deps -= DEP_BLACKLIST
deps |= getExtraDeps(mod)
return deps
MOD_IDS = {
"PowerCrystalsCore": ["PowerCrystalsCore"],
}
def fixModIDs(mod, ids):
for k, v in MOD_IDS.iteritems():
if mod.startswith(k):
return v
return ids
COREMODS = ["PowerCrystalsCore", "immibis-microblocks"]
def isCoremod(fn):
for k in COREMODS:
if fn.startswith(k):
return True
return False
| #!/usr/bin/python
# Fixes and mod-specific data for various mods' mcmod.info files
DEP_BLACKLIST = set((
"mod_MinecraftForge", # we always have Forge
"Forge", # typo for mod_MinecraftForge
"Industrialcraft", # typo for IC2
"GUI_Api", # typo for GuiAPI and not needed on server
"EurysCore", # replaced by SlimevoidLib?
))
DEP_ADDITIONS = {
"gregtech": ["IC2"],
"MineFactoryReloaded": ["PowerCrystalsCore"],
"NetherOres": ["PowerCrystalsCore"],
"PowerConverters": ["PowerCrystalsCore"],
"FlatBedrock": ["PowerCrystalsCore"],
"immibis-microblocks": ["ImmibisCore"],
"SlopesAndCorners": ["SlimevoidLib"],
}
def getExtraDeps(mod):
for k, v in DEP_ADDITIONS.iteritems():
if mod.startswith(k):
return set(v)
return set()
def fixDeps(mod, deps):
deps = set(deps)
deps -= DEP_BLACKLIST
deps |= getExtraDeps(mod)
return deps
MOD_IDS = {
"PowerCrystalsCore": ["PowerCrystalsCore"],
}
def fixModIDs(mod, ids):
for k, v in MOD_IDS.iteritems():
if mod.startswith(k):
return v
return ids
COREMODS = ["PowerCrystalsCore", "immibis-microblocks"]
def isCoremod(fn):
for k in COREMODS:
if fn.startswith(k):
return True
return False
| Add mcmod.info fix for SlopesAndCorners SlimevoidLib dependency | Add mcmod.info fix for SlopesAndCorners SlimevoidLib dependency
| Python | bsd-3-clause | agaricusb/ModAnalyzer,agaricusb/ModAnalyzer | #!/usr/bin/python
# Fixes and mod-specific data for various mods' mcmod.info files
DEP_BLACKLIST = set((
"mod_MinecraftForge", # we always have Forge
"Forge", # typo for mod_MinecraftForge
"Industrialcraft", # typo for IC2
"GUI_Api", # typo for GuiAPI and not needed on server
))
DEP_ADDITIONS = {
"gregtech": ["IC2"],
"MineFactoryReloaded": ["PowerCrystalsCore"],
"NetherOres": ["PowerCrystalsCore"],
"PowerConverters": ["PowerCrystalsCore"],
"FlatBedrock": ["PowerCrystalsCore"],
"immibis-microblocks": ["ImmibisCore"],
}
def getExtraDeps(mod):
for k, v in DEP_ADDITIONS.iteritems():
if mod.startswith(k):
return set(v)
return set()
def fixDeps(mod, deps):
deps = set(deps)
deps -= DEP_BLACKLIST
deps |= getExtraDeps(mod)
return deps
MOD_IDS = {
"PowerCrystalsCore": ["PowerCrystalsCore"],
}
def fixModIDs(mod, ids):
for k, v in MOD_IDS.iteritems():
if mod.startswith(k):
return v
return ids
COREMODS = ["PowerCrystalsCore", "immibis-microblocks"]
def isCoremod(fn):
for k in COREMODS:
if fn.startswith(k):
return True
return False
Add mcmod.info fix for SlopesAndCorners SlimevoidLib dependency | #!/usr/bin/python
# Fixes and mod-specific data for various mods' mcmod.info files
DEP_BLACKLIST = set((
"mod_MinecraftForge", # we always have Forge
"Forge", # typo for mod_MinecraftForge
"Industrialcraft", # typo for IC2
"GUI_Api", # typo for GuiAPI and not needed on server
"EurysCore", # replaced by SlimevoidLib?
))
DEP_ADDITIONS = {
"gregtech": ["IC2"],
"MineFactoryReloaded": ["PowerCrystalsCore"],
"NetherOres": ["PowerCrystalsCore"],
"PowerConverters": ["PowerCrystalsCore"],
"FlatBedrock": ["PowerCrystalsCore"],
"immibis-microblocks": ["ImmibisCore"],
"SlopesAndCorners": ["SlimevoidLib"],
}
def getExtraDeps(mod):
for k, v in DEP_ADDITIONS.iteritems():
if mod.startswith(k):
return set(v)
return set()
def fixDeps(mod, deps):
deps = set(deps)
deps -= DEP_BLACKLIST
deps |= getExtraDeps(mod)
return deps
MOD_IDS = {
"PowerCrystalsCore": ["PowerCrystalsCore"],
}
def fixModIDs(mod, ids):
for k, v in MOD_IDS.iteritems():
if mod.startswith(k):
return v
return ids
COREMODS = ["PowerCrystalsCore", "immibis-microblocks"]
def isCoremod(fn):
for k in COREMODS:
if fn.startswith(k):
return True
return False
| <commit_before>#!/usr/bin/python
# Fixes and mod-specific data for various mods' mcmod.info files
DEP_BLACKLIST = set((
"mod_MinecraftForge", # we always have Forge
"Forge", # typo for mod_MinecraftForge
"Industrialcraft", # typo for IC2
"GUI_Api", # typo for GuiAPI and not needed on server
))
DEP_ADDITIONS = {
"gregtech": ["IC2"],
"MineFactoryReloaded": ["PowerCrystalsCore"],
"NetherOres": ["PowerCrystalsCore"],
"PowerConverters": ["PowerCrystalsCore"],
"FlatBedrock": ["PowerCrystalsCore"],
"immibis-microblocks": ["ImmibisCore"],
}
def getExtraDeps(mod):
for k, v in DEP_ADDITIONS.iteritems():
if mod.startswith(k):
return set(v)
return set()
def fixDeps(mod, deps):
deps = set(deps)
deps -= DEP_BLACKLIST
deps |= getExtraDeps(mod)
return deps
MOD_IDS = {
"PowerCrystalsCore": ["PowerCrystalsCore"],
}
def fixModIDs(mod, ids):
for k, v in MOD_IDS.iteritems():
if mod.startswith(k):
return v
return ids
COREMODS = ["PowerCrystalsCore", "immibis-microblocks"]
def isCoremod(fn):
for k in COREMODS:
if fn.startswith(k):
return True
return False
<commit_msg>Add mcmod.info fix for SlopesAndCorners SlimevoidLib dependency<commit_after> | #!/usr/bin/python
# Fixes and mod-specific data for various mods' mcmod.info files
DEP_BLACKLIST = set((
"mod_MinecraftForge", # we always have Forge
"Forge", # typo for mod_MinecraftForge
"Industrialcraft", # typo for IC2
"GUI_Api", # typo for GuiAPI and not needed on server
"EurysCore", # replaced by SlimevoidLib?
))
DEP_ADDITIONS = {
"gregtech": ["IC2"],
"MineFactoryReloaded": ["PowerCrystalsCore"],
"NetherOres": ["PowerCrystalsCore"],
"PowerConverters": ["PowerCrystalsCore"],
"FlatBedrock": ["PowerCrystalsCore"],
"immibis-microblocks": ["ImmibisCore"],
"SlopesAndCorners": ["SlimevoidLib"],
}
def getExtraDeps(mod):
for k, v in DEP_ADDITIONS.iteritems():
if mod.startswith(k):
return set(v)
return set()
def fixDeps(mod, deps):
deps = set(deps)
deps -= DEP_BLACKLIST
deps |= getExtraDeps(mod)
return deps
MOD_IDS = {
"PowerCrystalsCore": ["PowerCrystalsCore"],
}
def fixModIDs(mod, ids):
for k, v in MOD_IDS.iteritems():
if mod.startswith(k):
return v
return ids
COREMODS = ["PowerCrystalsCore", "immibis-microblocks"]
def isCoremod(fn):
for k in COREMODS:
if fn.startswith(k):
return True
return False
| #!/usr/bin/python
# Fixes and mod-specific data for various mods' mcmod.info files
DEP_BLACKLIST = set((
"mod_MinecraftForge", # we always have Forge
"Forge", # typo for mod_MinecraftForge
"Industrialcraft", # typo for IC2
"GUI_Api", # typo for GuiAPI and not needed on server
))
DEP_ADDITIONS = {
"gregtech": ["IC2"],
"MineFactoryReloaded": ["PowerCrystalsCore"],
"NetherOres": ["PowerCrystalsCore"],
"PowerConverters": ["PowerCrystalsCore"],
"FlatBedrock": ["PowerCrystalsCore"],
"immibis-microblocks": ["ImmibisCore"],
}
def getExtraDeps(mod):
for k, v in DEP_ADDITIONS.iteritems():
if mod.startswith(k):
return set(v)
return set()
def fixDeps(mod, deps):
deps = set(deps)
deps -= DEP_BLACKLIST
deps |= getExtraDeps(mod)
return deps
MOD_IDS = {
"PowerCrystalsCore": ["PowerCrystalsCore"],
}
def fixModIDs(mod, ids):
for k, v in MOD_IDS.iteritems():
if mod.startswith(k):
return v
return ids
COREMODS = ["PowerCrystalsCore", "immibis-microblocks"]
def isCoremod(fn):
for k in COREMODS:
if fn.startswith(k):
return True
return False
Add mcmod.info fix for SlopesAndCorners SlimevoidLib dependency#!/usr/bin/python
# Fixes and mod-specific data for various mods' mcmod.info files
DEP_BLACKLIST = set((
"mod_MinecraftForge", # we always have Forge
"Forge", # typo for mod_MinecraftForge
"Industrialcraft", # typo for IC2
"GUI_Api", # typo for GuiAPI and not needed on server
"EurysCore", # replaced by SlimevoidLib?
))
DEP_ADDITIONS = {
"gregtech": ["IC2"],
"MineFactoryReloaded": ["PowerCrystalsCore"],
"NetherOres": ["PowerCrystalsCore"],
"PowerConverters": ["PowerCrystalsCore"],
"FlatBedrock": ["PowerCrystalsCore"],
"immibis-microblocks": ["ImmibisCore"],
"SlopesAndCorners": ["SlimevoidLib"],
}
def getExtraDeps(mod):
for k, v in DEP_ADDITIONS.iteritems():
if mod.startswith(k):
return set(v)
return set()
def fixDeps(mod, deps):
deps = set(deps)
deps -= DEP_BLACKLIST
deps |= getExtraDeps(mod)
return deps
MOD_IDS = {
"PowerCrystalsCore": ["PowerCrystalsCore"],
}
def fixModIDs(mod, ids):
for k, v in MOD_IDS.iteritems():
if mod.startswith(k):
return v
return ids
COREMODS = ["PowerCrystalsCore", "immibis-microblocks"]
def isCoremod(fn):
for k in COREMODS:
if fn.startswith(k):
return True
return False
| <commit_before>#!/usr/bin/python
# Fixes and mod-specific data for various mods' mcmod.info files
DEP_BLACKLIST = set((
"mod_MinecraftForge", # we always have Forge
"Forge", # typo for mod_MinecraftForge
"Industrialcraft", # typo for IC2
"GUI_Api", # typo for GuiAPI and not needed on server
))
DEP_ADDITIONS = {
"gregtech": ["IC2"],
"MineFactoryReloaded": ["PowerCrystalsCore"],
"NetherOres": ["PowerCrystalsCore"],
"PowerConverters": ["PowerCrystalsCore"],
"FlatBedrock": ["PowerCrystalsCore"],
"immibis-microblocks": ["ImmibisCore"],
}
def getExtraDeps(mod):
for k, v in DEP_ADDITIONS.iteritems():
if mod.startswith(k):
return set(v)
return set()
def fixDeps(mod, deps):
deps = set(deps)
deps -= DEP_BLACKLIST
deps |= getExtraDeps(mod)
return deps
MOD_IDS = {
"PowerCrystalsCore": ["PowerCrystalsCore"],
}
def fixModIDs(mod, ids):
for k, v in MOD_IDS.iteritems():
if mod.startswith(k):
return v
return ids
COREMODS = ["PowerCrystalsCore", "immibis-microblocks"]
def isCoremod(fn):
for k in COREMODS:
if fn.startswith(k):
return True
return False
<commit_msg>Add mcmod.info fix for SlopesAndCorners SlimevoidLib dependency<commit_after>#!/usr/bin/python
# Fixes and mod-specific data for various mods' mcmod.info files
DEP_BLACKLIST = set((
"mod_MinecraftForge", # we always have Forge
"Forge", # typo for mod_MinecraftForge
"Industrialcraft", # typo for IC2
"GUI_Api", # typo for GuiAPI and not needed on server
"EurysCore", # replaced by SlimevoidLib?
))
DEP_ADDITIONS = {
"gregtech": ["IC2"],
"MineFactoryReloaded": ["PowerCrystalsCore"],
"NetherOres": ["PowerCrystalsCore"],
"PowerConverters": ["PowerCrystalsCore"],
"FlatBedrock": ["PowerCrystalsCore"],
"immibis-microblocks": ["ImmibisCore"],
"SlopesAndCorners": ["SlimevoidLib"],
}
def getExtraDeps(mod):
for k, v in DEP_ADDITIONS.iteritems():
if mod.startswith(k):
return set(v)
return set()
def fixDeps(mod, deps):
deps = set(deps)
deps -= DEP_BLACKLIST
deps |= getExtraDeps(mod)
return deps
MOD_IDS = {
"PowerCrystalsCore": ["PowerCrystalsCore"],
}
def fixModIDs(mod, ids):
for k, v in MOD_IDS.iteritems():
if mod.startswith(k):
return v
return ids
COREMODS = ["PowerCrystalsCore", "immibis-microblocks"]
def isCoremod(fn):
for k in COREMODS:
if fn.startswith(k):
return True
return False
|
60a90722fbd5fc047fee5e9f7377f03e11f6a654 | examples/root_finding/test_funcs.py | examples/root_finding/test_funcs.py | import math
def f1(x):
"""
Test function 1
"""
return x*x*x - math.pi*x + math.e/100
| import numpy as npy
def f1(x):
"""
Test function 1
"""
return x*x*x - npy.pi*x + npy.e/100
def f2(x):
"""
Test function 2
"""
return -1.13 + npy.tanh(x-2) + 4*npy.exp(-x)*npy.sin((1/8.)*x**3) \
*x + .1*npy.exp((1/35.)*x) | Use numpy instead of math to allow vectorization | Use numpy instead of math to allow vectorization
| Python | bsd-3-clause | robclewley/fovea,akuefler/fovea | import math
def f1(x):
"""
Test function 1
"""
return x*x*x - math.pi*x + math.e/100
Use numpy instead of math to allow vectorization | import numpy as npy
def f1(x):
"""
Test function 1
"""
return x*x*x - npy.pi*x + npy.e/100
def f2(x):
"""
Test function 2
"""
return -1.13 + npy.tanh(x-2) + 4*npy.exp(-x)*npy.sin((1/8.)*x**3) \
*x + .1*npy.exp((1/35.)*x) | <commit_before>import math
def f1(x):
"""
Test function 1
"""
return x*x*x - math.pi*x + math.e/100
<commit_msg>Use numpy instead of math to allow vectorization<commit_after> | import numpy as npy
def f1(x):
"""
Test function 1
"""
return x*x*x - npy.pi*x + npy.e/100
def f2(x):
"""
Test function 2
"""
return -1.13 + npy.tanh(x-2) + 4*npy.exp(-x)*npy.sin((1/8.)*x**3) \
*x + .1*npy.exp((1/35.)*x) | import math
def f1(x):
"""
Test function 1
"""
return x*x*x - math.pi*x + math.e/100
Use numpy instead of math to allow vectorizationimport numpy as npy
def f1(x):
"""
Test function 1
"""
return x*x*x - npy.pi*x + npy.e/100
def f2(x):
"""
Test function 2
"""
return -1.13 + npy.tanh(x-2) + 4*npy.exp(-x)*npy.sin((1/8.)*x**3) \
*x + .1*npy.exp((1/35.)*x) | <commit_before>import math
def f1(x):
"""
Test function 1
"""
return x*x*x - math.pi*x + math.e/100
<commit_msg>Use numpy instead of math to allow vectorization<commit_after>import numpy as npy
def f1(x):
"""
Test function 1
"""
return x*x*x - npy.pi*x + npy.e/100
def f2(x):
"""
Test function 2
"""
return -1.13 + npy.tanh(x-2) + 4*npy.exp(-x)*npy.sin((1/8.)*x**3) \
*x + .1*npy.exp((1/35.)*x) |
cd030a1ed2c3c7f0bf7d9a5d86f9cc81f802fcba | corehq/mobile_flags.py | corehq/mobile_flags.py | from collections import namedtuple
TAG_DIMAGI_ONLY = 'Dimagi Only'
MobileFlag = namedtuple('MobileFlag', 'slug label tags')
SUPERUSER = MobileFlag(
'superuser',
'Enable superuser-only features',
tags=(TAG_DIMAGI_ONLY,)
)
| from collections import namedtuple
MobileFlag = namedtuple('MobileFlag', 'slug label')
SUPERUSER = MobileFlag(
'superuser',
'Enable superuser-only features'
)
| Add tags for mobile flags when you need them | Add tags for mobile flags when you need them
| Python | bsd-3-clause | dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq | from collections import namedtuple
TAG_DIMAGI_ONLY = 'Dimagi Only'
MobileFlag = namedtuple('MobileFlag', 'slug label tags')
SUPERUSER = MobileFlag(
'superuser',
'Enable superuser-only features',
tags=(TAG_DIMAGI_ONLY,)
)
Add tags for mobile flags when you need them | from collections import namedtuple
MobileFlag = namedtuple('MobileFlag', 'slug label')
SUPERUSER = MobileFlag(
'superuser',
'Enable superuser-only features'
)
| <commit_before>from collections import namedtuple
TAG_DIMAGI_ONLY = 'Dimagi Only'
MobileFlag = namedtuple('MobileFlag', 'slug label tags')
SUPERUSER = MobileFlag(
'superuser',
'Enable superuser-only features',
tags=(TAG_DIMAGI_ONLY,)
)
<commit_msg>Add tags for mobile flags when you need them<commit_after> | from collections import namedtuple
MobileFlag = namedtuple('MobileFlag', 'slug label')
SUPERUSER = MobileFlag(
'superuser',
'Enable superuser-only features'
)
| from collections import namedtuple
TAG_DIMAGI_ONLY = 'Dimagi Only'
MobileFlag = namedtuple('MobileFlag', 'slug label tags')
SUPERUSER = MobileFlag(
'superuser',
'Enable superuser-only features',
tags=(TAG_DIMAGI_ONLY,)
)
Add tags for mobile flags when you need themfrom collections import namedtuple
MobileFlag = namedtuple('MobileFlag', 'slug label')
SUPERUSER = MobileFlag(
'superuser',
'Enable superuser-only features'
)
| <commit_before>from collections import namedtuple
TAG_DIMAGI_ONLY = 'Dimagi Only'
MobileFlag = namedtuple('MobileFlag', 'slug label tags')
SUPERUSER = MobileFlag(
'superuser',
'Enable superuser-only features',
tags=(TAG_DIMAGI_ONLY,)
)
<commit_msg>Add tags for mobile flags when you need them<commit_after>from collections import namedtuple
MobileFlag = namedtuple('MobileFlag', 'slug label')
SUPERUSER = MobileFlag(
'superuser',
'Enable superuser-only features'
)
|
54bb12bdeec33e98451451837dce90665413bd67 | mgsv_names.py | mgsv_names.py | from __future__ import unicode_literals, print_function
import sqlite3, os, random
_select_random = 'select {0} from {1} limit 1 offset abs(random()) % (select count({0}) from {1});'
_select_uncommon = 'select value from uncommons where key=?;'
def generate_name():
conn = sqlite3.connect(os.path.join(os.path.dirname(__file__), 'names.db'))
cursor = conn.cursor()
adj = cursor.execute(_select_random.format('adjective', 'adjectives')).fetchone()[0]
anim = cursor.execute(_select_random.format('animal', 'animals')).fetchone()[0]
rare = cursor.execute(_select_random.format('name', 'rares')).fetchone()[0]
uncommon_anim = cursor.execute(_select_uncommon, [adj]).fetchone()
uncommon_adj = cursor.execute(_select_uncommon, [anim]).fetchone()
conn.close()
r = random.random()
if r < 0.001 or r >= 0.999:
return rare
elif r < 0.3 and uncommon_anim is not None:
return ' '.join((adj, uncommon_anim[0]))
elif r >= 0.7 and uncommon_adj is not None:
return ' '.join((uncommon_adj[0], anim))
return ' '.join((adj, anim))
if __name__ == '__main__':
for _ in range(20):
print(generate_name())
| from __future__ import unicode_literals, print_function
import sqlite3, os, random
_select_random = 'select {0} from {1} limit 1 offset abs(random()) % (select count({0}) from {1});'
_select_uncommon = 'select value from uncommons where key=?;'
def generate_name():
conn = sqlite3.connect(os.path.join(os.path.dirname(__file__), 'names.db'))
cursor = conn.cursor()
adj = cursor.execute(_select_random.format('adjective', 'adjectives')).fetchone()[0]
anim = cursor.execute(_select_random.format('animal', 'animals')).fetchone()[0]
rare = cursor.execute(_select_random.format('name', 'rares')).fetchone()[0]
uncommon_anim = cursor.execute(_select_uncommon, [adj]).fetchone()
uncommon_adj = cursor.execute(_select_uncommon, [anim]).fetchone()
conn.close()
r = random.random()
if r < 0.001 or r >= 0.999:
return rare
elif r < 0.3 and uncommon_anim is not None:
return ' '.join((adj, uncommon_anim[0]))
elif r >= 0.7 and uncommon_adj is not None:
return ' '.join((uncommon_adj[0], anim))
return ' '.join((adj, anim))
if __name__ == '__main__':
print(generate_name())
| Print one name at a time. | Print one name at a time.
| Python | unlicense | rotated8/mgsv_names | from __future__ import unicode_literals, print_function
import sqlite3, os, random
_select_random = 'select {0} from {1} limit 1 offset abs(random()) % (select count({0}) from {1});'
_select_uncommon = 'select value from uncommons where key=?;'
def generate_name():
conn = sqlite3.connect(os.path.join(os.path.dirname(__file__), 'names.db'))
cursor = conn.cursor()
adj = cursor.execute(_select_random.format('adjective', 'adjectives')).fetchone()[0]
anim = cursor.execute(_select_random.format('animal', 'animals')).fetchone()[0]
rare = cursor.execute(_select_random.format('name', 'rares')).fetchone()[0]
uncommon_anim = cursor.execute(_select_uncommon, [adj]).fetchone()
uncommon_adj = cursor.execute(_select_uncommon, [anim]).fetchone()
conn.close()
r = random.random()
if r < 0.001 or r >= 0.999:
return rare
elif r < 0.3 and uncommon_anim is not None:
return ' '.join((adj, uncommon_anim[0]))
elif r >= 0.7 and uncommon_adj is not None:
return ' '.join((uncommon_adj[0], anim))
return ' '.join((adj, anim))
if __name__ == '__main__':
for _ in range(20):
print(generate_name())
Print one name at a time. | from __future__ import unicode_literals, print_function
import sqlite3, os, random
_select_random = 'select {0} from {1} limit 1 offset abs(random()) % (select count({0}) from {1});'
_select_uncommon = 'select value from uncommons where key=?;'
def generate_name():
conn = sqlite3.connect(os.path.join(os.path.dirname(__file__), 'names.db'))
cursor = conn.cursor()
adj = cursor.execute(_select_random.format('adjective', 'adjectives')).fetchone()[0]
anim = cursor.execute(_select_random.format('animal', 'animals')).fetchone()[0]
rare = cursor.execute(_select_random.format('name', 'rares')).fetchone()[0]
uncommon_anim = cursor.execute(_select_uncommon, [adj]).fetchone()
uncommon_adj = cursor.execute(_select_uncommon, [anim]).fetchone()
conn.close()
r = random.random()
if r < 0.001 or r >= 0.999:
return rare
elif r < 0.3 and uncommon_anim is not None:
return ' '.join((adj, uncommon_anim[0]))
elif r >= 0.7 and uncommon_adj is not None:
return ' '.join((uncommon_adj[0], anim))
return ' '.join((adj, anim))
if __name__ == '__main__':
print(generate_name())
| <commit_before>from __future__ import unicode_literals, print_function
import sqlite3, os, random
_select_random = 'select {0} from {1} limit 1 offset abs(random()) % (select count({0}) from {1});'
_select_uncommon = 'select value from uncommons where key=?;'
def generate_name():
conn = sqlite3.connect(os.path.join(os.path.dirname(__file__), 'names.db'))
cursor = conn.cursor()
adj = cursor.execute(_select_random.format('adjective', 'adjectives')).fetchone()[0]
anim = cursor.execute(_select_random.format('animal', 'animals')).fetchone()[0]
rare = cursor.execute(_select_random.format('name', 'rares')).fetchone()[0]
uncommon_anim = cursor.execute(_select_uncommon, [adj]).fetchone()
uncommon_adj = cursor.execute(_select_uncommon, [anim]).fetchone()
conn.close()
r = random.random()
if r < 0.001 or r >= 0.999:
return rare
elif r < 0.3 and uncommon_anim is not None:
return ' '.join((adj, uncommon_anim[0]))
elif r >= 0.7 and uncommon_adj is not None:
return ' '.join((uncommon_adj[0], anim))
return ' '.join((adj, anim))
if __name__ == '__main__':
for _ in range(20):
print(generate_name())
<commit_msg>Print one name at a time.<commit_after> | from __future__ import unicode_literals, print_function
import sqlite3, os, random
_select_random = 'select {0} from {1} limit 1 offset abs(random()) % (select count({0}) from {1});'
_select_uncommon = 'select value from uncommons where key=?;'
def generate_name():
conn = sqlite3.connect(os.path.join(os.path.dirname(__file__), 'names.db'))
cursor = conn.cursor()
adj = cursor.execute(_select_random.format('adjective', 'adjectives')).fetchone()[0]
anim = cursor.execute(_select_random.format('animal', 'animals')).fetchone()[0]
rare = cursor.execute(_select_random.format('name', 'rares')).fetchone()[0]
uncommon_anim = cursor.execute(_select_uncommon, [adj]).fetchone()
uncommon_adj = cursor.execute(_select_uncommon, [anim]).fetchone()
conn.close()
r = random.random()
if r < 0.001 or r >= 0.999:
return rare
elif r < 0.3 and uncommon_anim is not None:
return ' '.join((adj, uncommon_anim[0]))
elif r >= 0.7 and uncommon_adj is not None:
return ' '.join((uncommon_adj[0], anim))
return ' '.join((adj, anim))
if __name__ == '__main__':
print(generate_name())
| from __future__ import unicode_literals, print_function
import sqlite3, os, random
_select_random = 'select {0} from {1} limit 1 offset abs(random()) % (select count({0}) from {1});'
_select_uncommon = 'select value from uncommons where key=?;'
def generate_name():
conn = sqlite3.connect(os.path.join(os.path.dirname(__file__), 'names.db'))
cursor = conn.cursor()
adj = cursor.execute(_select_random.format('adjective', 'adjectives')).fetchone()[0]
anim = cursor.execute(_select_random.format('animal', 'animals')).fetchone()[0]
rare = cursor.execute(_select_random.format('name', 'rares')).fetchone()[0]
uncommon_anim = cursor.execute(_select_uncommon, [adj]).fetchone()
uncommon_adj = cursor.execute(_select_uncommon, [anim]).fetchone()
conn.close()
r = random.random()
if r < 0.001 or r >= 0.999:
return rare
elif r < 0.3 and uncommon_anim is not None:
return ' '.join((adj, uncommon_anim[0]))
elif r >= 0.7 and uncommon_adj is not None:
return ' '.join((uncommon_adj[0], anim))
return ' '.join((adj, anim))
if __name__ == '__main__':
for _ in range(20):
print(generate_name())
Print one name at a time.from __future__ import unicode_literals, print_function
import sqlite3, os, random
_select_random = 'select {0} from {1} limit 1 offset abs(random()) % (select count({0}) from {1});'
_select_uncommon = 'select value from uncommons where key=?;'
def generate_name():
conn = sqlite3.connect(os.path.join(os.path.dirname(__file__), 'names.db'))
cursor = conn.cursor()
adj = cursor.execute(_select_random.format('adjective', 'adjectives')).fetchone()[0]
anim = cursor.execute(_select_random.format('animal', 'animals')).fetchone()[0]
rare = cursor.execute(_select_random.format('name', 'rares')).fetchone()[0]
uncommon_anim = cursor.execute(_select_uncommon, [adj]).fetchone()
uncommon_adj = cursor.execute(_select_uncommon, [anim]).fetchone()
conn.close()
r = random.random()
if r < 0.001 or r >= 0.999:
return rare
elif r < 0.3 and uncommon_anim is not None:
return ' '.join((adj, uncommon_anim[0]))
elif r >= 0.7 and uncommon_adj is not None:
return ' '.join((uncommon_adj[0], anim))
return ' '.join((adj, anim))
if __name__ == '__main__':
print(generate_name())
| <commit_before>from __future__ import unicode_literals, print_function
import sqlite3, os, random
_select_random = 'select {0} from {1} limit 1 offset abs(random()) % (select count({0}) from {1});'
_select_uncommon = 'select value from uncommons where key=?;'
def generate_name():
conn = sqlite3.connect(os.path.join(os.path.dirname(__file__), 'names.db'))
cursor = conn.cursor()
adj = cursor.execute(_select_random.format('adjective', 'adjectives')).fetchone()[0]
anim = cursor.execute(_select_random.format('animal', 'animals')).fetchone()[0]
rare = cursor.execute(_select_random.format('name', 'rares')).fetchone()[0]
uncommon_anim = cursor.execute(_select_uncommon, [adj]).fetchone()
uncommon_adj = cursor.execute(_select_uncommon, [anim]).fetchone()
conn.close()
r = random.random()
if r < 0.001 or r >= 0.999:
return rare
elif r < 0.3 and uncommon_anim is not None:
return ' '.join((adj, uncommon_anim[0]))
elif r >= 0.7 and uncommon_adj is not None:
return ' '.join((uncommon_adj[0], anim))
return ' '.join((adj, anim))
if __name__ == '__main__':
for _ in range(20):
print(generate_name())
<commit_msg>Print one name at a time.<commit_after>from __future__ import unicode_literals, print_function
import sqlite3, os, random
_select_random = 'select {0} from {1} limit 1 offset abs(random()) % (select count({0}) from {1});'
_select_uncommon = 'select value from uncommons where key=?;'
def generate_name():
conn = sqlite3.connect(os.path.join(os.path.dirname(__file__), 'names.db'))
cursor = conn.cursor()
adj = cursor.execute(_select_random.format('adjective', 'adjectives')).fetchone()[0]
anim = cursor.execute(_select_random.format('animal', 'animals')).fetchone()[0]
rare = cursor.execute(_select_random.format('name', 'rares')).fetchone()[0]
uncommon_anim = cursor.execute(_select_uncommon, [adj]).fetchone()
uncommon_adj = cursor.execute(_select_uncommon, [anim]).fetchone()
conn.close()
r = random.random()
if r < 0.001 or r >= 0.999:
return rare
elif r < 0.3 and uncommon_anim is not None:
return ' '.join((adj, uncommon_anim[0]))
elif r >= 0.7 and uncommon_adj is not None:
return ' '.join((uncommon_adj[0], anim))
return ' '.join((adj, anim))
if __name__ == '__main__':
print(generate_name())
|
ba16b14203af704f1fa0a6eb3111d0537e0cc399 | mail_inline_css/models/mail_template.py | mail_inline_css/models/mail_template.py | # Copyright 2017 David BEAL @ Akretion
# Copyright 2019 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html).
from odoo import models
try:
from premailer import Premailer
except (ImportError, IOError) as err: # pragma: no cover
import logging
_logger = logging.getLogger(__name__)
_logger.debug(err)
class MailTemplate(models.Model):
_inherit = "mail.template"
def generate_email(self, res_ids, fields=None):
"""Use `premailer` to convert styles to inline styles."""
result = super().generate_email(res_ids, fields=fields)
if isinstance(res_ids, int):
result["body_html"] = self._premailer_apply_transform(result["body_html"])
else:
for __, data in result.items():
data["body_html"] = self._premailer_apply_transform(data["body_html"])
return result
def _premailer_apply_transform(self, data_html):
premailer = Premailer(html=data_html, **self._get_premailer_options())
return premailer.transform()
def _get_premailer_options(self):
return {}
| # Copyright 2017 David BEAL @ Akretion
# Copyright 2019 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html).
from odoo import models
try:
from premailer import Premailer
except (ImportError, IOError) as err: # pragma: no cover
import logging
_logger = logging.getLogger(__name__)
_logger.debug(err)
class MailTemplate(models.Model):
_inherit = "mail.template"
def generate_email(self, res_ids, fields=None):
"""Use `premailer` to convert styles to inline styles."""
result = super().generate_email(res_ids, fields=fields)
if isinstance(res_ids, int):
result["body_html"] = self._premailer_apply_transform(result["body_html"])
else:
for __, data in result.items():
data["body_html"] = self._premailer_apply_transform(data["body_html"])
return result
def _premailer_apply_transform(self, data_html):
if not data_html:
return data_html
premailer = Premailer(html=data_html, **self._get_premailer_options())
return premailer.transform()
def _get_premailer_options(self):
return {}
| Fix issue on empty template with premailer | Fix issue on empty template with premailer
If premailer receives an empty value, such as an empty string,
on parsing, it returns None and fails when trying to call
'etree.fromstring()' on this None result.
We should avoid to call premailer on an empty string, as the result
will anyway not change.
We may have an empty template for instance when a template could not
compile due to a mistake.
| Python | agpl-3.0 | OCA/social,OCA/social,OCA/social | # Copyright 2017 David BEAL @ Akretion
# Copyright 2019 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html).
from odoo import models
try:
from premailer import Premailer
except (ImportError, IOError) as err: # pragma: no cover
import logging
_logger = logging.getLogger(__name__)
_logger.debug(err)
class MailTemplate(models.Model):
_inherit = "mail.template"
def generate_email(self, res_ids, fields=None):
"""Use `premailer` to convert styles to inline styles."""
result = super().generate_email(res_ids, fields=fields)
if isinstance(res_ids, int):
result["body_html"] = self._premailer_apply_transform(result["body_html"])
else:
for __, data in result.items():
data["body_html"] = self._premailer_apply_transform(data["body_html"])
return result
def _premailer_apply_transform(self, data_html):
premailer = Premailer(html=data_html, **self._get_premailer_options())
return premailer.transform()
def _get_premailer_options(self):
return {}
Fix issue on empty template with premailer
If premailer receives an empty value, such as an empty string,
on parsing, it returns None and fails when trying to call
'etree.fromstring()' on this None result.
We should avoid to call premailer on an empty string, as the result
will anyway not change.
We may have an empty template for instance when a template could not
compile due to a mistake. | # Copyright 2017 David BEAL @ Akretion
# Copyright 2019 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html).
from odoo import models
try:
from premailer import Premailer
except (ImportError, IOError) as err: # pragma: no cover
import logging
_logger = logging.getLogger(__name__)
_logger.debug(err)
class MailTemplate(models.Model):
_inherit = "mail.template"
def generate_email(self, res_ids, fields=None):
"""Use `premailer` to convert styles to inline styles."""
result = super().generate_email(res_ids, fields=fields)
if isinstance(res_ids, int):
result["body_html"] = self._premailer_apply_transform(result["body_html"])
else:
for __, data in result.items():
data["body_html"] = self._premailer_apply_transform(data["body_html"])
return result
def _premailer_apply_transform(self, data_html):
if not data_html:
return data_html
premailer = Premailer(html=data_html, **self._get_premailer_options())
return premailer.transform()
def _get_premailer_options(self):
return {}
| <commit_before># Copyright 2017 David BEAL @ Akretion
# Copyright 2019 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html).
from odoo import models
try:
from premailer import Premailer
except (ImportError, IOError) as err: # pragma: no cover
import logging
_logger = logging.getLogger(__name__)
_logger.debug(err)
class MailTemplate(models.Model):
_inherit = "mail.template"
def generate_email(self, res_ids, fields=None):
"""Use `premailer` to convert styles to inline styles."""
result = super().generate_email(res_ids, fields=fields)
if isinstance(res_ids, int):
result["body_html"] = self._premailer_apply_transform(result["body_html"])
else:
for __, data in result.items():
data["body_html"] = self._premailer_apply_transform(data["body_html"])
return result
def _premailer_apply_transform(self, data_html):
premailer = Premailer(html=data_html, **self._get_premailer_options())
return premailer.transform()
def _get_premailer_options(self):
return {}
<commit_msg>Fix issue on empty template with premailer
If premailer receives an empty value, such as an empty string,
on parsing, it returns None and fails when trying to call
'etree.fromstring()' on this None result.
We should avoid to call premailer on an empty string, as the result
will anyway not change.
We may have an empty template for instance when a template could not
compile due to a mistake.<commit_after> | # Copyright 2017 David BEAL @ Akretion
# Copyright 2019 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html).
from odoo import models
try:
from premailer import Premailer
except (ImportError, IOError) as err: # pragma: no cover
import logging
_logger = logging.getLogger(__name__)
_logger.debug(err)
class MailTemplate(models.Model):
_inherit = "mail.template"
def generate_email(self, res_ids, fields=None):
"""Use `premailer` to convert styles to inline styles."""
result = super().generate_email(res_ids, fields=fields)
if isinstance(res_ids, int):
result["body_html"] = self._premailer_apply_transform(result["body_html"])
else:
for __, data in result.items():
data["body_html"] = self._premailer_apply_transform(data["body_html"])
return result
def _premailer_apply_transform(self, data_html):
if not data_html:
return data_html
premailer = Premailer(html=data_html, **self._get_premailer_options())
return premailer.transform()
def _get_premailer_options(self):
return {}
| # Copyright 2017 David BEAL @ Akretion
# Copyright 2019 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html).
from odoo import models
try:
from premailer import Premailer
except (ImportError, IOError) as err: # pragma: no cover
import logging
_logger = logging.getLogger(__name__)
_logger.debug(err)
class MailTemplate(models.Model):
_inherit = "mail.template"
def generate_email(self, res_ids, fields=None):
"""Use `premailer` to convert styles to inline styles."""
result = super().generate_email(res_ids, fields=fields)
if isinstance(res_ids, int):
result["body_html"] = self._premailer_apply_transform(result["body_html"])
else:
for __, data in result.items():
data["body_html"] = self._premailer_apply_transform(data["body_html"])
return result
def _premailer_apply_transform(self, data_html):
premailer = Premailer(html=data_html, **self._get_premailer_options())
return premailer.transform()
def _get_premailer_options(self):
return {}
Fix issue on empty template with premailer
If premailer receives an empty value, such as an empty string,
on parsing, it returns None and fails when trying to call
'etree.fromstring()' on this None result.
We should avoid to call premailer on an empty string, as the result
will anyway not change.
We may have an empty template for instance when a template could not
compile due to a mistake.# Copyright 2017 David BEAL @ Akretion
# Copyright 2019 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html).
from odoo import models
try:
from premailer import Premailer
except (ImportError, IOError) as err: # pragma: no cover
import logging
_logger = logging.getLogger(__name__)
_logger.debug(err)
class MailTemplate(models.Model):
_inherit = "mail.template"
def generate_email(self, res_ids, fields=None):
"""Use `premailer` to convert styles to inline styles."""
result = super().generate_email(res_ids, fields=fields)
if isinstance(res_ids, int):
result["body_html"] = self._premailer_apply_transform(result["body_html"])
else:
for __, data in result.items():
data["body_html"] = self._premailer_apply_transform(data["body_html"])
return result
def _premailer_apply_transform(self, data_html):
if not data_html:
return data_html
premailer = Premailer(html=data_html, **self._get_premailer_options())
return premailer.transform()
def _get_premailer_options(self):
return {}
| <commit_before># Copyright 2017 David BEAL @ Akretion
# Copyright 2019 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html).
from odoo import models
try:
from premailer import Premailer
except (ImportError, IOError) as err: # pragma: no cover
import logging
_logger = logging.getLogger(__name__)
_logger.debug(err)
class MailTemplate(models.Model):
_inherit = "mail.template"
def generate_email(self, res_ids, fields=None):
"""Use `premailer` to convert styles to inline styles."""
result = super().generate_email(res_ids, fields=fields)
if isinstance(res_ids, int):
result["body_html"] = self._premailer_apply_transform(result["body_html"])
else:
for __, data in result.items():
data["body_html"] = self._premailer_apply_transform(data["body_html"])
return result
def _premailer_apply_transform(self, data_html):
premailer = Premailer(html=data_html, **self._get_premailer_options())
return premailer.transform()
def _get_premailer_options(self):
return {}
<commit_msg>Fix issue on empty template with premailer
If premailer receives an empty value, such as an empty string,
on parsing, it returns None and fails when trying to call
'etree.fromstring()' on this None result.
We should avoid to call premailer on an empty string, as the result
will anyway not change.
We may have an empty template for instance when a template could not
compile due to a mistake.<commit_after># Copyright 2017 David BEAL @ Akretion
# Copyright 2019 Camptocamp SA
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html).
from odoo import models
try:
from premailer import Premailer
except (ImportError, IOError) as err: # pragma: no cover
import logging
_logger = logging.getLogger(__name__)
_logger.debug(err)
class MailTemplate(models.Model):
_inherit = "mail.template"
def generate_email(self, res_ids, fields=None):
"""Use `premailer` to convert styles to inline styles."""
result = super().generate_email(res_ids, fields=fields)
if isinstance(res_ids, int):
result["body_html"] = self._premailer_apply_transform(result["body_html"])
else:
for __, data in result.items():
data["body_html"] = self._premailer_apply_transform(data["body_html"])
return result
def _premailer_apply_transform(self, data_html):
if not data_html:
return data_html
premailer = Premailer(html=data_html, **self._get_premailer_options())
return premailer.transform()
def _get_premailer_options(self):
return {}
|
0298e8b6abcd7cea99df4cb235c73a49e340521a | tests/query_test/test_decimal_queries.py | tests/query_test/test_decimal_queries.py | #!/usr/bin/env python
# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Targeted tests for decimal type.
#
import logging
import pytest
from copy import copy
from tests.common.test_vector import *
from tests.common.impala_test_suite import *
class TestDecimalQueries(ImpalaTestSuite):
BATCH_SIZES = [0, 1]
@classmethod
def get_workload(cls):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestDecimalQueries, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(
TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
# On CDH4, hive does not support decimal so we can't run these tests against
# the other file formats. Enable them on C5.
cls.TestMatrix.add_constraint(lambda v:\
(v.get_value('table_format').file_format == 'text' and
v.get_value('table_format').compression_codec == 'none') or
v.get_value('table_format').file_format == 'parquet')
def test_queries(self, vector):
new_vector = copy(vector)
new_vector.get_value('exec_option')['batch_size'] = vector.get_value('batch_size')
self.run_test_case('QueryTest/decimal', new_vector)
| #!/usr/bin/env python
# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Targeted tests for decimal type.
#
import logging
import pytest
from copy import copy
from tests.common.test_vector import *
from tests.common.impala_test_suite import *
class TestDecimalQueries(ImpalaTestSuite):
BATCH_SIZES = [0, 1]
@classmethod
def get_workload(cls):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestDecimalQueries, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(
TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
# On CDH4, hive does not support decimal so we can't run these tests against
# the other file formats. Enable them on C5.
cls.TestMatrix.add_constraint(lambda v:\
(v.get_value('table_format').file_format == 'text' and
v.get_value('table_format').compression_codec == 'none') or
v.get_value('table_format').file_format == 'parquet')
def test_queries(self, vector):
if os.environ.get('ASAN_OPTIONS') == 'handle_segv=0':
pytest.xfail(reason="IMPALA-959: Sum on a decimal column fails ASAN")
new_vector = copy(vector)
new_vector.get_value('exec_option')['batch_size'] = vector.get_value('batch_size')
self.run_test_case('QueryTest/decimal', new_vector)
| Fix the ASAN build by xfailing test_decimal when ASAN_OPTIONS is set. | Fix the ASAN build by xfailing test_decimal when ASAN_OPTIONS is set.
Adding decimal columns crashes an ASAN built impalad. This change skips the test.
Change-Id: Ic94055a3f0d00f89354177de18bc27d2f4cecec2
Reviewed-on: http://gerrit.ent.cloudera.com:8080/2532
Reviewed-by: Ishaan Joshi <d1d1e60202ec9f2503deb1b724986485a125d802@cloudera.com>
Tested-by: jenkins
Reviewed-on: http://gerrit.ent.cloudera.com:8080/2594
| Python | apache-2.0 | cloudera/Impala,michaelhkw/incubator-impala,michaelhkw/incubator-impala,michaelhkw/incubator-impala,michaelhkw/incubator-impala,michaelhkw/incubator-impala,cloudera/Impala,cloudera/Impala,cloudera/Impala,michaelhkw/incubator-impala,cloudera/Impala,cloudera/Impala,michaelhkw/incubator-impala,cloudera/Impala | #!/usr/bin/env python
# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Targeted tests for decimal type.
#
import logging
import pytest
from copy import copy
from tests.common.test_vector import *
from tests.common.impala_test_suite import *
class TestDecimalQueries(ImpalaTestSuite):
BATCH_SIZES = [0, 1]
@classmethod
def get_workload(cls):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestDecimalQueries, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(
TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
# On CDH4, hive does not support decimal so we can't run these tests against
# the other file formats. Enable them on C5.
cls.TestMatrix.add_constraint(lambda v:\
(v.get_value('table_format').file_format == 'text' and
v.get_value('table_format').compression_codec == 'none') or
v.get_value('table_format').file_format == 'parquet')
def test_queries(self, vector):
new_vector = copy(vector)
new_vector.get_value('exec_option')['batch_size'] = vector.get_value('batch_size')
self.run_test_case('QueryTest/decimal', new_vector)
Fix the ASAN build by xfailing test_decimal when ASAN_OPTIONS is set.
Adding decimal columns crashes an ASAN built impalad. This change skips the test.
Change-Id: Ic94055a3f0d00f89354177de18bc27d2f4cecec2
Reviewed-on: http://gerrit.ent.cloudera.com:8080/2532
Reviewed-by: Ishaan Joshi <d1d1e60202ec9f2503deb1b724986485a125d802@cloudera.com>
Tested-by: jenkins
Reviewed-on: http://gerrit.ent.cloudera.com:8080/2594 | #!/usr/bin/env python
# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Targeted tests for decimal type.
#
import logging
import pytest
from copy import copy
from tests.common.test_vector import *
from tests.common.impala_test_suite import *
class TestDecimalQueries(ImpalaTestSuite):
BATCH_SIZES = [0, 1]
@classmethod
def get_workload(cls):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestDecimalQueries, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(
TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
# On CDH4, hive does not support decimal so we can't run these tests against
# the other file formats. Enable them on C5.
cls.TestMatrix.add_constraint(lambda v:\
(v.get_value('table_format').file_format == 'text' and
v.get_value('table_format').compression_codec == 'none') or
v.get_value('table_format').file_format == 'parquet')
def test_queries(self, vector):
if os.environ.get('ASAN_OPTIONS') == 'handle_segv=0':
pytest.xfail(reason="IMPALA-959: Sum on a decimal column fails ASAN")
new_vector = copy(vector)
new_vector.get_value('exec_option')['batch_size'] = vector.get_value('batch_size')
self.run_test_case('QueryTest/decimal', new_vector)
| <commit_before>#!/usr/bin/env python
# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Targeted tests for decimal type.
#
import logging
import pytest
from copy import copy
from tests.common.test_vector import *
from tests.common.impala_test_suite import *
class TestDecimalQueries(ImpalaTestSuite):
BATCH_SIZES = [0, 1]
@classmethod
def get_workload(cls):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestDecimalQueries, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(
TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
# On CDH4, hive does not support decimal so we can't run these tests against
# the other file formats. Enable them on C5.
cls.TestMatrix.add_constraint(lambda v:\
(v.get_value('table_format').file_format == 'text' and
v.get_value('table_format').compression_codec == 'none') or
v.get_value('table_format').file_format == 'parquet')
def test_queries(self, vector):
new_vector = copy(vector)
new_vector.get_value('exec_option')['batch_size'] = vector.get_value('batch_size')
self.run_test_case('QueryTest/decimal', new_vector)
<commit_msg>Fix the ASAN build by xfailing test_decimal when ASAN_OPTIONS is set.
Adding decimal columns crashes an ASAN built impalad. This change skips the test.
Change-Id: Ic94055a3f0d00f89354177de18bc27d2f4cecec2
Reviewed-on: http://gerrit.ent.cloudera.com:8080/2532
Reviewed-by: Ishaan Joshi <d1d1e60202ec9f2503deb1b724986485a125d802@cloudera.com>
Tested-by: jenkins
Reviewed-on: http://gerrit.ent.cloudera.com:8080/2594<commit_after> | #!/usr/bin/env python
# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Targeted tests for decimal type.
#
import logging
import pytest
from copy import copy
from tests.common.test_vector import *
from tests.common.impala_test_suite import *
class TestDecimalQueries(ImpalaTestSuite):
BATCH_SIZES = [0, 1]
@classmethod
def get_workload(cls):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestDecimalQueries, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(
TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
# On CDH4, hive does not support decimal so we can't run these tests against
# the other file formats. Enable them on C5.
cls.TestMatrix.add_constraint(lambda v:\
(v.get_value('table_format').file_format == 'text' and
v.get_value('table_format').compression_codec == 'none') or
v.get_value('table_format').file_format == 'parquet')
def test_queries(self, vector):
if os.environ.get('ASAN_OPTIONS') == 'handle_segv=0':
pytest.xfail(reason="IMPALA-959: Sum on a decimal column fails ASAN")
new_vector = copy(vector)
new_vector.get_value('exec_option')['batch_size'] = vector.get_value('batch_size')
self.run_test_case('QueryTest/decimal', new_vector)
| #!/usr/bin/env python
# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Targeted tests for decimal type.
#
import logging
import pytest
from copy import copy
from tests.common.test_vector import *
from tests.common.impala_test_suite import *
class TestDecimalQueries(ImpalaTestSuite):
BATCH_SIZES = [0, 1]
@classmethod
def get_workload(cls):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestDecimalQueries, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(
TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
# On CDH4, hive does not support decimal so we can't run these tests against
# the other file formats. Enable them on C5.
cls.TestMatrix.add_constraint(lambda v:\
(v.get_value('table_format').file_format == 'text' and
v.get_value('table_format').compression_codec == 'none') or
v.get_value('table_format').file_format == 'parquet')
def test_queries(self, vector):
new_vector = copy(vector)
new_vector.get_value('exec_option')['batch_size'] = vector.get_value('batch_size')
self.run_test_case('QueryTest/decimal', new_vector)
Fix the ASAN build by xfailing test_decimal when ASAN_OPTIONS is set.
Adding decimal columns crashes an ASAN built impalad. This change skips the test.
Change-Id: Ic94055a3f0d00f89354177de18bc27d2f4cecec2
Reviewed-on: http://gerrit.ent.cloudera.com:8080/2532
Reviewed-by: Ishaan Joshi <d1d1e60202ec9f2503deb1b724986485a125d802@cloudera.com>
Tested-by: jenkins
Reviewed-on: http://gerrit.ent.cloudera.com:8080/2594#!/usr/bin/env python
# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Targeted tests for decimal type.
#
import logging
import pytest
from copy import copy
from tests.common.test_vector import *
from tests.common.impala_test_suite import *
class TestDecimalQueries(ImpalaTestSuite):
BATCH_SIZES = [0, 1]
@classmethod
def get_workload(cls):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestDecimalQueries, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(
TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
# On CDH4, hive does not support decimal so we can't run these tests against
# the other file formats. Enable them on C5.
cls.TestMatrix.add_constraint(lambda v:\
(v.get_value('table_format').file_format == 'text' and
v.get_value('table_format').compression_codec == 'none') or
v.get_value('table_format').file_format == 'parquet')
def test_queries(self, vector):
if os.environ.get('ASAN_OPTIONS') == 'handle_segv=0':
pytest.xfail(reason="IMPALA-959: Sum on a decimal column fails ASAN")
new_vector = copy(vector)
new_vector.get_value('exec_option')['batch_size'] = vector.get_value('batch_size')
self.run_test_case('QueryTest/decimal', new_vector)
| <commit_before>#!/usr/bin/env python
# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Targeted tests for decimal type.
#
import logging
import pytest
from copy import copy
from tests.common.test_vector import *
from tests.common.impala_test_suite import *
class TestDecimalQueries(ImpalaTestSuite):
BATCH_SIZES = [0, 1]
@classmethod
def get_workload(cls):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestDecimalQueries, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(
TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
# On CDH4, hive does not support decimal so we can't run these tests against
# the other file formats. Enable them on C5.
cls.TestMatrix.add_constraint(lambda v:\
(v.get_value('table_format').file_format == 'text' and
v.get_value('table_format').compression_codec == 'none') or
v.get_value('table_format').file_format == 'parquet')
def test_queries(self, vector):
new_vector = copy(vector)
new_vector.get_value('exec_option')['batch_size'] = vector.get_value('batch_size')
self.run_test_case('QueryTest/decimal', new_vector)
<commit_msg>Fix the ASAN build by xfailing test_decimal when ASAN_OPTIONS is set.
Adding decimal columns crashes an ASAN built impalad. This change skips the test.
Change-Id: Ic94055a3f0d00f89354177de18bc27d2f4cecec2
Reviewed-on: http://gerrit.ent.cloudera.com:8080/2532
Reviewed-by: Ishaan Joshi <d1d1e60202ec9f2503deb1b724986485a125d802@cloudera.com>
Tested-by: jenkins
Reviewed-on: http://gerrit.ent.cloudera.com:8080/2594<commit_after>#!/usr/bin/env python
# Copyright (c) 2012 Cloudera, Inc. All rights reserved.
# Targeted tests for decimal type.
#
import logging
import pytest
from copy import copy
from tests.common.test_vector import *
from tests.common.impala_test_suite import *
class TestDecimalQueries(ImpalaTestSuite):
BATCH_SIZES = [0, 1]
@classmethod
def get_workload(cls):
return 'functional-query'
@classmethod
def add_test_dimensions(cls):
super(TestDecimalQueries, cls).add_test_dimensions()
cls.TestMatrix.add_dimension(
TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
# On CDH4, hive does not support decimal so we can't run these tests against
# the other file formats. Enable them on C5.
cls.TestMatrix.add_constraint(lambda v:\
(v.get_value('table_format').file_format == 'text' and
v.get_value('table_format').compression_codec == 'none') or
v.get_value('table_format').file_format == 'parquet')
def test_queries(self, vector):
if os.environ.get('ASAN_OPTIONS') == 'handle_segv=0':
pytest.xfail(reason="IMPALA-959: Sum on a decimal column fails ASAN")
new_vector = copy(vector)
new_vector.get_value('exec_option')['batch_size'] = vector.get_value('batch_size')
self.run_test_case('QueryTest/decimal', new_vector)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.