commit
stringlengths 40
40
| old_file
stringlengths 4
150
| new_file
stringlengths 4
150
| old_contents
stringlengths 0
3.26k
| new_contents
stringlengths 1
4.43k
| subject
stringlengths 15
501
| message
stringlengths 15
4.06k
| lang
stringclasses 4
values | license
stringclasses 13
values | repos
stringlengths 5
91.5k
| diff
stringlengths 0
4.35k
|
|---|---|---|---|---|---|---|---|---|---|---|
9be9ddd64d192240d45dea5b0c3dfbe3a2d3e261
|
gae2django/utils.py
|
gae2django/utils.py
|
#
# Copyright 2008 Andi Albrecht <albrecht.andi@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
class CallableString(str):
def __call__(self):
return self
def id(self):
try:
return int(self.split('_')[-1])
except:
return None
|
#
# Copyright 2008 Andi Albrecht <albrecht.andi@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
class CallableString(str):
def __call__(self):
return str(self)
def id(self):
try:
return int(self.split('_')[-1])
except:
return None
|
Return str(self) instead of self.
|
Return str(self) instead of self.
|
Python
|
apache-2.0
|
bubenkoff/bubenkoff-gae2django,andialbrecht/django-gae2django,andialbrecht/django-gae2django,bubenkoff/bubenkoff-gae2django
|
---
+++
@@ -17,7 +17,7 @@
class CallableString(str):
def __call__(self):
- return self
+ return str(self)
def id(self):
try:
|
3328a58e7c81fb86af2424b851de05e7b409ec00
|
asyncio/__init__.py
|
asyncio/__init__.py
|
"""The asyncio package, tracking PEP 3156."""
import sys
# This relies on each of the submodules having an __all__ variable.
from .futures import *
from .events import *
from .locks import *
from .transports import *
from .protocols import *
from .streams import *
from .tasks import *
if sys.platform == 'win32': # pragma: no cover
from .windows_events import *
else:
from .unix_events import * # pragma: no cover
__all__ = (futures.__all__ +
events.__all__ +
locks.__all__ +
transports.__all__ +
protocols.__all__ +
streams.__all__ +
tasks.__all__)
|
"""The asyncio package, tracking PEP 3156."""
import sys
# The selectors module is in the stdlib in Python 3.4 but not in 3.3.
# Do this first, so the other submodules can use "from . import selectors".
try:
import selectors # Will also be exported.
except ImportError:
from . import selectors
# This relies on each of the submodules having an __all__ variable.
from .futures import *
from .events import *
from .locks import *
from .transports import *
from .protocols import *
from .streams import *
from .tasks import *
if sys.platform == 'win32': # pragma: no cover
from .windows_events import *
else:
from .unix_events import * # pragma: no cover
__all__ = (futures.__all__ +
events.__all__ +
locks.__all__ +
transports.__all__ +
protocols.__all__ +
streams.__all__ +
tasks.__all__)
|
Add fakery so "from asyncio import selectors" always works.
|
Add fakery so "from asyncio import selectors" always works.
|
Python
|
apache-2.0
|
jashandeep-sohi/asyncio,gsb-eng/asyncio,ajdavis/asyncio,vxgmichel/asyncio,gvanrossum/asyncio,manipopopo/asyncio,gsb-eng/asyncio,fallen/asyncio,1st1/asyncio,Martiusweb/asyncio,vxgmichel/asyncio,ajdavis/asyncio,haypo/trollius,ajdavis/asyncio,fallen/asyncio,manipopopo/asyncio,1st1/asyncio,fallen/asyncio,Martiusweb/asyncio,Martiusweb/asyncio,1st1/asyncio,gvanrossum/asyncio,manipopopo/asyncio,jashandeep-sohi/asyncio,vxgmichel/asyncio,haypo/trollius,gsb-eng/asyncio,gvanrossum/asyncio,jashandeep-sohi/asyncio,haypo/trollius
|
---
+++
@@ -1,6 +1,13 @@
"""The asyncio package, tracking PEP 3156."""
import sys
+
+# The selectors module is in the stdlib in Python 3.4 but not in 3.3.
+# Do this first, so the other submodules can use "from . import selectors".
+try:
+ import selectors # Will also be exported.
+except ImportError:
+ from . import selectors
# This relies on each of the submodules having an __all__ variable.
from .futures import *
|
09e4dd8736d6e829b779dd14b882e0e1d7f5abb9
|
tester/register/prepare_test.py
|
tester/register/prepare_test.py
|
import sys
import os
import argparse
def write_csv(filename, nb_users):
with open(filename, "w") as csv_file:
csv_file.write("SEQUENTIAL\n")
for x in xrange(nb_users):
line = "{uname};localhost;[authentication username={uname} password={uname}];\n".format(uname=str(1000+x))
csv_file.write(line)
def write_sql(filename, nb_users):
with open(filename, "w") as sql_file:
header = """DROP DATABASE tests;
CREATE DATABASE tests;
USE tests;
CREATE TABLE accounts (user VARCHAR(20),password VARCHAR(20));"""
sql_file.write(header)
for x in xrange(nb_users):
line = """INSERT into accounts (user, password) VALUES ("{uname}", "{uname}");\n""".format(uname=str(1000+x))
sql_file.write(line)
def main(argv=None):
if argv == None:
argv = sys.argv
argparser = argparse.ArgumentParser(description="Prepare load tests for Flexisip.")
argparser.add_argument('-N', '--users', help="How many different users should be registering to flexisip", dest="users", default=5000)
args, additional_args = argparser.parse_known_args()
write_csv("users.csv", args.users)
write_sql("users.sql", args.users)
if __name__ == '__main__':
main()
|
import sys
import os
import argparse
def write_csv(filename, nb_users):
with open(filename, "w") as csv_file:
csv_file.write("SEQUENTIAL\n")
for x in xrange(nb_users):
line = "{uname};localhost;[authentication username={uname} password={uname}];\n".format(uname=str(1000+x))
csv_file.write(line)
def write_sql(filename, nb_users):
with open(filename, "w") as sql_file:
header = """DROP DATABASE IF EXISTS tests;
CREATE DATABASE tests;
USE tests;
CREATE TABLE accounts (user VARCHAR(20),password VARCHAR(20));"""
sql_file.write(header)
for x in xrange(nb_users):
line = """INSERT into accounts (user, password) VALUES ("{uname}", "{uname}");\n""".format(uname=str(1000+x))
sql_file.write(line)
def main(argv=None):
if argv == None:
argv = sys.argv
argparser = argparse.ArgumentParser(description="Prepare load tests for Flexisip.")
argparser.add_argument('-N', '--users', help="How many different users should be registering to flexisip", dest="users", default=5000)
args, additional_args = argparser.parse_known_args()
write_csv("users.csv", args.users)
write_sql("users.sql", args.users)
if __name__ == '__main__':
main()
|
DROP DATABASE IF EXISTS in tests.
|
DROP DATABASE IF EXISTS in tests.
|
Python
|
agpl-3.0
|
BelledonneCommunications/flexisip,BelledonneCommunications/flexisip,BelledonneCommunications/flexisip,BelledonneCommunications/flexisip
|
---
+++
@@ -13,7 +13,7 @@
def write_sql(filename, nb_users):
with open(filename, "w") as sql_file:
- header = """DROP DATABASE tests;
+ header = """DROP DATABASE IF EXISTS tests;
CREATE DATABASE tests;
USE tests;
CREATE TABLE accounts (user VARCHAR(20),password VARCHAR(20));"""
|
8be856ed565d9e961a4d24da74a13240e25f4ded
|
cio/plugins/base.py
|
cio/plugins/base.py
|
class BasePlugin(object):
ext = None
def load(self, content):
"""
Return plugin data for content string
"""
return content
def save(self, data):
"""
Persist external plugin resources and return content string for plugin data
"""
return data
def delete(self, data):
"""
Delete external plugin resources
"""
pass
def render(self, data):
"""
Render plugin
"""
return data
|
from cio.conf import settings
class BasePlugin(object):
ext = None
@property
def settings(self):
return settings.get(self.ext.upper(), {})
def load(self, content):
"""
Return plugin data for content string
"""
return content
def save(self, data):
"""
Persist external plugin resources and return content string for plugin data
"""
return data
def delete(self, data):
"""
Delete external plugin resources
"""
pass
def render(self, data):
"""
Render plugin
"""
return data
|
Add support for plugin settings
|
Add support for plugin settings
|
Python
|
bsd-3-clause
|
5monkeys/content-io
|
---
+++
@@ -1,6 +1,13 @@
+from cio.conf import settings
+
+
class BasePlugin(object):
ext = None
+
+ @property
+ def settings(self):
+ return settings.get(self.ext.upper(), {})
def load(self, content):
"""
|
e4696a04cbc003737d7ba28d58b14775e9fc2682
|
tests/transport/test_asyncio.py
|
tests/transport/test_asyncio.py
|
from unittest import TestCase
class AsyncioTransportTestCase(TestCase):
pass
|
from asyncio import get_event_loop
from unittest import TestCase, mock
from rfxcom.transport import AsyncioTransport
from rfxcom.protocol import RESET_PACKET, STATUS_PACKET
class AsyncioTransportTestCase(TestCase):
def test_loop_once(self):
loop = get_event_loop()
def handler(*args, **kwargs):
pass
device = mock.MagicMock()
AsyncioTransport(device, loop, callback=handler)
loop._run_once()
device.write.assert_has_call(bytearray(RESET_PACKET))
device.write.assert_has_call(bytearray(STATUS_PACKET))
|
Add an initial ghetto asyncio test.
|
Add an initial ghetto asyncio test.
|
Python
|
bsd-3-clause
|
skimpax/python-rfxcom,kalfa/python-rfxcom,AndyA13/python-rfxcom,d0ugal-archive/python-rfxcom,kalfa/python-rfxcom,d0ugal-archive/python-rfxcom,AndyA13/python-rfxcom,skimpax/python-rfxcom
|
---
+++
@@ -1,5 +1,24 @@
-from unittest import TestCase
+from asyncio import get_event_loop
+from unittest import TestCase, mock
+
+from rfxcom.transport import AsyncioTransport
+
+from rfxcom.protocol import RESET_PACKET, STATUS_PACKET
class AsyncioTransportTestCase(TestCase):
- pass
+
+ def test_loop_once(self):
+
+ loop = get_event_loop()
+
+ def handler(*args, **kwargs):
+ pass
+
+ device = mock.MagicMock()
+
+ AsyncioTransport(device, loop, callback=handler)
+ loop._run_once()
+
+ device.write.assert_has_call(bytearray(RESET_PACKET))
+ device.write.assert_has_call(bytearray(STATUS_PACKET))
|
f97e1d88f0508521fb2841ef9e8c98ec77424daa
|
SettingsTemplate.py
|
SettingsTemplate.py
|
HOST = "irc.twitch.tv"
PORT = 6667
PASS = "oauth:##############################" # Your bot's oauth (https://twitchapps.com/tmi/)
IDENTITY = "my_bot" # Your bot's username. Lowercase!!
WHITELIST = ["some_authourized_account", "another one"] # People who may execute commands. Lower!!
CHANNEL = "channel" # The channel to join. Lowercase!!
JOIN_MESSAGE = "Hi, I'm a bot that just joined this channel." # Message from the bot when it joins a channel.
WOT_KEY = "" # Api key of WOT to check sites (mywot.com)
CHECK_LINKS = True # Should the bot check links via WOT?
|
HOST = "irc.twitch.tv"
PORT = 6667
PASS = "oauth:##############################" # Your bot's oauth (https://twitchapps.com/tmi/)
IDENTITY = "my_bot" # Your bot's username. Lowercase!!
WHITELIST = ["some_authourized_account", "another one"] # People who may execute commands. Lower!!
CHANNEL = "channel" # The channel to join. Lowercase!!
JOIN_MESSAGE = "Hi, I'm a bot that just joined this channel." # Message from the bot when it joins a channel.
WOT_KEY = "" # Api key of WOT to check sites (mywot.com)
CHECK_LINKS = False # Should the bot check links via WOT?
|
Set CHECK_LINKS to False for default
|
Set CHECK_LINKS to False for default
|
Python
|
apache-2.0
|
K00sKlust/K00sTwitchBot
|
---
+++
@@ -9,4 +9,4 @@
JOIN_MESSAGE = "Hi, I'm a bot that just joined this channel." # Message from the bot when it joins a channel.
WOT_KEY = "" # Api key of WOT to check sites (mywot.com)
-CHECK_LINKS = True # Should the bot check links via WOT?
+CHECK_LINKS = False # Should the bot check links via WOT?
|
793ad28cf8bae098d12d08d971f19fd9cc29f3dd
|
colorlog/logging.py
|
colorlog/logging.py
|
"""Wrappers around the logging module"""
from __future__ import absolute_import
import functools
import logging
from colorlog.colorlog import ColoredFormatter
BASIC_FORMAT = "%(log_color)s%(levelname)s%(reset)s:%(name)s:%(message)s"
def basicConfig(**kwargs):
"""This calls basicConfig() and then overrides the formatter it creates"""
logging.basicConfig(**kwargs)
logging._acquireLock()
try:
stream = logging.root.handlers[0]
stream.setFormatter(
ColoredFormatter(kwargs.get('format', BASIC_FORMAT)))
finally:
logging._releaseLock()
def ensure_configured(func):
"""This ensures basicConfig is called if no handlers exist"""
@functools.wraps(func)
def wrapper(*args, **kwargs):
if len(logging.root.handlers) == 0:
basicConfig()
return func(*args, **kwargs)
return wrapper
root = logging.root
getLogger = logging.getLogger
debug = ensure_configured(logging.debug)
info = ensure_configured(logging.info)
warning = ensure_configured(logging.warning)
error = ensure_configured(logging.error)
critical = ensure_configured(logging.critical)
log = ensure_configured(logging.log)
exception = ensure_configured(logging.exception)
|
"""Wrappers around the logging module"""
from __future__ import absolute_import
import functools
import logging
from colorlog.colorlog import ColoredFormatter
BASIC_FORMAT = "%(log_color)s%(levelname)s%(reset)s:%(name)s:%(message)s"
def basicConfig(**kwargs):
"""This calls basicConfig() and then overrides the formatter it creates"""
logging.basicConfig(**kwargs)
logging._acquireLock()
try:
stream = logging.root.handlers[0]
stream.setFormatter(
ColoredFormatter(kwargs.get('format', BASIC_FORMAT),
datefmt=kwargs.get('datefmt')))
finally:
logging._releaseLock()
def ensure_configured(func):
"""This ensures basicConfig is called if no handlers exist"""
@functools.wraps(func)
def wrapper(*args, **kwargs):
if len(logging.root.handlers) == 0:
basicConfig()
return func(*args, **kwargs)
return wrapper
root = logging.root
getLogger = logging.getLogger
debug = ensure_configured(logging.debug)
info = ensure_configured(logging.info)
warning = ensure_configured(logging.warning)
error = ensure_configured(logging.error)
critical = ensure_configured(logging.critical)
log = ensure_configured(logging.log)
exception = ensure_configured(logging.exception)
|
Add datefmt to arguments that will be passed within decorated basicConfig
|
Add datefmt to arguments that will be passed within decorated basicConfig
|
Python
|
mit
|
borntyping/python-colorlog
|
---
+++
@@ -17,7 +17,8 @@
try:
stream = logging.root.handlers[0]
stream.setFormatter(
- ColoredFormatter(kwargs.get('format', BASIC_FORMAT)))
+ ColoredFormatter(kwargs.get('format', BASIC_FORMAT),
+ datefmt=kwargs.get('datefmt')))
finally:
logging._releaseLock()
|
b2d06e068fbc7bad9ed0c8f22e751b6bb46d353d
|
dependencies/contrib/_django.py
|
dependencies/contrib/_django.py
|
from __future__ import absolute_import
from collections import OrderedDict
from dependencies import this
from django.views.generic import View
def view(injector):
"""Create Django class-based view from injector class."""
handler = create_handler(View)
apply_http_methods(handler, injector)
handler.http_method_names = list(handler.http_method_names.keys())
return injector.let(as_view=handler.as_view)
def create_handler(from_class):
class Handler(from_class):
http_method_names = OrderedDict.fromkeys(["head", "options"])
return Handler
def apply_http_methods(handler, injector):
for method in ["get", "post", "put", "patch", "delete", "head", "options", "trace"]:
if method in injector:
def __view(self, request, *args, **kwargs):
return injector.let(
view=self,
request=request,
args=args,
kwargs=kwargs,
user=this.request.user,
).trace()
handler.http_method_names[method] = None
setattr(handler, method, __view)
|
from __future__ import absolute_import
from collections import OrderedDict
from dependencies import this
from django.views.generic import View
def view(injector):
"""Create Django class-based view from injector class."""
handler = create_handler(View)
apply_http_methods(handler, injector)
finalize_http_methods(handler)
return injector.let(as_view=handler.as_view)
def create_handler(from_class):
class Handler(from_class):
http_method_names = OrderedDict.fromkeys(["head", "options"])
return Handler
def apply_http_methods(handler, injector):
for method in ["get", "post", "put", "patch", "delete", "head", "options", "trace"]:
if method in injector:
def __view(self, request, *args, **kwargs):
ns = injector.let(
view=self,
request=request,
args=args,
kwargs=kwargs,
user=this.request.user,
)
return getattr(ns, __view.method)()
__view.method = method
handler.http_method_names[method] = None
setattr(handler, method, __view)
def finalize_http_methods(handler):
handler.http_method_names = list(handler.http_method_names.keys())
|
Fix django view closure issue.
|
Fix django view closure issue.
|
Python
|
bsd-2-clause
|
proofit404/dependencies,proofit404/dependencies,proofit404/dependencies,proofit404/dependencies
|
---
+++
@@ -11,8 +11,7 @@
handler = create_handler(View)
apply_http_methods(handler, injector)
- handler.http_method_names = list(handler.http_method_names.keys())
-
+ finalize_http_methods(handler)
return injector.let(as_view=handler.as_view)
@@ -27,17 +26,23 @@
def apply_http_methods(handler, injector):
for method in ["get", "post", "put", "patch", "delete", "head", "options", "trace"]:
-
if method in injector:
def __view(self, request, *args, **kwargs):
- return injector.let(
+ ns = injector.let(
view=self,
request=request,
args=args,
kwargs=kwargs,
user=this.request.user,
- ).trace()
+ )
+ return getattr(ns, __view.method)()
+ __view.method = method
handler.http_method_names[method] = None
setattr(handler, method, __view)
+
+
+def finalize_http_methods(handler):
+
+ handler.http_method_names = list(handler.http_method_names.keys())
|
5d5cb362410896927b6216deeb9421adfc3331c4
|
hatarake/net.py
|
hatarake/net.py
|
'''
Wrappers around Python requests
This allows us to handle all the custom headers in a single place
'''
from __future__ import absolute_import
import requests
from functools import wraps
from hatarake import USER_AGENT
def add_args(func):
@wraps(func)
def wrapper(*args, **kwargs):
if 'headers' not in kwargs:
kwargs['headers'] = {}
kwargs['headers']['user-agent'] = USER_AGENT
return func(*args, **kwargs)
return wrapper
get = add_args(requests.get)
post = add_args(requests.post)
put = add_args(requests.put)
|
'''
Wrappers around Python requests
This allows us to handle all the custom headers in a single place
'''
from __future__ import absolute_import
import requests
from functools import wraps
from hatarake import USER_AGENT
def add_args(func):
@wraps(func)
def wrapper(*args, **kwargs):
try:
kwargs['headers']['user-agent'] = USER_AGENT
except KeyError:
kwargs['headers'] = {'user-agent': USER_AGENT}
if 'token' in kwargs:
token = kwargs.pop('token')
kwargs['headers']['Authorization'] = 'Token %s' % token
return func(*args, **kwargs)
return wrapper
get = add_args(requests.get)
post = add_args(requests.post)
put = add_args(requests.put)
|
Support token as an argument to our requests wrapper
|
Support token as an argument to our requests wrapper
|
Python
|
mit
|
kfdm/hatarake
|
---
+++
@@ -14,9 +14,15 @@
def add_args(func):
@wraps(func)
def wrapper(*args, **kwargs):
- if 'headers' not in kwargs:
- kwargs['headers'] = {}
- kwargs['headers']['user-agent'] = USER_AGENT
+ try:
+ kwargs['headers']['user-agent'] = USER_AGENT
+ except KeyError:
+ kwargs['headers'] = {'user-agent': USER_AGENT}
+
+ if 'token' in kwargs:
+ token = kwargs.pop('token')
+ kwargs['headers']['Authorization'] = 'Token %s' % token
+
return func(*args, **kwargs)
return wrapper
|
6d6d1af248ce555cca56521bba5e7c356817c74e
|
account/forms.py
|
account/forms.py
|
from django.contrib.auth.models import User
from django import forms
from account.models import UserProfile
attributes = {"class": "required"}
class RegistrationForm(forms.Form):
username = forms.RegexField(regex=r'^[\w.@+-]+$',
max_length=30,
widget=forms.TextInput(attrs=attributes),
label="Username",
error_message={'invalid': "This value may contain only letters, numbers and @.+- characters."}
)
email = forms.EmailField()
def clean_username(self):
username = self.cleaned_data["username"]
existing = User.objects.filter(username__iexact=username)
if existing.exists():
raise forms.ValidationError("A user with that username already exists.")
else:
return self.cleaned_data["username"]
class SettingsForm(forms.Form):
email = forms.EmailField()
xsede_username = forms.CharField(max_length=50,
required=False,
label="XSEDE Username")
new_ssh_keypair = forms.BooleanField(required=False)
def clean(self):
if "password1" in self.cleaned_data and "password2" in self.cleaned_data:
if self.cleaned_data["password1"] != self.cleaned_data["password2"]:
raise forms.ValidationError("The two password fields did not match.")
return self.cleaned_data
class UserProfileForm(forms.ModelForm):
private_key = forms.CharField(widget=forms.Textarea)
public_key = forms.CharField(widget=forms.Textarea)
class Meta:
model = UserProfile
fields = ("xsede_username", "public_key", "activation_key", "password_reset_key", "reset_expires")
|
from django.contrib.auth.models import User
from django import forms
from account.models import UserProfile
attributes = {"class": "required"}
class RegistrationForm(forms.Form):
username = forms.RegexField(regex=r'^[\w.@+-]+$',
max_length=30,
widget=forms.TextInput(attrs=attributes),
label="Username",
error_message={'invalid': "This value may contain only letters, numbers and @.+- characters."}
)
email = forms.EmailField()
def clean_username(self):
username = self.cleaned_data["username"]
existing = User.objects.filter(username__iexact=username)
if existing.exists():
raise forms.ValidationError("A user with that username already exists.")
else:
return self.cleaned_data["username"]
class SettingsForm(forms.Form):
email = forms.EmailField()
xsede_username = forms.CharField(max_length=50,
required=False,
label="XSEDE Username")
new_ssh_keypair = forms.BooleanField(required=False)
class UserProfileForm(forms.ModelForm):
private_key = forms.CharField(widget=forms.Textarea)
public_key = forms.CharField(widget=forms.Textarea)
class Meta:
model = UserProfile
fields = ("xsede_username", "public_key", "activation_key", "password_reset_key", "reset_expires")
|
Remove unused section of SettingsForm
|
Remove unused section of SettingsForm
|
Python
|
mit
|
crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp
|
---
+++
@@ -32,12 +32,6 @@
label="XSEDE Username")
new_ssh_keypair = forms.BooleanField(required=False)
- def clean(self):
- if "password1" in self.cleaned_data and "password2" in self.cleaned_data:
- if self.cleaned_data["password1"] != self.cleaned_data["password2"]:
- raise forms.ValidationError("The two password fields did not match.")
- return self.cleaned_data
-
class UserProfileForm(forms.ModelForm):
private_key = forms.CharField(widget=forms.Textarea)
|
8c174388aefa3907aeb8733bb3d4c77c770eefe7
|
DataModelAdapter.py
|
DataModelAdapter.py
|
class DataModelAdapter(object) :
def __init__(self, data) :
self._data = data
self._children = set()
pass
def numChildren(self) :
return len(self._children)
def hasData(self) :
return self._data is not None
def getData(self, key) :
return self._data[key]
def addChild(self, child) :
self._children.add(child)
|
class DataModelAdapter(object) :
def __init__(self, data) :
self._data = data
self._children = set()
self._parent = None
pass
def numChildren(self) :
return len(self._children)
def hasData(self) :
return self._data is not None
def getData(self, key) :
return self._data[key]
def addChild(self, child) :
self._children.add(child)
def setParent(self, parent) :
self._parent = parent
def parent(self) :
return self._parent
|
Add _parent field with setter/getter
|
Add _parent field with setter/getter
|
Python
|
apache-2.0
|
mattdeckard/wherewithal
|
---
+++
@@ -4,6 +4,7 @@
def __init__(self, data) :
self._data = data
self._children = set()
+ self._parent = None
pass
def numChildren(self) :
@@ -17,3 +18,9 @@
def addChild(self, child) :
self._children.add(child)
+
+ def setParent(self, parent) :
+ self._parent = parent
+
+ def parent(self) :
+ return self._parent
|
e27f04e9c8d5d74afdd9cd7d6990cad5ff6f6cb5
|
api/v330/docking_event/serializers.py
|
api/v330/docking_event/serializers.py
|
from api.v330.common.serializers import *
class SpacecraftFlightSerializerForDockingEvent(serializers.HyperlinkedModelSerializer):
spacecraft = SpacecraftSerializer(read_only=True, many=False)
class Meta:
model = SpacecraftFlight
fields = ('id', 'url', 'destination', 'splashdown', 'spacecraft')
class DockingEventSerializer(serializers.HyperlinkedModelSerializer):
flight_vehicle = SpacecraftFlightSerializerForDockingEvent(read_only=True)
docking_location = serializers.StringRelatedField(many=False, read_only=True)
class Meta:
model = DockingEvent
fields = ('id', 'url', 'docking', 'departure', 'flight_vehicle', 'docking_location')
class DockingEventDetailedSerializer(serializers.HyperlinkedModelSerializer):
flight_vehicle = SpacecraftFlightSerializerForDockingEvent(read_only=True, many=False)
docking_location = serializers.StringRelatedField(many=False, read_only=True)
class Meta:
model = DockingEvent
fields = ('id', 'url', 'docking', 'departure', 'flight_vehicle', 'docking_location')
|
from api.v330.common.serializers import *
class SpacecraftFlightSerializerForDockingEvent(serializers.HyperlinkedModelSerializer):
spacecraft = SpacecraftSerializer(read_only=True, many=False)
class Meta:
model = SpacecraftFlight
fields = ('id', 'url', 'destination', 'splashdown', 'spacecraft')
class SpaceStationSerializerForDockingEvent(serializers.HyperlinkedModelSerializer):
class Meta:
model = SpaceStation
fields = ('id', 'url', 'name', 'image_url')
class DockingEventSerializer(serializers.HyperlinkedModelSerializer):
flight_vehicle = SpacecraftFlightSerializerForDockingEvent(read_only=True)
docking_location = serializers.StringRelatedField(many=False, read_only=True)
class Meta:
model = DockingEvent
fields = ('id', 'url', 'docking', 'departure', 'flight_vehicle', 'docking_location')
class DockingEventDetailedSerializer(serializers.HyperlinkedModelSerializer):
flight_vehicle = SpacecraftFlightSerializerForDockingEvent(read_only=True, many=False)
docking_location = serializers.StringRelatedField(many=False, read_only=True)
space_station = SpaceStationSerializerForDockingEvent(many=False, read_only=True)
class Meta:
model = DockingEvent
fields = ('id', 'url', 'docking', 'departure', 'flight_vehicle', 'docking_location', 'space_station')
|
Add space_station field to detailed docking event
|
Add space_station field to detailed docking event
|
Python
|
apache-2.0
|
ItsCalebJones/SpaceLaunchNow-Server,ItsCalebJones/SpaceLaunchNow-Server,ItsCalebJones/SpaceLaunchNow-Server
|
---
+++
@@ -7,6 +7,12 @@
class Meta:
model = SpacecraftFlight
fields = ('id', 'url', 'destination', 'splashdown', 'spacecraft')
+
+
+class SpaceStationSerializerForDockingEvent(serializers.HyperlinkedModelSerializer):
+ class Meta:
+ model = SpaceStation
+ fields = ('id', 'url', 'name', 'image_url')
class DockingEventSerializer(serializers.HyperlinkedModelSerializer):
@@ -21,7 +27,8 @@
class DockingEventDetailedSerializer(serializers.HyperlinkedModelSerializer):
flight_vehicle = SpacecraftFlightSerializerForDockingEvent(read_only=True, many=False)
docking_location = serializers.StringRelatedField(many=False, read_only=True)
+ space_station = SpaceStationSerializerForDockingEvent(many=False, read_only=True)
class Meta:
model = DockingEvent
- fields = ('id', 'url', 'docking', 'departure', 'flight_vehicle', 'docking_location')
+ fields = ('id', 'url', 'docking', 'departure', 'flight_vehicle', 'docking_location', 'space_station')
|
e0becdd677c06c29834ecea73c28635553e18337
|
app/main/presenters/search_results.py
|
app/main/presenters/search_results.py
|
from flask import Markup
class SearchResults(object):
"""Provides access to the search results information"""
def __init__(self, response, lots_by_slug):
self.search_results = response['services']
self._lots = lots_by_slug
self._annotate()
self.total = response['meta']['total']
if 'page' in response['meta']['query']:
self.page = response['meta']['query']['page']
def _annotate(self):
for service in self.search_results:
self._replace_lot(service)
self._add_highlighting(service)
def _replace_lot(self, service):
# replace lot slug with reference to dict containing all the relevant lot data
service['lot'] = self._lots.get(service['lot'])
def _add_highlighting(self, service):
if 'highlight' in service:
if 'serviceSummary' in service['highlight']:
service['serviceSummary'] = Markup(
''.join(service['highlight']['serviceSummary'])
)
class AggregationResults(object):
"""Provides access to the aggregation results information"""
def __init__(self, response):
self.results = response['aggregations']
self.total = response['meta']['total']
if 'page' in response['meta']['query']:
self.page = response['meta']['query']['page']
|
from flask import Markup
class SearchResults(object):
"""Provides access to the search results information"""
def __init__(self, response, lots_by_slug):
self.search_results = response['services']
self._lots = lots_by_slug
self._annotate()
self.total = response['meta']['total']
if 'page' in response['meta']['query']:
self.page = response['meta']['query']['page']
def _annotate(self):
for service in self.search_results:
self._replace_lot(service)
self._add_highlighting(service)
def _replace_lot(self, service):
# replace lot slug with reference to dict containing all the relevant lot data
service['lot'] = self._lots.get(service['lot'])
def _add_highlighting(self, service):
if 'highlight' in service:
for highlighted_field in ['serviceSummary', 'serviceDescription']:
if highlighted_field in service['highlight']:
service[highlighted_field] = Markup(
''.join(service['highlight'][highlighted_field])
)
class AggregationResults(object):
"""Provides access to the aggregation results information"""
def __init__(self, response):
self.results = response['aggregations']
self.total = response['meta']['total']
if 'page' in response['meta']['query']:
self.page = response['meta']['query']['page']
|
Add static highlighting on serviceDescription field
|
Add static highlighting on serviceDescription field
|
Python
|
mit
|
alphagov/digitalmarketplace-buyer-frontend,alphagov/digitalmarketplace-buyer-frontend,alphagov/digitalmarketplace-buyer-frontend,alphagov/digitalmarketplace-buyer-frontend
|
---
+++
@@ -23,10 +23,11 @@
def _add_highlighting(self, service):
if 'highlight' in service:
- if 'serviceSummary' in service['highlight']:
- service['serviceSummary'] = Markup(
- ''.join(service['highlight']['serviceSummary'])
- )
+ for highlighted_field in ['serviceSummary', 'serviceDescription']:
+ if highlighted_field in service['highlight']:
+ service[highlighted_field] = Markup(
+ ''.join(service['highlight'][highlighted_field])
+ )
class AggregationResults(object):
|
df6339ad776aa989362089f54f3a1f675a86bfb0
|
adhocracy/lib/tiles/badge_tiles.py
|
adhocracy/lib/tiles/badge_tiles.py
|
def badge(badge):
from adhocracy.lib.templating import render_def
return render_def('/badge/tiles.html', 'badge', badge=badge,
cached=True)
def badges(badges):
from adhocracy.lib.templating import render_def
return render_def('/badge/tiles.html', 'badges', badges=badges,
cached=True)
def badge_styles():
'''
Render a <style>-block with dyamic badge styles
'''
from adhocracy.lib.templating import render_def
from adhocracy.model import Badge
badges = Badge.all()
return render_def('/badge/tiles.html', 'badge_styles', badges=badges,
cached=True)
|
def badge(badge):
from adhocracy.lib.templating import render_def
return render_def('/badge/tiles.html', 'badge', badge=badge,
cached=True)
def badges(badges):
from adhocracy.lib.templating import render_def
return render_def('/badge/tiles.html', 'badges', badges=badges,
cached=True)
def badge_styles():
'''
Render a <style>-block with dyamic badge styles
'''
from adhocracy.lib.templating import render_def
from adhocracy.model import Badge
badges = Badge.all_q().all()
return render_def('/badge/tiles.html', 'badge_styles', badges=badges,
cached=True)
|
Include style for all badges
|
badges: Include style for all badges
Not only global badges.
|
Python
|
agpl-3.0
|
phihag/adhocracy,DanielNeugebauer/adhocracy,liqd/adhocracy,DanielNeugebauer/adhocracy,alkadis/vcv,liqd/adhocracy,alkadis/vcv,DanielNeugebauer/adhocracy,alkadis/vcv,alkadis/vcv,alkadis/vcv,liqd/adhocracy,SysTheron/adhocracy,DanielNeugebauer/adhocracy,phihag/adhocracy,SysTheron/adhocracy,phihag/adhocracy,SysTheron/adhocracy,liqd/adhocracy,phihag/adhocracy,DanielNeugebauer/adhocracy,phihag/adhocracy
|
---
+++
@@ -16,6 +16,6 @@
'''
from adhocracy.lib.templating import render_def
from adhocracy.model import Badge
- badges = Badge.all()
+ badges = Badge.all_q().all()
return render_def('/badge/tiles.html', 'badge_styles', badges=badges,
cached=True)
|
dbba6f10c867e64031ae07adb3d21becfe4a4e5a
|
law/contrib/cms/__init__.py
|
law/contrib/cms/__init__.py
|
# coding: utf-8
"""
CMS-related contrib package. https://home.cern/about/experiments/cms
"""
__all__ = ["CMSJobDashboard", "BundleCMSSW"]
# provisioning imports
from law.contrib.cms.job import CMSJobDashboard
from law.contrib.cms.tasks import BundleCMSSW
|
# coding: utf-8
"""
CMS-related contrib package. https://home.cern/about/experiments/cms
"""
__all__ = ["CMSJobDashboard", "BundleCMSSW", "Site", "lfn_to_pfn"]
# provisioning imports
from law.contrib.cms.job import CMSJobDashboard
from law.contrib.cms.tasks import BundleCMSSW
from law.contrib.cms.util import Site, lfn_to_pfn
|
Load utils in cms contrib package.
|
Load utils in cms contrib package.
|
Python
|
bsd-3-clause
|
riga/law,riga/law
|
---
+++
@@ -5,9 +5,10 @@
"""
-__all__ = ["CMSJobDashboard", "BundleCMSSW"]
+__all__ = ["CMSJobDashboard", "BundleCMSSW", "Site", "lfn_to_pfn"]
# provisioning imports
from law.contrib.cms.job import CMSJobDashboard
from law.contrib.cms.tasks import BundleCMSSW
+from law.contrib.cms.util import Site, lfn_to_pfn
|
372a1cab73dad91daab5640f472eda4552be0adb
|
chatterbot/__init__.py
|
chatterbot/__init__.py
|
"""
ChatterBot is a machine learning, conversational dialog engine.
"""
from .chatterbot import ChatBot
__version__ = '1.0.0a1'
__author__ = 'Gunther Cox'
__email__ = 'gunthercx@gmail.com'
__url__ = 'https://github.com/gunthercox/ChatterBot'
__all__ = (
'ChatBot',
)
|
"""
ChatterBot is a machine learning, conversational dialog engine.
"""
from .chatterbot import ChatBot
__version__ = '1.0.0a2'
__author__ = 'Gunther Cox'
__email__ = 'gunthercx@gmail.com'
__url__ = 'https://github.com/gunthercox/ChatterBot'
__all__ = (
'ChatBot',
)
|
Set release version to 1.0.0a2
|
Set release version to 1.0.0a2
|
Python
|
bsd-3-clause
|
gunthercox/ChatterBot,vkosuri/ChatterBot
|
---
+++
@@ -3,7 +3,7 @@
"""
from .chatterbot import ChatBot
-__version__ = '1.0.0a1'
+__version__ = '1.0.0a2'
__author__ = 'Gunther Cox'
__email__ = 'gunthercx@gmail.com'
__url__ = 'https://github.com/gunthercox/ChatterBot'
|
3c1357627bf1921fdee114b60f96f42c328120b4
|
caramel/__init__.py
|
caramel/__init__.py
|
#! /usr/bin/env python
# vim: expandtab shiftwidth=4 softtabstop=4 tabstop=17 filetype=python :
from pyramid.config import Configurator
from sqlalchemy import engine_from_config
from .models import (
init_session,
)
def main(global_config, **settings):
"""This function returns a Pyramid WSGI application."""
engine = engine_from_config(settings, "sqlalchemy.")
init_session(engine)
config = Configurator(settings=settings)
config.add_route("ca", "/root.crt", request_method="GET")
config.add_route("cabundle", "/bundle.crt", request_method="GET")
config.add_route("csr", "/{sha256:[0-9a-f]{64}}", request_method="POST")
config.add_route("cert", "/{sha256:[0-9a-f]{64}}", request_method="GET")
config.scan()
return config.make_wsgi_app()
|
#! /usr/bin/env python
# vim: expandtab shiftwidth=4 softtabstop=4 tabstop=17 filetype=python :
from pyramid.config import Configurator
from sqlalchemy import engine_from_config
from .models import (
init_session,
)
def main(global_config, **settings):
"""This function returns a Pyramid WSGI application."""
engine = engine_from_config(settings, "sqlalchemy.")
init_session(engine)
config = Configurator(settings=settings)
config.include("pyramid_tm")
config.add_route("ca", "/root.crt", request_method="GET")
config.add_route("cabundle", "/bundle.crt", request_method="GET")
config.add_route("csr", "/{sha256:[0-9a-f]{64}}", request_method="POST")
config.add_route("cert", "/{sha256:[0-9a-f]{64}}", request_method="GET")
config.scan()
return config.make_wsgi_app()
|
Move pyramid_tm include to caramel.main
|
Caramel: Move pyramid_tm include to caramel.main
Move the setting to include pyramid_tm to caramel.main from ini files.
This is a vital setting that should never be changed by the user.
|
Python
|
agpl-3.0
|
ModioAB/caramel,ModioAB/caramel
|
---
+++
@@ -13,6 +13,7 @@
engine = engine_from_config(settings, "sqlalchemy.")
init_session(engine)
config = Configurator(settings=settings)
+ config.include("pyramid_tm")
config.add_route("ca", "/root.crt", request_method="GET")
config.add_route("cabundle", "/bundle.crt", request_method="GET")
config.add_route("csr", "/{sha256:[0-9a-f]{64}}", request_method="POST")
|
409ad4af8a4ad933667d91709822a04dbbda77ac
|
km3pipe/cmd.py
|
km3pipe/cmd.py
|
# coding=utf-8
# Filename: cmd.py
"""
KM3Pipe command line utility.
Usage:
km3pipe test
km3pipe tohdf5 [-s] -i FILE -o FILE
km3pipe (-h | --help)
km3pipe --version
Options:
-h --help Show this screen.
-i FILE Input file.
-o FILE Output file.
-s Write each event in a separate dataset.
"""
from __future__ import division, absolute_import, print_function
from km3pipe import version
def tohdf5(input_file, output_file, separate_events=False):
"""Convert ROOT file to HDF5 file"""
from km3pipe import Pipeline # noqa
from km3pipe.pumps import AanetPump, HDF5Sink, HDF5SinkLegacy # noqa
sink = HDF5Sink if separate_events else HDF5SinkLegacy
pipe = Pipeline()
pipe.attach(AanetPump, filename=input_file)
pipe.attach(sink,
filename=output_file,
separate_events=separate_events)
pipe.drain()
def main():
from docopt import docopt
arguments = docopt(__doc__, version=version)
if arguments['tohdf5']:
tohdf5(arguments['-i'], arguments['-o'], arguments['-s'])
|
# coding=utf-8
# Filename: cmd.py
"""
KM3Pipe command line utility.
Usage:
km3pipe test
km3pipe tohdf5 [-s] -i FILE -o FILE
km3pipe (-h | --help)
km3pipe --version
Options:
-h --help Show this screen.
-i FILE Input file.
-o FILE Output file.
-s Write each event in a separate dataset.
"""
from __future__ import division, absolute_import, print_function
from km3pipe import version
def tohdf5(input_file, output_file, use_tables=True):
"""Convert ROOT file to HDF5 file"""
from km3pipe import Pipeline # noqa
from km3pipe.pumps import AanetPump, HDF5Sink, HDF5TableSink # noqa
sink = HDF5TableSink if use_tables else HDF5Sink
pipe = Pipeline()
pipe.attach(AanetPump, filename=input_file)
pipe.attach(sink,
filename=output_file,
separate_events=separate_events)
pipe.drain()
def main():
from docopt import docopt
arguments = docopt(__doc__, version=version)
if arguments['tohdf5']:
tohdf5(arguments['-i'], arguments['-o'], arguments['-s'])
|
Set HDF5TableSink as default sink
|
Set HDF5TableSink as default sink
|
Python
|
mit
|
tamasgal/km3pipe,tamasgal/km3pipe
|
---
+++
@@ -21,12 +21,12 @@
from km3pipe import version
-def tohdf5(input_file, output_file, separate_events=False):
+def tohdf5(input_file, output_file, use_tables=True):
"""Convert ROOT file to HDF5 file"""
from km3pipe import Pipeline # noqa
- from km3pipe.pumps import AanetPump, HDF5Sink, HDF5SinkLegacy # noqa
+ from km3pipe.pumps import AanetPump, HDF5Sink, HDF5TableSink # noqa
- sink = HDF5Sink if separate_events else HDF5SinkLegacy
+ sink = HDF5TableSink if use_tables else HDF5Sink
pipe = Pipeline()
pipe.attach(AanetPump, filename=input_file)
|
16f531cb7e9d067725a4c25a4321773aada9616d
|
api/v2/views/tag.py
|
api/v2/views/tag.py
|
from core.models import Tag
from api.permissions import CloudAdminRequired
from api.v2.serializers.summaries import TagSummarySerializer
from api.v2.views.base import AuthReadOnlyViewSet
class TagViewSet(AuthReadOnlyViewSet):
"""
API endpoint that allows tags to be viewed or edited.
"""
queryset = Tag.objects.all()
serializer_class = TagSummarySerializer
max_paginate_by = 1000
def perform_create(self, serializer):
serializer.save(user=self.request.user)
def get_permissions(self):
if self.request.method in ["POST", "PUT", "DELETE"]:
self.permission_classes = (CloudAdminRequired,)
return super(TagViewSet, self).get_permissions()
|
from threepio import logger
from core.models import Tag
from api.permissions import ApiAuthRequired, CloudAdminRequired,\
InMaintenance
from api.v2.serializers.summaries import TagSummarySerializer
from api.v2.views.base import AuthOptionalViewSet
class TagViewSet(AuthOptionalViewSet):
"""
API endpoint that allows tags to be viewed or edited.
"""
queryset = Tag.objects.all()
serializer_class = TagSummarySerializer
max_paginate_by = 1000
def perform_create(self, serializer):
same_name_tags = Tag.objects.filter(
name__iexact=serializer.validated_data.get("name"))
if same_name_tags:
raise serializers.ValidationError(
"A tag with this name already exists: %s" %
same_name_tags.first().name)
serializer.save(user=self.request.user)
def get_permissions(self):
if self.request.method is "":
self.permission_classes = (ApiAuthRequired,
InMaintenance,)
if self.request.method in ["PUT", "DELETE"]:
self.permission_classes = (CloudAdminRequired,
InMaintenance,)
return super(TagViewSet, self).get_permissions()
|
Address @jchansen's requests. No dupes. POST for authorized users, PUT DELETE for cloud admins and staff.
|
Address @jchansen's requests. No dupes. POST for authorized users, PUT DELETE for cloud admins and staff.
modified: api/v2/views/tag.py
|
Python
|
apache-2.0
|
CCI-MOC/GUI-Backend,CCI-MOC/GUI-Backend,CCI-MOC/GUI-Backend,CCI-MOC/GUI-Backend
|
---
+++
@@ -1,11 +1,13 @@
+from threepio import logger
from core.models import Tag
-from api.permissions import CloudAdminRequired
+from api.permissions import ApiAuthRequired, CloudAdminRequired,\
+ InMaintenance
from api.v2.serializers.summaries import TagSummarySerializer
-from api.v2.views.base import AuthReadOnlyViewSet
+from api.v2.views.base import AuthOptionalViewSet
-class TagViewSet(AuthReadOnlyViewSet):
+class TagViewSet(AuthOptionalViewSet):
"""
API endpoint that allows tags to be viewed or edited.
"""
@@ -15,9 +17,19 @@
max_paginate_by = 1000
def perform_create(self, serializer):
+ same_name_tags = Tag.objects.filter(
+ name__iexact=serializer.validated_data.get("name"))
+ if same_name_tags:
+ raise serializers.ValidationError(
+ "A tag with this name already exists: %s" %
+ same_name_tags.first().name)
serializer.save(user=self.request.user)
def get_permissions(self):
- if self.request.method in ["POST", "PUT", "DELETE"]:
- self.permission_classes = (CloudAdminRequired,)
+ if self.request.method is "":
+ self.permission_classes = (ApiAuthRequired,
+ InMaintenance,)
+ if self.request.method in ["PUT", "DELETE"]:
+ self.permission_classes = (CloudAdminRequired,
+ InMaintenance,)
return super(TagViewSet, self).get_permissions()
|
00a497b21b9c788cb38da6c92a985e1b5c22801a
|
apps/survey/urls.py
|
apps/survey/urls.py
|
from django.conf.urls.defaults import *
from . import views
urlpatterns = patterns('',
url(r'^profile/$', views.profile_index, name='survey_profile'),
url(r'^profile/electric/$', views.profile_electric, name='survey_profile_electric'),
#url(r'^profile/intake/$', views.survey_intake, name='survey_profile_intake'),
url(r'^profile/surveys/$', views.survey_management, name='survey_management'),
url(r'^main/$', views.main_index),
url(r'^survey_management/$', views.survey_management, name='survey_management'),
#url(r'^survey_data/(?P<survey_shortname>.+)/(?P<id>\d+)/$', views.survey_data, name='survey_data'),
url(r'^intake/$', views.survey_data, name='survey_data'),
url(r'^monthly/(?P<id>\d+)/$', views.survey_data_monthly ,name='survey_data_monthly'),
url(r'^thanks_profile/$', views.thanks_profile, name='profile_thanks'),
#url(r'^select/$', views.select_user, name='survey_select_user'),
url(r'^$', views.index, name='survey_index'),
)
|
from django.conf.urls.defaults import *
from . import views
urlpatterns = patterns('',
url(r'^profile/$', views.profile_index, name='survey_profile'),
url(r'^profile/electric/$', views.profile_electric, name='survey_profile_electric'),
url(r'^profile/surveys/$', views.survey_management, name='survey_management'),
url(r'^main/$', views.main_index),
url(r'^survey_management/$', views.survey_management, name='survey_management'),
url(r'^intake/view/$', views.survey_intake_view, name='survey_intake_view'),
url(r'^intake/update/$', views.survey_intake_update, name='survey_intake_update'),
url(r'^monthly/(?P<id>\d+)/$', views.survey_monthly ,name='survey_monthly'),
url(r'^monthly/(?P<id>\d+)/update/$', views.survey_monthly_update ,name='survey_monthly_update'),
url(r'^thanks_profile/$', views.thanks_profile, name='profile_thanks'),
#url(r'^select/$', views.select_user, name='survey_select_user'),
url(r'^$', views.index, name='survey_index'),
)
|
Add view and update decorators
|
Add view and update decorators
|
Python
|
agpl-3.0
|
chispita/epiwork,chispita/epiwork,chispita/epiwork,chispita/epiwork,chispita/epiwork,chispita/epiwork,chispita/epiwork
|
---
+++
@@ -5,15 +5,15 @@
urlpatterns = patterns('',
url(r'^profile/$', views.profile_index, name='survey_profile'),
url(r'^profile/electric/$', views.profile_electric, name='survey_profile_electric'),
- #url(r'^profile/intake/$', views.survey_intake, name='survey_profile_intake'),
url(r'^profile/surveys/$', views.survey_management, name='survey_management'),
url(r'^main/$', views.main_index),
url(r'^survey_management/$', views.survey_management, name='survey_management'),
- #url(r'^survey_data/(?P<survey_shortname>.+)/(?P<id>\d+)/$', views.survey_data, name='survey_data'),
-
- url(r'^intake/$', views.survey_data, name='survey_data'),
- url(r'^monthly/(?P<id>\d+)/$', views.survey_data_monthly ,name='survey_data_monthly'),
-
+
+ url(r'^intake/view/$', views.survey_intake_view, name='survey_intake_view'),
+ url(r'^intake/update/$', views.survey_intake_update, name='survey_intake_update'),
+
+ url(r'^monthly/(?P<id>\d+)/$', views.survey_monthly ,name='survey_monthly'),
+ url(r'^monthly/(?P<id>\d+)/update/$', views.survey_monthly_update ,name='survey_monthly_update'),
url(r'^thanks_profile/$', views.thanks_profile, name='profile_thanks'),
#url(r'^select/$', views.select_user, name='survey_select_user'),
|
5a27329e32942523e73d2d01b43ba75ecd281622
|
dask/array/__init__.py
|
dask/array/__init__.py
|
from __future__ import absolute_import, division, print_function
from ..utils import ignoring
from .core import (Array, stack, concatenate, tensordot, transpose, from_array,
choose, coarsen, constant, fromfunction, compute, unique, store)
from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2,
ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod,
frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians,
sin, sinh, sqrt, tan, tanh, trunc, isnull, notnull)
from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm,
argmin, argmax,
nansum, nanmean, nanstd, nanvar, nanmin,
nanmax, nanargmin, nanargmax)
from .percentile import percentile
with ignoring(ImportError):
from .reductions import nanprod
from . import random, linalg, ghost
from .wrap import ones, zeros, empty
from ..context import set_options
|
from __future__ import absolute_import, division, print_function
from ..utils import ignoring
from .core import (Array, stack, concatenate, tensordot, transpose, from_array,
choose, where, coarsen, constant, fromfunction, compute, unique, store)
from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2,
ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod,
frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians,
sin, sinh, sqrt, tan, tanh, trunc, isnull, notnull)
from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm,
argmin, argmax,
nansum, nanmean, nanstd, nanvar, nanmin,
nanmax, nanargmin, nanargmax)
from .percentile import percentile
with ignoring(ImportError):
from .reductions import nanprod
from . import random, linalg, ghost
from .wrap import ones, zeros, empty
from ..context import set_options
|
Add "where" to top level dask.array module
|
Add "where" to top level dask.array module
|
Python
|
bsd-3-clause
|
vikhyat/dask,minrk/dask,esc/dask,gameduell/dask,ssanderson/dask,vikhyat/dask,jayhetee/dask,mrocklin/dask,wiso/dask,dask/dask,ContinuumIO/dask,mrocklin/dask,jcrist/dask,jayhetee/dask,blaze/dask,marianotepper/dask,ContinuumIO/dask,jakirkham/dask,pombredanne/dask,chrisbarber/dask,wiso/dask,PhE/dask,freeman-lab/dask,jcrist/dask,cpcloud/dask,pombredanne/dask,PhE/dask,ssanderson/dask,mraspaud/dask,clarkfitzg/dask,marianotepper/dask,hainm/dask,mikegraham/dask,esc/dask,minrk/dask,simudream/dask,dask/dask,jakirkham/dask,cowlicks/dask,simudream/dask,mraspaud/dask,blaze/dask,clarkfitzg/dask,hainm/dask,freeman-lab/dask
|
---
+++
@@ -2,7 +2,7 @@
from ..utils import ignoring
from .core import (Array, stack, concatenate, tensordot, transpose, from_array,
- choose, coarsen, constant, fromfunction, compute, unique, store)
+ choose, where, coarsen, constant, fromfunction, compute, unique, store)
from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2,
ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod,
frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians,
|
eef03e6c4eb6d80dd04ccbbea6b530d5679a8142
|
sydent/http/servlets/pubkeyservlets.py
|
sydent/http/servlets/pubkeyservlets.py
|
# -*- coding: utf-8 -*-
# Copyright 2014 matrix.org
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from twisted.web.resource import Resource
import json
import nacl.encoding
class Ed25519Servlet(Resource):
isLeaf = True
def __init__(self, syd):
self.sydent = syd
def render_GET(self, request):
pubKey = self.sydent.signers.ed25519.signing_key.verify_key
pubKeyHex = pubKey.encode(encoder=nacl.encoding.HexEncoder)
return json.dumps({'public_key':pubKeyHex})
|
# -*- coding: utf-8 -*-
# Copyright 2014 matrix.org
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from twisted.web.resource import Resource
import json
import nacl.encoding
class Ed25519Servlet(Resource):
isLeaf = True
def __init__(self, syd):
self.sydent = syd
def render_GET(self, request):
pubKey = self.sydent.keyring.ed25519.verify_key
pubKeyHex = pubKey.encode(encoder=nacl.encoding.HexEncoder)
return json.dumps({'public_key':pubKeyHex})
|
Update pubkey servlet to s/signers/keyring
|
Update pubkey servlet to s/signers/keyring
|
Python
|
apache-2.0
|
matrix-org/sydent,matrix-org/sydent,matrix-org/sydent
|
---
+++
@@ -26,7 +26,7 @@
self.sydent = syd
def render_GET(self, request):
- pubKey = self.sydent.signers.ed25519.signing_key.verify_key
+ pubKey = self.sydent.keyring.ed25519.verify_key
pubKeyHex = pubKey.encode(encoder=nacl.encoding.HexEncoder)
return json.dumps({'public_key':pubKeyHex})
|
e6f19cc58f32b855fc1f71086dac0ad56b697ed3
|
opps/articles/urls.py
|
opps/articles/urls.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
#
from django.conf.urls import patterns, url
from .views import OppsDetail, OppsList, Search
urlpatterns = patterns(
'',
url(r'^$', OppsList.as_view(), name='home'),
url(r'^search/', Search(), name='search'),
url(r'^(?P<channel__long_slug>[\w//-]+)/(?P<slug>[\w-]+)$',
OppsDetail.as_view(), name='open'),
url(r'^(?P<channel__long_slug>[\w\b//-]+)/$', OppsList.as_view(),
name='channel'),
)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.conf.urls import patterns, url
from django.views.decorators.cache import cache_page
from .views import OppsDetail, OppsList, Search
urlpatterns = patterns(
'',
url(r'^$', cache_page(60 * 2)(OppsList.as_view()), name='home'),
url(r'^search/', Search(), name='search'),
url(r'^(?P<channel__long_slug>[\w//-]+)/(?P<slug>[\w-]+)$',
cache_page(60 * 15)(OppsDetail.as_view()), name='open'),
url(r'^(?P<channel__long_slug>[\w\b//-]+)/$',
cache_page(60 * 2)(OppsList.as_view()), name='channel'),
)
|
Add cache on article page (via url)
|
Add cache on article page (via url)
|
Python
|
mit
|
opps/opps,YACOWS/opps,YACOWS/opps,williamroot/opps,opps/opps,jeanmask/opps,jeanmask/opps,jeanmask/opps,opps/opps,jeanmask/opps,YACOWS/opps,YACOWS/opps,williamroot/opps,opps/opps,williamroot/opps,williamroot/opps
|
---
+++
@@ -1,17 +1,17 @@
#!/usr/bin/env python
# -*- coding: utf-8 -*-
-#
from django.conf.urls import patterns, url
+from django.views.decorators.cache import cache_page
from .views import OppsDetail, OppsList, Search
urlpatterns = patterns(
'',
- url(r'^$', OppsList.as_view(), name='home'),
+ url(r'^$', cache_page(60 * 2)(OppsList.as_view()), name='home'),
url(r'^search/', Search(), name='search'),
url(r'^(?P<channel__long_slug>[\w//-]+)/(?P<slug>[\w-]+)$',
- OppsDetail.as_view(), name='open'),
- url(r'^(?P<channel__long_slug>[\w\b//-]+)/$', OppsList.as_view(),
- name='channel'),
+ cache_page(60 * 15)(OppsDetail.as_view()), name='open'),
+ url(r'^(?P<channel__long_slug>[\w\b//-]+)/$',
+ cache_page(60 * 2)(OppsList.as_view()), name='channel'),
)
|
cfa22fada64882a20f2daec4c8b83488920e8c3a
|
models/log_entry.py
|
models/log_entry.py
|
from database import db
from conversions import datetime_from_str
class LogEntry(db.Model):
id = db.Column(db.Integer, primary_key=True)
timestamp = db.Column(db.DateTime, index=True)
server = db.Column(db.String(100), index=True)
log_name = db.Column(db.String(760), index=True)
message = db.Column(db.Text(), nullable=True)
def __init__(self, timestamp, server, log_name, message):
self.timestamp = datetime_from_str(timestamp)
self.server = server
self.log_name = log_name
self.message = message
def to_dict(self):
return {
'timestamp': self.timestamp,
'server': self.server,
'log_name': self.log_name,
'message': self.message,
}
|
from database import db
from conversions import datetime_from_str
class LogEntry(db.Model):
id = db.Column(db.Integer, primary_key=True)
timestamp = db.Column(db.DateTime, index=True)
server = db.Column(db.String(100), index=True)
log_name = db.Column(db.String(255), index=True)
message = db.Column(db.Text(), nullable=True)
def __init__(self, timestamp, server, log_name, message):
self.timestamp = datetime_from_str(timestamp)
self.server = server
self.log_name = log_name
self.message = message
def to_dict(self):
return {
'timestamp': self.timestamp,
'server': self.server,
'log_name': self.log_name,
'message': self.message,
}
|
Reduce the size of log_name so it fits within mysql's limit.
|
Reduce the size of log_name so it fits within mysql's limit.
|
Python
|
agpl-3.0
|
izrik/sawmill,izrik/sawmill,izrik/sawmill
|
---
+++
@@ -8,7 +8,7 @@
id = db.Column(db.Integer, primary_key=True)
timestamp = db.Column(db.DateTime, index=True)
server = db.Column(db.String(100), index=True)
- log_name = db.Column(db.String(760), index=True)
+ log_name = db.Column(db.String(255), index=True)
message = db.Column(db.Text(), nullable=True)
def __init__(self, timestamp, server, log_name, message):
|
fb019fa4c277cf988d479333c6cba08a637f948a
|
flexget/plugins/output/dump_config.py
|
flexget/plugins/output/dump_config.py
|
from argparse import SUPPRESS
from loguru import logger
from rich.syntax import Syntax
from flexget import options, plugin
from flexget.event import event
from flexget.terminal import console
logger = logger.bind(name='dump_config')
class OutputDumpConfig:
"""
Dumps task config in STDOUT in yaml at exit or abort event.
"""
@plugin.priority(plugin.PRIORITY_LAST)
def on_task_start(self, task, config):
if task.options.dump_config:
import yaml
console.rule(f'config from task: {task.name}')
syntax = Syntax(yaml.safe_dump(task.config).strip(), 'yaml')
console(syntax)
console.rule()
task.abort(silent=True)
if task.options.dump_config_python:
console(task.config)
task.abort(silent=True)
@event('plugin.register')
def register_plugin():
plugin.register(OutputDumpConfig, 'dump_config', debug=True, builtin=True, api_ver=2)
@event('options.register')
def register_parser_arguments():
exec_parser = options.get_parser('execute')
exec_parser.add_argument(
'--dump-config',
action='store_true',
dest='dump_config',
default=False,
help='display the config of each feed after template merging/config generation occurs',
)
exec_parser.add_argument(
'--dump-config-python',
action='store_true',
dest='dump_config_python',
default=False,
help=SUPPRESS,
)
|
from argparse import SUPPRESS
import yaml
from loguru import logger
from rich.syntax import Syntax
from flexget import options, plugin
from flexget.event import event
from flexget.terminal import console
logger = logger.bind(name='dump_config')
class OutputDumpConfig:
"""
Dumps task config in STDOUT in yaml at exit or abort event.
"""
@plugin.priority(plugin.PRIORITY_LAST)
def on_task_start(self, task, config):
if task.options.dump_config:
console.rule(f'config from task: {task.name}')
syntax = Syntax(yaml.safe_dump(task.config).strip(), 'yaml+jinja', theme='native')
console(syntax)
console.rule()
task.abort(silent=True)
if task.options.dump_config_python:
console(task.config)
task.abort(silent=True)
@event('plugin.register')
def register_plugin():
plugin.register(OutputDumpConfig, 'dump_config', debug=True, builtin=True, api_ver=2)
@event('options.register')
def register_parser_arguments():
exec_parser = options.get_parser('execute')
exec_parser.add_argument(
'--dump-config',
action='store_true',
dest='dump_config',
default=False,
help='display the config of each feed after template merging/config generation occurs',
)
exec_parser.add_argument(
'--dump-config-python',
action='store_true',
dest='dump_config_python',
default=False,
help=SUPPRESS,
)
|
Adjust --dump-config color theme for better readability Add jinja highlighting to --dump-config
|
Adjust --dump-config color theme for better readability
Add jinja highlighting to --dump-config
|
Python
|
mit
|
crawln45/Flexget,Flexget/Flexget,Flexget/Flexget,Flexget/Flexget,Flexget/Flexget,crawln45/Flexget,crawln45/Flexget,crawln45/Flexget
|
---
+++
@@ -1,5 +1,6 @@
from argparse import SUPPRESS
+import yaml
from loguru import logger
from rich.syntax import Syntax
@@ -18,10 +19,8 @@
@plugin.priority(plugin.PRIORITY_LAST)
def on_task_start(self, task, config):
if task.options.dump_config:
- import yaml
-
console.rule(f'config from task: {task.name}')
- syntax = Syntax(yaml.safe_dump(task.config).strip(), 'yaml')
+ syntax = Syntax(yaml.safe_dump(task.config).strip(), 'yaml+jinja', theme='native')
console(syntax)
console.rule()
task.abort(silent=True)
|
fbc4247fc7b7d36286c3f25e6ae71dfb7ebb2d39
|
example/__init__.py
|
example/__init__.py
|
from pupa.scrape import Jurisdiction
from .people import PersonScraper
class Example(Jurisdiction):
jurisdiction_id = 'ocd-jurisdiction/country:us/state:ex/place:example'
def get_metadata(self):
return {
'name': 'Example Legislature',
'url': 'http://example.com',
'terms': [{
'name': '2013-2014',
'sessions': ['2013'],
'start_year': 2013,
'end_year': 2014
}],
'provides': ['people'],
'parties': [
{'name': 'Independent' },
{'name': 'Green' },
{'name': 'Bull-Moose'}
],
'session_details': {
'2013': {'_scraped_name': '2013'}
},
'feature_flags': [],
}
def get_scraper(self, term, session, scraper_type):
if scraper_type == 'people':
return PersonScraper
def scrape_session_list(self):
return ['2013']
|
from pupa.scrape import Jurisdiction
from .people import PersonScraper
class Example(Jurisdiction):
jurisdiction_id = 'ocd-jurisdiction/country:us/state:ex/place:example'
name = 'Example Legislature'
url = 'http://example.com'
terms = [{
'name': '2013-2014',
'sessions': ['2013'],
'start_year': 2013,
'end_year': 2014
}]
provides = ['people'],
parties = [
{'name': 'Independent' },
{'name': 'Green' },
{'name': 'Bull-Moose'}
]
session_details = {
'2013': {'_scraped_name': '2013'}
}
def get_scraper(self, term, session, scraper_type):
if scraper_type == 'people':
return PersonScraper
def scrape_session_list(self):
return ['2013']
|
Use new-style metadata in example
|
Use new-style metadata in example
|
Python
|
bsd-3-clause
|
rshorey/pupa,rshorey/pupa,mileswwatkins/pupa,opencivicdata/pupa,datamade/pupa,mileswwatkins/pupa,influence-usa/pupa,influence-usa/pupa,datamade/pupa,opencivicdata/pupa
|
---
+++
@@ -5,28 +5,23 @@
class Example(Jurisdiction):
jurisdiction_id = 'ocd-jurisdiction/country:us/state:ex/place:example'
-
- def get_metadata(self):
- return {
- 'name': 'Example Legislature',
- 'url': 'http://example.com',
- 'terms': [{
- 'name': '2013-2014',
- 'sessions': ['2013'],
- 'start_year': 2013,
- 'end_year': 2014
- }],
- 'provides': ['people'],
- 'parties': [
- {'name': 'Independent' },
- {'name': 'Green' },
- {'name': 'Bull-Moose'}
- ],
- 'session_details': {
- '2013': {'_scraped_name': '2013'}
- },
- 'feature_flags': [],
- }
+ name = 'Example Legislature'
+ url = 'http://example.com'
+ terms = [{
+ 'name': '2013-2014',
+ 'sessions': ['2013'],
+ 'start_year': 2013,
+ 'end_year': 2014
+ }]
+ provides = ['people'],
+ parties = [
+ {'name': 'Independent' },
+ {'name': 'Green' },
+ {'name': 'Bull-Moose'}
+ ]
+ session_details = {
+ '2013': {'_scraped_name': '2013'}
+ }
def get_scraper(self, term, session, scraper_type):
if scraper_type == 'people':
|
07da63a9ac95a054332297638df17fcf00ac4291
|
core/components/security/factor.py
|
core/components/security/factor.py
|
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
import json
from u2flib_server.u2f import (begin_registration,
begin_authentication,
complete_registration,
complete_authentication)
from components.eternity import config
facet = config.rss['link']
async def enroll(user):
enroll = begin_registration(facet, user.get('_u2f_devices_', []))
user['_u2f_enroll_'] = enroll.json
return user, json.dumps(enroll.data_for_client)
async def bind(user, data):
response = data['tokenResponse']
enroll = user.pop('_u2f_enroll_')
device, cert = complete_registration(enroll, response, [facet])
patch = device
patch['deviceName'] = data['deviceName']
patch['registerDate'] = data['date']
user.setdefault('_u2f_devices_', []).append(json.dumps(patch))
# cert = x509.load_der_x509_certificate(cert, default_backend())
return user, True
async def sign(user):
challenge = begin_authentication(facet, user.get('_u2f_devices_', []))
user['_u2f_challenge_'] = challenge.json
return user, json.dumps(challenge.data_for_client)
async def verify(user, data):
challenge = user.pop('_u2f_challenge_')
try:
complete_authentication(challenge, data, [facet])
except AttributeError:
return user, False
return user, True
|
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
import json
from u2flib_server.u2f import (begin_registration,
begin_authentication,
complete_registration,
complete_authentication)
from components.eternity import config
facet = config.rss['link']
async def enroll(user):
enroll = begin_registration(facet, user.get('_u2f_devices_', []))
user['_u2f_enroll_'] = enroll.json
return user, json.dumps(enroll.data_for_client)
async def bind(user, data):
response = data['tokenResponse']
enroll = user.pop('_u2f_enroll_')
device, cert = complete_registration(enroll, response, [facet])
patch = device
patch['deviceName'] = data['deviceName']
patch['registerDate'] = data['date']
user.setdefault('_u2f_devices_', []).append(json.dumps(patch))
# cert = x509.load_der_x509_certificate(cert, default_backend())
return user, True
async def sign(user):
challenge = begin_authentication(facet, user.get('_u2f_devices_', []))
user['_u2f_challenge_'] = challenge.json
return user, json.dumps(challenge.data_for_client)
async def verify(user, data):
print(user)
challenge = user.pop('_u2f_challenge_')
try:
complete_authentication(challenge, data, [facet])
except AttributeError:
return user, False
return user, True
|
Fix server error when login with u2f
|
Fix server error when login with u2f
|
Python
|
mit
|
chiaki64/Windless,chiaki64/Windless
|
---
+++
@@ -34,6 +34,7 @@
return user, json.dumps(challenge.data_for_client)
async def verify(user, data):
+ print(user)
challenge = user.pop('_u2f_challenge_')
try:
complete_authentication(challenge, data, [facet])
|
4c484a29480ec9d85a87ac7c2aaf09ced7d15457
|
nn/file/__init__.py
|
nn/file/__init__.py
|
import functools
import tensorflow as tf
from . import cnn_dailymail_rc
from ..flags import FLAGS
READERS = { "cnn_dailymail_rc": cnn_dailymail_rc.read_files }
def read_files(file_pattern, file_format):
return READERS[file_format](_file_pattern_to_names(file_pattern))
def _file_pattern_to_names(pattern):
return tf.train.string_input_producer(tf.train.match_filenames_once(pattern),
num_epochs=FLAGS.num_epochs,
capacity=FLAGS.filename_queue_capacity)
|
import functools
import tensorflow as tf
from . import cnn_dailymail_rc
from .. import collections
from ..flags import FLAGS
from ..util import func_scope
READERS = { "cnn_dailymail_rc": cnn_dailymail_rc.read_files }
@func_scope
def read_files(file_pattern, file_format):
return monitored_batch_queue(
*READERS[file_format](_file_pattern_to_names(file_pattern)))
@func_scope
def _file_pattern_to_names(pattern):
return tf.train.string_input_producer(tf.train.match_filenames_once(pattern),
num_epochs=FLAGS.num_epochs,
capacity=FLAGS.filename_queue_capacity)
@func_scope
def monitored_batch_queue(*tensors):
queue = tf.FIFOQueue(FLAGS.batch_queue_capacity,
[tensor.dtype for tensor in tensors])
collections.add_metric(queue.size(), "batches_in_queue")
tf.train.add_queue_runner(
tf.train.QueueRunner(queue, [queue.enqueue(tensors)]))
results = queue.dequeue()
for tensor, result in zip(tensors, results):
result.set_shape(tensor.get_shape())
return results
|
Monitor number of batches in a input batch queue
|
Monitor number of batches in a input batch queue
|
Python
|
unlicense
|
raviqqe/tensorflow-extenteten,raviqqe/tensorflow-extenteten
|
---
+++
@@ -2,18 +2,39 @@
import tensorflow as tf
from . import cnn_dailymail_rc
+from .. import collections
from ..flags import FLAGS
+from ..util import func_scope
READERS = { "cnn_dailymail_rc": cnn_dailymail_rc.read_files }
+@func_scope
def read_files(file_pattern, file_format):
- return READERS[file_format](_file_pattern_to_names(file_pattern))
+ return monitored_batch_queue(
+ *READERS[file_format](_file_pattern_to_names(file_pattern)))
+@func_scope
def _file_pattern_to_names(pattern):
return tf.train.string_input_producer(tf.train.match_filenames_once(pattern),
num_epochs=FLAGS.num_epochs,
capacity=FLAGS.filename_queue_capacity)
+
+
+@func_scope
+def monitored_batch_queue(*tensors):
+ queue = tf.FIFOQueue(FLAGS.batch_queue_capacity,
+ [tensor.dtype for tensor in tensors])
+ collections.add_metric(queue.size(), "batches_in_queue")
+
+ tf.train.add_queue_runner(
+ tf.train.QueueRunner(queue, [queue.enqueue(tensors)]))
+
+ results = queue.dequeue()
+ for tensor, result in zip(tensors, results):
+ result.set_shape(tensor.get_shape())
+
+ return results
|
c0c59a9c5d3aa2d7ed50e8e895f1a3e02a4ae380
|
Basic-Number-Guessing-Game-Challenge.py
|
Basic-Number-Guessing-Game-Challenge.py
|
import random
attempts = 1
number = str(random.randint(1, 100))
while True:
print number
if raw_input("Guess (1 - 100): ") == number:
print "Correct!"
print "It Only Took You", attempts, "Attempts!" if attempts > 1 else "Attempt!"
break
else:
print "Incorrect, Guess Again!"
attempts += 1
|
import random
attempts = 1
number = random.randint(1, 100)
while True:
guess = raw_input("Guess (1 - 100): ")
if guess.isdigit():
guess = int(guess)
if 1 <= guess and guess <= 100:
if guess == number:
print "Correct!"
print "It Only Took You", attempts, "Attempts!" if attempts > 1 else "Attempt!"
break
elif guess > number:
print "That Guess Is Too High!"
elif guess < number:
print "That Guess Is Too Low!"
else:
print "Guesses Must Be Between 1 And 100!"
else:
print "That's Not A Number!"
attempts += 1
|
Verify that the guess is a number, that it is between 1 and 100, and tells the user if the guessed number is too high or low.
|
Verify that the guess is a number, that it is between 1 and 100, and tells the user if the guessed number is too high or low.
|
Python
|
mit
|
RascalTwo/Basic-Number-Guessing-Game-Challenge
|
---
+++
@@ -1,13 +1,22 @@
import random
attempts = 1
-number = str(random.randint(1, 100))
+number = random.randint(1, 100)
while True:
- print number
- if raw_input("Guess (1 - 100): ") == number:
- print "Correct!"
- print "It Only Took You", attempts, "Attempts!" if attempts > 1 else "Attempt!"
- break
+ guess = raw_input("Guess (1 - 100): ")
+ if guess.isdigit():
+ guess = int(guess)
+ if 1 <= guess and guess <= 100:
+ if guess == number:
+ print "Correct!"
+ print "It Only Took You", attempts, "Attempts!" if attempts > 1 else "Attempt!"
+ break
+ elif guess > number:
+ print "That Guess Is Too High!"
+ elif guess < number:
+ print "That Guess Is Too Low!"
+ else:
+ print "Guesses Must Be Between 1 And 100!"
else:
- print "Incorrect, Guess Again!"
- attempts += 1
+ print "That's Not A Number!"
+ attempts += 1
|
21e3eac740f54194954d01d517aca0eb841ca1b3
|
wagtail/search/apps.py
|
wagtail/search/apps.py
|
from django.apps import AppConfig
from django.core.checks import Tags, Warning, register
from django.db import connection
from django.utils.translation import gettext_lazy as _
from wagtail.search.signal_handlers import register_signal_handlers
class WagtailSearchAppConfig(AppConfig):
name = 'wagtail.search'
label = 'wagtailsearch'
verbose_name = _("Wagtail search")
default_auto_field = 'django.db.models.AutoField'
def ready(self):
register_signal_handlers()
if connection.vendor == 'postgresql':
# Only PostgreSQL has support for tsvector weights
from wagtail.search.backends.database.postgres.weights import set_weights
set_weights()
from wagtail.search.models import IndexEntry
IndexEntry.add_generic_relations()
@register(Tags.compatibility, Tags.database)
def check_if_sqlite_version_is_supported(app_configs, **kwargs):
if connection.vendor == 'sqlite':
import sqlite3
if sqlite3.sqlite_version_info < (3, 19, 0):
return [Warning('Your SQLite version is older than 3.19.0. A fallback search backend will be used instead.', hint='Upgrade your SQLite version to at least 3.19.0', id='wagtailsearch.W002', obj=WagtailSearchAppConfig)]
return []
|
from django.apps import AppConfig
from django.core.checks import Tags, Warning, register
from django.db import connection
from django.utils.translation import gettext_lazy as _
from wagtail.search.signal_handlers import register_signal_handlers
class WagtailSearchAppConfig(AppConfig):
name = 'wagtail.search'
label = 'wagtailsearch'
verbose_name = _("Wagtail search")
default_auto_field = 'django.db.models.AutoField'
def ready(self):
register_signal_handlers()
if connection.vendor == 'postgresql':
# Only PostgreSQL has support for tsvector weights
from wagtail.search.backends.database.postgres.weights import set_weights
set_weights()
from wagtail.search.models import IndexEntry
IndexEntry.add_generic_relations()
@register(Tags.compatibility, Tags.database)
def check_if_sqlite_version_is_supported(app_configs, **kwargs):
if connection.vendor == 'sqlite':
import sqlite3
from wagtail.search.backends.database.sqlite.utils import fts5_available
if sqlite3.sqlite_version_info < (3, 19, 0):
return [Warning('Your SQLite version is older than 3.19.0. A fallback search backend will be used instead.', hint='Upgrade your SQLite version to at least 3.19.0', id='wagtailsearch.W002', obj=WagtailSearchAppConfig)]
elif not fts5_available():
return [Warning('Your SQLite installation is missing the fts5 extension. A fallback search backend will be used instead.', hint='Upgrade your SQLite installation to a version with fts5 enabled', id='wagtailsearch.W003', obj=WagtailSearchAppConfig)]
return []
|
Add alternative warning if sqlite is >=3.19 but is missing fts5 support
|
Add alternative warning if sqlite is >=3.19 but is missing fts5 support
|
Python
|
bsd-3-clause
|
torchbox/wagtail,torchbox/wagtail,torchbox/wagtail,torchbox/wagtail
|
---
+++
@@ -27,6 +27,11 @@
def check_if_sqlite_version_is_supported(app_configs, **kwargs):
if connection.vendor == 'sqlite':
import sqlite3
+
+ from wagtail.search.backends.database.sqlite.utils import fts5_available
+
if sqlite3.sqlite_version_info < (3, 19, 0):
return [Warning('Your SQLite version is older than 3.19.0. A fallback search backend will be used instead.', hint='Upgrade your SQLite version to at least 3.19.0', id='wagtailsearch.W002', obj=WagtailSearchAppConfig)]
+ elif not fts5_available():
+ return [Warning('Your SQLite installation is missing the fts5 extension. A fallback search backend will be used instead.', hint='Upgrade your SQLite installation to a version with fts5 enabled', id='wagtailsearch.W003', obj=WagtailSearchAppConfig)]
return []
|
cdfeac780643ddd2502c17f4cd7d949018de8b06
|
warehouse/__about__.py
|
warehouse/__about__.py
|
# Copyright 2013 Donald Stufft
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import absolute_import, division, print_function
from __future__ import unicode_literals
# This file is automatically generated, do not edit it
__all__ = [
"__title__", "__summary__", "__uri__", "__version__",
"__author__", "__email__", "__license__", "__copyright__",
]
__title__ = "warehouse"
__summary__ = "Next Generation Python Package Repository"
__uri__ = "https://github.com/dstufft/warehouse"
__version__ = "13.10.10"
__build__ = "<development>"
__author__ = "Donald Stufft"
__email__ = "donald@stufft.io"
__license__ = "Apache License, Version 2.0"
__copyright__ = "Copyright 2013 Donald Stufft"
|
# Copyright 2013 Donald Stufft
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import absolute_import, division, print_function
from __future__ import unicode_literals
# This file is automatically generated, do not edit it
__all__ = [
"__title__", "__summary__", "__uri__", "__version__",
"__author__", "__email__", "__license__", "__copyright__",
]
__title__ = "warehouse"
__summary__ = "Next Generation Python Package Repository"
__uri__ = "https://github.com/dstufft/warehouse"
__version__ = "13.10.10.dev0"
__build__ = "<development>"
__author__ = "Donald Stufft"
__email__ = "donald@stufft.io"
__license__ = "Apache License, Version 2.0"
__copyright__ = "Copyright 2013 Donald Stufft"
|
Add back the development marker
|
Add back the development marker
|
Python
|
apache-2.0
|
mattrobenolt/warehouse,robhudson/warehouse,techtonik/warehouse,robhudson/warehouse,mattrobenolt/warehouse,techtonik/warehouse,mattrobenolt/warehouse
|
---
+++
@@ -26,7 +26,7 @@
__summary__ = "Next Generation Python Package Repository"
__uri__ = "https://github.com/dstufft/warehouse"
-__version__ = "13.10.10"
+__version__ = "13.10.10.dev0"
__build__ = "<development>"
__author__ = "Donald Stufft"
|
7473a70893a0f31ba717e26a3f508d0adc5026f9
|
osbrain/__init__.py
|
osbrain/__init__.py
|
import Pyro4
Pyro4.config.SERIALIZERS_ACCEPTED.add('pickle')
Pyro4.config.SERIALIZER = 'pickle'
Pyro4.config.THREADPOOL_SIZE = 16
Pyro4.config.SERVERTYPE = 'multiplex'
# TODO: should we set COMMTIMEOUT as well?
Pyro4.config.DETAILED_TRACEBACK = True
__version__ = '0.2.1'
from .core import BaseAgent, Agent, run_agent
from .nameserver import random_nameserver, run_nameserver
from .proxy import Proxy, NSProxy
from .address import SocketAddress, AgentAddress
from .logging import Logger, run_logger
|
import Pyro4
Pyro4.config.SERIALIZERS_ACCEPTED.add('pickle')
Pyro4.config.SERIALIZER = 'pickle'
Pyro4.config.THREADPOOL_SIZE = 16
Pyro4.config.SERVERTYPE = 'multiplex'
Pyro4.config.REQUIRE_EXPOSE = False
# TODO: should we set COMMTIMEOUT as well?
Pyro4.config.DETAILED_TRACEBACK = True
__version__ = '0.2.1'
from .core import BaseAgent, Agent, run_agent
from .nameserver import random_nameserver, run_nameserver
from .proxy import Proxy, NSProxy
from .address import SocketAddress, AgentAddress
from .logging import Logger, run_logger
|
Make osBrain compatible with latest Pyro4 changes
|
Make osBrain compatible with latest Pyro4 changes
|
Python
|
apache-2.0
|
opensistemas-hub/osbrain
|
---
+++
@@ -3,6 +3,7 @@
Pyro4.config.SERIALIZER = 'pickle'
Pyro4.config.THREADPOOL_SIZE = 16
Pyro4.config.SERVERTYPE = 'multiplex'
+Pyro4.config.REQUIRE_EXPOSE = False
# TODO: should we set COMMTIMEOUT as well?
Pyro4.config.DETAILED_TRACEBACK = True
|
e12de19ae37a6f3fa0335ecfd0db00b18badf730
|
website/files/utils.py
|
website/files/utils.py
|
def copy_files(src, target_node, parent=None, name=None):
"""Copy the files from src to the target node
:param Folder src: The source to copy children from
:param Node target_node: The node to copy files to
:param Folder parent: The parent of to attach the clone of src to, if applicable
"""
assert not parent or not parent.is_file, 'Parent must be a folder'
cloned = src.clone()
cloned.parent = parent
cloned.target = target_node
cloned.name = name or cloned.name
cloned.copied_from = src
cloned.save()
if src.is_file and src.versions.exists():
fileversions = src.versions.select_related('region').order_by('-created')
most_recent_fileversion = fileversions.first()
if most_recent_fileversion.region != target_node.osfstorage_region:
# add all original version except the most recent
cloned.versions.add(*fileversions[1:])
# create a new most recent version and update the region before adding
new_fileversion = most_recent_fileversion.clone()
new_fileversion.region = target_node.osfstorage_region
new_fileversion.save()
cloned.versions.add(new_fileversion)
else:
cloned.versions.add(*src.versions.all())
if not src.is_file:
for child in src.children:
copy_files(child, target_node, parent=cloned)
return cloned
|
def copy_files(src, target_node, parent=None, name=None):
"""Copy the files from src to the target node
:param Folder src: The source to copy children from
:param Node target_node: The node to copy files to
:param Folder parent: The parent of to attach the clone of src to, if applicable
"""
assert not parent or not parent.is_file, 'Parent must be a folder'
cloned = src.clone()
cloned.parent = parent
cloned.target = target_node
cloned.name = name or cloned.name
cloned.copied_from = src
cloned.save()
if src.is_file and src.versions.exists():
fileversions = src.versions.select_related('region').order_by('-created')
most_recent_fileversion = fileversions.first()
if most_recent_fileversion.region and most_recent_fileversion.region != target_node.osfstorage_region:
# add all original version except the most recent
cloned.versions.add(*fileversions[1:])
# create a new most recent version and update the region before adding
new_fileversion = most_recent_fileversion.clone()
new_fileversion.region = target_node.osfstorage_region
new_fileversion.save()
cloned.versions.add(new_fileversion)
else:
cloned.versions.add(*src.versions.all())
if not src.is_file:
for child in src.children:
copy_files(child, target_node, parent=cloned)
return cloned
|
Check for a fileversion region before copying them to the source region
|
Check for a fileversion region before copying them to the source region
This is mainly for test fileversions that are created without regions by default
|
Python
|
apache-2.0
|
aaxelb/osf.io,felliott/osf.io,brianjgeiger/osf.io,HalcyonChimera/osf.io,HalcyonChimera/osf.io,baylee-d/osf.io,mattclark/osf.io,saradbowman/osf.io,Johnetordoff/osf.io,brianjgeiger/osf.io,cslzchen/osf.io,cslzchen/osf.io,mfraezz/osf.io,HalcyonChimera/osf.io,adlius/osf.io,felliott/osf.io,Johnetordoff/osf.io,caseyrollins/osf.io,mattclark/osf.io,mfraezz/osf.io,CenterForOpenScience/osf.io,adlius/osf.io,felliott/osf.io,adlius/osf.io,brianjgeiger/osf.io,mfraezz/osf.io,baylee-d/osf.io,mattclark/osf.io,felliott/osf.io,HalcyonChimera/osf.io,baylee-d/osf.io,mfraezz/osf.io,caseyrollins/osf.io,pattisdr/osf.io,Johnetordoff/osf.io,brianjgeiger/osf.io,aaxelb/osf.io,aaxelb/osf.io,pattisdr/osf.io,pattisdr/osf.io,saradbowman/osf.io,cslzchen/osf.io,adlius/osf.io,CenterForOpenScience/osf.io,cslzchen/osf.io,caseyrollins/osf.io,CenterForOpenScience/osf.io,CenterForOpenScience/osf.io,aaxelb/osf.io,Johnetordoff/osf.io
|
---
+++
@@ -18,7 +18,7 @@
if src.is_file and src.versions.exists():
fileversions = src.versions.select_related('region').order_by('-created')
most_recent_fileversion = fileversions.first()
- if most_recent_fileversion.region != target_node.osfstorage_region:
+ if most_recent_fileversion.region and most_recent_fileversion.region != target_node.osfstorage_region:
# add all original version except the most recent
cloned.versions.add(*fileversions[1:])
# create a new most recent version and update the region before adding
|
7e046e4999959e0cfa3527780ca04e581698b328
|
cbagent/collectors/libstats/psstats.py
|
cbagent/collectors/libstats/psstats.py
|
from fabric.api import run
from cbagent.collectors.libstats.remotestats import (
RemoteStats, multi_node_task)
class PSStats(RemoteStats):
METRICS = (
("rss", 1024), # kB -> B
("vsize", 1024),
)
def __init__(self, hosts, user, password):
super(PSStats, self).__init__(hosts, user, password)
self.ps_cmd = "ps -eo pid,rss,vsize,comm | " \
"grep {} | grep -v grep | sort -n -k 2 | tail -n 1"
self.top_cmd = "top -bn2d1 -p {} | grep {}"
@multi_node_task
def get_samples(self, process):
samples = {}
stdout = run(self.ps_cmd.format(process))
if stdout:
for i, value in enumerate(stdout.split()[1:1+len(self.METRICS)]):
metric, multiplier = self.METRICS[i]
title = "{}_{}".format(process, metric)
samples[title] = float(value) * multiplier
pid = stdout.split()[0]
else:
return samples
stdout = run(self.top_cmd.format(pid, process))
if stdout:
title = "{}_cpu".format(process)
samples[title] = float(stdout.split()[8])
return samples
|
from fabric.api import run
from cbagent.collectors.libstats.remotestats import (
RemoteStats, multi_node_task)
class PSStats(RemoteStats):
METRICS = (
("rss", 1024), # kB -> B
("vsize", 1024),
)
def __init__(self, hosts, user, password):
super(PSStats, self).__init__(hosts, user, password)
self.ps_cmd = "ps -eo pid,rss,vsize,comm | " \
"grep {} | grep -v grep | sort -n -k 2 | tail -n 1"
self.top_cmd = "top -bn2 -d1 -p {} | grep {}"
@multi_node_task
def get_samples(self, process):
samples = {}
stdout = run(self.ps_cmd.format(process))
if stdout:
for i, value in enumerate(stdout.split()[1:1+len(self.METRICS)]):
metric, multiplier = self.METRICS[i]
title = "{}_{}".format(process, metric)
samples[title] = float(value) * multiplier
pid = stdout.split()[0]
else:
return samples
stdout = run(self.top_cmd.format(pid, process))
if stdout:
title = "{}_cpu".format(process)
samples[title] = float(stdout.split()[8])
return samples
|
Make `top` work on Ubuntu 12.04
|
MB-13234: Make `top` work on Ubuntu 12.04
The `top` in Ubuntu12.04 seems to do different command line parsing
than the one on CentOS. Separating the parameters should work on
both.
Change-Id: I8f9ec022bcb8e0158316fdaac226acbfb0d9d004
Reviewed-on: http://review.couchbase.org/50126
Reviewed-by: Dave Rigby <a09264da4832c7ff1d3bf1608a19f4b870f93750@couchbase.com>
Reviewed-by: Michael Wiederhold <a17fed27eaa842282862ff7c1b9c8395a26ac320@couchbase.com>
Tested-by: Volker Mische <fb414f8ac0dbbf87663550ae4ef5fc95b1041941@gmail.com>
|
Python
|
apache-2.0
|
mikewied/cbagent,couchbase/cbagent
|
---
+++
@@ -15,7 +15,7 @@
super(PSStats, self).__init__(hosts, user, password)
self.ps_cmd = "ps -eo pid,rss,vsize,comm | " \
"grep {} | grep -v grep | sort -n -k 2 | tail -n 1"
- self.top_cmd = "top -bn2d1 -p {} | grep {}"
+ self.top_cmd = "top -bn2 -d1 -p {} | grep {}"
@multi_node_task
def get_samples(self, process):
|
b4482c257c5333f902569b40bf4e61c8003dbacc
|
www/config/__init__.py
|
www/config/__init__.py
|
from __future__ import unicode_literals
try:
from local import *
except ImportError:
try:
from dev import *
except ImportError:
pass
try:
DEBUG
TEMPLATE_DEBUG
DATABASES['default']
CELERY_BROKER
except NameError:
raise NameError('Required config values not found. Abort !')
|
from __future__ import unicode_literals, absolute_import
try:
from .local import *
except ImportError:
try:
from dev import *
except ImportError:
pass
try:
DEBUG
TEMPLATE_DEBUG
DATABASES['default']
CELERY_BROKER
except NameError:
raise NameError('Required config values not found. Abort !')
|
Use absolute import to correctly import local.py config file
|
Use absolute import to correctly import local.py config file
|
Python
|
agpl-3.0
|
UrLab/beta402,UrLab/beta402,UrLab/DocHub,UrLab/DocHub,UrLab/beta402,UrLab/DocHub,UrLab/DocHub
|
---
+++
@@ -1,7 +1,7 @@
-from __future__ import unicode_literals
+from __future__ import unicode_literals, absolute_import
try:
- from local import *
+ from .local import *
except ImportError:
try:
from dev import *
|
bf7626df74d78f2811f20173fb21c36a96cc9500
|
packages/gtk-doc.py
|
packages/gtk-doc.py
|
GnomePackage ('gtk-doc', version_major = '1', version_minor = '17', sources = [
'http://ftp.gnome.org/pub/gnome/sources/%{name}/%{version}/%{name}-%{version}.tar.bz2'
])
|
GnomePackage ('gtk-doc', version_major = '1', version_minor = '17', configure_flags = ['--with-xml-catalog="%{prefix}/etc/xml/catalog"'], sources = [
'http://ftp.gnome.org/pub/gnome/sources/%{name}/%{version}/%{name}-%{version}.tar.bz2'
])
|
Set xml catalog to the right prefix.
|
Set xml catalog to the right prefix.
|
Python
|
mit
|
BansheeMediaPlayer/bockbuild,mono/bockbuild,mono/bockbuild,BansheeMediaPlayer/bockbuild,BansheeMediaPlayer/bockbuild
|
---
+++
@@ -1,3 +1,3 @@
-GnomePackage ('gtk-doc', version_major = '1', version_minor = '17', sources = [
+GnomePackage ('gtk-doc', version_major = '1', version_minor = '17', configure_flags = ['--with-xml-catalog="%{prefix}/etc/xml/catalog"'], sources = [
'http://ftp.gnome.org/pub/gnome/sources/%{name}/%{version}/%{name}-%{version}.tar.bz2'
])
|
5be3d07995803d81e6238a561c772e856b81a367
|
icekit/templatetags/search_tags.py
|
icekit/templatetags/search_tags.py
|
from django.template import Library, Node
from django.test.client import RequestFactory
register = Library()
factory = RequestFactory()
class FakeRequestNode(Node):
def render(self, context):
req = factory.get('/')
req.notifications = []
context['request'] = req
return ''
@register.tag
def fake_request(parser, token):
"""
Create a fake request object in the context
"""
return FakeRequestNode()
|
from django.contrib.auth.models import AnonymousUser
from django.template import Library, Node
from django.test.client import RequestFactory
register = Library()
factory = RequestFactory()
class FakeRequestNode(Node):
def render(self, context):
req = factory.get('/')
req.notifications = []
req.user = AnonymousUser()
context['request'] = req
return ''
@register.tag
def fake_request(parser, token):
"""
Create a fake request object in the context
"""
return FakeRequestNode()
|
Add anonymous user to fake request object.
|
Add anonymous user to fake request object.
|
Python
|
mit
|
ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit
|
---
+++
@@ -1,3 +1,4 @@
+from django.contrib.auth.models import AnonymousUser
from django.template import Library, Node
from django.test.client import RequestFactory
@@ -10,6 +11,7 @@
def render(self, context):
req = factory.get('/')
req.notifications = []
+ req.user = AnonymousUser()
context['request'] = req
return ''
@@ -21,4 +23,3 @@
Create a fake request object in the context
"""
return FakeRequestNode()
-
|
94b6b97dc1e706a6560092aa29cbe4e21f052924
|
froide/account/apps.py
|
froide/account/apps.py
|
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
class AccountConfig(AppConfig):
name = 'froide.account'
verbose_name = _("Account")
def ready(self):
from froide.bounce.signals import user_email_bounced
user_email_bounced.connect(deactivate_user_after_bounce)
def deactivate_user_after_bounce(sender, bounce, should_deactivate=False, **kwargs):
if not should_deactivate:
return
if not bounce.user:
return
bounce.user.deactivate()
|
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
from django.urls import reverse
from .menu import menu_registry, MenuItem
class AccountConfig(AppConfig):
name = 'froide.account'
verbose_name = _("Account")
def ready(self):
from froide.bounce.signals import user_email_bounced
user_email_bounced.connect(deactivate_user_after_bounce)
menu_registry.register(get_settings_menu_item)
menu_registry.register(get_request_menu_item)
def deactivate_user_after_bounce(sender, bounce, should_deactivate=False, **kwargs):
if not should_deactivate:
return
if not bounce.user:
return
bounce.user.deactivate()
def get_request_menu_item(request):
return MenuItem(
section='before_request', order=999,
url=reverse('account-show'),
label=_('My requests')
)
def get_settings_menu_item(request):
return MenuItem(
section='after_settings', order=-1,
url=reverse('account-settings'),
label=_('Settings')
)
|
Make settings and requests menu items
|
Make settings and requests menu items
|
Python
|
mit
|
fin/froide,fin/froide,fin/froide,stefanw/froide,stefanw/froide,fin/froide,stefanw/froide,stefanw/froide,stefanw/froide
|
---
+++
@@ -1,5 +1,8 @@
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
+from django.urls import reverse
+
+from .menu import menu_registry, MenuItem
class AccountConfig(AppConfig):
@@ -11,6 +14,9 @@
user_email_bounced.connect(deactivate_user_after_bounce)
+ menu_registry.register(get_settings_menu_item)
+ menu_registry.register(get_request_menu_item)
+
def deactivate_user_after_bounce(sender, bounce, should_deactivate=False, **kwargs):
if not should_deactivate:
@@ -18,3 +24,19 @@
if not bounce.user:
return
bounce.user.deactivate()
+
+
+def get_request_menu_item(request):
+ return MenuItem(
+ section='before_request', order=999,
+ url=reverse('account-show'),
+ label=_('My requests')
+ )
+
+
+def get_settings_menu_item(request):
+ return MenuItem(
+ section='after_settings', order=-1,
+ url=reverse('account-settings'),
+ label=_('Settings')
+ )
|
2538ca440620de8ed08510e0ae902c82184a9daa
|
consts/auth_type.py
|
consts/auth_type.py
|
class AuthType(object):
"""
An auth type defines what write privileges an authenticated agent has.
"""
MATCH_VIDEO = 0
EVENT_TEAMS = 1
EVENT_MATCHES = 2
EVENT_RANKINGS = 3
EVENT_ALLIANCES = 4
EVENT_AWARDS = 5
type_names = {
MATCH_VIDEO: "match video",
EVENT_TEAMS: "event teams",
EVENT_MATCHES: "event matches",
EVENT_RANKINGS: "event rankings",
EVENT_ALLIANCES: "event alliances",
EVENT_AWARDS: "event awrads"
}
|
class AuthType(object):
"""
An auth type defines what write privileges an authenticated agent has.
"""
EVENT_DATA = 0
MATCH_VIDEO = 1
EVENT_TEAMS = 2
EVENT_MATCHES = 3
EVENT_RANKINGS = 4
EVENT_ALLIANCES = 5
EVENT_AWARDS = 6
type_names = {
EVENT_DATA: "event data",
MATCH_VIDEO: "match video",
EVENT_TEAMS: "event teams",
EVENT_MATCHES: "event matches",
EVENT_RANKINGS: "event rankings",
EVENT_ALLIANCES: "event alliances",
EVENT_AWARDS: "event awrads"
}
|
Revert "Remove AuthType.EVENT_DATA and renumber"
|
Revert "Remove AuthType.EVENT_DATA and renumber"
This reverts commit 38248941eb47a04b82fe52e1adca7387dafcf7f3.
|
Python
|
mit
|
phil-lopreiato/the-blue-alliance,fangeugene/the-blue-alliance,bdaroz/the-blue-alliance,bvisness/the-blue-alliance,tsteward/the-blue-alliance,josephbisch/the-blue-alliance,jaredhasenklein/the-blue-alliance,verycumbersome/the-blue-alliance,phil-lopreiato/the-blue-alliance,the-blue-alliance/the-blue-alliance,bdaroz/the-blue-alliance,bdaroz/the-blue-alliance,bdaroz/the-blue-alliance,josephbisch/the-blue-alliance,synth3tk/the-blue-alliance,jaredhasenklein/the-blue-alliance,josephbisch/the-blue-alliance,jaredhasenklein/the-blue-alliance,nwalters512/the-blue-alliance,bvisness/the-blue-alliance,synth3tk/the-blue-alliance,tsteward/the-blue-alliance,jaredhasenklein/the-blue-alliance,nwalters512/the-blue-alliance,fangeugene/the-blue-alliance,verycumbersome/the-blue-alliance,bvisness/the-blue-alliance,tsteward/the-blue-alliance,fangeugene/the-blue-alliance,synth3tk/the-blue-alliance,fangeugene/the-blue-alliance,phil-lopreiato/the-blue-alliance,bvisness/the-blue-alliance,tsteward/the-blue-alliance,fangeugene/the-blue-alliance,fangeugene/the-blue-alliance,tsteward/the-blue-alliance,josephbisch/the-blue-alliance,synth3tk/the-blue-alliance,synth3tk/the-blue-alliance,bdaroz/the-blue-alliance,the-blue-alliance/the-blue-alliance,nwalters512/the-blue-alliance,bvisness/the-blue-alliance,jaredhasenklein/the-blue-alliance,nwalters512/the-blue-alliance,the-blue-alliance/the-blue-alliance,the-blue-alliance/the-blue-alliance,verycumbersome/the-blue-alliance,bvisness/the-blue-alliance,verycumbersome/the-blue-alliance,synth3tk/the-blue-alliance,nwalters512/the-blue-alliance,phil-lopreiato/the-blue-alliance,phil-lopreiato/the-blue-alliance,jaredhasenklein/the-blue-alliance,nwalters512/the-blue-alliance,tsteward/the-blue-alliance,phil-lopreiato/the-blue-alliance,josephbisch/the-blue-alliance,the-blue-alliance/the-blue-alliance,bdaroz/the-blue-alliance,the-blue-alliance/the-blue-alliance,verycumbersome/the-blue-alliance,josephbisch/the-blue-alliance,verycumbersome/the-blue-alliance
|
---
+++
@@ -2,14 +2,16 @@
"""
An auth type defines what write privileges an authenticated agent has.
"""
- MATCH_VIDEO = 0
- EVENT_TEAMS = 1
- EVENT_MATCHES = 2
- EVENT_RANKINGS = 3
- EVENT_ALLIANCES = 4
- EVENT_AWARDS = 5
+ EVENT_DATA = 0
+ MATCH_VIDEO = 1
+ EVENT_TEAMS = 2
+ EVENT_MATCHES = 3
+ EVENT_RANKINGS = 4
+ EVENT_ALLIANCES = 5
+ EVENT_AWARDS = 6
type_names = {
+ EVENT_DATA: "event data",
MATCH_VIDEO: "match video",
EVENT_TEAMS: "event teams",
EVENT_MATCHES: "event matches",
|
e7df1e7f3e9d8afd5cf1892df2f136751b276136
|
aio_pika/transaction.py
|
aio_pika/transaction.py
|
import asyncio
from enum import Enum
import aiormq
class TransactionStates(Enum):
created = "created"
commited = "commited"
rolled_back = "rolled back"
started = "started"
class Transaction:
def __str__(self):
return self.state.value
def __init__(self, channel):
self.loop = channel.loop
self._channel = channel
self.state = TransactionStates.created # type: TransactionStates
@property
def channel(self) -> aiormq.Channel:
if self._channel is None:
raise RuntimeError("Channel not opened")
if self._channel.is_closed:
raise RuntimeError("Closed channel")
return self._channel
async def select(self, timeout=None) -> aiormq.spec.Tx.SelectOk:
result = await asyncio.wait_for(
self.channel.tx_select(), timeout=timeout,
)
self.state = TransactionStates.started
return result
async def rollback(self, timeout=None):
result = await asyncio.wait_for(
self.channel.tx_rollback(), timeout=timeout,
)
self.state = TransactionStates.rolled_back
return result
async def commit(self, timeout=None):
result = await asyncio.wait_for(
self.channel.tx_commit(), timeout=timeout,
)
self.state = TransactionStates.commited
return result
async def __aenter__(self):
return await self.select()
async def __aexit__(self, exc_type, exc_val, exc_tb):
if exc_type:
await self.rollback()
else:
await self.commit()
|
import asyncio
from enum import Enum
import aiormq
class TransactionStates(Enum):
created = "created"
commited = "commited"
rolled_back = "rolled back"
started = "started"
class Transaction:
def __str__(self):
return self.state.value
def __init__(self, channel):
self.loop = channel.loop
self._channel = channel
self.state = TransactionStates.created # type: TransactionStates
@property
def channel(self) -> aiormq.Channel:
if self._channel is None:
raise RuntimeError("Channel not opened")
if self._channel.is_closed:
raise RuntimeError("Closed channel")
return self._channel
async def select(self, timeout=None) -> aiormq.spec.Tx.SelectOk:
result = await asyncio.wait_for(
self.channel.tx_select(), timeout=timeout,
)
self.state = TransactionStates.started
return result
async def rollback(self, timeout=None):
result = await asyncio.wait_for(
self.channel.tx_rollback(), timeout=timeout,
)
self.state = TransactionStates.rolled_back
return result
async def commit(self, timeout=None):
result = await asyncio.wait_for(
self.channel.tx_commit(), timeout=timeout,
)
self.state = TransactionStates.commited
return result
async def __aenter__(self):
await self.select()
return self
async def __aexit__(self, exc_type, exc_val, exc_tb):
if exc_type:
await self.rollback()
else:
await self.commit()
|
Return self instead of select result in __aenter__
|
Return self instead of select result in __aenter__
|
Python
|
apache-2.0
|
mosquito/aio-pika
|
---
+++
@@ -54,7 +54,8 @@
return result
async def __aenter__(self):
- return await self.select()
+ await self.select()
+ return self
async def __aexit__(self, exc_type, exc_val, exc_tb):
if exc_type:
|
e0990fcdb7e5e1c90762a71ced7492e28f24c903
|
raven/utils/compat.py
|
raven/utils/compat.py
|
"""
raven.utils.compat
~~~~~~~~~~~~~~~~~~~~~~~~~~
:copyright: (c) 2010-2012 by the Sentry Team, see AUTHORS for more details.
:license: BSD, see LICENSE for more details.
"""
try:
from urllib.error import HTTPError
except ImportError:
from urllib2 import HTTPError
try:
from urllib.request import Request, urlopen
except ImportError:
from urllib2 import Request, urlopen
try:
from urllib import quote as urllib_quote
except ImportError:
from urllib.parse import quote as urllib_quote
try:
from queue import Queue
except ImportError:
from Queue import Queue
try:
import urlparse as _urlparse
except ImportError:
from urllib import parse as _urlparse
try:
from unittest import TestCase, skipIf
except ImportError:
from unittest2 import TestCase, skipIf
|
"""
raven.utils.compat
~~~~~~~~~~~~~~~~~~~~~~~~~~
:copyright: (c) 2010-2012 by the Sentry Team, see AUTHORS for more details.
:license: BSD, see LICENSE for more details.
"""
try:
from urllib.error import HTTPError
except ImportError:
from urllib2 import HTTPError
try:
from urllib.request import Request, urlopen
except ImportError:
from urllib2 import Request, urlopen
try:
from urllib import quote as urllib_quote
except ImportError:
from urllib.parse import quote as urllib_quote
try:
from queue import Queue
except ImportError:
from Queue import Queue
try:
import urlparse as _urlparse
except ImportError:
from urllib import parse as _urlparse
try:
from unittest2 import TestCase, skipIf
except ImportError:
from unittest import TestCase, skipIf
|
Use unittest2 if available and gracefully falls back to unittest.
|
Use unittest2 if available and gracefully falls back to unittest.
|
Python
|
bsd-3-clause
|
getsentry/raven-python,lepture/raven-python,recht/raven-python,nikolas/raven-python,nikolas/raven-python,inspirehep/raven-python,dbravender/raven-python,smarkets/raven-python,percipient/raven-python,Photonomie/raven-python,ronaldevers/raven-python,danriti/raven-python,someonehan/raven-python,smarkets/raven-python,Photonomie/raven-python,recht/raven-python,icereval/raven-python,akalipetis/raven-python,jbarbuto/raven-python,inspirehep/raven-python,danriti/raven-python,jmagnusson/raven-python,icereval/raven-python,hzy/raven-python,dbravender/raven-python,percipient/raven-python,jbarbuto/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,akalipetis/raven-python,ewdurbin/raven-python,recht/raven-python,jmp0xf/raven-python,akheron/raven-python,someonehan/raven-python,someonehan/raven-python,lepture/raven-python,ewdurbin/raven-python,johansteffner/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,nikolas/raven-python,hzy/raven-python,arthurlogilab/raven-python,nikolas/raven-python,ronaldevers/raven-python,danriti/raven-python,arthurlogilab/raven-python,Photonomie/raven-python,jbarbuto/raven-python,arthurlogilab/raven-python,getsentry/raven-python,arthurlogilab/raven-python,inspirehep/raven-python,smarkets/raven-python,icereval/raven-python,akheron/raven-python,lepture/raven-python,icereval/raven-python,jmagnusson/raven-python,jmagnusson/raven-python,akheron/raven-python,jmp0xf/raven-python,ronaldevers/raven-python,smarkets/raven-python,ewdurbin/raven-python,johansteffner/raven-python,inspirehep/raven-python,percipient/raven-python,jbarbuto/raven-python,getsentry/raven-python,johansteffner/raven-python,jmp0xf/raven-python,akalipetis/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,dbravender/raven-python,hzy/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python
|
---
+++
@@ -37,7 +37,7 @@
try:
+ from unittest2 import TestCase, skipIf
+except ImportError:
from unittest import TestCase, skipIf
-except ImportError:
- from unittest2 import TestCase, skipIf
|
cbb0ed5ed66571feba22413472a5fe1a20824dbd
|
shared_export.py
|
shared_export.py
|
import bpy
def find_seqs(scene, select_marker):
sequences = {}
sequence_flags = {}
for marker in scene.timeline_markers:
if ":" not in marker.name or (select_marker and not marker.select):
continue
name, what = marker.name.rsplit(":", 1)
if name not in sequences:
sequences[name] = {}
if what in sequences[name]:
print("Warning: Got duplicate '{}' marker for sequence '{}' at frame {} (first was at frame {}), ignoring".format(what, name, marker.frame, sequences[name][what].frame))
continue
sequences[name][what] = marker
if "Sequences" in bpy.data.texts:
for line in bpy.data.texts["Sequences"].as_string().split("\n"):
line = line.strip()
if not line:
continue
if ":" not in line:
print("Invalid line in 'Sequences':", line)
continue
name, flags = line.split(":", 1)
if flags.lstrip():
flags = tuple(map(lambda f: f.strip(), flags.split(",")))
else:
flags = ()
sequence_flags[name] = flags
return sequences, sequence_flags
|
import bpy
def find_seqs(scene, select_marker):
sequences = {}
sequence_flags = {}
for marker in scene.timeline_markers:
if ":" not in marker.name or (select_marker and not marker.select):
continue
name, what = marker.name.rsplit(":", 1)
what = what.lower()
if name not in sequences:
sequences[name] = {}
if what in sequences[name]:
print("Warning: Got duplicate '{}' marker for sequence '{}' at frame {} (first was at frame {}), ignoring".format(what, name, marker.frame, sequences[name][what].frame))
continue
sequences[name][what] = marker
if "Sequences" in bpy.data.texts:
for line in bpy.data.texts["Sequences"].as_string().split("\n"):
line = line.strip()
if not line:
continue
if ":" not in line:
print("Invalid line in 'Sequences':", line)
continue
name, flags = line.split(":", 1)
if flags.lstrip():
flags = tuple(map(lambda f: f.strip(), flags.split(",")))
else:
flags = ()
sequence_flags[name] = flags
return sequences, sequence_flags
|
Make sequence marker types (:type) case insensitive
|
Make sequence marker types (:type) case insensitive
|
Python
|
mit
|
qoh/io_scene_dts,portify/io_scene_dts
|
---
+++
@@ -9,6 +9,7 @@
continue
name, what = marker.name.rsplit(":", 1)
+ what = what.lower()
if name not in sequences:
sequences[name] = {}
|
7ceba1f2b83628a2b89ffbdd30e435970e8c5e91
|
tests/test_kafka_streams.py
|
tests/test_kafka_streams.py
|
"""
Test the top-level Kafka Streams class
"""
import pytest
from winton_kafka_streams import kafka_config
from winton_kafka_streams.errors.kafka_streams_error import KafkaStreamsError
from winton_kafka_streams.kafka_streams import KafkaStreams
from winton_kafka_streams.processor.processor import BaseProcessor
from winton_kafka_streams.processor.topology import TopologyBuilder
class MyTestProcessor(BaseProcessor):
pass
def test_Given_StreamAlreadyStarted_When_CallStartAgain_Then_RaiseError():
kafka_config.NUM_STREAM_THREADS = 0
topology_builder = TopologyBuilder()
topology_builder.source('my-source', ['my-input-topic-1'])
topology_builder.processor('my-processor', MyTestProcessor, 'my-source')
topology_builder.sink('my-sink', 'my-output-topic-1', 'my-processor')
topology = topology_builder.build()
kafka_streams = KafkaStreams(topology, kafka_config)
kafka_streams.start()
with pytest.raises(KafkaStreamsError, message='KafkaStreams already started.'):
kafka_streams.start()
|
"""
Test the top-level Kafka Streams class
"""
import pytest
from winton_kafka_streams import kafka_config
from winton_kafka_streams.errors.kafka_streams_error import KafkaStreamsError
from winton_kafka_streams.kafka_streams import KafkaStreams
from winton_kafka_streams.processor.processor import BaseProcessor
from winton_kafka_streams.processor.topology import TopologyBuilder
class MyTestProcessor(BaseProcessor):
pass
def test__given__stream_already_started__when__call_start_again__then__raise_error():
kafka_config.NUM_STREAM_THREADS = 0
topology_builder = TopologyBuilder()
topology_builder.source('my-source', ['my-input-topic-1'])
topology_builder.processor('my-processor', MyTestProcessor, 'my-source')
topology_builder.sink('my-sink', 'my-output-topic-1', 'my-processor')
topology = topology_builder.build()
kafka_streams = KafkaStreams(topology, kafka_config)
kafka_streams.start()
with pytest.raises(KafkaStreamsError, message='KafkaStreams already started.'):
kafka_streams.start()
|
Use more Pythonic name for test.
|
Use more Pythonic name for test.
|
Python
|
apache-2.0
|
wintoncode/winton-kafka-streams
|
---
+++
@@ -16,7 +16,7 @@
pass
-def test_Given_StreamAlreadyStarted_When_CallStartAgain_Then_RaiseError():
+def test__given__stream_already_started__when__call_start_again__then__raise_error():
kafka_config.NUM_STREAM_THREADS = 0
topology_builder = TopologyBuilder()
|
3fb1800548ad421520bf3f2845aad4f51f6f5839
|
rapidsms_multimodem/tests/__init__.py
|
rapidsms_multimodem/tests/__init__.py
|
from test_utils import * # noqa
from test_views import * # noqa
|
from test_outgoing import * # noqa
from test_utils import * # noqa
from test_views import * # noqa
|
Add import for older versions of Django
|
Add import for older versions of Django
|
Python
|
bsd-3-clause
|
caktus/rapidsms-multimodem
|
---
+++
@@ -1,2 +1,3 @@
+from test_outgoing import * # noqa
from test_utils import * # noqa
from test_views import * # noqa
|
075b11aa830c9a5961e9ee63e42484192990f7d3
|
tools/misc/python/test-data-in-out.py
|
tools/misc/python/test-data-in-out.py
|
# TOOL test-data-in-out.py: "Test data input and output in Python" (Data input output test.)
# INPUT input TYPE GENERIC
# OUTPUT output
# OUTPUT OPTIONAL missing_output.txt
import shutil
shutil.copyfile('input', 'output')
|
# TOOL test-data-in-out.py: "Test data input and output in Python" (Data input output test.)
# INPUT input TYPE GENERIC
# OUTPUT output
# PARAMETER delay: Delay TYPE INTEGER FROM 0 TO 1000 DEFAULT 1 (Delay in seconds)
import shutil
import time
time.sleep(delay)
shutil.copyfile('input', 'output')
|
Add delay to input-output test
|
Add delay to input-output test
|
Python
|
mit
|
chipster/chipster-tools,chipster/chipster-tools,chipster/chipster-tools,chipster/chipster-tools
|
---
+++
@@ -1,9 +1,11 @@
# TOOL test-data-in-out.py: "Test data input and output in Python" (Data input output test.)
# INPUT input TYPE GENERIC
# OUTPUT output
-# OUTPUT OPTIONAL missing_output.txt
+# PARAMETER delay: Delay TYPE INTEGER FROM 0 TO 1000 DEFAULT 1 (Delay in seconds)
import shutil
+import time
+time.sleep(delay)
shutil.copyfile('input', 'output')
|
943d2648c17facb9dbfd4f26d335beef341e9c49
|
fabfile.py
|
fabfile.py
|
from fabric.api import local
__author__ = 'derek'
def deploy(version):
local('python runtests.py')
local("git tag -a %s -m %s" % (version, version))
local('python setup.py sdist upload')
|
from fabric.api import local
__author__ = 'derek'
def deploy(version):
local('python runtests.py')
local("git tag -a %s -m %s" % (version, version))
local("git push origin --tags")
local('python setup.py sdist upload')
|
Make sure to push the tags
|
Make sure to push the tags
|
Python
|
mit
|
winfieldco/django-mail-queue,Goury/django-mail-queue,Goury/django-mail-queue,dstegelman/django-mail-queue,dstegelman/django-mail-queue,styrmis/django-mail-queue
|
---
+++
@@ -6,4 +6,5 @@
def deploy(version):
local('python runtests.py')
local("git tag -a %s -m %s" % (version, version))
+ local("git push origin --tags")
local('python setup.py sdist upload')
|
3e23d60857461b7806f3616cf41b2cd9c812fa7b
|
fabfile.py
|
fabfile.py
|
from fabric.api import cd, sudo, env
import os
expected_vars = [
'PROJECT',
]
for var in expected_vars:
if var not in os.environ:
raise Exception('Please specify %s environment variable' % (
var,))
PROJECT = os.environ['PROJECT']
USER = os.environ.get('USER', 'jmbo')
env.path = os.path.join('var', 'praekelt', PROJECT)
def restart():
sudo('/etc/init.d/nginx restart')
sudo('supervisorctl reload')
def deploy():
with cd(env.path):
sudo('git pull', user=USER)
sudo('ve/bin/python manage.py syncdb --migrate --noinput',
user=USER)
sudo('ve/bin/python manage.py collectstatic --noinput',
user=USER)
def install_packages(force=False):
with cd(env.path):
sudo('ve/bin/pip install %s -r requirements.pip' % (
'--upgrade' if force else '',), user=USER)
|
from fabric.api import cd, sudo, env
import os
expected_vars = [
'PROJECT',
]
for var in expected_vars:
if var not in os.environ:
raise Exception('Please specify %s environment variable' % (
var,))
PROJECT = os.environ['PROJECT']
USER = os.environ.get('USER', 'jmbo')
env.path = os.path.join('/', 'var', 'praekelt', PROJECT)
def restart():
sudo('/etc/init.d/nginx restart')
sudo('supervisorctl reload')
def deploy():
with cd(env.path):
sudo('git pull', user=USER)
sudo('ve/bin/python manage.py syncdb --migrate --noinput',
user=USER)
sudo('ve/bin/python manage.py collectstatic --noinput',
user=USER)
def install_packages(force=False):
with cd(env.path):
sudo('ve/bin/pip install %s -r requirements.pip' % (
'--upgrade' if force else '',), user=USER)
|
Make env.path to /var/praekelt/<PROJECT> an absolute path.
|
Make env.path to /var/praekelt/<PROJECT> an absolute path.
|
Python
|
bsd-3-clause
|
praekelt/go-rts-zambia
|
---
+++
@@ -13,7 +13,7 @@
PROJECT = os.environ['PROJECT']
USER = os.environ.get('USER', 'jmbo')
-env.path = os.path.join('var', 'praekelt', PROJECT)
+env.path = os.path.join('/', 'var', 'praekelt', PROJECT)
def restart():
|
e660d8e05e54adbd0ea199a02cc188dc7007089a
|
fabfile.py
|
fabfile.py
|
from fabric.api import cd, sudo, env
import os
expected_vars = [
'PROJECT',
]
for var in expected_vars:
if var not in os.environ:
raise Exception('Please specify %s environment variable' % (
var,))
PROJECT = os.environ['PROJECT']
USER = os.environ.get('USER', 'jmbo')
env.path = os.path.join('/', 'var', 'praekelt', PROJECT)
def restart():
sudo('/etc/init.d/nginx restart')
sudo('supervisorctl reload')
def deploy():
with cd(env.path):
sudo('git pull', user=USER)
sudo('ve/bin/python manage.py syncdb --migrate --noinput',
user=USER)
sudo('ve/bin/python manage.py collectstatic --noinput',
user=USER)
def install_packages(force=False):
with cd(env.path):
sudo('ve/bin/pip install %s -r requirements.pip' % (
'--upgrade' if force else '',), user=USER)
|
from fabric.api import cd, sudo, env
import os
PROJECT = os.environ.get('PROJECT', 'go-rts-zambia')
DEPLOY_USER = os.environ.get('DEPLOY_USER', 'jmbo')
env.path = os.path.join('/', 'var', 'praekelt', PROJECT)
def restart():
sudo('/etc/init.d/nginx restart')
sudo('supervisorctl reload')
def deploy():
with cd(env.path):
sudo('git pull', user=DEPLOY_USER)
sudo('ve/bin/python manage.py syncdb --migrate --noinput',
user=DEPLOY_USER)
sudo('ve/bin/python manage.py collectstatic --noinput',
user=DEPLOY_USER)
def install_packages(force=False):
with cd(env.path):
sudo('ve/bin/pip install %s -r requirements.pip' % (
'--upgrade' if force else '',), user=DEPLOY_USER)
|
Fix USER in fabric file (the deploy user and ssh user aren't necessarily the same). Set default value for PROJECT (this is the go-rts-zambia repo after all).
|
Fix USER in fabric file (the deploy user and ssh user aren't necessarily the same). Set default value for PROJECT (this is the go-rts-zambia repo after all).
|
Python
|
bsd-3-clause
|
praekelt/go-rts-zambia
|
---
+++
@@ -1,17 +1,8 @@
from fabric.api import cd, sudo, env
import os
-expected_vars = [
- 'PROJECT',
-]
-
-for var in expected_vars:
- if var not in os.environ:
- raise Exception('Please specify %s environment variable' % (
- var,))
-
-PROJECT = os.environ['PROJECT']
-USER = os.environ.get('USER', 'jmbo')
+PROJECT = os.environ.get('PROJECT', 'go-rts-zambia')
+DEPLOY_USER = os.environ.get('DEPLOY_USER', 'jmbo')
env.path = os.path.join('/', 'var', 'praekelt', PROJECT)
@@ -23,14 +14,14 @@
def deploy():
with cd(env.path):
- sudo('git pull', user=USER)
+ sudo('git pull', user=DEPLOY_USER)
sudo('ve/bin/python manage.py syncdb --migrate --noinput',
- user=USER)
+ user=DEPLOY_USER)
sudo('ve/bin/python manage.py collectstatic --noinput',
- user=USER)
+ user=DEPLOY_USER)
def install_packages(force=False):
with cd(env.path):
sudo('ve/bin/pip install %s -r requirements.pip' % (
- '--upgrade' if force else '',), user=USER)
+ '--upgrade' if force else '',), user=DEPLOY_USER)
|
04e8206a8610666c6027fc0f4be5e786e4bd5513
|
fabfile.py
|
fabfile.py
|
set(
fab_hosts = ['startthedark.com'],
fab_user = 'startthedark',
)
def unlink_nginx():
'Un-link nginx rules for startthedark.'
sudo('rm -f /etc/nginx/sites-enabled/startthedark.com')
sudo('/etc/init.d/nginx reload')
def link_nginx():
'Link nginx rules for startthedark'
sudo('ln -s /etc/nginx/sites-available/startthedark.com /etc/nginx/sites-enabled/startthedark.com')
sudo('/etc/init.d/nginx reload')
def deploy():
'Deploy startthedark.'
local('bash make_prod_css.sh')
set(fab_fail = 'ignore')
local('git commit -a -m "Rebuilt Prod CSS For Commit"')
local('git push origin master')
set(fab_fail = 'abort')
run('cd /var/www/startthedark.com/startthedark; git pull;')
run('cd /var/www/startthedark.com/startthedark; /usr/bin/python manage.py syncdb')
sudo('/etc/init.d/apache2 reload')
|
set(
fab_hosts = ['startthedark.com'],
fab_user = 'startthedark',
)
def unlink_nginx():
'Un-link nginx rules for startthedark.'
sudo('rm -f /etc/nginx/sites-enabled/startthedark.com')
sudo('/etc/init.d/nginx reload')
def link_nginx():
'Link nginx rules for startthedark'
sudo('ln -s /etc/nginx/sites-available/startthedark.com /etc/nginx/sites-enabled/startthedark.com')
sudo('/etc/init.d/nginx reload')
def rebuild_prod_css():
local('bash make_prod_css.sh')
local('git commit -a -m "Rebuilt Prod CSS For Commit"')
local('git push origin master')
def deploy():
'Deploy startthedark.'
run('cd /var/www/startthedark.com/startthedark; git pull;')
run('cd /var/www/startthedark.com/startthedark; /usr/bin/python manage.py syncdb')
sudo('/etc/init.d/apache2 reload')
|
Split out the css rebuilding into its own fab method.
|
Split out the css rebuilding into its own fab method.
|
Python
|
bsd-3-clause
|
mvayngrib/startthedark,mvayngrib/startthedark,ericflo/startthedark,ericflo/startthedark,mvayngrib/startthedark,ericflo/startthedark
|
---
+++
@@ -13,13 +13,13 @@
sudo('ln -s /etc/nginx/sites-available/startthedark.com /etc/nginx/sites-enabled/startthedark.com')
sudo('/etc/init.d/nginx reload')
+def rebuild_prod_css():
+ local('bash make_prod_css.sh')
+ local('git commit -a -m "Rebuilt Prod CSS For Commit"')
+ local('git push origin master')
+
def deploy():
'Deploy startthedark.'
- local('bash make_prod_css.sh')
- set(fab_fail = 'ignore')
- local('git commit -a -m "Rebuilt Prod CSS For Commit"')
- local('git push origin master')
- set(fab_fail = 'abort')
run('cd /var/www/startthedark.com/startthedark; git pull;')
run('cd /var/www/startthedark.com/startthedark; /usr/bin/python manage.py syncdb')
sudo('/etc/init.d/apache2 reload')
|
b8b72be48328ba0bc6e946a4ecf15c00f5f8b3b6
|
director/director/config/dev.py
|
director/director/config/dev.py
|
import imp
import os
import sys
import traceback
from os.path import dirname
from configurations import values
from .common import Common, BASE_DIR, external_keys
SECRETS_DIR = os.path.join(dirname(BASE_DIR), "secrets")
class Dev(Common):
DEBUG = True
INSTALLED_APPS = Common.INSTALLED_APPS + [
'django_extensions'
]
EMAIL_BACKEND = 'django.core.mail.backends.console.EmailBackend'
@classmethod
def setup(cls):
super(Dev, cls).setup()
filename = "director_dev_secrets.py"
try:
dev_secrets = imp.load_source(
"dev_secrets", os.path.join(SECRETS_DIR, filename))
except ImportError as e:
print("Could not import %s: %s" % (filename, e))
return
for key in external_keys:
if not hasattr(cls, key) and hasattr(dev_secrets, key):
setattr(cls, key, getattr(dev_secrets, key))
|
import imp
import os
import sys
import traceback
from os.path import dirname
from configurations import values
from .common import Common, BASE_DIR, external_keys
SECRETS_DIR = os.path.join(dirname(BASE_DIR), "secrets")
class Dev(Common):
DEBUG = True
INSTALLED_APPS = Common.INSTALLED_APPS + [
'django_extensions'
]
EMAIL_BACKEND = 'django.core.mail.backends.console.EmailBackend'
@classmethod
def setup(cls):
super(Dev, cls).setup()
filename = "director_dev_secrets.py"
path = os.path.join(SECRETS_DIR, filename)
try:
dev_secrets = imp.load_source("dev_secrets", path)
except ImportError as e:
print("Could not import %s: %s" % (filename, e))
return
except FileNotFoundError as e:
print("File %s not found" % path)
return
for key in external_keys:
if not hasattr(cls, key) and hasattr(dev_secrets, key):
setattr(cls, key, getattr(dev_secrets, key))
|
Print message when secrets file is missing
|
Print message when secrets file is missing
|
Python
|
apache-2.0
|
stencila/hub,stencila/hub,stencila/hub,stencila/hub,stencila/hub
|
---
+++
@@ -24,11 +24,14 @@
def setup(cls):
super(Dev, cls).setup()
filename = "director_dev_secrets.py"
+ path = os.path.join(SECRETS_DIR, filename)
try:
- dev_secrets = imp.load_source(
- "dev_secrets", os.path.join(SECRETS_DIR, filename))
+ dev_secrets = imp.load_source("dev_secrets", path)
except ImportError as e:
print("Could not import %s: %s" % (filename, e))
+ return
+ except FileNotFoundError as e:
+ print("File %s not found" % path)
return
for key in external_keys:
|
9e90fe6523e7d61d6b94f9bb37a2dbf711cd3b83
|
example/__init__.py
|
example/__init__.py
|
"""An example application using Confit for configuration."""
from __future__ import print_function
from __future__ import unicode_literals
import confit
import argparse
config = confit.LazyConfig('ConfitExample', __name__)
def main():
parser = argparse.ArgumentParser(description='example Confit program')
parser.add_argument('--library', '-l', dest='library', metavar='LIBPATH',
help='library database file')
parser.add_argument('--directory', '-d', dest='directory',
metavar='DIRECTORY',
help='destination music directory')
parser.add_argument('--verbose', '-v', dest='verbose', action='store_true',
help='print debugging messages')
args = parser.parse_args()
config.set_args(args)
print('configuration directory is', config.config_dir())
# Use a boolean flag and the transient overlay.
if config['verbose']:
print('verbose mode')
config['log']['level'] = 2
else:
config['log']['level'] = 0
print('logging level is', config['log']['level'].validate(int))
# Some validated/converted values.
print('directory is', config['directory'].as_filename())
print('library is', config['library'].as_filename())
|
"""An example application using Confit for configuration."""
from __future__ import print_function
from __future__ import unicode_literals
import confit
import argparse
config = confit.LazyConfig('ConfitExample', __name__)
def main():
parser = argparse.ArgumentParser(description='example Confit program')
parser.add_argument('--library', '-l', dest='library', metavar='LIBPATH',
help='library database file')
parser.add_argument('--directory', '-d', dest='directory',
metavar='DIRECTORY',
help='destination music directory')
parser.add_argument('--verbose', '-v', dest='verbose', action='store_true',
help='print debugging messages')
args = parser.parse_args()
config.set_args(args)
print('configuration directory is', config.config_dir())
# Use a boolean flag and the transient overlay.
if config['verbose']:
print('verbose mode')
config['log']['level'] = 2
else:
config['log']['level'] = 0
print('logging level is', config['log']['level'].get(int))
# Some validated/converted values.
print('directory is', config['directory'].as_filename())
print('library is', config['library'].as_filename())
|
Fix example to use get
|
Fix example to use get
|
Python
|
mit
|
sampsyo/confit,sampsyo/confuse
|
---
+++
@@ -29,7 +29,7 @@
config['log']['level'] = 2
else:
config['log']['level'] = 0
- print('logging level is', config['log']['level'].validate(int))
+ print('logging level is', config['log']['level'].get(int))
# Some validated/converted values.
print('directory is', config['directory'].as_filename())
|
b575099c0d1f23916038172d46852a264a5f5a95
|
bluebottle/utils/staticfiles_finders.py
|
bluebottle/utils/staticfiles_finders.py
|
from django.utils._os import safe_join
import os
from django.conf import settings
from django.contrib.staticfiles.finders import FileSystemFinder
from bluebottle.clients.models import Client
class TenantStaticFilesFinder(FileSystemFinder):
def find(self, path, all=False):
"""
Looks for files in the client static directories.
static/assets/greatbarier/images/logo.jpg
will translate to
MULTI_TENANT_DIR/greatbarier/static/images/logo.jpg
"""
tenants = Client.objects.all()
tenant_dir = getattr(settings, 'MULTI_TENANT_DIR', None)
if not tenant_dir:
return
for tenant in tenants:
if "{0}/".format(tenant.client_name) in path:
tenant_path = path.replace('{0}/'.format(tenant.client_name),
'{0}/static/'.format(tenant.client_name))
local_path = safe_join(tenant_dir, tenant_path)
if os.path.exists(local_path):
return local_path
return
|
from django.utils._os import safe_join
import os
from django.conf import settings
from django.contrib.staticfiles.finders import FileSystemFinder
from bluebottle.clients.models import Client
class TenantStaticFilesFinder(FileSystemFinder):
def find(self, path, all=False):
"""
Looks for files in the client static directories.
static/assets/greatbarier/images/logo.jpg
will translate to
MULTI_TENANT_DIR/greatbarier/static/images/logo.jpg
"""
tenants = Client.objects.all()
tenant_dir = getattr(settings, 'MULTI_TENANT_DIR', None)
if not tenant_dir:
return []
for tenant in tenants:
if "{0}/".format(tenant.client_name) in path:
tenant_path = path.replace('{0}/'.format(tenant.client_name),
'{0}/static/'.format(tenant.client_name))
local_path = safe_join(tenant_dir, tenant_path)
if os.path.exists(local_path):
if all:
return [local_path]
return local_path
return []
|
Fix static files finder errors
|
Fix static files finder errors
|
Python
|
bsd-3-clause
|
jfterpstra/bluebottle,onepercentclub/bluebottle,jfterpstra/bluebottle,onepercentclub/bluebottle,jfterpstra/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,jfterpstra/bluebottle
|
---
+++
@@ -17,8 +17,9 @@
"""
tenants = Client.objects.all()
tenant_dir = getattr(settings, 'MULTI_TENANT_DIR', None)
+
if not tenant_dir:
- return
+ return []
for tenant in tenants:
if "{0}/".format(tenant.client_name) in path:
@@ -26,5 +27,7 @@
'{0}/static/'.format(tenant.client_name))
local_path = safe_join(tenant_dir, tenant_path)
if os.path.exists(local_path):
+ if all:
+ return [local_path]
return local_path
- return
+ return []
|
d52d8ce18745d5ec0e722340cf09735938c8a0c0
|
src/BaseUtils.py
|
src/BaseUtils.py
|
'''
Base NLP utilities
'''
def get_words(sentence):
''' Return all the words found in a sentence.
Ignore whitespace and all punctuation
>>> get_words('a most interesting piece')
>>> ['a', 'most', 'interesting', 'piece']
>>> get_words('a, most$ **interesting piece')
>>> ['a', 'most', 'interesting', 'piece']
'''
clean_sentence = ''.join([char for char in sentence if ord(char) in
xrange(97, 123) or ord(char) in xrange(75, 91)
or ord(char) == 32])
segments = clean_sentence.split(' ')
words = [word for word in segments if not word == '']
return words
|
'''
Base NLP utilities
'''
def get_words(sentence):
''' Return all the words found in a sentence.
Ignore whitespace and all punctuation
>>> get_words('a most interesting piece')
>>> ['a', 'most', 'interesting', 'piece']
>>> get_words('a, most$ **interesting piece')
>>> ['a', 'most', 'interesting', 'piece']
'''
clean_sentence = ''.join([char for char in sentence if char.isalpha()
or char.isspace()])
segments = clean_sentence.split(' ')
words = [word for word in segments if not word == '']
return words
|
Replace ASCII checking chars and space with library methods
|
Replace ASCII checking chars and space with library methods
|
Python
|
bsd-2-clause
|
ambidextrousTx/RNLTK
|
---
+++
@@ -11,9 +11,8 @@
>>> get_words('a, most$ **interesting piece')
>>> ['a', 'most', 'interesting', 'piece']
'''
- clean_sentence = ''.join([char for char in sentence if ord(char) in
- xrange(97, 123) or ord(char) in xrange(75, 91)
- or ord(char) == 32])
+ clean_sentence = ''.join([char for char in sentence if char.isalpha()
+ or char.isspace()])
segments = clean_sentence.split(' ')
words = [word for word in segments if not word == '']
return words
|
716d967971d9ea23ab54d327231ba873b681a7c7
|
isserviceup/services/models/service.py
|
isserviceup/services/models/service.py
|
from enum import Enum
class Status(Enum):
ok = 1 # green
maintenance = 2 # blue
minor = 3 # yellow
major = 4 # orange
critical = 5 # red
unavailable = 6 # gray
class Service(object):
@property
def id(self):
return self.__class__.__name__
@property
def status_url(self):
raise NotImplemented()
@property
def name(self):
raise NotImplemented()
def get_status(self):
raise NotImplemented()
|
from enum import Enum
class Status(Enum):
ok = 1 # green
maintenance = 2 # blue
minor = 3 # yellow
major = 4 # orange
critical = 5 # red
unavailable = 6 # gray
class Service(object):
@property
def id(self):
return self.__class__.__name__
@property
def status_url(self):
raise NotImplemented()
@property
def icon_url(self):
raise NotImplemented()
@property
def name(self):
raise NotImplemented()
def get_status(self):
raise NotImplemented()
|
Add icon_url as abstract property
|
Add icon_url as abstract property
|
Python
|
apache-2.0
|
marcopaz/is-service-up,marcopaz/is-service-up,marcopaz/is-service-up
|
---
+++
@@ -21,6 +21,10 @@
raise NotImplemented()
@property
+ def icon_url(self):
+ raise NotImplemented()
+
+ @property
def name(self):
raise NotImplemented()
|
5233e6d7f7d4f494f62576206ede87d13e8f760d
|
calexicon/calendars/tests/test_other.py
|
calexicon/calendars/tests/test_other.py
|
from datetime import date as vanilla_date
from calendar_testing import CalendarTest
from calexicon.calendars.other import JulianDayNumber
class TestJulianDayNumber(CalendarTest):
def setUp(self):
self.calendar = JulianDayNumber()
def test_make_date(self):
vd = vanilla_date(2010, 8, 1)
d = self.calendar.from_date(vd)
self.assertIsNotNone(d)
def test_first_date(self):
vd = vanilla_date(1, 1, 1)
d = self.calendar.from_date(vd)
self.assertEqual(str(d), 'Day 1721423 (Julian Day Number)')
def compare_date_and_number(self, year, month, day, number):
vd = vanilla_date(year, month, day)
d = self.calendar.from_date(vd)
self.assertEqual(d.native_representation(), {'day_number': number})
def test_other_date(self):
self.compare_date_and_number(2013, 1, 1, 2456293)
|
from datetime import date as vanilla_date
from calendar_testing import CalendarTest
from calexicon.calendars.other import JulianDayNumber
class TestJulianDayNumber(CalendarTest):
def setUp(self):
self.calendar = JulianDayNumber()
def test_make_date(self):
vd = vanilla_date(2010, 8, 1)
d = self.calendar.from_date(vd)
self.assertIsNotNone(d)
def test_first_date(self):
vd = vanilla_date(1, 1, 1)
d = self.calendar.from_date(vd)
self.assertEqual(str(d), 'Day 1721423 (Julian Day Number)')
def compare_date_and_number(self, year, month, day, number):
vd = vanilla_date(year, month, day)
d = self.calendar.from_date(vd)
self.assertEqual(d.native_representation(), {'day_number': number})
def test_every_400_years(self):
days_in_400_years = 400 * 365 + 97
for i in range(25):
self.compare_date_and_number(1 + 400 * i, 1, 1, 1721423 + days_in_400_years * i)
def test_another_date(self):
self.compare_date_and_number(2013, 1, 1, 2456293)
|
Add a test to check the right number of days in 400 year cycles.
|
Add a test to check the right number of days in 400 year cycles.
|
Python
|
apache-2.0
|
jwg4/calexicon,jwg4/qual
|
---
+++
@@ -23,5 +23,10 @@
d = self.calendar.from_date(vd)
self.assertEqual(d.native_representation(), {'day_number': number})
- def test_other_date(self):
+ def test_every_400_years(self):
+ days_in_400_years = 400 * 365 + 97
+ for i in range(25):
+ self.compare_date_and_number(1 + 400 * i, 1, 1, 1721423 + days_in_400_years * i)
+
+ def test_another_date(self):
self.compare_date_and_number(2013, 1, 1, 2456293)
|
03fb3e68f9ec9432a25c60bda06fcd49d604befc
|
src/__init__.py
|
src/__init__.py
|
from ..override_audit import reload
reload("src", ["core", "events", "contexts", "browse", "settings_proxy"])
reload("src.commands")
from . import core
from .core import *
from .events import *
from .contexts import *
from .settings_proxy import *
from .commands import *
__all__ = [
# core
"core",
# browse
"browse",
# settings_proxy
"OverrideAuditOpenFileCommand",
"OverrideAuditEditSettingsCommand",
# events/contexts
"OverrideAuditEventListener",
"OverrideAuditContextListener",
# commands/*
"OverrideAuditPackageReportCommand",
"OverrideAuditOverrideReportCommand",
"OverrideAuditDiffReportCommand",
"OverrideAuditRefreshReportCommand",
"OverrideAuditToggleOverrideCommand",
"OverrideAuditCreateOverrideCommand",
"OverrideAuditDiffOverrideCommand",
"OverrideAuditRevertOverrideCommand",
"OverrideAuditDiffExternallyCommand",
"OverrideAuditEditOverrideCommand",
"OverrideAuditDeleteOverrideCommand",
"OverrideAuditFreshenOverrideCommand",
"OverrideAuditDiffPackageCommand",
"OverrideAuditFreshenPackageCommand",
"OverrideAuditDiffSingleCommand",
"OverrideAuditModifyMarkCommand"
]
|
from ..override_audit import reload
reload("src", ["core", "events", "contexts", "browse", "settings_proxy"])
reload("src.commands")
from . import core
from .core import *
from .events import *
from .contexts import *
from .settings_proxy import *
from .commands import *
__all__ = [
# core
"core",
# browse
"browse",
# settings_proxy
"OverrideAuditOpenFileCommand",
"OverrideAuditEditSettingsCommand",
# events/contexts
"OverrideAuditEventListener",
"CreateOverrideEventListener",
"OverrideAuditContextListener",
# commands/*
"OverrideAuditPackageReportCommand",
"OverrideAuditOverrideReportCommand",
"OverrideAuditDiffReportCommand",
"OverrideAuditRefreshReportCommand",
"OverrideAuditToggleOverrideCommand",
"OverrideAuditCreateOverrideCommand",
"OverrideAuditDiffOverrideCommand",
"OverrideAuditRevertOverrideCommand",
"OverrideAuditDiffExternallyCommand",
"OverrideAuditEditOverrideCommand",
"OverrideAuditDeleteOverrideCommand",
"OverrideAuditFreshenOverrideCommand",
"OverrideAuditDiffPackageCommand",
"OverrideAuditFreshenPackageCommand",
"OverrideAuditDiffSingleCommand",
"OverrideAuditModifyMarkCommand"
]
|
Add missing event listener for new overrides
|
Add missing event listener for new overrides
While rolling the test code for creating overrides into the base code,
we remembered to make sure that we put the event handler used to handle
override creation in place but forgot to export them to to the base
plugin so that it would actually do something.
|
Python
|
mit
|
OdatNurd/OverrideAudit
|
---
+++
@@ -23,6 +23,7 @@
# events/contexts
"OverrideAuditEventListener",
+ "CreateOverrideEventListener",
"OverrideAuditContextListener",
# commands/*
|
f46770697d668e31518ada41d31fdb59a84f3cf6
|
kokki/cookbooks/aws/recipes/default.py
|
kokki/cookbooks/aws/recipes/default.py
|
from kokki import *
Package("python-boto")
# Mount volumes and format is necessary
for vol in env.config.aws.volumes:
env.cookbooks.aws.EBSVolume(vol['volume_id'],
availability_zone = env.config.aws.availability_zone,
device = vol['device'],
action = "attach")
if vol.get('fstype'):
if vol['fstype'] == "xfs":
Package("xfsprogs")
Execute("mkfs.%(fstype)s -f %(device)s" % vol,
not_if = """if [ "`file -s %(device)s`" = "%(device)s: data" ]; then exit 1; fi""" % vol)
if vol.get('mount_point'):
Mount(vol['mount_point'],
device = vol['device'],
fstype = vol.get('fstype'),
options = vol.get('fsoptions', ["noatime"]),
action = ["mount", "enable"])
|
import os
from kokki import *
# Package("python-boto")
Execute("pip install git+http://github.com/boto/boto.git#egg=boto",
not_if = 'python -c "import boto"')
Execute("mv /usr/lib/pymodules/python2.6/boto /tmp/boto.orig",
only_if = os.path.exists("/usr/lib/pymodules/python2.6/boto"))
# Mount volumes and format is necessary
for vol in env.config.aws.volumes:
env.cookbooks.aws.EBSVolume(vol['volume_id'],
availability_zone = env.config.aws.availability_zone,
device = vol['device'],
action = "attach")
if vol.get('fstype'):
if vol['fstype'] == "xfs":
Package("xfsprogs")
Execute("mkfs.%(fstype)s -f %(device)s" % vol,
not_if = """if [ "`file -s %(device)s`" = "%(device)s: data" ]; then exit 1; fi""" % vol)
if vol.get('mount_point'):
Mount(vol['mount_point'],
device = vol['device'],
fstype = vol.get('fstype'),
options = vol.get('fsoptions', ["noatime"]),
action = ["mount", "enable"])
|
Install github verison of boto in aws cookbook (for now)
|
Install github verison of boto in aws cookbook (for now)
|
Python
|
bsd-3-clause
|
samuel/kokki
|
---
+++
@@ -1,7 +1,12 @@
+import os
from kokki import *
-Package("python-boto")
+# Package("python-boto")
+Execute("pip install git+http://github.com/boto/boto.git#egg=boto",
+ not_if = 'python -c "import boto"')
+Execute("mv /usr/lib/pymodules/python2.6/boto /tmp/boto.orig",
+ only_if = os.path.exists("/usr/lib/pymodules/python2.6/boto"))
# Mount volumes and format is necessary
|
183f9455425fa63b6ca43c5d4fe650bcf2179da5
|
ironic/drivers/modules/storage/noop.py
|
ironic/drivers/modules/storage/noop.py
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from ironic.drivers import base
class NoopStorage(base.StorageInterface):
"""No-op Storage Interface."""
def validate(self, task):
pass
def get_properties(self):
return {}
def attach_volumes(self, task):
pass
def detach_volumes(self, task):
pass
def should_write_image(self, task):
return True
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from ironic.drivers import base
class NoopStorage(base.StorageInterface):
"""No-op Storage Interface."""
def validate(self, task):
pass
def get_properties(self):
return {}
def attach_volumes(self, task):
pass
def detach_volumes(self, task):
pass
def should_write_image(self, task):
return False
|
Update should_write_image to return False
|
Update should_write_image to return False
Update should_write_image to return False
|
Python
|
apache-2.0
|
SauloAislan/ironic,SauloAislan/ironic
|
---
+++
@@ -29,4 +29,4 @@
pass
def should_write_image(self, task):
- return True
+ return False
|
ae689c9de698daeaf8ab5275c384183cb665c903
|
neutron_classifier/common/constants.py
|
neutron_classifier/common/constants.py
|
# Copyright (c) 2015 Mirantis, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
CLASSIFIER_TYPES = ['ip_classifier', 'ipv4_classifier', 'ipv6_classifier',
'transport_classifier', 'ethernet_classifier',
'encapsulation_classifier', 'neutron_port_classifier']
# TODO(sc68cal) add more protocols`
PROTOCOLS = ['tcp', 'udp', 'icmp', 'icmpv6']
ENCAPSULATION_TYPES = ['vxlan', 'gre']
NEUTRON_SERVICES = ['neutron-fwaas', 'networking-sfc', 'security-group']
|
# Copyright (c) 2015 Mirantis, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
# TODO(sc68cal) add more protocols`
PROTOCOLS = ['tcp', 'udp', 'icmp', 'icmpv6']
ENCAPSULATION_TYPES = ['vxlan', 'gre']
NEUTRON_SERVICES = ['neutron-fwaas', 'networking-sfc', 'security-group']
|
Remove CLASSIFIER_TYPES constant - it was never used
|
Remove CLASSIFIER_TYPES constant - it was never used
Change-Id: Ia6ba4453f6bc9b9de0da1e83d2dc75147fb91882
|
Python
|
apache-2.0
|
openstack/neutron-classifier,openstack/neutron-classifier
|
---
+++
@@ -12,10 +12,6 @@
# License for the specific language governing permissions and limitations
# under the License.
-CLASSIFIER_TYPES = ['ip_classifier', 'ipv4_classifier', 'ipv6_classifier',
- 'transport_classifier', 'ethernet_classifier',
- 'encapsulation_classifier', 'neutron_port_classifier']
-
# TODO(sc68cal) add more protocols`
PROTOCOLS = ['tcp', 'udp', 'icmp', 'icmpv6']
|
15db774538b4fa18c0653fb741ba14c0373867c8
|
main/admin/forms.py
|
main/admin/forms.py
|
from django import forms
from main.models import Profile, LanProfile
class AdminLanProfileForm(forms.ModelForm):
class Meta:
model = LanProfile
fields = '__all__'
class AdminProfileForm(forms.ModelForm):
class Meta:
model = Profile
fields = '__all__'
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
instance = kwargs.get('instance', None)
if instance:
self.fields['grade'].choices += ((instance.grade, instance.grade), ('none', 'Ukendt'))
|
from django import forms
from main.models import Profile, LanProfile, GRADES
class AdminLanProfileForm(forms.ModelForm):
class Meta:
model = LanProfile
fields = '__all__'
class AdminProfileForm(forms.ModelForm):
class Meta:
model = Profile
fields = '__all__'
grade = forms.ChoiceField(GRADES, required=True, label='Klasse')
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
instance = kwargs.get('instance', None)
if instance:
self.fields['grade'].choices += ((instance.grade, instance.grade), ('none', 'Ukendt'))
|
Make that work with admin too
|
Make that work with admin too
|
Python
|
mit
|
bomjacob/htxaarhuslan,bomjacob/htxaarhuslan,bomjacob/htxaarhuslan
|
---
+++
@@ -1,6 +1,6 @@
from django import forms
-from main.models import Profile, LanProfile
+from main.models import Profile, LanProfile, GRADES
class AdminLanProfileForm(forms.ModelForm):
@@ -14,6 +14,8 @@
model = Profile
fields = '__all__'
+ grade = forms.ChoiceField(GRADES, required=True, label='Klasse')
+
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
|
c557058a7a7206167108535129bc0b160e4fe62b
|
nipype/testing/tests/test_utils.py
|
nipype/testing/tests/test_utils.py
|
# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*-
# vi: set ft=python sts=4 ts=4 sw=4 et:
"""Test testing utilities
"""
from nipype.testing.utils import TempFATFS
from nose.tools import assert_true
def test_tempfatfs():
with TempFATFS() as tmpdir:
yield assert_true, tmpdir is not None
|
# emacs: -*- mode: python; py-indent-offset: 4; indent-tabs-mode: nil -*-
# vi: set ft=python sts=4 ts=4 sw=4 et:
"""Test testing utilities
"""
import os
import warnings
from nipype.testing.utils import TempFATFS
from nose.tools import assert_true
def test_tempfatfs():
try:
fatfs = TempFATFS()
except IOError:
warnings.warn("Cannot mount FAT filesystems with FUSE")
else:
with fatfs as tmpdir:
yield assert_true, os.path.exists(tmpdir)
|
Add warning for TempFATFS test
|
TEST: Add warning for TempFATFS test
|
Python
|
bsd-3-clause
|
mick-d/nipype,carolFrohlich/nipype,FCP-INDI/nipype,sgiavasis/nipype,mick-d/nipype,FCP-INDI/nipype,FCP-INDI/nipype,carolFrohlich/nipype,carolFrohlich/nipype,FCP-INDI/nipype,mick-d/nipype,sgiavasis/nipype,mick-d/nipype,carolFrohlich/nipype,sgiavasis/nipype,sgiavasis/nipype
|
---
+++
@@ -3,10 +3,17 @@
"""Test testing utilities
"""
+import os
+import warnings
from nipype.testing.utils import TempFATFS
from nose.tools import assert_true
def test_tempfatfs():
- with TempFATFS() as tmpdir:
- yield assert_true, tmpdir is not None
+ try:
+ fatfs = TempFATFS()
+ except IOError:
+ warnings.warn("Cannot mount FAT filesystems with FUSE")
+ else:
+ with fatfs as tmpdir:
+ yield assert_true, os.path.exists(tmpdir)
|
44547695f662c957f5242f7cfefd328b33d99830
|
sso/backends.py
|
sso/backends.py
|
from django.contrib.auth.backends import ModelBackend
from django.contrib.auth.models import User
from hashlib import sha1
class SimpleHashModelBackend(ModelBackend):
def authenticate(self, username=None, password=None):
try:
user = User.objects.get(username=username)
except User.DoesNotExist:
return None
if '$' in user.password:
if user.check_password(password):
user.password = sha1(password).hexdigest()
user.save()
return user
else:
if user.password == sha1(password).hexdigest():
return user
return None
|
from django.contrib.auth.backends import ModelBackend
from django.contrib.auth.models import User
from hashlib import sha1
class SimpleHashModelBackend(ModelBackend):
supports_anonymous_user = False
supports_object_permissions = False
supports_inactive_user = False
def authenticate(self, username=None, password=None):
try:
user = User.objects.get(username=username)
except User.DoesNotExist:
return None
if '$' in user.password:
if user.check_password(password):
user.password = sha1(password).hexdigest()
user.save()
return user
else:
if user.password == sha1(password).hexdigest():
return user
return None
|
Update the authentication backend with upcoming features
|
Update the authentication backend with upcoming features
|
Python
|
bsd-3-clause
|
nikdoof/test-auth
|
---
+++
@@ -4,6 +4,10 @@
class SimpleHashModelBackend(ModelBackend):
+
+ supports_anonymous_user = False
+ supports_object_permissions = False
+ supports_inactive_user = False
def authenticate(self, username=None, password=None):
try:
|
2515b3402c671c2949e5f3c712cb284777f2accf
|
examples/boilerplates/base_test_case.py
|
examples/boilerplates/base_test_case.py
|
'''
You can use this as a boilerplate for your test framework.
Define your customized library methods in a master class like this.
Then have all your test classes inherit it.
BaseTestCase will inherit SeleniumBase methods from BaseCase.
'''
from seleniumbase import BaseCase
class BaseTestCase(BaseCase):
def setUp(self):
super(BaseTestCase, self).setUp()
# Add custom setUp code for your tests AFTER the super().setUp()
def tearDown(self):
# Add custom tearDown code for your tests BEFORE the super().tearDown()
super(BaseTestCase, self).tearDown()
def login(self):
# <<< Placeholder. Add your code here. >>>
# Reduce duplicate code in tests by having reusable methods like this.
# If the UI changes, the fix can be applied in one place.
pass
def example_method(self):
# <<< Placeholder. Add your code here. >>>
pass
'''
# Now you can do something like this in your test files:
from base_test_case import BaseTestCase
class MyTests(BaseTestCase):
def test_example(self):
self.login()
self.example_method()
'''
|
'''
You can use this as a boilerplate for your test framework.
Define your customized library methods in a master class like this.
Then have all your test classes inherit it.
BaseTestCase will inherit SeleniumBase methods from BaseCase.
'''
from seleniumbase import BaseCase
class BaseTestCase(BaseCase):
def setUp(self):
super(BaseTestCase, self).setUp()
# <<< Add custom setUp code for tests AFTER the super().setUp() >>>
def tearDown(self):
self.save_teardown_screenshot()
# <<< Add custom tearDown code BEFORE the super().tearDown() >>>
super(BaseTestCase, self).tearDown()
def login(self):
# <<< Placeholder. Add your code here. >>>
# Reduce duplicate code in tests by having reusable methods like this.
# If the UI changes, the fix can be applied in one place.
pass
def example_method(self):
# <<< Placeholder. Add your code here. >>>
pass
'''
# Now you can do something like this in your test files:
from base_test_case import BaseTestCase
class MyTests(BaseTestCase):
def test_example(self):
self.login()
self.example_method()
'''
|
Update boilerplate to save a screenshot before the tearDown()
|
Update boilerplate to save a screenshot before the tearDown()
|
Python
|
mit
|
seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase
|
---
+++
@@ -12,10 +12,11 @@
def setUp(self):
super(BaseTestCase, self).setUp()
- # Add custom setUp code for your tests AFTER the super().setUp()
+ # <<< Add custom setUp code for tests AFTER the super().setUp() >>>
def tearDown(self):
- # Add custom tearDown code for your tests BEFORE the super().tearDown()
+ self.save_teardown_screenshot()
+ # <<< Add custom tearDown code BEFORE the super().tearDown() >>>
super(BaseTestCase, self).tearDown()
def login(self):
|
1b338aa716311b4c91281993e35e9beda376735a
|
addons/osfstorage/settings/defaults.py
|
addons/osfstorage/settings/defaults.py
|
# encoding: utf-8
import importlib
import os
import logging
from website import settings
logger = logging.getLogger(__name__)
WATERBUTLER_CREDENTIALS = {
'storage': {}
}
WATERBUTLER_SETTINGS = {
'storage': {
'provider': 'filesystem',
'folder': os.path.join(settings.BASE_PATH, 'osfstoragecache'),
}
}
WATERBUTLER_RESOURCE = 'folder'
DISK_SAVING_MODE = settings.DISK_SAVING_MODE
try:
importlib.import_module('.{}'.format(settings.MIGRATION_ENV))
except:
logger.warn('No migration settings loaded for OSFStorage, falling back to local dev.')
|
# encoding: utf-8
import importlib
import os
import logging
from website import settings
logger = logging.getLogger(__name__)
WATERBUTLER_CREDENTIALS = {
'storage': {}
}
WATERBUTLER_SETTINGS = {
'storage': {
'provider': 'filesystem',
'folder': os.path.join(settings.BASE_PATH, 'osfstoragecache'),
}
}
WATERBUTLER_RESOURCE = 'folder'
DISK_SAVING_MODE = settings.DISK_SAVING_MODE
try:
mod = importlib.import_module('.{}'.format(settings.MIGRATION_ENV), package='addons.osfstorage.settings')
globals().update({k: getattr(mod, k) for k in dir(mod)})
except Exception as ex:
logger.warn('No migration settings loaded for OSFStorage, falling back to local dev. {}'.format(ex))
|
Make environmental override of osfstorage settings work
|
Make environmental override of osfstorage settings work
Signed-off-by: Chris Wisecarver <5fccdd17c1f7bcc7e393d2cb5e2fad37705ca69f@cos.io>
|
Python
|
apache-2.0
|
CenterForOpenScience/osf.io,binoculars/osf.io,monikagrabowska/osf.io,caneruguz/osf.io,TomBaxter/osf.io,chrisseto/osf.io,monikagrabowska/osf.io,sloria/osf.io,crcresearch/osf.io,erinspace/osf.io,caseyrollins/osf.io,Nesiehr/osf.io,TomBaxter/osf.io,leb2dg/osf.io,pattisdr/osf.io,baylee-d/osf.io,adlius/osf.io,felliott/osf.io,mfraezz/osf.io,chrisseto/osf.io,acshi/osf.io,CenterForOpenScience/osf.io,cslzchen/osf.io,CenterForOpenScience/osf.io,cslzchen/osf.io,acshi/osf.io,HalcyonChimera/osf.io,mattclark/osf.io,icereval/osf.io,brianjgeiger/osf.io,chennan47/osf.io,Nesiehr/osf.io,hmoco/osf.io,monikagrabowska/osf.io,felliott/osf.io,baylee-d/osf.io,leb2dg/osf.io,leb2dg/osf.io,felliott/osf.io,mfraezz/osf.io,caseyrollins/osf.io,caneruguz/osf.io,sloria/osf.io,monikagrabowska/osf.io,aaxelb/osf.io,acshi/osf.io,aaxelb/osf.io,cwisecarver/osf.io,adlius/osf.io,mattclark/osf.io,aaxelb/osf.io,saradbowman/osf.io,HalcyonChimera/osf.io,hmoco/osf.io,brianjgeiger/osf.io,pattisdr/osf.io,laurenrevere/osf.io,adlius/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,Johnetordoff/osf.io,cwisecarver/osf.io,saradbowman/osf.io,TomBaxter/osf.io,hmoco/osf.io,cwisecarver/osf.io,monikagrabowska/osf.io,HalcyonChimera/osf.io,mattclark/osf.io,icereval/osf.io,chrisseto/osf.io,leb2dg/osf.io,Johnetordoff/osf.io,binoculars/osf.io,hmoco/osf.io,crcresearch/osf.io,Nesiehr/osf.io,brianjgeiger/osf.io,icereval/osf.io,mfraezz/osf.io,baylee-d/osf.io,pattisdr/osf.io,laurenrevere/osf.io,acshi/osf.io,chennan47/osf.io,binoculars/osf.io,cwisecarver/osf.io,brianjgeiger/osf.io,mfraezz/osf.io,chennan47/osf.io,chrisseto/osf.io,caseyrollins/osf.io,adlius/osf.io,felliott/osf.io,Johnetordoff/osf.io,Nesiehr/osf.io,erinspace/osf.io,HalcyonChimera/osf.io,CenterForOpenScience/osf.io,acshi/osf.io,cslzchen/osf.io,caneruguz/osf.io,aaxelb/osf.io,laurenrevere/osf.io,erinspace/osf.io,caneruguz/osf.io,sloria/osf.io,crcresearch/osf.io
|
---
+++
@@ -24,6 +24,7 @@
try:
- importlib.import_module('.{}'.format(settings.MIGRATION_ENV))
-except:
- logger.warn('No migration settings loaded for OSFStorage, falling back to local dev.')
+ mod = importlib.import_module('.{}'.format(settings.MIGRATION_ENV), package='addons.osfstorage.settings')
+ globals().update({k: getattr(mod, k) for k in dir(mod)})
+except Exception as ex:
+ logger.warn('No migration settings loaded for OSFStorage, falling back to local dev. {}'.format(ex))
|
5a45a312eebe9e432b066b99d914b49a2adb920c
|
openfaas/yaml2json/function/handler.py
|
openfaas/yaml2json/function/handler.py
|
# Author: Milos Buncic
# Date: 2017/10/14
# Description: Convert YAML to JSON and vice versa (OpenFaaS function)
import os
import sys
import json
import yaml
def handle(data, **parms):
def yaml2json(ydata):
"""
Convert YAML to JSON (output: JSON)
"""
try:
d = yaml.load(ydata, Loader=yaml.BaseLoader)
except Exception as e:
d = {'error': '{}'.format(e)}
return json.dumps(d)
def json2yaml(jdata):
"""
Convert JSON to YAML (output: YAML)
"""
try:
d = json.loads(jdata)
except Exception as e:
d = {'error': '{}'.format(e)}
return yaml.dump(d, default_flow_style=False)
if parms.get('reverse') == 'true':
print(json2yaml(data))
else:
print(yaml2json(data))
|
# Author: Milos Buncic
# Date: 2017/10/14
# Description: Convert YAML to JSON and vice versa (OpenFaaS function)
import os
import sys
import json
import yaml
def yaml2json(data):
"""
Convert YAML to JSON (output: JSON)
"""
try:
d = yaml.load(data, Loader=yaml.BaseLoader)
except Exception as e:
d = {'error': '{}'.format(e)}
return json.dumps(d)
def json2yaml(data):
"""
Convert JSON to YAML (output: YAML)
"""
try:
d = json.loads(data)
except Exception as e:
d = {'error': '{}'.format(e)}
return yaml.dump(d, default_flow_style=False)
def handle(data, **parms):
if parms.get('reverse') == 'true':
print(json2yaml(data))
else:
print(yaml2json(data))
|
Make handle function to behave similar as main function
|
Make handle function to behave similar as main function
|
Python
|
mit
|
psyhomb/serverless
|
---
+++
@@ -8,31 +8,31 @@
import yaml
-def handle(data, **parms):
- def yaml2json(ydata):
- """
- Convert YAML to JSON (output: JSON)
- """
- try:
- d = yaml.load(ydata, Loader=yaml.BaseLoader)
- except Exception as e:
- d = {'error': '{}'.format(e)}
+def yaml2json(data):
+ """
+ Convert YAML to JSON (output: JSON)
+ """
+ try:
+ d = yaml.load(data, Loader=yaml.BaseLoader)
+ except Exception as e:
+ d = {'error': '{}'.format(e)}
- return json.dumps(d)
+ return json.dumps(d)
- def json2yaml(jdata):
- """
- Convert JSON to YAML (output: YAML)
- """
- try:
- d = json.loads(jdata)
- except Exception as e:
- d = {'error': '{}'.format(e)}
+def json2yaml(data):
+ """
+ Convert JSON to YAML (output: YAML)
+ """
+ try:
+ d = json.loads(data)
+ except Exception as e:
+ d = {'error': '{}'.format(e)}
- return yaml.dump(d, default_flow_style=False)
+ return yaml.dump(d, default_flow_style=False)
+def handle(data, **parms):
if parms.get('reverse') == 'true':
print(json2yaml(data))
else:
|
bd9d08870ec3db09c41c825029c6a513ecc4d1c7
|
packs/asserts/actions/object_equals.py
|
packs/asserts/actions/object_equals.py
|
import pprint
import sys
from st2actions.runners.pythonrunner import Action
__all__ = [
'AssertObjectEquals'
]
class AssertObjectEquals(Action):
def run(self, object, expected):
ret = cmp(object, expected)
if ret == 0:
sys.stdout.write('EQUAL.')
else:
pprint.pprint('Input: \n%s' % object, stream=sys.stderr)
pprint.pprint('Expected: \n%s' % expected, stream=sys.stderr)
raise ValueError('Objects not equal. Input: %s, Expected: %s.' % (object, expected))
|
import pprint
import sys
from st2actions.runners.pythonrunner import Action
__all__ = [
'AssertObjectEquals'
]
def cmp(x, y):
return (x > y) - (x < y)
class AssertObjectEquals(Action):
def run(self, object, expected):
ret = cmp(object, expected)
if ret == 0:
sys.stdout.write('EQUAL.')
else:
pprint.pprint('Input: \n%s' % object, stream=sys.stderr)
pprint.pprint('Expected: \n%s' % expected, stream=sys.stderr)
raise ValueError('Objects not equal. Input: %s, Expected: %s.' % (object, expected))
|
Make action python 3 compatible
|
Make action python 3 compatible
|
Python
|
apache-2.0
|
StackStorm/st2tests,StackStorm/st2tests,StackStorm/st2tests
|
---
+++
@@ -6,6 +6,10 @@
__all__ = [
'AssertObjectEquals'
]
+
+
+def cmp(x, y):
+ return (x > y) - (x < y)
class AssertObjectEquals(Action):
|
22412b3f46451177286f8fc58509a69bb2d95731
|
numpy/testing/setup.py
|
numpy/testing/setup.py
|
#!/usr/bin/env python3
def configuration(parent_package='',top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('testing', parent_package, top_path)
config.add_subpackage('_private')
config.add_subpackage('tests')
config.add_data_files('*.pyi')
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(maintainer="NumPy Developers",
maintainer_email="numpy-dev@numpy.org",
description="NumPy test module",
url="https://www.numpy.org",
license="NumPy License (BSD Style)",
configuration=configuration,
)
|
#!/usr/bin/env python3
def configuration(parent_package='',top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('testing', parent_package, top_path)
config.add_subpackage('_private')
config.add_subpackage('tests')
config.add_data_files('*.pyi')
config.add_data_files('_private/*.pyi')
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(maintainer="NumPy Developers",
maintainer_email="numpy-dev@numpy.org",
description="NumPy test module",
url="https://www.numpy.org",
license="NumPy License (BSD Style)",
configuration=configuration,
)
|
Add support for *.pyi data-files to `np.testing._private`
|
BLD: Add support for *.pyi data-files to `np.testing._private`
|
Python
|
bsd-3-clause
|
rgommers/numpy,numpy/numpy,endolith/numpy,pdebuyl/numpy,simongibbons/numpy,simongibbons/numpy,mhvk/numpy,seberg/numpy,mattip/numpy,jakirkham/numpy,jakirkham/numpy,mattip/numpy,mhvk/numpy,charris/numpy,seberg/numpy,anntzer/numpy,rgommers/numpy,jakirkham/numpy,mhvk/numpy,charris/numpy,anntzer/numpy,pdebuyl/numpy,rgommers/numpy,anntzer/numpy,charris/numpy,jakirkham/numpy,mattip/numpy,endolith/numpy,endolith/numpy,simongibbons/numpy,endolith/numpy,numpy/numpy,mhvk/numpy,charris/numpy,mattip/numpy,jakirkham/numpy,numpy/numpy,simongibbons/numpy,pdebuyl/numpy,simongibbons/numpy,seberg/numpy,anntzer/numpy,mhvk/numpy,rgommers/numpy,numpy/numpy,pdebuyl/numpy,seberg/numpy
|
---
+++
@@ -7,6 +7,7 @@
config.add_subpackage('_private')
config.add_subpackage('tests')
config.add_data_files('*.pyi')
+ config.add_data_files('_private/*.pyi')
return config
if __name__ == '__main__':
|
3c2663d4c8ca523d072b6e82bf872f412aba9321
|
mrgeo-python/src/main/python/pymrgeo/rastermapop.py
|
mrgeo-python/src/main/python/pymrgeo/rastermapop.py
|
import copy
import json
from py4j.java_gateway import JavaClass, java_import
from pymrgeo.instance import is_instance_of as iio
class RasterMapOp(object):
mapop = None
gateway = None
context = None
job = None
def __init__(self, gateway=None, context=None, mapop=None, job=None):
self.gateway = gateway
self.context = context
self.mapop = mapop
self.job = job
@staticmethod
def nan():
return float('nan')
def clone(self):
return copy.copy(self)
def is_instance_of(self, java_object, java_class):
return iio(self.gateway, java_object, java_class)
def metadata(self):
if self.mapop is None:
return None
jvm = self.gateway.jvm
java_import(jvm, "org.mrgeo.mapalgebra.raster.RasterMapOp")
java_import(jvm, "org.mrgeo.image.MrsPyramidMetadata")
meta = self.mapop.metadata().getOrElse(None)
if meta is None:
return None
java_import(jvm, "com.fasterxml.jackson.databind.ObjectMapper")
mapper = jvm.com.fasterxml.jackson.databind.ObjectMapper()
jsonstr = mapper.writeValueAsString(meta)
# print(mapper.writerWithDefaultPrettyPrinter().writeValueAsString(meta))
return json.loads(jsonstr)
|
import copy
import json
from py4j.java_gateway import java_import
from pymrgeo.instance import is_instance_of as iio
class RasterMapOp(object):
mapop = None
gateway = None
context = None
job = None
def __init__(self, gateway=None, context=None, mapop=None, job=None):
self.gateway = gateway
self.context = context
self.mapop = mapop
self.job = job
@staticmethod
def nan():
return float('nan')
def clone(self):
return copy.copy(self)
def is_instance_of(self, java_object, java_class):
return iio(self.gateway, java_object, java_class)
def metadata(self):
if self.mapop is None:
return None
jvm = self.gateway.jvm
java_import(jvm, "org.mrgeo.mapalgebra.raster.RasterMapOp")
java_import(jvm, "org.mrgeo.image.MrsPyramidMetadata")
if self.mapop.metadata().isEmpty():
return None
meta = self.mapop.metadata().get()
java_import(jvm, "com.fasterxml.jackson.databind.ObjectMapper")
mapper = jvm.com.fasterxml.jackson.databind.ObjectMapper()
jsonstr = mapper.writeValueAsString(meta)
# print(mapper.writerWithDefaultPrettyPrinter().writeValueAsString(meta))
return json.loads(jsonstr)
|
Implement empty metadata a little differently
|
Implement empty metadata a little differently
|
Python
|
apache-2.0
|
ngageoint/mrgeo,ngageoint/mrgeo,ngageoint/mrgeo
|
---
+++
@@ -1,7 +1,7 @@
import copy
import json
-from py4j.java_gateway import JavaClass, java_import
+from py4j.java_gateway import java_import
from pymrgeo.instance import is_instance_of as iio
class RasterMapOp(object):
@@ -35,9 +35,10 @@
java_import(jvm, "org.mrgeo.mapalgebra.raster.RasterMapOp")
java_import(jvm, "org.mrgeo.image.MrsPyramidMetadata")
- meta = self.mapop.metadata().getOrElse(None)
- if meta is None:
+ if self.mapop.metadata().isEmpty():
return None
+
+ meta = self.mapop.metadata().get()
java_import(jvm, "com.fasterxml.jackson.databind.ObjectMapper")
|
b56446e0dc7c0c0fe45557461dd4fffa7f1da8d5
|
django_project/setup.py
|
django_project/setup.py
|
# coding=utf-8
"""Setup file for distutils / pypi."""
try:
from ez_setup import use_setuptools
use_setuptools()
except ImportError:
pass
from setuptools import setup, find_packages
setup(
name='django-wms-client',
version='0.1.1',
author='Tim Sutton',
author_email='tim@kartoza.com',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
scripts=[],
url='http://pypi.python.org/pypi/django-wms-client/',
license='../LICENSE.txt',
description=(
'An app to let you include browsable OGC WMS '
'maps on your django web site.'),
long_description=open('README.md').read(),
install_requires=[
"Django==1.7",
"django-leaflet==0.14.1",
"psycopg2==2.5.4",
"factory-boy==2.4.1",
],
test_suite='wms_client.run_tests.run',
)
|
# coding=utf-8
"""Setup file for distutils / pypi."""
try:
from ez_setup import use_setuptools
use_setuptools()
except ImportError:
pass
from setuptools import setup, find_packages
setup(
name='django-wms-client',
version='0.1.1',
author='Tim Sutton',
author_email='tim@kartoza.com',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
scripts=[],
url='http://pypi.python.org/pypi/django-wms-client/',
license='../LICENSE.txt',
description=(
'An app to let you include browsable OGC WMS '
'maps on your django web site.'),
long_description=open('README.md').read(),
install_requires=[
"Django",
"django-leaflet",
"psycopg2",
"factory-boy",
],
test_suite='wms_client.run_tests.run',
)
|
Change requirements to look for abstract (latest) versions.
|
Change requirements to look for abstract (latest) versions.
|
Python
|
bsd-2-clause
|
kartoza/django-wms-client,kartoza/django-wms-client,kartoza/django-wms-client,kartoza/django-wms-client
|
---
+++
@@ -24,10 +24,10 @@
'maps on your django web site.'),
long_description=open('README.md').read(),
install_requires=[
- "Django==1.7",
- "django-leaflet==0.14.1",
- "psycopg2==2.5.4",
- "factory-boy==2.4.1",
+ "Django",
+ "django-leaflet",
+ "psycopg2",
+ "factory-boy",
],
test_suite='wms_client.run_tests.run',
)
|
a78f56e5c4dedc4148ff3503a05705a8d343b638
|
qmpy/web/views/analysis/calculation.py
|
qmpy/web/views/analysis/calculation.py
|
from django.shortcuts import render_to_response
from django.template import RequestContext
import os.path
from qmpy.models import Calculation
from ..tools import get_globals
from bokeh.embed import components
def calculation_view(request, calculation_id):
calculation = Calculation.objects.get(pk=calculation_id)
data = get_globals()
data['calculation'] = calculation
data['stdout'] = ''
data['stderr'] = ''
if os.path.exists(calculation.path+'/stdout.txt'):
data['stdout'] = open(calculation.path+'/stdout.txt').read()
if os.path.exists(calculation.path+'/stderr.txt'):
data['stderr'] = open(calculation.path+'/stderr.txt').read()
#if not calculation.dos is None:
# data['dos'] = calculation.dos.plot.get_flot_script()
if not calculation.dos is None:
script, div = components(calculation.dos.bokeh_plot)
data['dos'] = script
data['dosdiv'] = div
## Get exact INCAR settings from INCAR file
data['incar'] = ''.join(calculation.read_incar())
return render_to_response('analysis/calculation.html',
data, RequestContext(request))
|
from django.shortcuts import render_to_response
from django.template import RequestContext
import os
from qmpy.models import Calculation
from ..tools import get_globals
from bokeh.embed import components
def calculation_view(request, calculation_id):
calculation = Calculation.objects.get(pk=calculation_id)
data = get_globals()
data['calculation'] = calculation
data['stdout'] = ''
data['stderr'] = ''
if os.path.exists(os.path.join(calculation.path, 'stdout.txt')):
with open(os.path.join(calculation.path, 'stdout.txt')) as fr:
data['stdout'] = fr.read()
if os.path.exists(os.path.join(calculation.path, 'stderr.txt')):
with open(os.path.join(calculation.path, 'stderr.txt')) as fr:
data['stderr'] = fr.read()
try:
data['incar'] = ''.join(calculation.read_incar())
except FileNotFoundError:
data['incar'] = 'Could not read INCAR'
if not calculation.dos is None:
script, div = components(calculation.dos.bokeh_plot)
data['dos'] = script
data['dosdiv'] = div
return render_to_response('analysis/calculation.html',
data, RequestContext(request))
|
Handle missing INCAR files gracefully
|
Handle missing INCAR files gracefully
|
Python
|
mit
|
wolverton-research-group/qmpy,wolverton-research-group/qmpy,wolverton-research-group/qmpy,wolverton-research-group/qmpy,wolverton-research-group/qmpy
|
---
+++
@@ -1,6 +1,6 @@
from django.shortcuts import render_to_response
from django.template import RequestContext
-import os.path
+import os
from qmpy.models import Calculation
from ..tools import get_globals
@@ -14,20 +14,21 @@
data['stdout'] = ''
data['stderr'] = ''
- if os.path.exists(calculation.path+'/stdout.txt'):
- data['stdout'] = open(calculation.path+'/stdout.txt').read()
- if os.path.exists(calculation.path+'/stderr.txt'):
- data['stderr'] = open(calculation.path+'/stderr.txt').read()
- #if not calculation.dos is None:
- # data['dos'] = calculation.dos.plot.get_flot_script()
+ if os.path.exists(os.path.join(calculation.path, 'stdout.txt')):
+ with open(os.path.join(calculation.path, 'stdout.txt')) as fr:
+ data['stdout'] = fr.read()
+ if os.path.exists(os.path.join(calculation.path, 'stderr.txt')):
+ with open(os.path.join(calculation.path, 'stderr.txt')) as fr:
+ data['stderr'] = fr.read()
+ try:
+ data['incar'] = ''.join(calculation.read_incar())
+ except FileNotFoundError:
+ data['incar'] = 'Could not read INCAR'
if not calculation.dos is None:
script, div = components(calculation.dos.bokeh_plot)
data['dos'] = script
data['dosdiv'] = div
- ## Get exact INCAR settings from INCAR file
- data['incar'] = ''.join(calculation.read_incar())
-
- return render_to_response('analysis/calculation.html',
+ return render_to_response('analysis/calculation.html',
data, RequestContext(request))
|
a64f8aaa2822ccd280f252e0937be5027d5ec012
|
censusreporter/config/prod/settings.py
|
censusreporter/config/prod/settings.py
|
from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = [
'censusreporter.org',
'www.censusreporter.org',
'censusreporter.dokku.censusreporter.org',
]
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL'),
}
}
|
from censusreporter.config.base.settings import *
import os
DEBUG = False
ROOT_URLCONF = 'censusreporter.config.prod.urls'
WSGI_APPLICATION = "censusreporter.config.prod.wsgi.application"
ALLOWED_HOSTS = [
'censusreporter.org',
'www.censusreporter.org',
'censusreporter.dokku.censusreporter.org',
]
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
|
Fix Dockerfile so it can cache requirements
|
Fix Dockerfile so it can cache requirements
|
Python
|
mit
|
censusreporter/censusreporter,censusreporter/censusreporter,censusreporter/censusreporter,censusreporter/censusreporter
|
---
+++
@@ -15,6 +15,6 @@
CACHES = {
'default': {
'BACKEND': 'redis_cache.RedisCache',
- 'LOCATION': os.environ.get('REDIS_URL'),
+ 'LOCATION': os.environ.get('REDIS_URL', ''),
}
}
|
eb453010915f6700edd1baa0febcc634deec81dc
|
src/viewsapp/views.py
|
src/viewsapp/views.py
|
from decorator_plus import (
require_form_methods, require_safe_methods)
from django.shortcuts import (
get_object_or_404, redirect, render)
from .forms import ExampleForm
from .models import ExampleModel
@require_safe_methods
def model_detail(request, *args, **kwargs):
request_slug = kwargs.get('slug')
example_obj = get_object_or_404(
ExampleModel, slug=request_slug)
return render(
request,
'viewsapp/detail.html',
{'object': example_obj})
@require_form_methods
def model_create(request, *args, **kwargs):
if request.method == 'POST':
form = ExampleForm(request.POST)
if form.is_valid():
new_obj = form.save()
return redirect(new_obj)
else:
form = ExampleForm()
return render(
request,
'viewsapp/form.html',
{'form': form})
|
from decorator_plus import require_http_methods
from django.shortcuts import (
get_object_or_404, redirect, render)
from .forms import ExampleForm
from .models import ExampleModel
@require_http_methods(['GET'])
def model_detail(request, *args, **kwargs):
request_slug = kwargs.get('slug')
example_obj = get_object_or_404(
ExampleModel, slug=request_slug)
return render(
request,
'viewsapp/detail.html',
{'object': example_obj})
@require_http_methods(['GET', 'POST'])
def model_create(request, *args, **kwargs):
if request.method == 'POST':
form = ExampleForm(request.POST)
if form.is_valid():
new_obj = form.save()
return redirect(new_obj)
else:
form = ExampleForm()
return render(
request,
'viewsapp/form.html',
{'form': form})
|
Switch to using require_http_methods decorator.
|
Switch to using require_http_methods decorator.
|
Python
|
bsd-2-clause
|
jambonrose/djangocon2015-views,jambonrose/djangocon2015-views
|
---
+++
@@ -1,5 +1,4 @@
-from decorator_plus import (
- require_form_methods, require_safe_methods)
+from decorator_plus import require_http_methods
from django.shortcuts import (
get_object_or_404, redirect, render)
@@ -7,7 +6,7 @@
from .models import ExampleModel
-@require_safe_methods
+@require_http_methods(['GET'])
def model_detail(request, *args, **kwargs):
request_slug = kwargs.get('slug')
example_obj = get_object_or_404(
@@ -18,7 +17,7 @@
{'object': example_obj})
-@require_form_methods
+@require_http_methods(['GET', 'POST'])
def model_create(request, *args, **kwargs):
if request.method == 'POST':
form = ExampleForm(request.POST)
|
94e44e18832311bc063830c0a6bfe23e04e40a8d
|
srsly/_msgpack_api.py
|
srsly/_msgpack_api.py
|
# coding: utf8
from __future__ import unicode_literals
import gc
from . import msgpack
from .util import force_path
def msgpack_dumps(data):
return msgpack.dumps(data, use_bin_type=True)
def msgpack_loads(data):
# msgpack-python docs suggest disabling gc before unpacking large messages
gc.disable()
msg = msgpack.loads(data, raw=False)
gc.enable()
return msg
def read_msgpack(location):
file_path = force_path(location)
with file_path.open("rb") as f:
gc.disable()
msg = msgpack.load(f, raw=False)
gc.enable()
return msg
def write_msgpack(location, data):
file_path = force_path(location, require_exists=False)
with file_path.open("wb") as f:
msgpack.dump(data, f, use_bin_type=True)
|
# coding: utf8
from __future__ import unicode_literals
import gc
from . import msgpack
from .util import force_path
def msgpack_dumps(data):
return msgpack.dumps(data, use_bin_type=True)
def msgpack_loads(data, use_list=True):
# msgpack-python docs suggest disabling gc before unpacking large messages
gc.disable()
msg = msgpack.loads(data, raw=False, use_list=use_list)
gc.enable()
return msg
def read_msgpack(location):
file_path = force_path(location)
with file_path.open("rb") as f:
gc.disable()
msg = msgpack.load(f, raw=False)
gc.enable()
return msg
def write_msgpack(location, data):
file_path = force_path(location, require_exists=False)
with file_path.open("wb") as f:
msgpack.dump(data, f, use_bin_type=True)
|
Allow passing use_list to msgpack_loads
|
Allow passing use_list to msgpack_loads
|
Python
|
mit
|
explosion/srsly,explosion/srsly,explosion/srsly,explosion/srsly
|
---
+++
@@ -11,10 +11,10 @@
return msgpack.dumps(data, use_bin_type=True)
-def msgpack_loads(data):
+def msgpack_loads(data, use_list=True):
# msgpack-python docs suggest disabling gc before unpacking large messages
gc.disable()
- msg = msgpack.loads(data, raw=False)
+ msg = msgpack.loads(data, raw=False, use_list=use_list)
gc.enable()
return msg
|
e78b3f53150a5f1c170b860f8719e982cf1c6f9e
|
integration/main.py
|
integration/main.py
|
import os
import sys
from spec import Spec, skip
from invoke import run
class Integration(Spec):
def setup(self):
from tessera.application import db
# Ensure we have a clean db target.
self.dbpath = db.engine.url.database
msg = "You seem to have a db in the default location ({0}) - please (re)move it before running tests to avoid collisions."
assert not os.path.exists(self.dbpath), msg.format(self.dbpath)
def teardown(self):
# Teardown only runs if setup completed, so the below will not nuke
# pre-existing dbs that cause setup's check to fail.
if os.path.exists(self.dbpath):
os.remove(self.dbpath)
def is_importable(self):
import tessera
assert tessera.app
assert tessera.db
def can_initdb(self):
from tessera.application import db
from tessera.model.database import Dashboard
# Make sure we can create and look at the DB
db.create_all()
assert len(Dashboard.query.all()) == 0
|
import os
import sys
from spec import Spec, skip, eq_
from invoke import run
class Integration(Spec):
def setup(self):
from tessera.application import db
# Ensure we have a clean db target.
self.dbpath = db.engine.url.database
msg = "You seem to have a db in the default location ({0}) - please (re)move it before running tests to avoid collisions."
assert not os.path.exists(self.dbpath), msg.format(self.dbpath)
def teardown(self):
from tessera.application import db
# Teardown only runs if setup completed, so the below will not nuke
# pre-existing dbs that cause setup's check to fail.
if os.path.exists(self.dbpath):
os.remove(self.dbpath)
# Ensure no cached session crap
db.session.close_all()
def is_importable(self):
import tessera
assert tessera.app
assert tessera.db
def can_initdb(self):
from tessera.application import db
from tessera.model.database import Dashboard
# Make sure we can create and look at the DB
db.create_all()
eq_(len(Dashboard.query.all()), 0)
def can_import_fixtures(self):
from tessera.application import db
from tessera.importer.json import JsonImporter
from tessera.model.database import Dashboard
db.create_all()
path = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', 'demo', 'demo-gallery.json'
))
JsonImporter.import_file(path)
eq_(len(Dashboard.query.all()), 1)
|
Fix state bleed, add fixture import test
|
Fix state bleed, add fixture import test
|
Python
|
apache-2.0
|
section-io/tessera,urbanairship/tessera,tessera-metrics/tessera,jmptrader/tessera,section-io/tessera,urbanairship/tessera,aalpern/tessera,section-io/tessera,jmptrader/tessera,filippog/tessera,aalpern/tessera,aalpern/tessera,Slach/tessera,Slach/tessera,urbanairship/tessera,urbanairship/tessera,Slach/tessera,jmptrader/tessera,jmptrader/tessera,tessera-metrics/tessera,aalpern/tessera,section-io/tessera,urbanairship/tessera,tessera-metrics/tessera,jmptrader/tessera,tessera-metrics/tessera,Slach/tessera,filippog/tessera,filippog/tessera,aalpern/tessera,tessera-metrics/tessera
|
---
+++
@@ -1,7 +1,7 @@
import os
import sys
-from spec import Spec, skip
+from spec import Spec, skip, eq_
from invoke import run
@@ -14,10 +14,14 @@
assert not os.path.exists(self.dbpath), msg.format(self.dbpath)
def teardown(self):
+ from tessera.application import db
# Teardown only runs if setup completed, so the below will not nuke
# pre-existing dbs that cause setup's check to fail.
if os.path.exists(self.dbpath):
os.remove(self.dbpath)
+ # Ensure no cached session crap
+ db.session.close_all()
+
def is_importable(self):
import tessera
@@ -29,4 +33,15 @@
from tessera.model.database import Dashboard
# Make sure we can create and look at the DB
db.create_all()
- assert len(Dashboard.query.all()) == 0
+ eq_(len(Dashboard.query.all()), 0)
+
+ def can_import_fixtures(self):
+ from tessera.application import db
+ from tessera.importer.json import JsonImporter
+ from tessera.model.database import Dashboard
+ db.create_all()
+ path = os.path.abspath(os.path.join(
+ os.path.dirname(__file__), '..', 'demo', 'demo-gallery.json'
+ ))
+ JsonImporter.import_file(path)
+ eq_(len(Dashboard.query.all()), 1)
|
0ad29aa9945448236500b221fc489c1627e6693b
|
api.py
|
api.py
|
from collections import namedtuple
import requests
QUERY_TEMPLATE = '?token={}&domain={}'
BASE_URL = 'https://pddimp.yandex.ru/api2/'
Connection = namedtuple('Connection', ['auth', 'domain'])
def list_emails(connection):
url = '{}admin/email/list'.format(BASE_URL) + QUERY_TEMPLATE.format(*connection)
ret = requests.get(url)
return ret.json()
|
import json
import random
import string
from collections import namedtuple
import requests
QUERY_TEMPLATE = '?token={}&domain={}'
BASE_URL = 'https://pddimp.yandex.ru/api2/'
Connection = namedtuple('Connection', ['auth', 'domain'])
class YandexException(Exception):
pass
rndstr = lambda: ''.join(random.sample(string.ascii_letters + string.hexdigits, 17))
def _check_error(ret_json):
if ret_json.get('success') == 'error':
raise YandexException(ret_json['error'])
def list_emails(connection):
url = '{}admin/email/list'.format(BASE_URL) + QUERY_TEMPLATE.format(*connection)
ret = requests.get(url).json()
_check_error(ret)
return ret
def create_email(connection, email, password=None):
if not password:
password = rndstr()
url = '{}admin/email/add'.format(BASE_URL) + QUERY_TEMPLATE.format(*connection)
url += '&login={}&password={}'.format(email, password)
ret = requests.post(url).json()
_check_error(ret)
return ret, password
def delete_email(connection, email=None, uid=None):
if not email and uid:
raise YandexException('Must specify email or uid')
url = '{}admin/email/del'.format(BASE_URL) + QUERY_TEMPLATE.format(*connection)
if email:
url += '&login={}'.format(email)
else:
url += '&uid={}'.format(uid)
ret = requests.post(url).json()
_check_error(ret)
return ret
|
Add create email and delete email methods
|
Add create email and delete email methods
|
Python
|
mit
|
chrisseto/dinosaurs.sexy,chrisseto/dinosaurs.sexy
|
---
+++
@@ -1,3 +1,7 @@
+import json
+import random
+import string
+
from collections import namedtuple
import requests
@@ -9,7 +13,53 @@
Connection = namedtuple('Connection', ['auth', 'domain'])
+class YandexException(Exception):
+ pass
+
+
+rndstr = lambda: ''.join(random.sample(string.ascii_letters + string.hexdigits, 17))
+
+
+def _check_error(ret_json):
+ if ret_json.get('success') == 'error':
+ raise YandexException(ret_json['error'])
+
+
def list_emails(connection):
url = '{}admin/email/list'.format(BASE_URL) + QUERY_TEMPLATE.format(*connection)
- ret = requests.get(url)
- return ret.json()
+ ret = requests.get(url).json()
+ _check_error(ret)
+ return ret
+
+
+def create_email(connection, email, password=None):
+ if not password:
+ password = rndstr()
+
+ url = '{}admin/email/add'.format(BASE_URL) + QUERY_TEMPLATE.format(*connection)
+
+ url += '&login={}&password={}'.format(email, password)
+
+ ret = requests.post(url).json()
+
+ _check_error(ret)
+
+ return ret, password
+
+
+def delete_email(connection, email=None, uid=None):
+ if not email and uid:
+ raise YandexException('Must specify email or uid')
+
+ url = '{}admin/email/del'.format(BASE_URL) + QUERY_TEMPLATE.format(*connection)
+
+ if email:
+ url += '&login={}'.format(email)
+ else:
+ url += '&uid={}'.format(uid)
+
+ ret = requests.post(url).json()
+
+ _check_error(ret)
+
+ return ret
|
df325e7caee163f754c1b85cf71dab7810700933
|
src/waldur_keycloak_rancher/tasks.py
|
src/waldur_keycloak_rancher/tasks.py
|
import logging
from celery import shared_task
from django.conf import settings
from waldur_keycloak.models import ProjectGroup
from waldur_rancher.models import Cluster, ClusterRole
logger = logging.getLogger(__name__)
@shared_task(name='waldur_keycloak_rancher.sync_groups')
def sync_groups():
if not settings.WALDUR_KEYCLOAK['ENABLED']:
logger.debug('Skipping Keycloak synchronization because plugin is disabled.')
return
for project_group in ProjectGroup.objects.all():
project = project_group.project
for cluster in Cluster.objects.filter(project=project):
backend = cluster.get_backend()
try:
backend.get_or_create_cluster_group_role(
f'keycloakoidc_group://{project.name}',
cluster.backend_id,
ClusterRole.CLUSTER_MEMBER,
)
except Exception:
logger.warning(
'Unable to create cluster group for project %s and cluster %s',
project,
cluster,
)
|
import logging
from celery import shared_task
from django.conf import settings
from waldur_keycloak.models import ProjectGroup
from waldur_rancher.enums import ClusterRoles
from waldur_rancher.models import Cluster
logger = logging.getLogger(__name__)
@shared_task(name='waldur_keycloak_rancher.sync_groups')
def sync_groups():
if not settings.WALDUR_KEYCLOAK['ENABLED']:
logger.debug('Skipping Keycloak synchronization because plugin is disabled.')
return
for project_group in ProjectGroup.objects.all():
project = project_group.project
for cluster in Cluster.objects.filter(project=project):
backend = cluster.get_backend()
try:
backend.get_or_create_cluster_group_role(
f'keycloakoidc_group://{project.name}',
cluster.backend_id,
ClusterRoles.cluster_member,
)
except Exception:
logger.warning(
'Unable to create cluster group for project %s and cluster %s',
project,
cluster,
)
|
Fix cluster group membership synchronization.
|
Fix cluster group membership synchronization.
|
Python
|
mit
|
opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/waldur-mastermind
|
---
+++
@@ -4,7 +4,8 @@
from django.conf import settings
from waldur_keycloak.models import ProjectGroup
-from waldur_rancher.models import Cluster, ClusterRole
+from waldur_rancher.enums import ClusterRoles
+from waldur_rancher.models import Cluster
logger = logging.getLogger(__name__)
@@ -23,7 +24,7 @@
backend.get_or_create_cluster_group_role(
f'keycloakoidc_group://{project.name}',
cluster.backend_id,
- ClusterRole.CLUSTER_MEMBER,
+ ClusterRoles.cluster_member,
)
except Exception:
logger.warning(
|
cbddfe308f4e0da728974777f10b245a966520b6
|
summarize/__init__.py
|
summarize/__init__.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from itertools import combinations
from operator import itemgetter
from distance import jaccard
from networkx import Graph, pagerank
from nltk import tokenize
from .utils import get_stopwords, get_words
def summarize(text, sentence_count=5, language='english'):
stopwords = get_stopwords(language)
sentence_list = tokenize.sent_tokenize(text, language)
wordsets = [get_words(sentence, stopwords) for sentence in sentence_list]
graph = Graph()
pairs = combinations(enumerate(filter(None, wordsets)), 2)
for (index_a, words_a), (index_b, words_b) in pairs:
similarity = 1 - jaccard(words_a, words_b)
if similarity > 0:
graph.add_edge(index_a, index_b, weight=similarity)
ranked_sentence_indexes = pagerank(graph).items()
sentences_by_rank = sorted(
ranked_sentence_indexes, key=itemgetter(1), reverse=True)
best_sentences = map(itemgetter(0), sentences_by_rank[:sentence_count])
best_sentences_in_order = sorted(best_sentences)
return ' '.join(sentence_list[index] for index in best_sentences_in_order)
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from itertools import combinations
from operator import itemgetter
from distance import jaccard
from networkx import Graph, pagerank
from nltk import tokenize
from .utils import get_stopwords, get_words
def summarize(text, sentence_count=5, language='english'):
stopwords = get_stopwords(language)
sentence_list = tokenize.sent_tokenize(text, language)
wordsets = [get_words(sentence, stopwords) for sentence in sentence_list]
graph = Graph()
pairs = combinations(enumerate(wordsets), 2)
for (index_a, words_a), (index_b, words_b) in pairs:
if words_a and words_b:
similarity = 1 - jaccard(words_a, words_b)
if similarity > 0:
graph.add_edge(index_a, index_b, weight=similarity)
ranked_sentence_indexes = pagerank(graph).items()
sentences_by_rank = sorted(
ranked_sentence_indexes, key=itemgetter(1), reverse=True)
best_sentences = map(itemgetter(0), sentences_by_rank[:sentence_count])
best_sentences_in_order = sorted(best_sentences)
return ' '.join(sentence_list[index] for index in best_sentences_in_order)
|
Fix sentence index shifting with empty sentences
|
Fix sentence index shifting with empty sentences
|
Python
|
mit
|
despawnerer/summarize
|
---
+++
@@ -17,11 +17,12 @@
wordsets = [get_words(sentence, stopwords) for sentence in sentence_list]
graph = Graph()
- pairs = combinations(enumerate(filter(None, wordsets)), 2)
+ pairs = combinations(enumerate(wordsets), 2)
for (index_a, words_a), (index_b, words_b) in pairs:
- similarity = 1 - jaccard(words_a, words_b)
- if similarity > 0:
- graph.add_edge(index_a, index_b, weight=similarity)
+ if words_a and words_b:
+ similarity = 1 - jaccard(words_a, words_b)
+ if similarity > 0:
+ graph.add_edge(index_a, index_b, weight=similarity)
ranked_sentence_indexes = pagerank(graph).items()
sentences_by_rank = sorted(
|
dc68813d5f555a01f1bdd2511d9d2de820369573
|
conditional/blueprints/spring_evals.py
|
conditional/blueprints/spring_evals.py
|
from flask import Blueprint
from flask import render_template
from flask import request
spring_evals_bp = Blueprint('spring_evals_bp', __name__)
@spring_evals_bp.route('/spring_evals/')
def display_spring_evals():
# get user data
user_name = request.headers.get('x-webauth-user')
members = [
{
'name': "Liam Middlebrook",
'uid': 'loothelion',
'committee_meetings': 24,
'house_meetings_missed': [{'date': "aprial fools fayas ads", 'reason': "I was playing videogames"}],
'major_project': 'open_container',
'major_project_passed': True,
'social_events': "",
'comments': "please don't fail me",
'result': 'Pending'
},
{
'name': "Julien Eid",
'uid': 'jeid',
'committee_meetings': 69,
'house_meetings_missed': [],
'major_project': 'wii-u shit',
'major_project_passed': True,
'social_events': "Manipulation and Opportunism",
'comments': "imdabes",
'result': 'Passed'
}
]
# return names in 'first last (username)' format
return render_template('spring_evals.html',
username = user_name,
members = members)
|
from flask import Blueprint
from flask import render_template
from flask import request
spring_evals_bp = Blueprint('spring_evals_bp', __name__)
@spring_evals_bp.route('/spring_evals/')
def display_spring_evals():
# get user data
user_name = request.headers.get('x-webauth-user')
members = [
{
'name': "Liam Middlebrook",
'uid': 'loothelion',
'committee_meetings': 24,
'house_meetings_missed': [{'date': "aprial fools fayas ads", 'reason': "I was playing videogames"}],
'major_projects': [
{
'name': "open container",
'status': "Passed",
'description': "Riding With A Flask"
}],
'major_projects_len': 1,
'major_project_passed': True,
'social_events': "",
'comments': "please don't fail me",
'result': 'Pending'
},
{
'name': "Julien Eid",
'uid': 'jeid',
'committee_meetings': 69,
'house_meetings_missed': [],
'major_projects': [
{
'name': "wii-u shit",
'status': "Failed",
'description': "Rot 3 Encryption"
}],
'major_projects_len': 1,
'major_project_passed': False,
'social_events': "Manipulation and Opportunism",
'comments': "imdabes",
'result': 'Passed'
}
]
# return names in 'first last (username)' format
return render_template('spring_evals.html',
username = user_name,
members = members)
|
Restructure major projects for spring evals 👷
|
Restructure major projects for spring evals 👷
|
Python
|
mit
|
RamZallan/conditional,ComputerScienceHouse/conditional,RamZallan/conditional,ComputerScienceHouse/conditional,RamZallan/conditional,ComputerScienceHouse/conditional
|
---
+++
@@ -16,7 +16,13 @@
'uid': 'loothelion',
'committee_meetings': 24,
'house_meetings_missed': [{'date': "aprial fools fayas ads", 'reason': "I was playing videogames"}],
- 'major_project': 'open_container',
+ 'major_projects': [
+ {
+ 'name': "open container",
+ 'status': "Passed",
+ 'description': "Riding With A Flask"
+ }],
+ 'major_projects_len': 1,
'major_project_passed': True,
'social_events': "",
'comments': "please don't fail me",
@@ -27,8 +33,14 @@
'uid': 'jeid',
'committee_meetings': 69,
'house_meetings_missed': [],
- 'major_project': 'wii-u shit',
- 'major_project_passed': True,
+ 'major_projects': [
+ {
+ 'name': "wii-u shit",
+ 'status': "Failed",
+ 'description': "Rot 3 Encryption"
+ }],
+ 'major_projects_len': 1,
+ 'major_project_passed': False,
'social_events': "Manipulation and Opportunism",
'comments': "imdabes",
'result': 'Passed'
|
3ffea173c206d4c8a58953eceb34d80fb66d609b
|
utils/tasks.py
|
utils/tasks.py
|
from celery.task import task
from utils import send_templated_email
@task
def send_templated_email_async(to, subject, body_template, body_dict,
from_email=None, ct="html", fail_silently=False):
return send_templated_email(to,subject, body_template, body_dict,
from_email=None, ct="html", fail_silently=False)
|
from celery.task import task
from utils import send_templated_email
@task
def send_templated_email_async(to, subject, body_template, body_dict,
from_email=None, ct="html", fail_silently=False, check_user_preference=True):
return send_templated_email(to,subject, body_template, body_dict,
from_email=None, ct="html", fail_silently=False)
|
Allow check_user_preferences for email to be passed from async templated email send
|
Allow check_user_preferences for email to be passed from async templated email send
|
Python
|
agpl-3.0
|
ReachingOut/unisubs,norayr/unisubs,wevoice/wesub,eloquence/unisubs,ujdhesa/unisubs,ujdhesa/unisubs,ReachingOut/unisubs,pculture/unisubs,eloquence/unisubs,wevoice/wesub,ReachingOut/unisubs,ofer43211/unisubs,ofer43211/unisubs,ofer43211/unisubs,wevoice/wesub,ujdhesa/unisubs,pculture/unisubs,wevoice/wesub,ujdhesa/unisubs,norayr/unisubs,ReachingOut/unisubs,pculture/unisubs,pculture/unisubs,norayr/unisubs,eloquence/unisubs,eloquence/unisubs,ofer43211/unisubs,norayr/unisubs
|
---
+++
@@ -3,6 +3,6 @@
@task
def send_templated_email_async(to, subject, body_template, body_dict,
- from_email=None, ct="html", fail_silently=False):
+ from_email=None, ct="html", fail_silently=False, check_user_preference=True):
return send_templated_email(to,subject, body_template, body_dict,
from_email=None, ct="html", fail_silently=False)
|
45963022a39f8c0f3d57199017adc78b39005d6a
|
openspending/lib/unicode_dict_reader.py
|
openspending/lib/unicode_dict_reader.py
|
# work around python2's csv.py's difficulty with utf8
# partly cribbed from http://stackoverflow.com/questions/5478659/python-module-like-csv-dictreader-with-full-utf8-support
class EmptyCSVError(Exception):
pass
def UnicodeDictReader(file_or_str, encoding='utf8', **kwargs):
import csv
def decode(s, encoding):
if s is None:
return None
return s.decode(encoding)
csv_reader = csv.DictReader(file_or_str, **kwargs)
if not csv_reader.fieldnames:
raise EmptyCSVError("No fieldnames in CSV reader: empty file?")
keymap = dict((k, k.decode(encoding)) for k in csv_reader.fieldnames)
for row in csv_reader:
yield dict((keymap[k], decode(v, encoding)) for k, v in row.iteritems())
|
# work around python2's csv.py's difficulty with utf8
# partly cribbed from http://stackoverflow.com/questions/5478659/python-module-like-csv-dictreader-with-full-utf8-support
import csv
class EmptyCSVError(Exception):
pass
class UnicodeDictReader(object):
def __init__(self, file_or_str, encoding='utf8', **kwargs):
self.encoding = encoding
self.reader = csv.DictReader(file_or_str, **kwargs)
if not self.reader.fieldnames:
raise EmptyCSVError("No fieldnames in CSV reader: empty file?")
self.keymap = dict((k, k.decode(encoding)) for k in self.reader.fieldnames)
def __iter__(self):
return (self._decode_row(row) for row in self.reader)
def _decode_row(self, row):
return dict(
(self.keymap[k], self._decode_str(v)) for k, v in row.iteritems()
)
def _decode_str(self, s):
if s is None:
return None
return s.decode(self.encoding)
|
Convert UnicodeDictReader to an iterator class, so that EmptyCSVError will get thrown on instantiation, rather than on iteration.
|
Convert UnicodeDictReader to an iterator class, so that EmptyCSVError will get thrown on instantiation, rather than on iteration.
|
Python
|
agpl-3.0
|
pudo/spendb,nathanhilbert/FPA_Core,USStateDept/FPA_Core,openspending/spendb,CivicVision/datahub,nathanhilbert/FPA_Core,CivicVision/datahub,spendb/spendb,johnjohndoe/spendb,johnjohndoe/spendb,spendb/spendb,USStateDept/FPA_Core,nathanhilbert/FPA_Core,pudo/spendb,openspending/spendb,johnjohndoe/spendb,openspending/spendb,pudo/spendb,spendb/spendb,CivicVision/datahub,USStateDept/FPA_Core
|
---
+++
@@ -1,22 +1,31 @@
# work around python2's csv.py's difficulty with utf8
# partly cribbed from http://stackoverflow.com/questions/5478659/python-module-like-csv-dictreader-with-full-utf8-support
+
+import csv
class EmptyCSVError(Exception):
pass
-def UnicodeDictReader(file_or_str, encoding='utf8', **kwargs):
- import csv
- def decode(s, encoding):
+class UnicodeDictReader(object):
+ def __init__(self, file_or_str, encoding='utf8', **kwargs):
+ self.encoding = encoding
+ self.reader = csv.DictReader(file_or_str, **kwargs)
+
+ if not self.reader.fieldnames:
+ raise EmptyCSVError("No fieldnames in CSV reader: empty file?")
+
+ self.keymap = dict((k, k.decode(encoding)) for k in self.reader.fieldnames)
+
+ def __iter__(self):
+ return (self._decode_row(row) for row in self.reader)
+
+ def _decode_row(self, row):
+ return dict(
+ (self.keymap[k], self._decode_str(v)) for k, v in row.iteritems()
+ )
+
+ def _decode_str(self, s):
if s is None:
return None
- return s.decode(encoding)
-
- csv_reader = csv.DictReader(file_or_str, **kwargs)
-
- if not csv_reader.fieldnames:
- raise EmptyCSVError("No fieldnames in CSV reader: empty file?")
-
- keymap = dict((k, k.decode(encoding)) for k in csv_reader.fieldnames)
- for row in csv_reader:
- yield dict((keymap[k], decode(v, encoding)) for k, v in row.iteritems())
+ return s.decode(self.encoding)
|
db935a152efc8ab730491fc860db6d4c9cf65c5f
|
test/test_packages.py
|
test/test_packages.py
|
import pytest
@pytest.mark.parametrize("name", [
("apt-file"),
("apt-transport-https"),
("atom"),
("blktrace"),
("ca-certificates"),
("chromium-browser"),
("cron"),
("curl"),
("diod"),
("docker-ce"),
("fonts-font-awesome"),
("git"),
("gnupg"),
("gnupg2"),
("gnupg-agent"),
("handbrake"),
("handbrake-cli"),
("haveged"),
("htop"),
("i3"),
("iotop"),
("language-pack-en-base"),
("laptop-mode-tools"),
("nfs-common"),
("ntop"),
("ntp"),
("openssh-client"),
("openssh-server"),
("openssh-sftp-server"),
("openssl"),
("pinta"),
("python"),
("python-pip"),
("scrot"),
("software-properties-common"),
("suckless-tools"),
("sysstat"),
("tree"),
("vagrant"),
("vim"),
("virtualbox"),
("vlc"),
("wget"),
("whois"),
("x264"),
("xfce4-terminal"),
("xfonts-terminus"),
("xinit"),
])
def test_packages(Package, name):
assert Package(name).is_installed
|
import pytest
@pytest.mark.parametrize("name", [
("apt-file"),
("apt-transport-https"),
("atom"),
("blktrace"),
("ca-certificates"),
("chromium-browser"),
("cron"),
("curl"),
("diod"),
("docker-ce"),
("fonts-font-awesome"),
("git"),
("gnupg"),
("gnupg2"),
("gnupg-agent"),
("handbrake"),
("handbrake-cli"),
("haveged"),
("htop"),
("i3"),
("iotop"),
("language-pack-en-base"),
("laptop-mode-tools"),
("nfs-common"),
("ntop"),
("ntp"),
("openssh-client"),
("openssh-server"),
("openssh-sftp-server"),
("openssl"),
("pinta"),
("python"),
("python-pip"),
("scrot"),
("software-properties-common"),
("suckless-tools"),
("sysstat"),
("tree"),
("vagrant"),
("vim"),
("virtualbox"),
("vlc"),
("wget"),
("whois"),
("x264"),
("xfce4-terminal"),
("xfonts-terminus"),
("xinit"),
])
def test_packages(host, name):
pkg = host.package(name)
assert pkg.is_installed
|
Change test function as existing method deprecated
|
Change test function as existing method deprecated
|
Python
|
mit
|
wicksy/laptop-build,wicksy/laptop-build,wicksy/laptop-build,wicksy/laptop-build
|
---
+++
@@ -51,5 +51,6 @@
("xinit"),
])
-def test_packages(Package, name):
- assert Package(name).is_installed
+def test_packages(host, name):
+ pkg = host.package(name)
+ assert pkg.is_installed
|
7d621db3618db90679461550fb0c952417616402
|
bot.py
|
bot.py
|
import asyncio
import configparser
import discord
from discord.ext import commands
class Bot(commands.Bot):
def __init__(self, config_filepath, command_prefix):
super().__init__(command_prefix)
self._load_config_data(config_filepath)
def _load_config_data(self, filepath):
config = configparser.ConfigParser()
config.read(filepath)
self.email = config['LOGIN']['email']
self.password = config['LOGIN']['password']
self.owner_ID = config['OWNER']['id']
self.twitch_ID = config['TWITCH']['client_id']
self.carbon_key = config['CARBON']['key']
def run(self):
pass
|
import asyncio
import configparser
import discord
from discord.ext import commands
class Bot(commands.Bot):
def __init__(self, config_filepath, command_prefix):
super().__init__(command_prefix)
self._load_config_data(config_filepath)
def _load_config_data(self, filepath):
config = configparser.ConfigParser()
config.read(filepath)
self.email = config['LOGIN']['email']
self.password = config['LOGIN']['password']
self.owner_ID = config['OWNER']['id']
self.twitch_ID = config['TWITCH']['client_id']
self.carbon_key = config['CARBON']['key']
async def on_ready(self):
print("Logged in as {}".format(self.user.name))
print("User ID: {}".format(self.user.id))
print("Library: {} - {}".format(discord.__title__, discord.__version__))
def run(self):
try:
self.loop.run_until_complete(self.start(self.email, self.password))
except KeyboardInterrupt:
self.loop.run_until_complete(self.logout())
finally:
self.loop.close()
|
Add Bot.run and an on_ready message
|
Add Bot.run and an on_ready message
|
Python
|
mit
|
MagiChau/ZonBot
|
---
+++
@@ -17,5 +17,15 @@
self.twitch_ID = config['TWITCH']['client_id']
self.carbon_key = config['CARBON']['key']
+ async def on_ready(self):
+ print("Logged in as {}".format(self.user.name))
+ print("User ID: {}".format(self.user.id))
+ print("Library: {} - {}".format(discord.__title__, discord.__version__))
+
def run(self):
- pass
+ try:
+ self.loop.run_until_complete(self.start(self.email, self.password))
+ except KeyboardInterrupt:
+ self.loop.run_until_complete(self.logout())
+ finally:
+ self.loop.close()
|
ee2ed250fdf42d0e4616f0e783ff1ace0a201514
|
scripts/spat_conlisk_univariate_fig2.py
|
scripts/spat_conlisk_univariate_fig2.py
|
"""
Purpose: to recreate the univariate pdfs of Conlisk et al. (2007)
in figure 2. This provides a test that the univariate pdfs are
working correctly
"""
import numpy as np
import mete
import csv
n0 = 617
c = 256
sing_pdfs = np.zeros((4, 7))
psi = [0.01, 0.25, 0.5, 0.75]
for i in range(0, len(psi)):
sing_pdfs[i, : ] = [mete.single_prob(n, n0, psi[i], c) for n in range(0, 7)]
writer = open('./data/conlisk_data_fig2a.csv', 'wb')
datawriter = csv.writer(writer)
for i in range(0, np.shape(sing_pdfs)[0]):
datawriter.writerow(sing_pdfs[i, ])
writer.close()
|
"""
Purpose: to recreate the univariate pdfs of Conlisk et al. (2007)
in figure 2. This provides a test that the univariate pdfs are
working correctly
"""
import numpy as np
import mete
import matplotlib.pyplot as plt
n0 = 617
c = 256
sing_pdfs = np.zeros((4, 8))
psi = [0.01, 0.25, 0.5, 0.75]
for i in range(0, len(psi)):
sing_pdfs[i, : ] = [mete.single_prob(n, n0, psi[i], c) for n in range(0, 8)]
n = range(0, 8)
plt.plot(n, sing_pdfs[0, :], color='black', linewidth=1)
plt.plot(n, sing_pdfs[1, :], color='black', linewidth=1, ls='--')
plt.plot(n, sing_pdfs[2, :], color='black', linewidth=2)
plt.plot(n, sing_pdfs[3, :], color='lightgrey', linewidth=2)
plt.axis([0, 7, 0, 0.9])
plt.xlabel('n')
plt.ylabel('P(n)')
plt.legend(['psi = 0.01', 'psi = 0.25', 'psi = 0.50','psi = 0.75'])
plt.savefig('../figs/conlisk_univaritate_fig2a.pdf')
|
Create a pdf of the univariate pdfs instead of writing to a .csv file.
|
Create a pdf of the univariate pdfs instead of writing to a .csv file.
|
Python
|
mit
|
weecology/mete-spatial,weecology/mete-spatial,weecology/mete-spatial,weecology/mete-spatial
|
---
+++
@@ -5,21 +5,26 @@
"""
import numpy as np
import mete
-import csv
+import matplotlib.pyplot as plt
n0 = 617
c = 256
-sing_pdfs = np.zeros((4, 7))
+sing_pdfs = np.zeros((4, 8))
psi = [0.01, 0.25, 0.5, 0.75]
for i in range(0, len(psi)):
- sing_pdfs[i, : ] = [mete.single_prob(n, n0, psi[i], c) for n in range(0, 7)]
+ sing_pdfs[i, : ] = [mete.single_prob(n, n0, psi[i], c) for n in range(0, 8)]
-writer = open('./data/conlisk_data_fig2a.csv', 'wb')
-datawriter = csv.writer(writer)
-
-for i in range(0, np.shape(sing_pdfs)[0]):
- datawriter.writerow(sing_pdfs[i, ])
+n = range(0, 8)
-writer.close()
+plt.plot(n, sing_pdfs[0, :], color='black', linewidth=1)
+plt.plot(n, sing_pdfs[1, :], color='black', linewidth=1, ls='--')
+plt.plot(n, sing_pdfs[2, :], color='black', linewidth=2)
+plt.plot(n, sing_pdfs[3, :], color='lightgrey', linewidth=2)
+plt.axis([0, 7, 0, 0.9])
+plt.xlabel('n')
+plt.ylabel('P(n)')
+plt.legend(['psi = 0.01', 'psi = 0.25', 'psi = 0.50','psi = 0.75'])
+plt.savefig('../figs/conlisk_univaritate_fig2a.pdf')
+
|
54e715f26ed62e62e8794d8084110091c8db580b
|
oauth_provider/utils.py
|
oauth_provider/utils.py
|
import oauth.oauth as oauth
from django.conf import settings
from django.http import HttpResponse
from stores import DataStore
OAUTH_REALM_KEY_NAME = 'OAUTH_REALM_KEY_NAME'
def initialize_server_request(request):
"""Shortcut for initialization."""
oauth_request = oauth.OAuthRequest.from_request(request.method,
request.build_absolute_uri(),
headers=request.META,
parameters=dict(request.REQUEST.items()),
query_string=request.environ.get('QUERY_STRING', ''))
if oauth_request:
oauth_server = oauth.OAuthServer(DataStore(oauth_request))
oauth_server.add_signature_method(oauth.OAuthSignatureMethod_PLAINTEXT())
oauth_server.add_signature_method(oauth.OAuthSignatureMethod_HMAC_SHA1())
else:
oauth_server = None
return oauth_server, oauth_request
def send_oauth_error(err=None):
"""Shortcut for sending an error."""
# send a 401 error
response = HttpResponse(err.message.encode('utf-8'), mimetype="text/plain")
response.status_code = 401
# return the authenticate header
realm = getattr(settings, OAUTH_REALM_KEY_NAME, '')
header = oauth.build_authenticate_header(realm=realm)
for k, v in header.iteritems():
response[k] = v
return response
|
import oauth.oauth as oauth
from django.conf import settings
from django.http import HttpResponse
from stores import DataStore
OAUTH_REALM_KEY_NAME = 'OAUTH_REALM_KEY_NAME'
def initialize_server_request(request):
"""Shortcut for initialization."""
# Django converts Authorization header in HTTP_AUTHORIZATION
# Warning: it doesn't happen in tests but it's useful, do not remove!
auth_header = {}
if 'Authorization' in request.META:
auth_header = {'Authorization': request.META['Authorization']}
elif 'HTTP_AUTHORIZATION' in request.META:
auth_header = {'Authorization': request.META['HTTP_AUTHORIZATION']}
oauth_request = oauth.OAuthRequest.from_request(request.method,
request.build_absolute_uri(),
headers=auth_header,
parameters=dict(request.REQUEST.items()),
query_string=request.environ.get('QUERY_STRING', ''))
if oauth_request:
oauth_server = oauth.OAuthServer(DataStore(oauth_request))
oauth_server.add_signature_method(oauth.OAuthSignatureMethod_PLAINTEXT())
oauth_server.add_signature_method(oauth.OAuthSignatureMethod_HMAC_SHA1())
else:
oauth_server = None
return oauth_server, oauth_request
def send_oauth_error(err=None):
"""Shortcut for sending an error."""
# send a 401 error
response = HttpResponse(err.message.encode('utf-8'), mimetype="text/plain")
response.status_code = 401
# return the authenticate header
realm = getattr(settings, OAUTH_REALM_KEY_NAME, '')
header = oauth.build_authenticate_header(realm=realm)
for k, v in header.iteritems():
response[k] = v
return response
|
Fix a bug introduced in the latest revision, testing auth header in initialize_server_request now, thanks Chris McMichael for the report and patch
|
Fix a bug introduced in the latest revision, testing auth header in initialize_server_request now, thanks Chris McMichael for the report and patch
|
Python
|
bsd-3-clause
|
lukegb/django-oauth-plus,amrox/django-oauth-plus
|
---
+++
@@ -9,9 +9,17 @@
def initialize_server_request(request):
"""Shortcut for initialization."""
+ # Django converts Authorization header in HTTP_AUTHORIZATION
+ # Warning: it doesn't happen in tests but it's useful, do not remove!
+ auth_header = {}
+ if 'Authorization' in request.META:
+ auth_header = {'Authorization': request.META['Authorization']}
+ elif 'HTTP_AUTHORIZATION' in request.META:
+ auth_header = {'Authorization': request.META['HTTP_AUTHORIZATION']}
+
oauth_request = oauth.OAuthRequest.from_request(request.method,
request.build_absolute_uri(),
- headers=request.META,
+ headers=auth_header,
parameters=dict(request.REQUEST.items()),
query_string=request.environ.get('QUERY_STRING', ''))
if oauth_request:
|
73fe535416dbf744752d745f0186d5406bd15d8c
|
test/client/local_recognizer_test.py
|
test/client/local_recognizer_test.py
|
import unittest
import os
from speech_recognition import WavFile
from mycroft.client.speech.listener import RecognizerLoop
__author__ = 'seanfitz'
DATA_DIR = os.path.join(os.path.abspath(os.path.dirname(__file__)), "data")
class LocalRecognizerTest(unittest.TestCase):
def setUp(self):
self.recognizer = RecognizerLoop.create_mycroft_recognizer(16000,
"en-us")
def testRecognizerWrapper(self):
source = WavFile(os.path.join(DATA_DIR, "hey_mycroft.wav"))
with source as audio:
hyp = self.recognizer.transcribe(audio.stream.read())
assert self.recognizer.key_phrase in hyp.hypstr.lower()
source = WavFile(os.path.join(DATA_DIR, "mycroft.wav"))
with source as audio:
hyp = self.recognizer.transcribe(audio.stream.read())
assert self.recognizer.key_phrase in hyp.hypstr.lower()
def testRecognitionInLongerUtterance(self):
source = WavFile(os.path.join(DATA_DIR, "weather_mycroft.wav"))
with source as audio:
hyp = self.recognizer.transcribe(audio.stream.read())
assert self.recognizer.key_phrase in hyp.hypstr.lower()
|
import unittest
import os
from speech_recognition import WavFile
from mycroft.client.speech.listener import RecognizerLoop
__author__ = 'seanfitz'
DATA_DIR = os.path.join(os.path.abspath(os.path.dirname(__file__)), "data")
class LocalRecognizerTest(unittest.TestCase):
def setUp(self):
rl = RecognizerLoop()
self.recognizer = RecognizerLoop.create_mycroft_recognizer(rl,
16000,
"en-us")
def testRecognizerWrapper(self):
source = WavFile(os.path.join(DATA_DIR, "hey_mycroft.wav"))
with source as audio:
hyp = self.recognizer.transcribe(audio.stream.read())
assert self.recognizer.key_phrase in hyp.hypstr.lower()
source = WavFile(os.path.join(DATA_DIR, "mycroft.wav"))
with source as audio:
hyp = self.recognizer.transcribe(audio.stream.read())
assert self.recognizer.key_phrase in hyp.hypstr.lower()
def testRecognitionInLongerUtterance(self):
source = WavFile(os.path.join(DATA_DIR, "weather_mycroft.wav"))
with source as audio:
hyp = self.recognizer.transcribe(audio.stream.read())
assert self.recognizer.key_phrase in hyp.hypstr.lower()
|
Fix init of local recognizer
|
Fix init of local recognizer
|
Python
|
apache-2.0
|
linuxipho/mycroft-core,aatchison/mycroft-core,MycroftAI/mycroft-core,aatchison/mycroft-core,forslund/mycroft-core,linuxipho/mycroft-core,Dark5ide/mycroft-core,Dark5ide/mycroft-core,MycroftAI/mycroft-core,forslund/mycroft-core
|
---
+++
@@ -12,7 +12,9 @@
class LocalRecognizerTest(unittest.TestCase):
def setUp(self):
- self.recognizer = RecognizerLoop.create_mycroft_recognizer(16000,
+ rl = RecognizerLoop()
+ self.recognizer = RecognizerLoop.create_mycroft_recognizer(rl,
+ 16000,
"en-us")
def testRecognizerWrapper(self):
|
eb06e85c7dcb93febe22d20cd7e3e694939449ba
|
tests/test_xelatex.py
|
tests/test_xelatex.py
|
from latex.build import LatexMkBuilder
def test_xelatex():
min_latex = r"""
\documentclass{article}
\begin{document}
Hello, world!
\end{document}
"""
builder = LatexMkBuilder(variant='xelatex')
pdf = builder.build_pdf(min_latex)
assert pdf
|
from latex.build import LatexMkBuilder
def test_xelatex():
# the example below should not compile on pdflatex, but on xelatex
min_latex = r"""
\documentclass[12pt]{article}
\usepackage{fontspec}
\setmainfont{Times New Roman}
\title{Sample font document}
\author{Hubert Farnsworth}
\date{this month, 2014}
\begin{document}
\maketitle
This an \textit{example} of document compiled
with \textbf{xelatex} compiler. LuaLaTeX should
work fine also.
\end{document}
"""
builder = LatexMkBuilder(variant='xelatex')
pdf = builder.build_pdf(min_latex)
assert pdf
|
Make XeLaTeX test harder (to actually test xelatex being used).
|
Make XeLaTeX test harder (to actually test xelatex being used).
|
Python
|
bsd-3-clause
|
mbr/latex
|
---
+++
@@ -2,10 +2,25 @@
def test_xelatex():
+ # the example below should not compile on pdflatex, but on xelatex
min_latex = r"""
-\documentclass{article}
+\documentclass[12pt]{article}
+\usepackage{fontspec}
+
+\setmainfont{Times New Roman}
+
+ \title{Sample font document}
+ \author{Hubert Farnsworth}
+ \date{this month, 2014}
+
\begin{document}
-Hello, world!
+
+ \maketitle
+
+ This an \textit{example} of document compiled
+ with \textbf{xelatex} compiler. LuaLaTeX should
+ work fine also.
+
\end{document}
"""
builder = LatexMkBuilder(variant='xelatex')
|
d95d71f996483bdde4f0b27d9d9c023aef706c65
|
freebasics/tests/test_env_variables.py
|
freebasics/tests/test_env_variables.py
|
from django.test import TestCase, RequestFactory
from molo.core.tests.base import MoloTestCaseMixin
from freebasics.views import HomeView
from freebasics.templatetags import freebasics_tags
class EnvTestCase(TestCase, MoloTestCaseMixin):
def setUp(self):
self.mk_main()
def test_block_ordering(self):
with self.settings(BLOCK_POSITION_BANNER=4,
BLOCK_POSITION_LATEST=3,
BLOCK_POSITION_QUESTIONS=2,
BLOCK_POSITION_SECTIONS=1):
factory = RequestFactory()
request = factory.get('/')
request.site = self.site
home = HomeView()
home.request = request
context = home.get_context_data()
self.assertEquals(context['blocks'][0], (
'blocks/sections.html', 1))
self.assertEquals(context['blocks'][1], (
'blocks/questions.html', 2))
self.assertEquals(context['blocks'][2], ('blocks/latest.html', 3))
self.assertEquals(context['blocks'][3], ('blocks/banners.html', 4))
def test_css_vars(self):
with self.settings(CUSTOM_CSS_BLOCK_TEXT_TRANSFORM="lowercase",
CUSTOM_CSS_ACCENT_2="red"):
styles = freebasics_tags.custom_css(context='')
self.assertEquals(styles['accent_2'], 'red')
self.assertEquals(styles['text_transform'], 'lowercase')
|
from django.test import TestCase, RequestFactory
from molo.core.tests.base import MoloTestCaseMixin
from freebasics.views import HomeView
from freebasics.templatetags import freebasics_tags
class EnvTestCase(TestCase, MoloTestCaseMixin):
def setUp(self):
self.mk_main()
def test_block_ordering(self):
with self.settings(BLOCK_POSITION_BANNER=4,
BLOCK_POSITION_LATEST=3,
BLOCK_POSITION_QUESTIONS=2,
BLOCK_POSITION_SECTIONS=1):
factory = RequestFactory()
request = factory.get('/')
request.site = self.site
home = HomeView()
home.request = request
context = home.get_context_data()
self.assertEquals(context['blocks'][0], (
'blocks/sections.html', 1))
self.assertEquals(context['blocks'][1], (
'blocks/questions.html', 2))
self.assertEquals(context['blocks'][2], ('blocks/latest.html', 3))
self.assertEquals(context['blocks'][3], ('blocks/banners.html', 4))
def test_css_vars(self):
with self.settings(CUSTOM_CSS_BLOCK_TEXT_TRANSFORM="lowercase",
CUSTOM_CSS_ACCENT_2="red"):
styles = freebasics_tags.custom_css(context='')
self.assertEquals(styles['accent_2'], 'red')
self.assertEquals(styles['text_transform'], 'lowercase')
def test_custom_css(self):
response = self.client.get('/')
self.assertContains(response, '.fb-body .base-bcolor')
self.assertContains(response, '.fb-body .block-heading')
self.assertContains(response, '.section-nav__items')
|
Add test for custom css
|
Add test for custom css
|
Python
|
bsd-2-clause
|
praekelt/molo-freebasics,praekelt/molo-freebasics,praekelt/molo-freebasics,praekelt/molo-freebasics
|
---
+++
@@ -33,3 +33,9 @@
styles = freebasics_tags.custom_css(context='')
self.assertEquals(styles['accent_2'], 'red')
self.assertEquals(styles['text_transform'], 'lowercase')
+
+ def test_custom_css(self):
+ response = self.client.get('/')
+ self.assertContains(response, '.fb-body .base-bcolor')
+ self.assertContains(response, '.fb-body .block-heading')
+ self.assertContains(response, '.section-nav__items')
|
59c66d0e172b23ea7106a70866871d20bbcabe5b
|
timezones/__init__.py
|
timezones/__init__.py
|
import pytz
TIMEZONE_CHOICES = zip(pytz.all_timezones, pytz.all_timezones)
|
import pytz
TIMEZONE_CHOICES = zip(pytz.common_timezones, pytz.common_timezones)
|
Use common timezones and not all timezones pytz provides.
|
Use common timezones and not all timezones pytz provides.
git-svn-id: 13ea2b4cf383b32e0a7498d153ee882d068671f7@13 86ebb30f-654e-0410-bc0d-7bf82786d749
|
Python
|
bsd-2-clause
|
mfogel/django-timezone-field,brosner/django-timezones
|
---
+++
@@ -1,4 +1,4 @@
import pytz
-TIMEZONE_CHOICES = zip(pytz.all_timezones, pytz.all_timezones)
+TIMEZONE_CHOICES = zip(pytz.common_timezones, pytz.common_timezones)
|
8bd738972cebd27b068250bd52db8aacea6c7876
|
src/condor_tests/ornithology/plugin.py
|
src/condor_tests/ornithology/plugin.py
|
import pytest
from .scripts import SCRIPTS
# This module is loaded as a "plugin" by pytest by a setting in conftest.py
# Any fixtures defined here will be globally available in tests,
# as if they were defined in conftest.py itself.
@pytest.fixture(scope="session")
def path_to_sleep():
return SCRIPTS["sleep"]
|
import sys
import pytest
from .scripts import SCRIPTS
# This module is loaded as a "plugin" by pytest by a setting in conftest.py
# Any fixtures defined here will be globally available in tests,
# as if they were defined in conftest.py itself.
@pytest.fixture(scope="session")
def path_to_sleep():
return SCRIPTS["sleep"]
@pytest.fixture(scope="session")
def path_to_python():
return sys.executable
|
Add path_to_python fixture to make writing multiplatform job scripts easier.
|
Add path_to_python fixture to make writing multiplatform job scripts easier.
|
Python
|
apache-2.0
|
htcondor/htcondor,htcondor/htcondor,htcondor/htcondor,htcondor/htcondor,htcondor/htcondor,htcondor/htcondor,htcondor/htcondor,htcondor/htcondor
|
---
+++
@@ -1,3 +1,4 @@
+import sys
import pytest
from .scripts import SCRIPTS
@@ -9,3 +10,8 @@
@pytest.fixture(scope="session")
def path_to_sleep():
return SCRIPTS["sleep"]
+
+
+@pytest.fixture(scope="session")
+def path_to_python():
+ return sys.executable
|
775170d69862aaff63231b669639a872596ed2cd
|
test_interpreter.py
|
test_interpreter.py
|
import unittest
import brainfuck
test_cases = [("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")]
class InterpreterTestCase(unittest.TestCase):
def setUp(self):
self.interpreter = brainfuck.BrainfuckInterpreter()
def runTest(self):
for case in test_cases:
self.assertEqual(case[1], self.interpreter.eval(case[0]))
|
import unittest
import brainfuck
hello_case = ("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")
class InterpreterTestCase(unittest.TestCase):
def setUp(self):
self.interpreter = brainfuck.BrainfuckInterpreter()
def test_hello_world(self):
self.assertEqual(hello_case[1], self.interpreter.eval(hello_case[0]))
def test_missing_parenthesis(self):
self.assertRaises(SyntaxError, self.interpreter.eval, '[++]+]')
|
Add unittest for missing parenthesis
|
Add unittest for missing parenthesis
|
Python
|
bsd-3-clause
|
handrake/brainfuck
|
---
+++
@@ -1,11 +1,12 @@
import unittest
import brainfuck
-test_cases = [("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")]
+hello_case = ("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")
class InterpreterTestCase(unittest.TestCase):
def setUp(self):
self.interpreter = brainfuck.BrainfuckInterpreter()
- def runTest(self):
- for case in test_cases:
- self.assertEqual(case[1], self.interpreter.eval(case[0]))
+ def test_hello_world(self):
+ self.assertEqual(hello_case[1], self.interpreter.eval(hello_case[0]))
+ def test_missing_parenthesis(self):
+ self.assertRaises(SyntaxError, self.interpreter.eval, '[++]+]')
|
131454ffaec010442c17f748365ab491668d947f
|
plumeria/plugins/bing_images.py
|
plumeria/plugins/bing_images.py
|
from aiohttp import BasicAuth
from plumeria import config, scoped_config
from plumeria.command import commands, CommandError
from plumeria.config.common import nsfw
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "images", "i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image and returns a URL to that image.
Example::
/image socially awkward penguin
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Adult', "'Off'" if scoped_config.get(nsfw, message.channel) else "'Strict'"),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
|
import random
from aiohttp import BasicAuth
from plumeria import config, scoped_config
from plumeria.command import commands, CommandError
from plumeria.config.common import nsfw
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "images", "i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image and returns a URL to that image.
Example::
/image socially awkward penguin
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '20'),
('Adult', "'Off'" if scoped_config.get(nsfw, message.channel) else "'Strict'"),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(random.choice(data['results'])['MediaUrl'])
else:
raise CommandError("no results found")
|
Make the Bing images search pick a random top 20 image.
|
Make the Bing images search pick a random top 20 image.
|
Python
|
mit
|
sk89q/Plumeria,sk89q/Plumeria,sk89q/Plumeria
|
---
+++
@@ -1,3 +1,5 @@
+import random
+
from aiohttp import BasicAuth
from plumeria import config, scoped_config
from plumeria.command import commands, CommandError
@@ -28,12 +30,12 @@
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
- ('$top', '10'),
+ ('$top', '20'),
('Adult', "'Off'" if scoped_config.get(nsfw, message.channel) else "'Strict'"),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
- return Response(data['results'][0]['MediaUrl'])
+ return Response(random.choice(data['results'])['MediaUrl'])
else:
raise CommandError("no results found")
|
f7caec9d0c0058b5d760992172b434b461f70d90
|
molecule/default/tests/test_default.py
|
molecule/default/tests/test_default.py
|
import os
import pytest
import testinfra.utils.ansible_runner
testinfra_hosts = testinfra.utils.ansible_runner.AnsibleRunner(
os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all')
def test_docker_service_enabled(host):
service = host.service('docker')
assert service.is_enabled
def test_docker_service_running(host):
service = host.service('docker')
assert service.is_running
@pytest.mark.parametrize('socket_def', [
# listening on localhost, for tcp sockets on port 1883 (MQQT)
('tcp://127.0.0.1:1883'),
# all IPv4 tcp sockets on port 8883 (MQQTS)
('tcp://8883'),
])
def test_listening_sockets(host, socket_def):
socket = host.socket(socket_def)
assert socket.is_listening
# Tests to write:
# - using the localhost listener:
# - verify that a message can be published
# - verify that a published message can be subscribed to
|
import os
import pytest
import testinfra.utils.ansible_runner
testinfra_hosts = testinfra.utils.ansible_runner.AnsibleRunner(
os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all')
def test_docker_service_enabled(host):
service = host.service('docker')
assert service.is_enabled
def test_docker_service_running(host):
service = host.service('docker')
assert service.is_running
@pytest.mark.parametrize('socket_def', [
# all IPv4 tcp sockets on port 8883 (MQQTS)
('tcp://8883'),
])
def test_listening_sockets(host, socket_def):
socket = host.socket(socket_def)
assert socket.is_listening
# Tests to write:
# - using the localhost listener:
# - verify that a message can be published
# - verify that a published message can be subscribed to
|
Remove test enforcing listening on port 1883
|
Remove test enforcing listening on port 1883
|
Python
|
mit
|
triplepoint/ansible-mosquitto
|
---
+++
@@ -17,8 +17,6 @@
@pytest.mark.parametrize('socket_def', [
- # listening on localhost, for tcp sockets on port 1883 (MQQT)
- ('tcp://127.0.0.1:1883'),
# all IPv4 tcp sockets on port 8883 (MQQTS)
('tcp://8883'),
])
|
3dc06581d07a204a3044e3a78deb84950a6ebf74
|
mtp_transaction_uploader/api_client.py
|
mtp_transaction_uploader/api_client.py
|
from urllib.parse import urljoin
from oauthlib.oauth2 import LegacyApplicationClient
from requests_oauthlib import OAuth2Session
import slumber
from . import settings
REQUEST_TOKEN_URL = urljoin(settings.API_URL, '/oauth2/token/')
def get_authenticated_connection():
"""
Returns:
an authenticated slumber connection
"""
session = OAuth2Session(
client=LegacyApplicationClient(
client_id=settings.API_CLIENT_ID
)
)
session.fetch_token(
token_url=REQUEST_TOKEN_URL,
username=settings.API_USERNAME,
password=settings.API_PASSWORD,
client_id=settings.API_CLIENT_ID,
client_secret=settings.API_CLIENT_SECRET
)
return slumber.API(
base_url=settings.API_URL, session=session
)
|
from urllib.parse import urljoin
from oauthlib.oauth2 import LegacyApplicationClient
from requests.auth import HTTPBasicAuth
from requests_oauthlib import OAuth2Session
import slumber
from . import settings
REQUEST_TOKEN_URL = urljoin(settings.API_URL, '/oauth2/token/')
def get_authenticated_connection():
"""
Returns:
an authenticated slumber connection
"""
session = OAuth2Session(
client=LegacyApplicationClient(
client_id=settings.API_CLIENT_ID
)
)
session.fetch_token(
token_url=REQUEST_TOKEN_URL,
username=settings.API_USERNAME,
password=settings.API_PASSWORD,
auth=HTTPBasicAuth(settings.API_CLIENT_ID, settings.API_CLIENT_SECRET)
)
return slumber.API(
base_url=settings.API_URL, session=session
)
|
Use HTTPBasicAuth when connecting to the API
|
Use HTTPBasicAuth when connecting to the API
|
Python
|
mit
|
ministryofjustice/money-to-prisoners-transaction-uploader
|
---
+++
@@ -1,6 +1,7 @@
from urllib.parse import urljoin
from oauthlib.oauth2 import LegacyApplicationClient
+from requests.auth import HTTPBasicAuth
from requests_oauthlib import OAuth2Session
import slumber
@@ -24,8 +25,7 @@
token_url=REQUEST_TOKEN_URL,
username=settings.API_USERNAME,
password=settings.API_PASSWORD,
- client_id=settings.API_CLIENT_ID,
- client_secret=settings.API_CLIENT_SECRET
+ auth=HTTPBasicAuth(settings.API_CLIENT_ID, settings.API_CLIENT_SECRET)
)
return slumber.API(
|
e6933a46086f83913de307fb8803ccbfd3c55114
|
mysite/mysite/tests/test_middleware.py
|
mysite/mysite/tests/test_middleware.py
|
from django.contrib.auth.models import User
from django.test import TestCase
from DjangoLibrary.middleware import FactoryBoyMiddleware
from mock import Mock
import json
class TestFactoryBoyMiddleware(TestCase):
def setUp(self):
self.middleware = FactoryBoyMiddleware()
self.request = Mock()
self.request.session = {}
def test_process_request_creates_object(self):
self.request.configure_mock(
**{
'GET': {
'FACTORY_BOY_MODEL_PATH': 'mysite.tests.factories.UserFactory', # noqa
'FACTORY_BOY_ARGS': ''
}
}
)
response = self.middleware.process_request(self.request)
self.assertEqual(201, response.status_code)
self.assertEqual(
'johndoe',
json.loads(response.content).get('username')
)
self.assertEqual(1, len(User.objects.values()))
self.assertEqual('johndoe', User.objects.values()[0]['username'])
|
from django.contrib.auth.models import User
from django.test import TestCase
from DjangoLibrary.middleware import FactoryBoyMiddleware
from mock import Mock
import json
class TestFactoryBoyMiddleware(TestCase):
def setUp(self):
self.middleware = FactoryBoyMiddleware()
self.request = Mock()
self.request.session = {}
def test_process_request_creates_object(self):
self.request.configure_mock(
**{
'GET': {
'FACTORY_BOY_MODEL_PATH': 'mysite.tests.factories.UserFactory', # noqa
'FACTORY_BOY_ARGS': ''
}
}
)
response = self.middleware.process_request(self.request)
self.assertEqual(201, response.status_code)
self.assertEqual(
'johndoe',
json.loads(response.content.decode('utf-8')).get('username')
)
self.assertEqual(1, len(User.objects.values()))
self.assertEqual('johndoe', User.objects.values()[0]['username'])
|
Fix integration tests for Python 3.
|
Fix integration tests for Python 3.
|
Python
|
apache-2.0
|
kitconcept/robotframework-djangolibrary
|
---
+++
@@ -28,7 +28,7 @@
self.assertEqual(201, response.status_code)
self.assertEqual(
'johndoe',
- json.loads(response.content).get('username')
+ json.loads(response.content.decode('utf-8')).get('username')
)
self.assertEqual(1, len(User.objects.values()))
self.assertEqual('johndoe', User.objects.values()[0]['username'])
|
47e5fdce2d248e6bf78addcb46e1a8f12fcc07d6
|
tests/app/main/test_form_validators.py
|
tests/app/main/test_form_validators.py
|
import mock
import pytest
from flask.ext.wtf import Form
from wtforms.fields.core import Field
from wtforms.validators import StopValidation
from app.main.forms import AdminEmailAddressValidator
@mock.patch('app.main.forms.data_api_client')
class TestAdminEmailAddressValidator(object):
def setup_method(self):
self.form_mock = mock.MagicMock(Form)
self.field_mock = mock.MagicMock(Field, data='the_email_address')
self.validator = AdminEmailAddressValidator(message='The message passed to validator')
def test_admin_email_address_validator_calls_api(self, data_api_client):
self.validator(self.form_mock, self.field_mock)
data_api_client.email_is_valid_for_admin_user.assert_called_once_with('the_email_address')
def test_admin_email_address_validator_raises_with_invalid_response(self, data_api_client):
data_api_client.email_is_valid_for_admin_user.return_value = False
with pytest.raises(StopValidation, match='The message passed to validator'):
self.validator(self.form_mock, self.field_mock)
def test_admin_email_address_validator_passes_with_valid_response(self, data_api_client):
data_api_client.email_is_valid_for_admin_user.return_value = True
assert self.validator(self.form_mock, self.field_mock) is None
|
import mock
import pytest
from flask_wtf import Form
from wtforms.fields.core import Field
from wtforms.validators import StopValidation
from app.main.forms import AdminEmailAddressValidator
@mock.patch('app.main.forms.data_api_client')
class TestAdminEmailAddressValidator(object):
def setup_method(self):
self.form_mock = mock.MagicMock(Form)
self.field_mock = mock.MagicMock(Field, data='the_email_address')
self.validator = AdminEmailAddressValidator(message='The message passed to validator')
def test_admin_email_address_validator_calls_api(self, data_api_client):
self.validator(self.form_mock, self.field_mock)
data_api_client.email_is_valid_for_admin_user.assert_called_once_with('the_email_address')
def test_admin_email_address_validator_raises_with_invalid_response(self, data_api_client):
data_api_client.email_is_valid_for_admin_user.return_value = False
with pytest.raises(StopValidation, match='The message passed to validator'):
self.validator(self.form_mock, self.field_mock)
def test_admin_email_address_validator_passes_with_valid_response(self, data_api_client):
data_api_client.email_is_valid_for_admin_user.return_value = True
assert self.validator(self.form_mock, self.field_mock) is None
|
Tidy up test imports a bit
|
Tidy up test imports a bit
|
Python
|
mit
|
alphagov/digitalmarketplace-admin-frontend,alphagov/digitalmarketplace-admin-frontend,alphagov/digitalmarketplace-admin-frontend,alphagov/digitalmarketplace-admin-frontend
|
---
+++
@@ -1,7 +1,6 @@
import mock
import pytest
-
-from flask.ext.wtf import Form
+from flask_wtf import Form
from wtforms.fields.core import Field
from wtforms.validators import StopValidation
|
e68c38428c055f7c001011c6cc325593d2a26a81
|
pyFxTrader/strategy/__init__.py
|
pyFxTrader/strategy/__init__.py
|
# -*- coding: utf-8 -*-
class Strategy(object):
TIMEFRAMES = [] # e.g. ['M30', 'H2']
def __init__(self, instrument):
self.instrument = instrument
if not self.TIMEFRAMES:
raise ValueError('Please define TIMEFRAMES variable.')
def start(self):
"""Called on strategy start."""
raise NotImplementedError()
def new_bar(self, instrument, cur_index):
"""Called on every bar of every instrument that client is subscribed on."""
raise NotImplementedError()
def execute(self, engine, instruments, cur_index):
"""Called on after all indicators have been updated for this bar's index"""
raise NotImplementedError()
def end(self, engine):
"""Called on strategy stop."""
raise NotImplementedError()
|
# -*- coding: utf-8 -*-
from collections import deque
from logbook import Logger
log = Logger('pyFxTrader')
class Strategy(object):
TIMEFRAMES = [] # e.g. ['M30', 'H2']
BUFFER_SIZE = 500
feeds = {}
def __init__(self, instrument):
self.instrument = instrument
if not self.TIMEFRAMES:
raise ValueError('Please define TIMEFRAMES variable.')
for tf in self.TIMEFRAMES:
self.feeds[tf] = deque(maxlen=self.BUFFER_SIZE)
log.info('Initialized %s feed for %s' % (tf, self.instrument))
def start(self):
"""Called on strategy start."""
raise NotImplementedError()
def new_bar(self, instrument, cur_index):
"""Called on every bar of every instrument that client is subscribed on."""
raise NotImplementedError()
def execute(self, engine, instruments, cur_index):
"""Called on after all indicators have been updated for this bar's index"""
raise NotImplementedError()
def end(self, engine):
"""Called on strategy stop."""
raise NotImplementedError()
|
Add default BUFFER_SIZE for feeds
|
Add default BUFFER_SIZE for feeds
|
Python
|
mit
|
jmelett/pyfx,jmelett/pyFxTrader,jmelett/pyfx
|
---
+++
@@ -1,12 +1,27 @@
# -*- coding: utf-8 -*-
+
+from collections import deque
+
+from logbook import Logger
+
+log = Logger('pyFxTrader')
+
class Strategy(object):
TIMEFRAMES = [] # e.g. ['M30', 'H2']
+ BUFFER_SIZE = 500
+
+ feeds = {}
def __init__(self, instrument):
self.instrument = instrument
+
if not self.TIMEFRAMES:
raise ValueError('Please define TIMEFRAMES variable.')
+
+ for tf in self.TIMEFRAMES:
+ self.feeds[tf] = deque(maxlen=self.BUFFER_SIZE)
+ log.info('Initialized %s feed for %s' % (tf, self.instrument))
def start(self):
"""Called on strategy start."""
|
73d9049dea55ddfa32e4cb09f969b6ff083fee2c
|
tests/redisdl_test.py
|
tests/redisdl_test.py
|
import redisdl
import unittest
import json
import os.path
class RedisdlTest(unittest.TestCase):
def setUp(self):
import redis
self.r = redis.Redis()
for key in self.r.keys('*'):
self.r.delete(key)
def test_roundtrip(self):
path = os.path.join(os.path.dirname(__file__), 'fixtures', 'dump.json')
with open(path) as f:
dump = f.read()
redisdl.loads(dump)
redump = redisdl.dumps()
expected = json.loads(dump)
actual = json.loads(redump)
self.assertEqual(expected, actual)
|
import redisdl
import unittest
import json
import os.path
class RedisdlTest(unittest.TestCase):
def setUp(self):
import redis
self.r = redis.Redis()
for key in self.r.keys('*'):
self.r.delete(key)
def test_roundtrip(self):
path = os.path.join(os.path.dirname(__file__), 'fixtures', 'dump.json')
with open(path) as f:
dump = f.read()
redisdl.loads(dump)
redump = redisdl.dumps()
expected = json.loads(dump)
actual = json.loads(redump)
self.assertEqual(expected, actual)
def test_dump_string_value(self):
self.r.set('key', 'value')
dump = redisdl.dumps()
actual = json.loads(dump)
expected = {'key': {'type': 'string', 'value': 'value'}}
self.assertEqual(expected, actual)
def test_dump_unicode_value(self):
self.r.set('key', u"\u041c\u043e\u0441\u043a\u0432\u0430")
dump = redisdl.dumps()
actual = json.loads(dump)
expected = {'key': {'type': 'string', 'value': u"\u041c\u043e\u0441\u043a\u0432\u0430"}}
self.assertEqual(expected, actual)
|
Add tests for dumping string and unicode values
|
Add tests for dumping string and unicode values
|
Python
|
bsd-2-clause
|
p/redis-dump-load,p/redis-dump-load,hyunchel/redis-dump-load,hyunchel/redis-dump-load
|
---
+++
@@ -21,5 +21,19 @@
expected = json.loads(dump)
actual = json.loads(redump)
-
+
self.assertEqual(expected, actual)
+
+ def test_dump_string_value(self):
+ self.r.set('key', 'value')
+ dump = redisdl.dumps()
+ actual = json.loads(dump)
+ expected = {'key': {'type': 'string', 'value': 'value'}}
+ self.assertEqual(expected, actual)
+
+ def test_dump_unicode_value(self):
+ self.r.set('key', u"\u041c\u043e\u0441\u043a\u0432\u0430")
+ dump = redisdl.dumps()
+ actual = json.loads(dump)
+ expected = {'key': {'type': 'string', 'value': u"\u041c\u043e\u0441\u043a\u0432\u0430"}}
+ self.assertEqual(expected, actual)
|
23a0db627060afc3e1563d298c733edd8bb106a1
|
src/ConfigLoader.py
|
src/ConfigLoader.py
|
import json
import sys
def load_config_file(out=sys.stdout):
default_filepath = "../resources/config/default-config.json"
user_filepath = "../resources/config/user-config.json"
try:
default_json = read_json(default_filepath)
user_json = read_json(user_filepath)
for property in user_json:
default_json[property] = user_json[property]
except FileNotFoundError as e:
out.write("Cannot find file: " + e.filename)
else:
out.write("Read styling config JSON correctly.")
return default_json
def read_json(filepath):
config_string = ''
with open(filepath) as f:
for line in f:
line = line.lstrip()
if not line.startswith("//"):
config_string += line
config_json = json.loads(config_string)
return config_json
if __name__ == "__main__":
load_config_file()
|
import json
import sys
def load_config_file(out=sys.stdout):
if sys.argv[0].endswith('nosetests'):
default_filepath = "./resources/config/default-config.json"
user_filepath = "./resources/config/user-config.json"
else:
default_filepath = "../resources/config/default-config.json"
user_filepath = "../resources/config/user-config.json"
try:
default_json = read_json(default_filepath)
user_json = read_json(user_filepath)
for property in user_json:
default_json[property] = user_json[property]
except FileNotFoundError as e:
out.write("Cannot find file: " + e.filename)
else:
out.write("Read styling config JSON correctly.")
return default_json
def read_json(filepath):
config_string = ''
with open(filepath) as f:
for line in f:
line = line.lstrip()
if not line.startswith("//"):
config_string += line
config_json = json.loads(config_string)
return config_json
if __name__ == "__main__":
load_config_file()
|
Fix nosetests for config file loading
|
Fix nosetests for config file loading
|
Python
|
bsd-3-clause
|
sky-uk/bslint
|
---
+++
@@ -3,8 +3,12 @@
def load_config_file(out=sys.stdout):
- default_filepath = "../resources/config/default-config.json"
- user_filepath = "../resources/config/user-config.json"
+ if sys.argv[0].endswith('nosetests'):
+ default_filepath = "./resources/config/default-config.json"
+ user_filepath = "./resources/config/user-config.json"
+ else:
+ default_filepath = "../resources/config/default-config.json"
+ user_filepath = "../resources/config/user-config.json"
try:
default_json = read_json(default_filepath)
|
4fb39abc5afef5b0ca87e5c3b40e3dc9c9c0b2ef
|
tests/functions_tests/test_accuracy.py
|
tests/functions_tests/test_accuracy.py
|
import unittest
import numpy
import six
import chainer
from chainer import cuda
from chainer import gradient_check
from chainer.testing import attr
if cuda.available:
cuda.init()
class TestAccuracy(unittest.TestCase):
def setUp(self):
self.x = numpy.random.uniform(-1, 1, (10, 3)).astype(numpy.float32)
self.t = numpy.random.randint(3, size=(10,)).astype(numpy.int32)
def check_forward(self, x_data, t_data):
x = chainer.Variable(x_data)
t = chainer.Variable(t_data)
y = chainer.functions.accuracy(x, t)
count = 0
for i in six.moves.range(self.t.size):
pred = self.x[i].argmax()
if pred == self.t[i]:
count += 1
expected = float(count) / self.t.size
gradient_check.assert_allclose(expected, cuda.to_cpu(y.data))
def test_forward_cpu(self):
self.check_forward(self.x, self.t)
@attr.gpu
def test_forward_gpu(self):
self.check_forward(cuda.to_gpu(self.x), cuda.to_gpu(self.t))
|
import unittest
import numpy
import six
import chainer
from chainer import cuda
from chainer import gradient_check
from chainer.testing import attr
if cuda.available:
cuda.init()
class TestAccuracy(unittest.TestCase):
def setUp(self):
self.x = numpy.random.uniform(-1, 1, (10, 3)).astype(numpy.float32)
self.t = numpy.random.randint(3, size=(10,)).astype(numpy.int32)
def check_forward(self, x_data, t_data):
x = chainer.Variable(x_data)
t = chainer.Variable(t_data)
y = chainer.functions.accuracy(x, t)
self.assertEqual((), y.data.shape)
count = 0
for i in six.moves.range(self.t.size):
pred = self.x[i].argmax()
if pred == self.t[i]:
count += 1
expected = float(count) / self.t.size
gradient_check.assert_allclose(expected, cuda.to_cpu(y.data))
def test_forward_cpu(self):
self.check_forward(self.x, self.t)
@attr.gpu
def test_forward_gpu(self):
self.check_forward(cuda.to_gpu(self.x), cuda.to_gpu(self.t))
|
Add test fot shape of result of accuracy function
|
Add test fot shape of result of accuracy function
|
Python
|
mit
|
elviswf/chainer,1986ks/chainer,cupy/cupy,ktnyt/chainer,hvy/chainer,tigerneil/chainer,rezoo/chainer,sou81821/chainer,tereka114/chainer,hvy/chainer,wavelets/chainer,jfsantos/chainer,chainer/chainer,ikasumi/chainer,okuta/chainer,aonotas/chainer,ysekky/chainer,kashif/chainer,ktnyt/chainer,niboshi/chainer,pfnet/chainer,ytoyama/yans_chainer_hackathon,cupy/cupy,jnishi/chainer,muupan/chainer,AlpacaDB/chainer,cemoody/chainer,chainer/chainer,anaruse/chainer,wkentaro/chainer,ronekko/chainer,kiyukuta/chainer,masia02/chainer,truongdq/chainer,kikusu/chainer,umitanuki/chainer,hidenori-t/chainer,cupy/cupy,t-abe/chainer,keisuke-umezawa/chainer,jnishi/chainer,jnishi/chainer,keisuke-umezawa/chainer,truongdq/chainer,jnishi/chainer,AlpacaDB/chainer,niboshi/chainer,laysakura/chainer,niboshi/chainer,okuta/chainer,benob/chainer,hvy/chainer,t-abe/chainer,keisuke-umezawa/chainer,ktnyt/chainer,kikusu/chainer,kuwa32/chainer,chainer/chainer,delta2323/chainer,wkentaro/chainer,woodshop/complex-chainer,tscohen/chainer,muupan/chainer,ktnyt/chainer,woodshop/chainer,wkentaro/chainer,bayerj/chainer,okuta/chainer,minhpqn/chainer,Kaisuke5/chainer,cupy/cupy,sinhrks/chainer,benob/chainer,okuta/chainer,wkentaro/chainer,hvy/chainer,tkerola/chainer,chainer/chainer,keisuke-umezawa/chainer,niboshi/chainer,sinhrks/chainer,yanweifu/chainer
|
---
+++
@@ -23,6 +23,7 @@
x = chainer.Variable(x_data)
t = chainer.Variable(t_data)
y = chainer.functions.accuracy(x, t)
+ self.assertEqual((), y.data.shape)
count = 0
for i in six.moves.range(self.t.size):
|
b458e34c0e6466afd3125fff6b1f36278572857b
|
pipes/iam/consts.py
|
pipes/iam/consts.py
|
"""Constant variables for Spinnaker IAM Pipe."""
API_URL = 'http://spinnaker.build.example.com:8084'
|
"""Constant variables for Spinnaker IAM Pipe."""
API_URL = 'http://gate-api.build.example.com:8084'
|
Update to use new gate api
|
Update to use new gate api
|
Python
|
apache-2.0
|
gogoair/foremast,gogoair/foremast
|
---
+++
@@ -1,2 +1,2 @@
"""Constant variables for Spinnaker IAM Pipe."""
-API_URL = 'http://spinnaker.build.example.com:8084'
+API_URL = 'http://gate-api.build.example.com:8084'
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.