commit stringlengths 40 40 | old_file stringlengths 4 150 | new_file stringlengths 4 150 | old_contents stringlengths 0 3.26k | new_contents stringlengths 1 4.43k | subject stringlengths 15 501 | message stringlengths 15 4.06k | lang stringclasses 4 values | license stringclasses 13 values | repos stringlengths 5 91.5k | diff stringlengths 0 4.35k |
|---|---|---|---|---|---|---|---|---|---|---|
a1de1f26d30a753c7e3ea66a600b04fb130b02ec | hellomama_registration/testsettings.py | hellomama_registration/testsettings.py | from hellomama_registration.settings import * # flake8: noqa
# SECURITY WARNING: keep the secret key used in production secret!
SECRET_KEY = 'TESTSEKRET'
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
TEMPLATE_DEBUG = True
CELERY_EAGER_PROPAGATES_EXCEPTIONS = True
CELERY_ALWAYS_EAGER = True
BROKER_BACKEND = 'memory'
CELERY_RESULT_BACKEND = 'djcelery.backends.database:DatabaseBackend'
METRICS_URL = "http://metrics-url"
METRICS_AUTH_TOKEN = "REPLACEME"
| from hellomama_registration.settings import * # flake8: noqa
# SECURITY WARNING: keep the secret key used in production secret!
SECRET_KEY = 'TESTSEKRET'
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
TEMPLATE_DEBUG = True
CELERY_EAGER_PROPAGATES_EXCEPTIONS = True
CELERY_ALWAYS_EAGER = True
BROKER_BACKEND = 'memory'
CELERY_RESULT_BACKEND = 'djcelery.backends.database:DatabaseBackend'
METRICS_URL = "http://metrics-url"
METRICS_AUTH_TOKEN = "REPLACEME"
PASSWORD_HASHERS = ('django.contrib.auth.hashers.MD5PasswordHasher',)
| Use MD5 password hasher for test runs | Use MD5 password hasher for test runs
| Python | bsd-3-clause | praekelt/hellomama-registration,praekelt/hellomama-registration | ---
+++
@@ -15,3 +15,5 @@
METRICS_URL = "http://metrics-url"
METRICS_AUTH_TOKEN = "REPLACEME"
+
+PASSWORD_HASHERS = ('django.contrib.auth.hashers.MD5PasswordHasher',) |
f7a9391c22a1b69bcd645a3efbe41f32e91e668c | hb_res/resources/ExplanationStorage.py | hb_res/resources/ExplanationStorage.py | __author__ = 'skird'
class ExplanationStorage:
"""
Interface of abstract readable/writeable resource
Every resource is a map from string to list of strings
It supports random access and provides iterator on its elements
"""
def entries(self):
return NotImplementedError
def add_entry(self, entry) -> None:
raise NotImplementedError
def clear(self) -> None:
raise NotImplementedError
def __getitem__(self, item):
raise NotImplementedError | __author__ = 'skird'
class ExplanationStorage:
"""
Interface of abstract readable/writeable resource
Every resource is a map from string to list of strings
It supports random access and provides iterator on its elements
"""
def entries(self):
raise NotImplementedError
def add_entry(self, entry) -> None:
raise NotImplementedError
def clear(self) -> None:
raise NotImplementedError
def __getitem__(self, item):
raise NotImplementedError
| Fix typo with return instead of raise | Fix typo with return instead of raise
| Python | mit | hatbot-team/hatbot_resources | ---
+++
@@ -9,7 +9,7 @@
"""
def entries(self):
- return NotImplementedError
+ raise NotImplementedError
def add_entry(self, entry) -> None:
raise NotImplementedError |
6629a3a238432522d77f840b465eb99a3745593f | django_base64field/tests/models.py | django_base64field/tests/models.py | from django.db import models
from django.utils.baseconv import base64
from django_base64field.fields import Base64Field
class Planet(models.Model):
ek = Base64Field()
name = models.CharField(
default='Fucker',
max_length=103
)
class Continent(models.Model):
ek = Base64Field()
name = models.CharField(
default='Suckers!',
max_length=13
)
planet = models.ForeignKey(Planet, to_field='ek')
class Helper(models.Model):
"""
base64 encoded value won't be available at first time creation.
It can ve accessible by getting the object from database after creation
mean when it get saved completely, But what if we don't want to get our base64
encoded key from our sweet model by retrieving it again from database?
It's easy, efficient, holly and molly!
"""
ek = Base64Field()
def _ek(self):
if self.ek: return self.ek
if not self.ek and self.pk:
return base64.encode(self.pk)
return self.ek
class CustomReceiver(models.Model):
"""
Passing custom receiver to generate `youyouid` with a custom receiver.
"""
youyouid = Base64Field(
encode_receiver='django_base64field.tests.receivers:custom_receiver'
)
| from django.db import models
from django.utils.baseconv import base64
from django_base64field.fields import Base64Field
class Planet(models.Model):
# Making `ek` unique just because it will be used as `FK`
# in other models.
ek = Base64Field(unique=True)
name = models.CharField(
default='Fucker',
max_length=103
)
class Continent(models.Model):
ek = Base64Field()
name = models.CharField(
default='Suckers!',
max_length=13
)
planet = models.ForeignKey(Planet, to_field='ek')
class Helper(models.Model):
"""
base64 encoded value won't be available at first time creation.
It can ve accessible by getting the object from database after creation
mean when it get saved completely, But what if we don't want to get our base64
encoded key from our sweet model by retrieving it again from database?
It's easy, efficient, holly and molly!
"""
ek = Base64Field()
def _ek(self):
if self.ek: return self.ek
if not self.ek and self.pk:
return base64.encode(self.pk)
return self.ek
class CustomReceiver(models.Model):
"""
Passing custom receiver to generate `youyouid` with a custom receiver.
"""
youyouid = Base64Field(
encode_receiver='django_base64field.tests.receivers:custom_receiver'
)
| Add little bit comments for Planet model | Add little bit comments for Planet model
| Python | bsd-3-clause | Alir3z4/django-base64field | ---
+++
@@ -4,7 +4,9 @@
class Planet(models.Model):
- ek = Base64Field()
+ # Making `ek` unique just because it will be used as `FK`
+ # in other models.
+ ek = Base64Field(unique=True)
name = models.CharField(
default='Fucker',
max_length=103 |
084eac5735404edeed62cee4e2b429c8f4f2a7a5 | app/dao/inbound_numbers_dao.py | app/dao/inbound_numbers_dao.py | from app import db
from app.dao.dao_utils import transactional
from app.models import InboundNumber
def dao_get_inbound_numbers():
return InboundNumber.query.all()
def dao_get_available_inbound_numbers():
return InboundNumber.query.filter(InboundNumber.active, InboundNumber.service_id.is_(None)).all()
def dao_get_inbound_number_for_service(service_id):
return InboundNumber.query.filter(InboundNumber.service_id == service_id).first()
def dao_get_inbound_number(inbound_number_id):
return InboundNumber.query.filter(InboundNumber.id == inbound_number_id).first()
@transactional
def dao_set_inbound_number_to_service(service_id, inbound_number):
inbound_number.service_id = service_id
db.session.add(inbound_number)
@transactional
def dao_set_inbound_number_active_flag(service_id, active):
inbound_number = InboundNumber.query.filter(InboundNumber.service_id == service_id).first()
inbound_number.active = active
db.session.add(inbound_number)
| from app import db
from app.dao.dao_utils import transactional
from app.models import InboundNumber
def dao_get_inbound_numbers():
return InboundNumber.query.order_by(InboundNumber.updated_at, InboundNumber.number).all()
def dao_get_available_inbound_numbers():
return InboundNumber.query.filter(InboundNumber.active, InboundNumber.service_id.is_(None)).all()
def dao_get_inbound_number_for_service(service_id):
return InboundNumber.query.filter(InboundNumber.service_id == service_id).first()
def dao_get_inbound_number(inbound_number_id):
return InboundNumber.query.filter(InboundNumber.id == inbound_number_id).first()
@transactional
def dao_set_inbound_number_to_service(service_id, inbound_number):
inbound_number.service_id = service_id
db.session.add(inbound_number)
@transactional
def dao_set_inbound_number_active_flag(service_id, active):
inbound_number = InboundNumber.query.filter(InboundNumber.service_id == service_id).first()
inbound_number.active = active
db.session.add(inbound_number)
| Update dao to order by updated_at, number | Update dao to order by updated_at, number
| Python | mit | alphagov/notifications-api,alphagov/notifications-api | ---
+++
@@ -4,7 +4,7 @@
def dao_get_inbound_numbers():
- return InboundNumber.query.all()
+ return InboundNumber.query.order_by(InboundNumber.updated_at, InboundNumber.number).all()
def dao_get_available_inbound_numbers(): |
e8ebb4e9be78e32bc59b1f03cd4854add1148de3 | extra.py | extra.py | Import('env')
env.Append(CXXFLAGS=["-std=c++11"], CPPPATH=["src/wakaama", "wakaama"])
| Import('env')
env.Append(CFLAGS=["-std=gnu11"], CXXFLAGS=["-std=c++11"], CPPPATH=["src/wakaama", "wakaama"])
| Make C11 default. pio should actually do this. | Make C11 default. pio should actually do this.
Signed-off-by: David Graeff <cddf4d21ffd604ccd3fdffe1267f1f9d72e179ea@web.de>
| Python | mit | Openhab-Nodes/libWakaamaEmb,Openhab-Nodes/libWakaamaEmb,Openhab-Nodes/libWakaamaEmb,Openhab-Nodes/libWakaamaEmb | ---
+++
@@ -1,2 +1,2 @@
Import('env')
-env.Append(CXXFLAGS=["-std=c++11"], CPPPATH=["src/wakaama", "wakaama"])
+env.Append(CFLAGS=["-std=gnu11"], CXXFLAGS=["-std=c++11"], CPPPATH=["src/wakaama", "wakaama"]) |
ec867f87441af657f0138ec723de4a52299284d8 | src/epiweb/apps/survey/urls.py | src/epiweb/apps/survey/urls.py | from django.conf.urls.defaults import *
urlpatterns = patterns('',
(r'^profile/$', 'epiweb.apps.survey.views.profile_index'),
(r'^thanks/$', 'epiweb.apps.survey.views.thanks'),
url(r'^people/$', 'epiweb.apps.survey.views.people', name='survey_people'),
url(r'^people/add/$', 'epiweb.apps.survey.views.people_add', name='survey_people_add'),
(r'^$', 'epiweb.apps.survey.views.index'),
)
| from django.conf.urls.defaults import *
from . import views
urlpatterns = patterns('',
(r'^profile/$', views.profile_index),
(r'^thanks/$', views.thanks),
url(r'^people/$', views.people, name='survey_people'),
url(r'^people/add/$', views.people_add, name='survey_people_add'),
(r'^$', views.index),
)
| Use view's function instead of its name | Use view's function instead of its name
| Python | agpl-3.0 | ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website | ---
+++
@@ -1,10 +1,12 @@
from django.conf.urls.defaults import *
+from . import views
+
urlpatterns = patterns('',
- (r'^profile/$', 'epiweb.apps.survey.views.profile_index'),
- (r'^thanks/$', 'epiweb.apps.survey.views.thanks'),
- url(r'^people/$', 'epiweb.apps.survey.views.people', name='survey_people'),
- url(r'^people/add/$', 'epiweb.apps.survey.views.people_add', name='survey_people_add'),
- (r'^$', 'epiweb.apps.survey.views.index'),
+ (r'^profile/$', views.profile_index),
+ (r'^thanks/$', views.thanks),
+ url(r'^people/$', views.people, name='survey_people'),
+ url(r'^people/add/$', views.people_add, name='survey_people_add'),
+ (r'^$', views.index),
)
|
e85e5ea6e2a8b188ff79d114ae0546c8d3ca4c73 | examples/MNIST/mnist.py | examples/MNIST/mnist.py | import os
import gzip
import pickle
import sys
# Python 2/3 compatibility.
try:
from urllib.request import urlretrieve
except ImportError:
from urllib import urlretrieve
'''Adapted from theano tutorial'''
def load_mnist(data_file = os.path.join(os.path.dirname(__file__), 'mnist.pkl.gz')):
if not os.path.exists(data_file):
origin = ('http://www.iro.umontreal.ca/~lisa/deep/data/mnist/mnist.pkl.gz')
print('Downloading data from %s' % origin)
urlretrieve(origin, data_file)
print('... loading data')
f = gzip.open(data_file, 'rb')
if sys.version_info[0] == 3:
train_set, valid_set, test_set = pickle.load(f, encoding='latin1')
else:
train_set, valid_set, test_set = pickle.load(f)
f.close()
train_set_x, train_set_y = train_set
valid_set_x, valid_set_y = valid_set
test_set_x, test_set_y = test_set
return (train_set_x, train_set_y), (valid_set_x, valid_set_y), (test_set_x, test_set_y)
| import os
import gzip
import pickle
import sys
# Python 2/3 compatibility.
try:
from urllib.request import urlretrieve
except ImportError:
from urllib import urlretrieve
'''Adapted from theano tutorial'''
def load_mnist(data_file = os.path.join(os.path.dirname(__file__), 'mnist.pkl.gz')):
if not os.path.exists(data_file):
origin = 'http://www.iro.umontreal.ca/~lisa/deep/data/mnist/mnist.pkl.gz'
print('Downloading data from {}'.format(origin))
urlretrieve(origin, data_file)
print('... loading data')
with gzip.open(data_file, 'rb') as f:
if sys.version_info[0] == 3:
return pickle.load(f, encoding='latin1')
else:
return pickle.load(f)
| Simplify code loading MNIST dataset. | Simplify code loading MNIST dataset.
| Python | mit | VisualComputingInstitute/Beacon8,lucasb-eyer/DeepFried2,elPistolero/DeepFried2,yobibyte/DeepFried2,Pandoro/DeepFried2 | ---
+++
@@ -16,21 +16,14 @@
def load_mnist(data_file = os.path.join(os.path.dirname(__file__), 'mnist.pkl.gz')):
if not os.path.exists(data_file):
- origin = ('http://www.iro.umontreal.ca/~lisa/deep/data/mnist/mnist.pkl.gz')
- print('Downloading data from %s' % origin)
+ origin = 'http://www.iro.umontreal.ca/~lisa/deep/data/mnist/mnist.pkl.gz'
+ print('Downloading data from {}'.format(origin))
urlretrieve(origin, data_file)
print('... loading data')
- f = gzip.open(data_file, 'rb')
- if sys.version_info[0] == 3:
- train_set, valid_set, test_set = pickle.load(f, encoding='latin1')
- else:
- train_set, valid_set, test_set = pickle.load(f)
- f.close()
-
- train_set_x, train_set_y = train_set
- valid_set_x, valid_set_y = valid_set
- test_set_x, test_set_y = test_set
-
- return (train_set_x, train_set_y), (valid_set_x, valid_set_y), (test_set_x, test_set_y)
+ with gzip.open(data_file, 'rb') as f:
+ if sys.version_info[0] == 3:
+ return pickle.load(f, encoding='latin1')
+ else:
+ return pickle.load(f) |
6ce0b5fabd3573ac3c3feb30e8fb48af16d2504f | apps/users/tests/test_profile_admin.py | apps/users/tests/test_profile_admin.py | import fudge
from django.test import TestCase, Client
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User
from users.models import Profile, Link
class ProfileAdmin(TestCase):
def setUp(self):
self.client = Client()
self.User = User.objects.create(
username=u'test_ross',
password=u'password2',
is_active=True
)
self.profile = Profile.objects.create(
user=self.User
)
@fudge.patch('django_browserid.auth.BrowserIDBackend.authenticate')
def test_edit_without_links(self, fake):
redirect = '/profile/%s/' % self.User.username
post_data = {
'name': 'Boozeniges',
'link_url': 'http://ross-eats.co.uk',
'link_name': 'ross eats'
}
response = self.client.post('/profile/edit/', post_data, follow=True)
self.assertRedirects(response, redirect, status_code=301)
| from contextlib import contextmanager
import fudge
from django.test import TestCase, Client
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User
from users.models import Profile, Link
@contextmanager
def given_user(fake_auth, user):
"""Context manager to respond to any login call with a specific user."""
fake_auth.expects_call().returns(user)
yield
class ProfileAdmin(TestCase):
def setUp(self):
self.client = Client()
self.User = User.objects.create(
username=u'test_ross',
password=u'password2',
is_active=True
)
self.profile = Profile.objects.create(
user=self.User
)
@fudge.patch('django_browserid.auth.BrowserIDBackend.authenticate')
def test_edit_without_links(self, fake):
redirect = '/profile/%s/' % self.User.username
post_data = {
'name': 'Boozeniges',
'link_url': 'http://ross-eats.co.uk',
'link_name': 'ross eats'
}
with given_user(fake, self.User):
self.client.login()
response = self.client.post('/profile/edit/', post_data,
follow=True)
try:
self.assertRedirects(response, redirect, status_code=301)
except AssertionError:
print response.redirect_chain
raise
| Fix the profile test to log in as the test user. | Fix the profile test to log in as the test user. | Python | bsd-3-clause | mozilla/mozilla-ignite,mozilla/mozilla-ignite,mozilla/mozilla-ignite,mozilla/mozilla-ignite | ---
+++
@@ -1,3 +1,5 @@
+from contextlib import contextmanager
+
import fudge
from django.test import TestCase, Client
@@ -5,6 +7,14 @@
from django.contrib.auth.models import User
from users.models import Profile, Link
+
+
+@contextmanager
+def given_user(fake_auth, user):
+ """Context manager to respond to any login call with a specific user."""
+ fake_auth.expects_call().returns(user)
+ yield
+
class ProfileAdmin(TestCase):
@@ -27,5 +37,14 @@
'link_url': 'http://ross-eats.co.uk',
'link_name': 'ross eats'
}
- response = self.client.post('/profile/edit/', post_data, follow=True)
- self.assertRedirects(response, redirect, status_code=301)
+
+ with given_user(fake, self.User):
+ self.client.login()
+ response = self.client.post('/profile/edit/', post_data,
+ follow=True)
+
+ try:
+ self.assertRedirects(response, redirect, status_code=301)
+ except AssertionError:
+ print response.redirect_chain
+ raise |
bcfb29272d727ee2775c9f212053725e4a562752 | pip_review/__init__.py | pip_review/__init__.py | from functools import partial
import subprocess
import requests
import multiprocessing
import json
def get_pkg_info(pkg_name, session):
r = session.get('http://pypi.python.org/pypi/%s/json' % (pkg_name,))
if r.status_code == requests.codes.ok:
return json.loads(r.text)
else:
raise ValueError('Package %r not found on PyPI.' % (pkg_name,))
def latest_version(pkg_name, session, silent=False):
try:
info = get_pkg_info(pkg_name, session)
except ValueError:
if silent:
return None
else:
raise
return info['info']['version']
def get_latest_versions(pkg_names):
with requests.session() as session:
pool = multiprocessing.Pool(min(12, len(pkg_names)))
get_latest = partial(latest_version, session=session, silent=True)
versions = pool.map(get_latest, pkg_names)
return zip(pkg_names, versions)
def get_installed_pkgs(editables=False):
for line in subprocess.check_output(['pip', 'freeze']).split('\n'):
if not line:
continue
if line.startswith('-e'):
if editables:
yield line.split('#egg=', 1)[1], None, True
else:
name, version = line.split('==')
yield name, version, False
| from functools import partial
import subprocess
import urllib2
import multiprocessing
import json
def get_pkg_info(pkg_name):
req = urllib2.Request('http://pypi.python.org/pypi/%s/json' % (pkg_name,))
handler = urllib2.urlopen(req)
status = handler.getcode()
if status == 200:
content = handler.read()
return json.loads(content)
else:
raise ValueError('Package %r not found on PyPI.' % (pkg_name,))
def latest_version(pkg_name, silent=False):
try:
info = get_pkg_info(pkg_name)
except ValueError:
if silent:
return None
else:
raise
return info['info']['version']
def get_latest_versions(pkg_names):
pool = multiprocessing.Pool(min(12, len(pkg_names)))
get_latest = partial(latest_version, silent=True)
versions = pool.map(get_latest, pkg_names)
return zip(pkg_names, versions)
def get_installed_pkgs(editables=False):
for line in subprocess.check_output(['pip', 'freeze']).split('\n'):
if not line:
continue
if line.startswith('-e'):
if editables:
yield line.split('#egg=', 1)[1], None, True
else:
name, version = line.split('==')
yield name, version, False
| Remove the requests dependency altogether. | Remove the requests dependency altogether.
(Makes no sense for such small a tool.)
| Python | bsd-2-clause | suutari-ai/prequ,suutari/prequ,suutari/prequ | ---
+++
@@ -1,21 +1,24 @@
from functools import partial
import subprocess
-import requests
+import urllib2
import multiprocessing
import json
-def get_pkg_info(pkg_name, session):
- r = session.get('http://pypi.python.org/pypi/%s/json' % (pkg_name,))
- if r.status_code == requests.codes.ok:
- return json.loads(r.text)
+def get_pkg_info(pkg_name):
+ req = urllib2.Request('http://pypi.python.org/pypi/%s/json' % (pkg_name,))
+ handler = urllib2.urlopen(req)
+ status = handler.getcode()
+ if status == 200:
+ content = handler.read()
+ return json.loads(content)
else:
raise ValueError('Package %r not found on PyPI.' % (pkg_name,))
-def latest_version(pkg_name, session, silent=False):
+def latest_version(pkg_name, silent=False):
try:
- info = get_pkg_info(pkg_name, session)
+ info = get_pkg_info(pkg_name)
except ValueError:
if silent:
return None
@@ -25,10 +28,9 @@
def get_latest_versions(pkg_names):
- with requests.session() as session:
- pool = multiprocessing.Pool(min(12, len(pkg_names)))
- get_latest = partial(latest_version, session=session, silent=True)
- versions = pool.map(get_latest, pkg_names)
+ pool = multiprocessing.Pool(min(12, len(pkg_names)))
+ get_latest = partial(latest_version, silent=True)
+ versions = pool.map(get_latest, pkg_names)
return zip(pkg_names, versions)
|
f9f12e89e526b2645f013fc5856488d105a39d5a | bouncer/sentry.py | bouncer/sentry.py | """Report exceptions to Sentry/Raven."""
from pyramid import tweens
import raven
from bouncer import __version__
def get_raven_client(request):
"""Return the Raven client for reporting crashes to Sentry."""
client = request.registry["raven.client"]
client.http_context({
"url": request.url,
"method": request.method,
})
request.add_finished_callback(
lambda request: client.context.clear())
return client
def includeme(config):
config.registry["raven.client"] = raven.Client(
release=__version__)
config.add_request_method(
get_raven_client,
name="raven",
reify=True)
| """Report exceptions to Sentry/Raven."""
import os
import raven
from bouncer import __version__
def get_raven_client(request):
"""Return the Raven client for reporting crashes to Sentry."""
client = request.registry["raven.client"]
client.http_context({
"url": request.url,
"method": request.method,
})
request.add_finished_callback(
lambda request: client.context.clear())
return client
def includeme(config):
environment = os.environ.get('ENV', 'dev')
config.registry["raven.client"] = raven.Client(environment=environment,
release=__version__)
config.add_request_method(
get_raven_client,
name="raven",
reify=True)
| Add environment name to Sentry client | Add environment name to Sentry client
Skyliner provides the ENV environment variable for our application,
which contains either "prod" or "qa" depending on the environment.
Sentry supports partitioning reports based on the environment within a
single application. Adding this metadata to the `Client` allows us to
migrate QA and Prod onto a single application in Sentry.
| Python | bsd-2-clause | hypothesis/bouncer,hypothesis/bouncer,hypothesis/bouncer | ---
+++
@@ -1,5 +1,7 @@
"""Report exceptions to Sentry/Raven."""
-from pyramid import tweens
+
+import os
+
import raven
from bouncer import __version__
@@ -21,8 +23,10 @@
def includeme(config):
- config.registry["raven.client"] = raven.Client(
- release=__version__)
+ environment = os.environ.get('ENV', 'dev')
+
+ config.registry["raven.client"] = raven.Client(environment=environment,
+ release=__version__)
config.add_request_method(
get_raven_client, |
79cb9edf45ed77cdaa851e45d71f10c69db41221 | benchexec/tools/yogar-cbmc-parallel.py | benchexec/tools/yogar-cbmc-parallel.py | """
BenchExec is a framework for reliable benchmarking.
This file is part of BenchExec.
Copyright (C) 2007-2015 Dirk Beyer
All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
import benchexec.util as util
yogar_cbmc = __import__("benchexec.tools.yogar-cbmc", fromlist=["Tool"])
class Tool(yogar_cbmc.Tool):
def executable(self):
return util.find_executable('yogar-cbmc-parallel')
def name(self):
return 'Yogar-CBMC-Parallel'
def cmdline(self, executable, options, tasks, propertyfile, rlimits):
return [executable] + options + tasks
| """
BenchExec is a framework for reliable benchmarking.
This file is part of BenchExec.
Copyright (C) 2007-2015 Dirk Beyer
All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
import benchexec.util as util
yogar_cbmc = __import__("benchexec.tools.yogar-cbmc", fromlist=["Tool"])
class Tool(yogar_cbmc.Tool):
REQUIRED_PATHS = [
"yogar-cbmc"
]
def executable(self):
return util.find_executable('yogar-cbmc-parallel')
def name(self):
return 'Yogar-CBMC-Parallel'
def cmdline(self, executable, options, tasks, propertyfile, rlimits):
return [executable] + options + tasks
| Add forgotten program file for deployment | Add forgotten program file for deployment
| Python | apache-2.0 | ultimate-pa/benchexec,sosy-lab/benchexec,sosy-lab/benchexec,sosy-lab/benchexec,dbeyer/benchexec,ultimate-pa/benchexec,ultimate-pa/benchexec,dbeyer/benchexec,ultimate-pa/benchexec,ultimate-pa/benchexec,sosy-lab/benchexec,ultimate-pa/benchexec,sosy-lab/benchexec,dbeyer/benchexec,dbeyer/benchexec,sosy-lab/benchexec | ---
+++
@@ -19,6 +19,10 @@
class Tool(yogar_cbmc.Tool):
+ REQUIRED_PATHS = [
+ "yogar-cbmc"
+ ]
+
def executable(self):
return util.find_executable('yogar-cbmc-parallel')
|
94c943b3dcabb1611a745caef0ecf6da10e807e4 | kiteconnect/__version__.py | kiteconnect/__version__.py | __title__ = "kiteconnect"
__description__ = "The official Python client for the Kite Connect trading API"
__url__ = "https://kite.trade"
__download_url__ = "https://github.com/zerodhatech/pykiteconnect"
__version__ = "3.7.0"
__author__ = "Zerodha Technology Pvt ltd. (India)"
__author_email__ = "talk@zerodha.tech"
__license__ = "MIT"
| __title__ = "kiteconnect"
__description__ = "The official Python client for the Kite Connect trading API"
__url__ = "https://kite.trade"
__download_url__ = "https://github.com/zerodhatech/pykiteconnect"
__version__ = "3.7.0.b1"
__author__ = "Zerodha Technology Pvt ltd. (India)"
__author_email__ = "talk@zerodha.tech"
__license__ = "MIT"
| Mark release as beta in setup.py | Mark release as beta in setup.py
| Python | mit | rainmattertech/pykiteconnect | ---
+++
@@ -2,7 +2,7 @@
__description__ = "The official Python client for the Kite Connect trading API"
__url__ = "https://kite.trade"
__download_url__ = "https://github.com/zerodhatech/pykiteconnect"
-__version__ = "3.7.0"
+__version__ = "3.7.0.b1"
__author__ = "Zerodha Technology Pvt ltd. (India)"
__author_email__ = "talk@zerodha.tech"
__license__ = "MIT" |
6b53e14ec1f5c71d2aa1f17a4108f4e1d88e8b89 | pywikibot/families/wikia_family.py | pywikibot/families/wikia_family.py | # -*- coding: utf-8 -*-
__version__ = '$Id$'
import family
# The Wikia Search family
# user-config.py: usernames['wikia']['wikia'] = 'User name'
class Family(family.Family):
def __init__(self):
family.Family.__init__(self)
self.name = u'wikia'
self.langs = {
u'wikia': None,
}
def hostname(self, code):
return u'www.wikia.com'
def version(self, code):
return "1.15.1"
def scriptpath(self, code):
return ''
def apipath(self, code):
return '/api.php'
| # -*- coding: utf-8 -*-
__version__ = '$Id$'
import family
# The Wikia Search family
# user-config.py: usernames['wikia']['wikia'] = 'User name'
class Family(family.Family):
def __init__(self):
family.Family.__init__(self)
self.name = u'wikia'
self.langs = {
u'wikia': None,
}
def hostname(self, code):
return u'www.wikia.com'
def version(self, code):
return "1.16.2"
def scriptpath(self, code):
return ''
def apipath(self, code):
return '/api.php'
| Update a version number from trunk r9016 | Update a version number from trunk r9016
git-svn-id: 9a050473c2aca1e14f53d73349e19b938c2cf203@9040 6a7f98fc-eeb0-4dc1-a6e2-c2c589a08aa6
| Python | mit | legoktm/pywikipedia-rewrite | ---
+++
@@ -20,7 +20,7 @@
return u'www.wikia.com'
def version(self, code):
- return "1.15.1"
+ return "1.16.2"
def scriptpath(self, code):
return '' |
98f1df191febd889fcde861f94a9ca126c60ea37 | tests/GIR/runalltests.py | tests/GIR/runalltests.py | # -*- Mode: Python -*-
import os
import glob
import sys
import shutil
import unittest
testLoader = unittest.TestLoader()
names = []
for filename in glob.iglob("test_*.py"):
names.append(filename[:-3])
names.sort()
testSuite = testLoader.loadTestsFromNames(names)
runner = unittest.TextTestRunner(verbosity=2)
result = runner.run(testSuite)
# Cleanup
try:
shutil.rmtree("./test_data/blobs")
except OSError:
pass
if os.path.isfile("./test_data/test_gir.db"):
os.remove("./test_data/test_gir.db")
if not result.wasSuccessful():
sys.exit(1)
| # -*- Mode: Python -*-
import os
import glob
import sys
import shutil
import unittest
if os.path.isfile("./test_data/test_gir.db"):
os.remove("./test_data/test_gir.db")
testLoader = unittest.TestLoader()
names = []
for filename in glob.iglob("test_*.py"):
names.append(filename[:-3])
names.sort()
testSuite = testLoader.loadTestsFromNames(names)
runner = unittest.TextTestRunner(verbosity=2)
result = runner.run(testSuite)
# Cleanup
try:
shutil.rmtree("./test_data/blobs")
except OSError:
pass
if not result.wasSuccessful():
sys.exit(1)
| Remove SQLite database before running tests | Remove SQLite database before running tests
| Python | lgpl-2.1 | midgardproject/midgard-core,midgardproject/midgard-core,midgardproject/midgard-core,midgardproject/midgard-core | ---
+++
@@ -6,6 +6,9 @@
import shutil
import unittest
+
+if os.path.isfile("./test_data/test_gir.db"):
+ os.remove("./test_data/test_gir.db")
testLoader = unittest.TestLoader()
@@ -24,8 +27,5 @@
except OSError:
pass
-if os.path.isfile("./test_data/test_gir.db"):
- os.remove("./test_data/test_gir.db")
-
if not result.wasSuccessful():
sys.exit(1) |
74dcf36c2eecab290c1c76c947b024e51d280ea7 | tests/test_rover_init.py | tests/test_rover_init.py | def test_rover_init_with_default_parameters():
from rover import Rover
rover = Rover()
assert rover.x == 0
assert rover.y == 0
assert rover.direction == 'N'
def test_rover_init_with_custom_parameters():
from rover import Rover
rover = Rover(3, 7, 'W')
assert rover.x == 3
assert rover.y == 7
assert rover.direction == 'W'
def test_rover_init_custom_grid():
from rover import Rover
rover = Rover(grid_x=100, grid_y=150)
assert rover.x == 0
assert rover.y == 0
assert rover.direction == 'N'
assert rover.grid_x == 100
assert rover.grid_y == 150
def test_rover_init_full_custom_grid():
from rover import Rover
rover = Rover(5, 9, 'E', 100, 150)
assert rover.x == 5
assert rover.y == 9
assert rover.direction == 'E'
assert rover.grid_x == 100
assert rover.grid_y == 150
| def test_rover_init_with_default_parameters():
from rover import Rover
rover = Rover()
assert rover.x == 0
assert rover.y == 0
assert rover.direction == 'N'
assert rover.grid_x == 50
assert rover.grid_y == 50
def test_rover_init_with_custom_parameters():
from rover import Rover
rover = Rover(3, 7, 'W')
assert rover.x == 3
assert rover.y == 7
assert rover.direction == 'W'
assert rover.grid_x == 50
assert rover.grid_y == 50
def test_rover_init_custom_grid():
from rover import Rover
rover = Rover(grid_x=100, grid_y=150)
assert rover.x == 0
assert rover.y == 0
assert rover.direction == 'N'
assert rover.grid_x == 100
assert rover.grid_y == 150
def test_rover_init_full_custom_grid():
from rover import Rover
rover = Rover(5, 9, 'E', 100, 150)
assert rover.x == 5
assert rover.y == 9
assert rover.direction == 'E'
assert rover.grid_x == 100
assert rover.grid_y == 150
| Add testing for default grid_* values | Add testing for default grid_* values
| Python | mit | authentik8/rover | ---
+++
@@ -5,12 +5,18 @@
assert rover.y == 0
assert rover.direction == 'N'
+ assert rover.grid_x == 50
+ assert rover.grid_y == 50
+
def test_rover_init_with_custom_parameters():
from rover import Rover
rover = Rover(3, 7, 'W')
assert rover.x == 3
assert rover.y == 7
assert rover.direction == 'W'
+
+ assert rover.grid_x == 50
+ assert rover.grid_y == 50
def test_rover_init_custom_grid(): |
a9a01b2bb07fadb6e9ab07228fc6857f28e4d444 | src/edge/context_processors.py | src/edge/context_processors.py | import os
def export_envs(request):
data = {}
data['FULLSTORY_ORG_ID'] = os.getenv('FULLSTORY_ORG_ID', '')
return data
| import os
def export_envs(request):
data = {}
data['FULLSTORY_ORG_ID'] = os.getenv('FULLSTORY_ORG_ID', '')
return data
| Fix flake8 error by adding extra line | Fix flake8 error by adding extra line
| Python | mit | ginkgobioworks/edge,ginkgobioworks/edge,ginkgobioworks/edge,ginkgobioworks/edge | ---
+++
@@ -1,4 +1,5 @@
import os
+
def export_envs(request):
data = {} |
fa7ffa32f0484c45a665de2766a97ae7a23a0b6d | pyicloud/utils.py | pyicloud/utils.py | import getpass
import keyring
from .exceptions import NoStoredPasswordAvailable
KEYRING_SYSTEM = 'pyicloud://icloud-password'
def get_password(username, interactive=True):
try:
return get_password_from_keyring(username)
except NoStoredPasswordAvailable:
if not interactive:
raise
return getpass.getpass(
'ICloud Password for {username}: '.format(
username=username,
)
)
def password_exists_in_keyring(username):
try:
get_password_from_keyring(username)
except NoStoredPasswordAvailable:
return False
return True
def get_password_from_keyring(username):
result = keyring.get_password(
KEYRING_SYSTEM,
username
)
if result is None:
raise NoStoredPasswordAvailable(
"No pyicloud password for {username} could be found "
"in the system keychain. Use the `--store-in-keyring` "
"command-line option for storing a password for this "
"username.".format(
username=username,
)
)
return result
def store_password_in_keyring(username, password):
return keyring.set_password(
KEYRING_SYSTEM,
username,
password,
)
def delete_password_in_keyring(username):
return keyring.delete_password(
KEYRING_SYSTEM,
username,
)
| import getpass
import keyring
from .exceptions import NoStoredPasswordAvailable
KEYRING_SYSTEM = 'pyicloud://icloud-password'
def get_password(username, interactive=True):
try:
return get_password_from_keyring(username)
except NoStoredPasswordAvailable:
if not interactive:
raise
return getpass.getpass(
'Enter iCloud password for {username}: '.format(
username=username,
)
)
def password_exists_in_keyring(username):
try:
get_password_from_keyring(username)
except NoStoredPasswordAvailable:
return False
return True
def get_password_from_keyring(username):
result = keyring.get_password(
KEYRING_SYSTEM,
username
)
if result is None:
raise NoStoredPasswordAvailable(
"No pyicloud password for {username} could be found "
"in the system keychain. Use the `--store-in-keyring` "
"command-line option for storing a password for this "
"username.".format(
username=username,
)
)
return result
def store_password_in_keyring(username, password):
return keyring.set_password(
KEYRING_SYSTEM,
username,
password,
)
def delete_password_in_keyring(username):
return keyring.delete_password(
KEYRING_SYSTEM,
username,
)
| Tweak wording of password prompt | Tweak wording of password prompt
iCloud is branded with a lower case 'i' like most other Apple products.
| Python | mit | picklepete/pyicloud,picklepete/pyicloud | ---
+++
@@ -15,7 +15,7 @@
raise
return getpass.getpass(
- 'ICloud Password for {username}: '.format(
+ 'Enter iCloud password for {username}: '.format(
username=username,
)
) |
cd189a5cbf8cbb567efaba0e92b3c31278817a39 | pnrg/filters.py | pnrg/filters.py | from jinja2._compat import text_type
import re
def do_right(value, width=80):
"""Right-justifies the value in a field of a given width."""
return text_type(value).rjust(width)
_LATEX_SUBS = (
(re.compile(r'\\'), r'\\textbackslash'),
(re.compile(r'([{}_#%&$])'), r'\\\1'),
(re.compile(r'~'), r'\~{}'),
(re.compile(r'\^'), r'\^{}'),
(re.compile(r'"'), r"''"),
(re.compile(r'\.\.\.+'), r'\\ldots'),
(re.compile(r'&'), r'&'),
(re.compile(r'LaTeX'), r'\\LaTeX')
)
def escape_tex(value):
"""
Escapes TeX characters to avoid breaking {La,Lua,Xe}Tex compilers.
Kang'd (with permission!) from http://flask.pocoo.org/snippets/55/
"""
newval = value
for pattern, replacement in _LATEX_SUBS:
newval = pattern.sub(replacement, newval)
return newval
| from jinja2._compat import text_type
import re
def do_right(value, width=80):
"""Right-justifies the value in a field of a given width."""
return text_type(value).rjust(width)
_LATEX_SUBS = (
(re.compile(r'\\'), r'\\textbackslash'),
(re.compile(r'([{}_#%&$])'), r'\\\1'),
(re.compile(r'~'), r'\~{}'),
(re.compile(r'\^'), r'\^{}'),
(re.compile(r'"'), r"''"),
(re.compile(r'\.\.\.+'), r'\\ldots'),
(re.compile(r'&'), r'&'),
(re.compile(r'LaTeX'), r'\\textrm{\\LaTeX}')
)
def escape_tex(value):
"""
Escapes TeX characters to avoid breaking {La,Lua,Xe}Tex compilers.
Kang'd (with permission!) from http://flask.pocoo.org/snippets/55/
"""
newval = value
for pattern, replacement in _LATEX_SUBS:
newval = pattern.sub(replacement, newval)
return newval
| Use Computer Modern for \LaTeX macro | Use Computer Modern for \LaTeX macro
Source Sans pro (and most othe sans-serif fonts) render the LaTeX macro
pretty weirdly, so the classic Computer Modern should be okay here. It
may wind up being an odd contrast in an otherwise sans-serif document
though, so this may eventually get reverted!
| Python | mit | sjbarag/poorly-named-resume-generator,sjbarag/poorly-named-resume-generator | ---
+++
@@ -13,7 +13,7 @@
(re.compile(r'"'), r"''"),
(re.compile(r'\.\.\.+'), r'\\ldots'),
(re.compile(r'&'), r'&'),
- (re.compile(r'LaTeX'), r'\\LaTeX')
+ (re.compile(r'LaTeX'), r'\\textrm{\\LaTeX}')
)
def escape_tex(value): |
d1a893514092a2495fcd2b14365b9b014fd39f59 | calaccess_processed/models/__init__.py | calaccess_processed/models/__init__.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
Import all of the models from submodules and thread them together.
"""
from calaccess_processed.models.campaign.entities import (
Candidate,
CandidateCommittee,
)
from calaccess_processed.models.campaign.filings import (
Form460,
Form460Version,
Schedule497,
Schedule497Version,
)
from calaccess_processed.models.campaign.contributions import (
MonetaryContribution,
MonetaryContributionVersion,
NonMonetaryContribution,
NonMonetaryContributionVersion,
LateContributionReceived,
LateContributionReceivedVersion,
LateContributionMade,
LateContributionMadeVersion,
)
from calaccess_processed.models.common import (
FilerIDValue,
FilingIDValue,
)
from calaccess_processed.models.tracking import (
ProcessedDataVersion,
ProcessedDataFile,
)
__all__ = (
'ProcessedDataVersion',
'ProcessedDataFile',
'Candidate',
'CandidateCommittee',
'Form460',
'Form460Version',
'Schedule497',
'Schedule497Version',
'LateContributionReceived',
'LateContributionReceivedVersion',
'LateContributionMade',
'LateContributionMadeVersion',
'FilerIDValue',
'FilingIDValue',
)
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
Import all of the models from submodules and thread them together.
"""
from calaccess_processed.models.campaign.entities import (
Candidate,
CandidateCommittee,
)
from calaccess_processed.models.campaign.filings import (
Form460,
Form460Version,
Schedule497,
Schedule497Version,
)
from calaccess_processed.models.campaign.contributions import (
MonetaryContribution,
MonetaryContributionVersion,
NonMonetaryContribution,
NonMonetaryContributionVersion,
LateContributionReceived,
LateContributionReceivedVersion,
LateContributionMade,
LateContributionMadeVersion,
)
from calaccess_processed.models.common import (
FilerIDValue,
FilingIDValue,
)
from calaccess_processed.models.tracking import (
ProcessedDataVersion,
ProcessedDataFile,
)
__all__ = (
'ProcessedDataVersion',
'ProcessedDataFile',
'Candidate',
'CandidateCommittee',
'Form460',
'Form460Version',
'Schedule497',
'Schedule497Version',
'MonetaryContribution',
'MonetaryContributionVersion',
'NonMonetaryContribution',
'NonMonetaryContributionVersion',
'LateContributionReceived',
'LateContributionReceivedVersion',
'LateContributionMade',
'LateContributionMadeVersion',
'FilerIDValue',
'FilingIDValue',
)
| Add missing models to __all__ list | Add missing models to __all__ list
| Python | mit | california-civic-data-coalition/django-calaccess-processed-data,california-civic-data-coalition/django-calaccess-processed-data | ---
+++
@@ -41,6 +41,10 @@
'Form460Version',
'Schedule497',
'Schedule497Version',
+ 'MonetaryContribution',
+ 'MonetaryContributionVersion',
+ 'NonMonetaryContribution',
+ 'NonMonetaryContributionVersion',
'LateContributionReceived',
'LateContributionReceivedVersion',
'LateContributionMade', |
64dd3ecb1cf5adbf68d59d231b67e7d30ace715f | scripts/create-hosted-graphite-dashboards.py | scripts/create-hosted-graphite-dashboards.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
Description:
To add
Usage:
scripts/create-hosted-graphite-dashboards.py <hosted_graphite_api_key>
Example:
scripts/create-hosted-graphite-dashboards.py apikey
"""
import os
import sys
import requests
from docopt import docopt
sys.path.insert(0, '.') # noqa
def generate_dashboards(api_key):
endpoint = "https://api.hostedgraphite.com/api/v2/grafana/dashboards/"
path = os.path.join(os.path.dirname(__file__), "../grafana/")
for filename in os.listdir(path):
with open(path + filename) as fp:
resp = requests.put(endpoint, auth=(api_key, ''), data=fp.read())
resp.raise_for_status()
if __name__ == "__main__":
arguments = docopt(__doc__)
generate_dashboards(arguments['<hosted_graphite_api_key>'])
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
Description:
This pushes dashboards defined in /grafana directory up to our hosted graphite.
You can get the <hosted_graphite_api_key> from the "Account Home" page of hosted graphite.
To change a dashboard a process you should follow is:
1. Set "editable": true in one our existing JSON files
2. Change the title (to avoid overwriting the existing dashboard as you try things out)
3. Run this script
4. Find the newly created dashboard in hosted graphite and make changes
5. Export the JSON for the new dashboard (click settings cog, "View JSON") and overwrite the existing JSON file
6. Set "editable": false and set the title back to the original one. Remove the new "id" key from the exported JSON
7. Run this script again
8. Check you're happy with the new dashboard
9. Delete the editable "test" dashboard you created in step 3 (click settings cog, "Delete dashboard")
Usage:
scripts/create-hosted-graphite-dashboards.py <hosted_graphite_api_key>
Example:
scripts/create-hosted-graphite-dashboards.py apikey
"""
import os
import sys
import requests
from docopt import docopt
sys.path.insert(0, '.') # noqa
def generate_dashboards(api_key):
endpoint = "https://api.hostedgraphite.com/api/v2/grafana/dashboards/"
path = os.path.join(os.path.dirname(__file__), "../grafana/")
for filename in os.listdir(path):
with open(path + filename) as fp:
resp = requests.put(endpoint, auth=(api_key, ''), data=fp.read())
resp.raise_for_status()
if __name__ == "__main__":
arguments = docopt(__doc__)
generate_dashboards(arguments['<hosted_graphite_api_key>'])
| Update docstring for creating hosted graphite dashboards | Update docstring for creating hosted graphite dashboards
| Python | mit | alphagov/digitalmarketplace-aws,alphagov/digitalmarketplace-aws,alphagov/digitalmarketplace-aws | ---
+++
@@ -2,7 +2,19 @@
# -*- coding: utf-8 -*-
"""
Description:
- To add
+ This pushes dashboards defined in /grafana directory up to our hosted graphite.
+ You can get the <hosted_graphite_api_key> from the "Account Home" page of hosted graphite.
+
+ To change a dashboard a process you should follow is:
+ 1. Set "editable": true in one our existing JSON files
+ 2. Change the title (to avoid overwriting the existing dashboard as you try things out)
+ 3. Run this script
+ 4. Find the newly created dashboard in hosted graphite and make changes
+ 5. Export the JSON for the new dashboard (click settings cog, "View JSON") and overwrite the existing JSON file
+ 6. Set "editable": false and set the title back to the original one. Remove the new "id" key from the exported JSON
+ 7. Run this script again
+ 8. Check you're happy with the new dashboard
+ 9. Delete the editable "test" dashboard you created in step 3 (click settings cog, "Delete dashboard")
Usage:
scripts/create-hosted-graphite-dashboards.py <hosted_graphite_api_key> |
08aa975b100b1dbf8c9594a2e57368ad866f98a4 | api/caching/tasks.py | api/caching/tasks.py | import urlparse
import celery
import requests
from celery.utils.log import get_task_logger
from api.base import settings
from framework.tasks import app as celery_app
logger = get_task_logger(__name__)
class VarnishTask(celery.Task):
abstract = True
max_retries = 5
def get_varnish_servers():
# TODO: this should get the varnish servers from HAProxy or a setting
return settings.VARNISH_SERVERS
@celery_app.task(base=VarnishTask, name='caching_tasks.ban_url')
def ban_url(url):
if settings.ENABLE_VARNISH:
parsed_url = urlparse.urlparse(url)
for host in get_varnish_servers():
varnish_parsed_url = urlparse.urlparse(host)
ban_url = '{scheme}://{netloc}{path}.*'.format(
scheme=varnish_parsed_url.scheme,
netloc=varnish_parsed_url.netloc,
path=parsed_url.path
)
response = requests.request('BAN', ban_url, headers=dict(
Host=parsed_url.hostname
))
if not response.ok:
logger.error('Banning {} failed: {}'.format(
url,
response.text
))
| import urlparse
import requests
from celery.utils.log import get_task_logger
from api.base import settings
logger = get_task_logger(__name__)
def get_varnish_servers():
# TODO: this should get the varnish servers from HAProxy or a setting
return settings.VARNISH_SERVERS
def ban_url(url):
timeout = 0.5 # 500ms timeout for bans
if settings.ENABLE_VARNISH:
parsed_url = urlparse.urlparse(url)
for host in get_varnish_servers():
varnish_parsed_url = urlparse.urlparse(host)
ban_url = '{scheme}://{netloc}{path}.*'.format(
scheme=varnish_parsed_url.scheme,
netloc=varnish_parsed_url.netloc,
path=parsed_url.path
)
response = requests.request('BAN', ban_url, timeout=timeout, headers=dict(
Host=parsed_url.hostname
))
if not response.ok:
logger.error('Banning {} failed: {}'.format(
url,
response.text
))
| Move from celery task to regular-old function | Move from celery task to regular-old function
| Python | apache-2.0 | sloria/osf.io,cwisecarver/osf.io,mattclark/osf.io,emetsger/osf.io,SSJohns/osf.io,erinspace/osf.io,asanfilippo7/osf.io,mluo613/osf.io,mluo613/osf.io,caseyrollins/osf.io,saradbowman/osf.io,cwisecarver/osf.io,icereval/osf.io,doublebits/osf.io,brianjgeiger/osf.io,mluo613/osf.io,adlius/osf.io,adlius/osf.io,aaxelb/osf.io,binoculars/osf.io,sloria/osf.io,DanielSBrown/osf.io,hmoco/osf.io,samchrisinger/osf.io,mattclark/osf.io,mluke93/osf.io,cslzchen/osf.io,brianjgeiger/osf.io,leb2dg/osf.io,mluo613/osf.io,DanielSBrown/osf.io,kch8qx/osf.io,TomHeatwole/osf.io,caneruguz/osf.io,zamattiac/osf.io,binoculars/osf.io,chrisseto/osf.io,zamattiac/osf.io,doublebits/osf.io,zachjanicki/osf.io,aaxelb/osf.io,doublebits/osf.io,laurenrevere/osf.io,cwisecarver/osf.io,TomBaxter/osf.io,mluke93/osf.io,chennan47/osf.io,rdhyee/osf.io,RomanZWang/osf.io,leb2dg/osf.io,mfraezz/osf.io,crcresearch/osf.io,acshi/osf.io,RomanZWang/osf.io,HalcyonChimera/osf.io,leb2dg/osf.io,icereval/osf.io,Johnetordoff/osf.io,acshi/osf.io,adlius/osf.io,monikagrabowska/osf.io,HalcyonChimera/osf.io,chennan47/osf.io,aaxelb/osf.io,SSJohns/osf.io,mluke93/osf.io,chrisseto/osf.io,amyshi188/osf.io,amyshi188/osf.io,chrisseto/osf.io,zachjanicki/osf.io,Johnetordoff/osf.io,Nesiehr/osf.io,CenterForOpenScience/osf.io,caneruguz/osf.io,baylee-d/osf.io,billyhunt/osf.io,alexschiller/osf.io,mluo613/osf.io,jnayak1/osf.io,billyhunt/osf.io,abought/osf.io,kwierman/osf.io,mattclark/osf.io,abought/osf.io,pattisdr/osf.io,Nesiehr/osf.io,zachjanicki/osf.io,cslzchen/osf.io,doublebits/osf.io,erinspace/osf.io,HalcyonChimera/osf.io,caseyrollins/osf.io,kch8qx/osf.io,TomHeatwole/osf.io,abought/osf.io,HalcyonChimera/osf.io,alexschiller/osf.io,jnayak1/osf.io,wearpants/osf.io,kch8qx/osf.io,rdhyee/osf.io,emetsger/osf.io,binoculars/osf.io,TomBaxter/osf.io,kwierman/osf.io,alexschiller/osf.io,erinspace/osf.io,RomanZWang/osf.io,TomHeatwole/osf.io,cwisecarver/osf.io,kch8qx/osf.io,DanielSBrown/osf.io,kwierman/osf.io,sloria/osf.io,icereval/osf.io,TomHeatwole/osf.io,kwierman/osf.io,emetsger/osf.io,saradbowman/osf.io,RomanZWang/osf.io,Johnetordoff/osf.io,Johnetordoff/osf.io,mfraezz/osf.io,zamattiac/osf.io,amyshi188/osf.io,wearpants/osf.io,RomanZWang/osf.io,Nesiehr/osf.io,brianjgeiger/osf.io,CenterForOpenScience/osf.io,wearpants/osf.io,acshi/osf.io,acshi/osf.io,rdhyee/osf.io,felliott/osf.io,alexschiller/osf.io,pattisdr/osf.io,asanfilippo7/osf.io,monikagrabowska/osf.io,acshi/osf.io,laurenrevere/osf.io,mfraezz/osf.io,crcresearch/osf.io,monikagrabowska/osf.io,emetsger/osf.io,jnayak1/osf.io,hmoco/osf.io,chennan47/osf.io,monikagrabowska/osf.io,samchrisinger/osf.io,zachjanicki/osf.io,monikagrabowska/osf.io,baylee-d/osf.io,brianjgeiger/osf.io,aaxelb/osf.io,caseyrollins/osf.io,caneruguz/osf.io,adlius/osf.io,doublebits/osf.io,caneruguz/osf.io,alexschiller/osf.io,laurenrevere/osf.io,cslzchen/osf.io,felliott/osf.io,wearpants/osf.io,DanielSBrown/osf.io,billyhunt/osf.io,amyshi188/osf.io,hmoco/osf.io,baylee-d/osf.io,samchrisinger/osf.io,SSJohns/osf.io,billyhunt/osf.io,asanfilippo7/osf.io,felliott/osf.io,CenterForOpenScience/osf.io,cslzchen/osf.io,mfraezz/osf.io,samchrisinger/osf.io,abought/osf.io,kch8qx/osf.io,zamattiac/osf.io,TomBaxter/osf.io,CenterForOpenScience/osf.io,asanfilippo7/osf.io,mluke93/osf.io,chrisseto/osf.io,billyhunt/osf.io,SSJohns/osf.io,pattisdr/osf.io,jnayak1/osf.io,felliott/osf.io,hmoco/osf.io,rdhyee/osf.io,crcresearch/osf.io,Nesiehr/osf.io,leb2dg/osf.io | ---
+++
@@ -1,24 +1,18 @@
import urlparse
-import celery
import requests
from celery.utils.log import get_task_logger
+
from api.base import settings
-from framework.tasks import app as celery_app
-
logger = get_task_logger(__name__)
-
-class VarnishTask(celery.Task):
- abstract = True
- max_retries = 5
def get_varnish_servers():
# TODO: this should get the varnish servers from HAProxy or a setting
return settings.VARNISH_SERVERS
-@celery_app.task(base=VarnishTask, name='caching_tasks.ban_url')
def ban_url(url):
+ timeout = 0.5 # 500ms timeout for bans
if settings.ENABLE_VARNISH:
parsed_url = urlparse.urlparse(url)
@@ -29,7 +23,7 @@
netloc=varnish_parsed_url.netloc,
path=parsed_url.path
)
- response = requests.request('BAN', ban_url, headers=dict(
+ response = requests.request('BAN', ban_url, timeout=timeout, headers=dict(
Host=parsed_url.hostname
))
if not response.ok: |
13d9edb6d643d0955663f78b610fee6106e7128e | KISSmetrics/client.py | KISSmetrics/client.py | # -*- coding: utf-8 -*-
import KISSmetrics
from KISSmetrics import request
from urllib3 import PoolManager
class Client:
def __init__(self, key, trk_host=KISSmetrics.TRACKING_HOSTNAME,
trk_proto=KISSmetrics.TRACKING_PROTOCOL):
self.key = key
if trk_proto not in ['http', 'https']:
raise ValueError('trk_proto must be one of (http, https)')
self.http = PoolManager()
self.trk_host = trk_host
self.trk_proto = trk_proto
def request(self, query, method="GET"):
url = '%s://%s/%s' % (self.trk_proto, self.trk_host, query)
return self.http.request(method, url)
def record(self, person, event, properties=None, timestamp=None,
uri=KISSmetrics.RECORD_URI):
this_request = request.record(self.key, person, event,
timestamp=timestamp,
properties=properties, uri=uri)
return self.request(this_request)
def set(self, person, properties=None, timestamp=None,
uri=KISSmetrics.SET_URI):
this_request = request.set(self.key, person, timestamp=timestamp,
properties=properties, uri=uri)
return self.request(this_request)
def alias(self, person, identity, uri=KISSmetrics.ALIAS_URI):
this_request = request.alias(self.key, person, identity, uri=uri)
return self.request(this_request)
| # -*- coding: utf-8 -*-
import KISSmetrics
from KISSmetrics import request
from urllib3 import PoolManager
class Client:
def __init__(self, key, trk_host=KISSmetrics.TRACKING_HOSTNAME,
trk_proto=KISSmetrics.TRACKING_PROTOCOL):
self.key = key
if trk_proto not in ('http', 'https'):
raise ValueError('trk_proto must be one of (http, https)')
self.http = PoolManager()
self.trk_host = trk_host
self.trk_proto = trk_proto
def request(self, query, method="GET"):
url = '%s://%s/%s' % (self.trk_proto, self.trk_host, query)
return self.http.request(method, url)
def record(self, person, event, properties=None, timestamp=None,
uri=KISSmetrics.RECORD_URI):
this_request = request.record(self.key, person, event,
timestamp=timestamp,
properties=properties, uri=uri)
return self.request(this_request)
def set(self, person, properties=None, timestamp=None,
uri=KISSmetrics.SET_URI):
this_request = request.set(self.key, person, timestamp=timestamp,
properties=properties, uri=uri)
return self.request(this_request)
def alias(self, person, identity, uri=KISSmetrics.ALIAS_URI):
this_request = request.alias(self.key, person, identity, uri=uri)
return self.request(this_request)
| Use a tuple instead of a list for protocol membership test | Use a tuple instead of a list for protocol membership test
| Python | mit | kissmetrics/py-KISSmetrics | ---
+++
@@ -10,7 +10,7 @@
def __init__(self, key, trk_host=KISSmetrics.TRACKING_HOSTNAME,
trk_proto=KISSmetrics.TRACKING_PROTOCOL):
self.key = key
- if trk_proto not in ['http', 'https']:
+ if trk_proto not in ('http', 'https'):
raise ValueError('trk_proto must be one of (http, https)')
self.http = PoolManager()
self.trk_host = trk_host |
1bba76808aa5c598f1558cd127d8ed4a006692e1 | tests/conftest.py | tests/conftest.py | import os
def pytest_configure(config):
if 'USE_QT_API' in os.environ:
os.environ['QT_API'] = os.environ['USE_QT_API'].lower()
def pytest_report_header(config):
versions = os.linesep
versions += 'PyQt4: '
try:
from PyQt4 import Qt
versions += "PyQt: {0} - Qt: {1}".format(Qt.PYQT_VERSION_STR, Qt.QT_VERSION_STR)
except ImportError:
versions += 'not installed'
except AttributeError:
versions += 'unknown version'
versions += os.linesep
versions += 'PyQt5: '
try:
from PyQt5 import Qt
versions += "PyQt: {0} - Qt: {1}".format(Qt.PYQT_VERSION_STR, Qt.QT_VERSION_STR)
except ImportError:
versions += 'not installed'
except AttributeError:
versions += 'unknown version'
versions += os.linesep
versions += 'PySide: '
try:
import PySide
from PySide import QtCore
versions += "PySide: {0} - Qt: {1}".format(PySide.__version__, QtCore.__version__)
except ImportError:
versions += 'not installed'
except AttributeError:
versions += 'unknown version'
versions += os.linesep
return versions
| import os
def pytest_configure(config):
if 'USE_QT_API' in os.environ:
os.environ['QT_API'] = os.environ['USE_QT_API'].lower()
# We need to import qtpy here to make sure that the API versions get set
# straight away.
import qtpy
def pytest_report_header(config):
versions = os.linesep
versions += 'PyQt4: '
try:
from PyQt4 import Qt
versions += "PyQt: {0} - Qt: {1}".format(Qt.PYQT_VERSION_STR, Qt.QT_VERSION_STR)
except ImportError:
versions += 'not installed'
except AttributeError:
versions += 'unknown version'
versions += os.linesep
versions += 'PyQt5: '
try:
from PyQt5 import Qt
versions += "PyQt: {0} - Qt: {1}".format(Qt.PYQT_VERSION_STR, Qt.QT_VERSION_STR)
except ImportError:
versions += 'not installed'
except AttributeError:
versions += 'unknown version'
versions += os.linesep
versions += 'PySide: '
try:
import PySide
from PySide import QtCore
versions += "PySide: {0} - Qt: {1}".format(PySide.__version__, QtCore.__version__)
except ImportError:
versions += 'not installed'
except AttributeError:
versions += 'unknown version'
versions += os.linesep
return versions
| Make sure we import qtpy before importing any Qt wrappers directly | Make sure we import qtpy before importing any Qt wrappers directly | Python | mit | spyder-ide/qtpy,goanpeca/qtpy,davvid/qtpy,davvid/qtpy,goanpeca/qtpy | ---
+++
@@ -2,8 +2,13 @@
def pytest_configure(config):
+
if 'USE_QT_API' in os.environ:
os.environ['QT_API'] = os.environ['USE_QT_API'].lower()
+
+ # We need to import qtpy here to make sure that the API versions get set
+ # straight away.
+ import qtpy
def pytest_report_header(config): |
8359d60480371a8f63bdd4ea1b7cf03f231c1350 | djangopress/settings_tinymce.py | djangopress/settings_tinymce.py | # if you want support for tinymce in the admin pages
# add tinymce to the installed apps (after installing if needed)
# and then import these settings, or copy and adjust as needed
TINYMCE_DEFAULT_CONFIG = {
'relative_urls': False,
'plugins': "table code image link colorpicker textcolor wordcount",
'tools': "inserttable",
'toolbar': "undo redo | styleselect | bold italic underline strikethrough | alignleft aligncenter alignright alignjustify | bullist numlist outdent indent | link image | forecolor backcolor",
'extended_valid_elements': 'script[language|type|src],events[template|start],#gallery[class|id|show_description|show_title|count|slider]'
}
TINYMCE_JS_URL = "/static/js/tinymce/tinymce.min.js" | # if you want support for tinymce in the admin pages
# add tinymce to the installed apps (after installing if needed)
# and then import these settings, or copy and adjust as needed
TINYMCE_DEFAULT_CONFIG = {
'relative_urls': False,
'plugins': "table code image link colorpicker textcolor wordcount",
'tools': "inserttable",
'toolbar': "undo redo | styleselect | bold italic underline strikethrough | alignleft aligncenter alignright alignjustify | bullist numlist outdent indent | link image | forecolor backcolor",
'extended_valid_elements': 'script[language|type|src],events[template|start],#gallery[class|id|show_description|show_title|count|slider],#show_blog_latest[class|id|words|images|blog|count]'
}
TINYMCE_JS_URL = "/static/js/tinymce/tinymce.min.js" | Update settings for tinymce to allow show_blog_latest tag | Update settings for tinymce to allow show_blog_latest tag
| Python | mit | codefisher/djangopress,codefisher/djangopress,codefisher/djangopress,codefisher/djangopress | ---
+++
@@ -7,6 +7,6 @@
'plugins': "table code image link colorpicker textcolor wordcount",
'tools': "inserttable",
'toolbar': "undo redo | styleselect | bold italic underline strikethrough | alignleft aligncenter alignright alignjustify | bullist numlist outdent indent | link image | forecolor backcolor",
- 'extended_valid_elements': 'script[language|type|src],events[template|start],#gallery[class|id|show_description|show_title|count|slider]'
+ 'extended_valid_elements': 'script[language|type|src],events[template|start],#gallery[class|id|show_description|show_title|count|slider],#show_blog_latest[class|id|words|images|blog|count]'
}
TINYMCE_JS_URL = "/static/js/tinymce/tinymce.min.js" |
3ca7c667cbf37499dc959b336b9ff0e88f5d4275 | dbarray/tests/run.py | dbarray/tests/run.py | """From http://stackoverflow.com/a/12260597/400691"""
import sys
from django.conf import settings
settings.configure(
DATABASES={
'default': {
'ENGINE': 'django.db.backends.postgresql_psycopg2',
'NAME': 'dbarray',
'HOST': 'localhost'
}
},
INSTALLED_APPS=(
'dbarray.tests',
# 'django.contrib.auth',
# 'django.contrib.contenttypes',
# 'django.contrib.sessions',
# 'django.contrib.admin',
),
)
try:
from django.test.runner import DiscoverRunner
except ImportError:
# Fallback for django < 1.6
from discover_runner import DiscoverRunner
test_runner = DiscoverRunner(verbosity=1)
failures = test_runner.run_tests(['dbarray'])
if failures:
sys.exit(1)
| """From http://stackoverflow.com/a/12260597/400691"""
import sys
from django.conf import settings
settings.configure(
DATABASES={
'default': {
'ENGINE': 'django.db.backends.postgresql_psycopg2',
'NAME': 'dbarray',
'HOST': 'localhost'
}
},
INSTALLED_APPS=('dbarray.tests',),
)
try:
from django.test.runner import DiscoverRunner
except ImportError:
# Fallback for django < 1.6
from discover_runner import DiscoverRunner
test_runner = DiscoverRunner(verbosity=1)
failures = test_runner.run_tests(['dbarray'])
if failures:
sys.exit(1)
| Remove commented code . | Remove commented code [ci skip].
| Python | bsd-3-clause | ecometrica/django-dbarray | ---
+++
@@ -12,14 +12,9 @@
'HOST': 'localhost'
}
},
- INSTALLED_APPS=(
- 'dbarray.tests',
- # 'django.contrib.auth',
- # 'django.contrib.contenttypes',
- # 'django.contrib.sessions',
- # 'django.contrib.admin',
- ),
+ INSTALLED_APPS=('dbarray.tests',),
)
+
try:
from django.test.runner import DiscoverRunner
except ImportError: |
a1bf03f69b9cadddcc7e0015788f23f9bad0f862 | apps/splash/views.py | apps/splash/views.py | import datetime
from django.shortcuts import render
from apps.splash.models import SplashEvent, SplashYear
def index(request):
# I'm really sorry ...
splash_year = SplashYear.objects.get(start_date__gt=str(datetime.date.today() - datetime.timedelta(180)))
return render(request, 'splash/base.html', {'splash_year': splash_year })
# And I'm really sorry for this ...
def _merge_events(splash_events):
events = []
for event in splash_events:
if len(events) > 0 and event.start_time.strftime('%d-%m') == events[-1][0].start_time.strftime('%d-%m'):
events[-1].append(event)
else:
events.append([event])
return events
| import datetime
from django.shortcuts import render
from apps.splash.models import SplashEvent, SplashYear
def index(request):
# I'm really sorry ...
splash_year = SplashYear.objects.get(start_date__gt=str(datetime.date.today() - datetime.timedelta(180)))
splash_year.events = _merge_events(splash_year.splash_events.all())
return render(request, 'splash/base.html', {'splash_year': splash_year })
# And I'm really sorry for this ...
def _merge_events(splash_events):
events = []
for event in splash_events:
if len(events) > 0 and event.start_time.strftime('%d-%m') == events[-1][0].start_time.strftime('%d-%m'):
events[-1].append(event)
else:
events.append([event])
return events
| Append event merging on splash_events | Append event merging on splash_events
| Python | mit | dotKom/onlineweb4,dotKom/onlineweb4,dotKom/onlineweb4,dotKom/onlineweb4 | ---
+++
@@ -6,6 +6,9 @@
def index(request):
# I'm really sorry ...
splash_year = SplashYear.objects.get(start_date__gt=str(datetime.date.today() - datetime.timedelta(180)))
+
+ splash_year.events = _merge_events(splash_year.splash_events.all())
+
return render(request, 'splash/base.html', {'splash_year': splash_year })
|
e722dbc13dcf1c30086fb3aff9afd89d2bddb409 | validator/jschematest.py | validator/jschematest.py | # Testing jsonschema error printing capabilities
import jsonschema
import json
# Custom printing of errors (skip no-information messages)
# TODO: solve repetition of same error (with different paths)
def print_errors(errors, indent=0):
next_indent = indent + 2
for error in errors:
msg = error.message
print(' '*indent, end='')
if error.validator in ['anyOf', 'oneOf', 'allOf']:
if 'questionid' in error.instance:
qid = error.instance['questionid']
print('Question with ID {} is not valid:'.format(qid))
elif 'chapterid' in error.instance:
chid = error.instance['chapterid']
print('Chapter with ID {} is not valid:'.format(chid))
else:
print(msg)
else:
print(msg)
new_errors = sorted(error.context, key=lambda e: e.schema_path)
print_errors(new_errors, next_indent)
with open('../schema.alt.json') as f:
schema = json.load(f)
with open('../datamodel/core/chapter1.json') as f:
instance = json.load(f)
v = jsonschema.Draft4Validator(schema)
errors = sorted(v.iter_errors(instance), key=lambda e: e.path)
print_errors(errors)
| # Testing jsonschema error printing capabilities
import jsonschema
import json
# Custom printing of errors (skip no-information messages)
# TODO: solve repetition of same error (with different paths)
def print_errors(errors, indent=0):
next_indent = indent + 2
for error in errors:
msg = error.message
print(' '*indent, end='')
if error.validator in ['anyOf', 'oneOf', 'allOf']:
if 'questionid' in error.instance:
qid = error.instance['questionid']
print('Question with ID {} is not valid:'.format(qid))
elif 'chapterid' in error.instance:
chid = error.instance['chapterid']
print('Chapter with ID {} is not valid:'.format(chid))
else:
print(msg)
else:
print(msg)
new_errors = sorted(error.context, key=lambda e: e.schema_path)
print_errors(new_errors, next_indent)
if __name__ == "__main__":
import os
with open('../schema-chapter.json') as f:
schema = json.load(f)
v = jsonschema.Draft4Validator(schema)
for root, dirs, files in os.walk('../datamodel'):
for name in files:
if name.endswith('.json'):
print (os.path.join(root, name))
with open(os.path.join(root, name)) as f:
instance = json.load(f)
errors = sorted(v.iter_errors(instance), key=lambda e: e.path)
print_errors(errors)
| Make chapter validator loop over all chapter files in the knowledge model | Make chapter validator loop over all chapter files in the knowledge model
| Python | apache-2.0 | DataStewardshipPortal/ds-km,CCMi-FIT/ds-km,CCMi-FIT/ds-km-core | ---
+++
@@ -1,7 +1,6 @@
# Testing jsonschema error printing capabilities
import jsonschema
import json
-
# Custom printing of errors (skip no-information messages)
# TODO: solve repetition of same error (with different paths)
@@ -24,10 +23,19 @@
new_errors = sorted(error.context, key=lambda e: e.schema_path)
print_errors(new_errors, next_indent)
-with open('../schema.alt.json') as f:
- schema = json.load(f)
-with open('../datamodel/core/chapter1.json') as f:
- instance = json.load(f)
-v = jsonschema.Draft4Validator(schema)
-errors = sorted(v.iter_errors(instance), key=lambda e: e.path)
-print_errors(errors)
+
+if __name__ == "__main__":
+ import os
+
+ with open('../schema-chapter.json') as f:
+ schema = json.load(f)
+ v = jsonschema.Draft4Validator(schema)
+
+ for root, dirs, files in os.walk('../datamodel'):
+ for name in files:
+ if name.endswith('.json'):
+ print (os.path.join(root, name))
+ with open(os.path.join(root, name)) as f:
+ instance = json.load(f)
+ errors = sorted(v.iter_errors(instance), key=lambda e: e.path)
+ print_errors(errors) |
1f4fea5d4bb67f84defa1693e0ea26295de489ff | helios/conf/settings.py | helios/conf/settings.py | # -*- coding: utf-8 -*-
from django.conf import settings
PAGINATE_BY = getattr(settings, 'STORE_PAGINATE_BY', 50)
IS_MULTILINGUAL = getattr(settings, 'STORE_IS_MULTILINGUAL', False)
HAS_CURRENCIES = getattr(settings, 'STORE_HAS_CURRENCIES', False)
USE_PAYPAL = getattr(settings, 'STORE_USE_PAYPAL', False)
PRODUCT_MODEL = getattr(settings, 'STORE_PRODUCT_MODEL', 'helios.store.models.Product')
CART = getattr(settings, 'STORE_CART', 'helios.store.cart')
| # -*- coding: utf-8 -*-
from django.conf import settings
DEBUG = getattr(settings, 'STORE_DEBUG', False)
PAGINATE_BY = getattr(settings, 'STORE_PAGINATE_BY', 50)
IS_MULTILINGUAL = getattr(settings, 'STORE_IS_MULTILINGUAL', False)
HAS_CURRENCIES = getattr(settings, 'STORE_HAS_CURRENCIES', False)
USE_PAYPAL = getattr(settings, 'STORE_USE_PAYPAL', False)
PRODUCT_MODEL = getattr(settings, 'STORE_PRODUCT_MODEL', 'helios.store.models.Product')
CART = getattr(settings, 'STORE_CART', 'helios.store.cart')
| Set the DEBUG store setting | Set the DEBUG store setting
| Python | bsd-3-clause | panosl/helios | ---
+++
@@ -2,6 +2,7 @@
from django.conf import settings
+DEBUG = getattr(settings, 'STORE_DEBUG', False)
PAGINATE_BY = getattr(settings, 'STORE_PAGINATE_BY', 50)
IS_MULTILINGUAL = getattr(settings, 'STORE_IS_MULTILINGUAL', False)
HAS_CURRENCIES = getattr(settings, 'STORE_HAS_CURRENCIES', False) |
61b59bfdf9581a280263f4049e7d257fa47cdad0 | teknologr/registration/views.py | teknologr/registration/views.py | from django.shortcuts import render, redirect
from django.conf import settings
from django.views import View
from members.programmes import DEGREE_PROGRAMME_CHOICES
from registration.forms import RegistrationForm
from registration.mailutils import mailApplicantSubmission
class BaseView(View):
context = {'DEBUG': settings.DEBUG}
class HomeView(BaseView):
template = 'registration.html'
def get(self, request):
self.context['programmes'] = DEGREE_PROGRAMME_CHOICES
self.context['form'] = RegistrationForm()
return render(request, self.template, self.context)
class SubmitView(BaseView):
template = 'submit.html'
def post(self, request):
form = RegistrationForm(request.POST)
if form.is_valid():
registration = form.instance
self.context['name'] = registration.preferred_name or registration.given_names.split(' ')[0]
self.context['email'] = registration.email
# FIXME: handle situation where email is not sent (e.g. admin log tool)
mailApplicantSubmission(self.context)
registration.save()
else:
self.context['form'] = form
return render(request, HomeView.template, self.context, status=400)
return render(request, self.template, self.context)
| from django.shortcuts import render, redirect
from django.conf import settings
from django.views import View
from members.programmes import DEGREE_PROGRAMME_CHOICES
from registration.forms import RegistrationForm
from registration.mailutils import mailApplicantSubmission
class BaseView(View):
context = {'DEBUG': settings.DEBUG}
class HomeView(BaseView):
template = 'registration.html'
def get(self, request):
self.context['programmes'] = DEGREE_PROGRAMME_CHOICES
self.context['form'] = RegistrationForm()
return render(request, self.template, self.context)
class SubmitView(BaseView):
template = 'submit.html'
def get(self, request, **kwargs):
previous_context = request.session.pop('context', None)
if not previous_context:
return redirect('registration.views.home')
return render(request, self.template, previous_context)
def post(self, request):
form = RegistrationForm(request.POST)
if form.is_valid():
registration = form.instance
self.context['name'] = registration.preferred_name or registration.given_names.split(' ')[0]
self.context['email'] = registration.email
# FIXME: handle situation where email is not sent (e.g. admin log tool)
mailApplicantSubmission(self.context)
registration.save()
else:
self.context['form'] = form
return render(request, HomeView.template, self.context, status=400)
request.session['context'] = self.context
return redirect('registration.views.submit')
| Fix registration form replay issue | Fix registration form replay issue
| Python | mit | Teknologforeningen/teknologr.io,Teknologforeningen/teknologr.io,Teknologforeningen/teknologr.io,Teknologforeningen/teknologr.io | ---
+++
@@ -22,6 +22,13 @@
class SubmitView(BaseView):
template = 'submit.html'
+ def get(self, request, **kwargs):
+ previous_context = request.session.pop('context', None)
+ if not previous_context:
+ return redirect('registration.views.home')
+
+ return render(request, self.template, previous_context)
+
def post(self, request):
form = RegistrationForm(request.POST)
if form.is_valid():
@@ -38,4 +45,5 @@
self.context['form'] = form
return render(request, HomeView.template, self.context, status=400)
- return render(request, self.template, self.context)
+ request.session['context'] = self.context
+ return redirect('registration.views.submit') |
3fad3c64a317956265c14c82c182b66979ba8554 | greatbigcrane/preferences/forms.py | greatbigcrane/preferences/forms.py | """
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from preferences.models import Preference
class PreferencesForm(forms.Form):
def __init__(self, *args, **kwargs):
super(PreferencesForm, self).__init__(*args, **kwargs)
for preference in Preference.objects.all():
self.fields[preference.name] = forms.CharField()
| """
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
import os.path
from django import forms
from preferences.models import Preference
class PreferencesForm(forms.Form):
def __init__(self, *args, **kwargs):
super(PreferencesForm, self).__init__(*args, **kwargs)
for preference in Preference.objects.all():
self.fields[preference.name] = forms.CharField()
def clean(self):
# Apparently clean__projects_directory doesn't work with dynamic forms
if not self.cleaned_data['projects_directory'].endswith(os.path.sep):
self.cleaned_data['projects_directory'] += os.path.sep
return self.cleaned_data
| Append a trailing newline to the project directory | Append a trailing newline to the project directory
| Python | apache-2.0 | pnomolos/greatbigcrane,pnomolos/greatbigcrane | ---
+++
@@ -13,6 +13,7 @@
See the License for the specific language governing permissions and
limitations under the License.
"""
+import os.path
from django import forms
from preferences.models import Preference
@@ -21,3 +22,9 @@
super(PreferencesForm, self).__init__(*args, **kwargs)
for preference in Preference.objects.all():
self.fields[preference.name] = forms.CharField()
+
+ def clean(self):
+ # Apparently clean__projects_directory doesn't work with dynamic forms
+ if not self.cleaned_data['projects_directory'].endswith(os.path.sep):
+ self.cleaned_data['projects_directory'] += os.path.sep
+ return self.cleaned_data |
62454216b7d0426c23d75ba4aafd761093447e63 | compiler.py | compiler.py | #!/usr/local/bin/python3
import http.client, urllib.parse, sys, argparse
def main():
parser = argparse.ArgumentParser()
parser.add_argument("--js", default='/dev/stdin', help="Input file")
parser.add_argument("--js_output_file", default='/dev/stdout', help="Output file")
parser.add_argument("--compilation_level", default='WHITESPACE_ONLY', choices=['WHITESPACE_ONLY', 'SIMPLE_OPTIMIZATIONS', 'ADVANCED_OPTIMIZATIONS'], help="Compilation level")
args = parser.parse_args()
js_code = open(args.js, 'r')
params = urllib.parse.urlencode([
('js_code', js_code.read()),
('compilation_level', args.compilation_level),
('output_format', 'text'),
('output_info', 'compiled_code'),
])
js_code.close()
headers = { "Content-Type": "application/x-www-form-urlencoded;charset=UTF-8" }
conn = http.client.HTTPSConnection('closure-compiler.appspot.com')
conn.request('POST', '/compile', params, headers)
response = conn.getresponse()
data = response.read()
output_code = open(args.js_output_file, 'w')
output_code.write(data.decode("utf-8"))
output_code.close()
conn.close()
if __name__ == "__main__":
main()
| #!/usr/local/bin/python3
import http.client, urllib.parse, sys, argparse
def main():
parser = argparse.ArgumentParser()
parser.add_argument("--js", default='/dev/stdin', help="Input file")
parser.add_argument("--js_output_file", default='/dev/stdout', help="Output file")
parser.add_argument("--compilation_level", default='WHITESPACE_ONLY', choices=['WHITESPACE_ONLY', 'SIMPLE_OPTIMIZATIONS', 'ADVANCED_OPTIMIZATIONS'], help="Compilation level")
parser.add_argument("--js_externs", default='', help="Declare some external js vars and functions separated with semicolon")
args = parser.parse_args()
js_code = open(args.js, 'r')
params = urllib.parse.urlencode([
('js_code', js_code.read()),
('compilation_level', args.compilation_level),
('output_format', 'text'),
('output_info', 'compiled_code'),
('js_externs', args.js_externs),
])
js_code.close()
headers = { "Content-Type": "application/x-www-form-urlencoded;charset=UTF-8" }
conn = http.client.HTTPSConnection('closure-compiler.appspot.com')
conn.request('POST', '/compile', params, headers)
response = conn.getresponse()
data = response.read()
output_code = open(args.js_output_file, 'w')
output_code.write(data.decode("utf-8"))
output_code.close()
conn.close()
if __name__ == "__main__":
main()
| Add js_externs parameter for API | Add js_externs parameter for API
| Python | mit | femtopixel/docker-google-closure-compiler-api,femtopixel/docker-google-closure-compiler-api | ---
+++
@@ -7,6 +7,7 @@
parser.add_argument("--js", default='/dev/stdin', help="Input file")
parser.add_argument("--js_output_file", default='/dev/stdout', help="Output file")
parser.add_argument("--compilation_level", default='WHITESPACE_ONLY', choices=['WHITESPACE_ONLY', 'SIMPLE_OPTIMIZATIONS', 'ADVANCED_OPTIMIZATIONS'], help="Compilation level")
+ parser.add_argument("--js_externs", default='', help="Declare some external js vars and functions separated with semicolon")
args = parser.parse_args()
js_code = open(args.js, 'r')
@@ -15,6 +16,7 @@
('compilation_level', args.compilation_level),
('output_format', 'text'),
('output_info', 'compiled_code'),
+ ('js_externs', args.js_externs),
])
js_code.close() |
d8344f9bea9cfbc8ab22b952f223d2365de907b4 | cli_helpers/tabular_output/terminaltables_adapter.py | cli_helpers/tabular_output/terminaltables_adapter.py | # -*- coding: utf-8 -*-
"""Format adapter for the terminaltables module."""
import terminaltables
from cli_helpers.utils import filter_dict_by_key
from .preprocessors import (bytes_to_string, align_decimals,
override_missing_value)
supported_formats = ('ascii', 'double', 'github')
preprocessors = (bytes_to_string, override_missing_value, align_decimals)
def adapter(data, headers, table_format=None, **kwargs):
"""Wrap terminaltables inside a function for TabularOutputFormatter."""
keys = ('title', )
table_format_handler = {
'ascii': terminaltables.AsciiTable,
'double': terminaltables.DoubleTable,
'github': terminaltables.GithubFlavoredMarkdownTable,
}
try:
table = table_format_handler[table_format]
except KeyError:
raise ValueError('unrecognized table format: {}'.format(table_format))
t = table([headers] + data, **filter_dict_by_key(kwargs, keys))
return t.table
| # -*- coding: utf-8 -*-
"""Format adapter for the terminaltables module."""
import terminaltables
from cli_helpers.utils import filter_dict_by_key
from .preprocessors import (bytes_to_string, align_decimals,
override_missing_value)
supported_formats = ('ascii', 'double', 'github')
preprocessors = (bytes_to_string, override_missing_value, align_decimals)
def adapter(data, headers, table_format=None, **kwargs):
"""Wrap terminaltables inside a function for TabularOutputFormatter."""
keys = ('title', )
table_format_handler = {
'ascii': terminaltables.AsciiTable,
'double': terminaltables.DoubleTable,
'github': terminaltables.GithubFlavoredMarkdownTable,
}
table = table_format_handler[table_format]
t = table([headers] + data, **filter_dict_by_key(kwargs, keys))
return t.table
| Remove try/except that should never be hit. | Remove try/except that should never be hit.
Increases code coverage.
| Python | bsd-3-clause | dbcli/cli_helpers,dbcli/cli_helpers | ---
+++
@@ -21,10 +21,7 @@
'github': terminaltables.GithubFlavoredMarkdownTable,
}
- try:
- table = table_format_handler[table_format]
- except KeyError:
- raise ValueError('unrecognized table format: {}'.format(table_format))
+ table = table_format_handler[table_format]
t = table([headers] + data, **filter_dict_by_key(kwargs, keys))
return t.table |
9a58d241e61301b9390b17e391e4b65a3ea85071 | squadron/libraries/apt/__init__.py | squadron/libraries/apt/__init__.py | import os
import subprocess
from string import find
def run_command(command):
p = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
out, err = p.communicate()
return out,err
def schema():
"""
This returns
"""
return { 'title': 'apt schema',
'type': 'string'
}
def verify(inputhashes):
"""
"""
failed = []
for package in inputhashes:
out = run_command(['dpkg-query', '-W', package])[0]
#We expect the output to contain the version
#Any error doesn't get captured, so out will be empty (yes this is weird)
if(find(out, package) == -1):
failed.append(package)
return failed
def apply(inputhashes, dry_run=True):
failed = []
for package in inputhashes:
out = run_command(['apt-get', 'install', '-y', package])
if(find(out[1], 'Permission denied') != -1):
failed.append(package) #Install failed because we're not root
if(find(out[0], ('Setting up ' + package)) != -1 and find(out[0], (package + ' already the newest version')) != -1):
#Something else happened, we weren't installed and we didn't get installed
failed.append(package)
print out
return failed
| import os
import subprocess
from string import find
def run_command(command):
p = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
out, err = p.communicate()
return out,err
def schema():
"""
This returns
"""
return { 'title': 'apt schema',
'type': 'string'
}
def verify(inputhashes):
"""
"""
failed = []
for package in inputhashes:
out = run_command(['dpkg-query', '-W', package])[0]
#We expect the output to contain the version
#Any error doesn't get captured, so out will be empty (yes this is weird)
if(find(out, package) == -1):
failed.append(package)
return failed
def apply(inputhashes, dry_run=True):
failed = []
for package in inputhashes:
out = run_command(['apt-get', 'install', '-y', package])
if(find(out[1], 'Permission denied') != -1):
failed.append(package) # Install failed because we're not root
if(find(out[0], ('Setting up ' + package)) != -1 and find(out[0], (package + ' already the newest version')) != -1):
# Something else happened, we weren't installed and we didn't get installed
failed.append(package)
return failed
| Remove extra print in apt | Remove extra print in apt
| Python | mit | gosquadron/squadron,gosquadron/squadron | ---
+++
@@ -33,11 +33,10 @@
for package in inputhashes:
out = run_command(['apt-get', 'install', '-y', package])
if(find(out[1], 'Permission denied') != -1):
- failed.append(package) #Install failed because we're not root
+ failed.append(package) # Install failed because we're not root
if(find(out[0], ('Setting up ' + package)) != -1 and find(out[0], (package + ' already the newest version')) != -1):
- #Something else happened, we weren't installed and we didn't get installed
+ # Something else happened, we weren't installed and we didn't get installed
failed.append(package)
- print out
return failed
|
aae84224b5a2d1689b1739da319d140474702c96 | zuice/bindings.py | zuice/bindings.py | class Bindings(object):
def __init__(self):
self._bindings = {}
def bind(self, key):
if key in self:
raise AlreadyBoundException("Cannot rebind key: %s" % key)
return Binder(key, self._bindings)
def copy(self):
copy = Bindings()
copy._bindings = self._bindings.copy()
return copy
def update(self, bindings):
for key in bindings._bindings:
if key in self._bindings:
raise AlreadyBoundException("Key already bound: %s" % key)
self._bindings.update(bindings._bindings)
def __contains__(self, key):
return key in self._bindings
def __getitem__(self, key):
return self._bindings[key]
class Binder(object):
def __init__(self, key, bindings):
self._bound = False
self._key = key
self._bindings = bindings
def to_instance(self, instance):
self.to_provider(lambda: instance)
def to_key(self, key):
if key is self._key:
raise TypeError("Cannot bind a key to itself")
self.to_provider(lambda injector: injector.get(key))
def to_type(self, key):
return self.to_key(key)
def to_provider(self, provider):
if self._bound:
raise AlreadyBoundException()
self._bound = True
self._bindings[self._key] = provider
class AlreadyBoundException(Exception):
pass
| class Bindings(object):
def __init__(self):
self._bindings = {}
def bind(self, key, provider=None):
if key in self:
raise AlreadyBoundException("Cannot rebind key: %s" % key)
if provider is None:
return Binder(key, self)
else:
self._bindings[key] = provider
def copy(self):
copy = Bindings()
copy._bindings = self._bindings.copy()
return copy
def update(self, bindings):
for key in bindings._bindings:
if key in self._bindings:
raise AlreadyBoundException("Key already bound: %s" % key)
self._bindings.update(bindings._bindings)
def __contains__(self, key):
return key in self._bindings
def __getitem__(self, key):
return self._bindings[key]
class Binder(object):
def __init__(self, key, bindings):
self._key = key
self._bindings = bindings
def to_instance(self, instance):
self.to_provider(lambda: instance)
def to_key(self, key):
if key is self._key:
raise TypeError("Cannot bind a key to itself")
self.to_provider(lambda injector: injector.get(key))
def to_type(self, key):
return self.to_key(key)
def to_provider(self, provider):
self._bindings.bind(self._key, provider)
class AlreadyBoundException(Exception):
pass
| Simplify detection of keys already bound | Simplify detection of keys already bound
| Python | bsd-2-clause | mwilliamson/zuice | ---
+++
@@ -2,10 +2,14 @@
def __init__(self):
self._bindings = {}
- def bind(self, key):
+ def bind(self, key, provider=None):
if key in self:
raise AlreadyBoundException("Cannot rebind key: %s" % key)
- return Binder(key, self._bindings)
+
+ if provider is None:
+ return Binder(key, self)
+ else:
+ self._bindings[key] = provider
def copy(self):
copy = Bindings()
@@ -26,7 +30,6 @@
class Binder(object):
def __init__(self, key, bindings):
- self._bound = False
self._key = key
self._bindings = bindings
@@ -42,10 +45,7 @@
return self.to_key(key)
def to_provider(self, provider):
- if self._bound:
- raise AlreadyBoundException()
- self._bound = True
- self._bindings[self._key] = provider
+ self._bindings.bind(self._key, provider)
class AlreadyBoundException(Exception):
pass |
66212e51341562f156353a0ae195d15b0d22b21b | scripts/import.py | scripts/import.py | #!/usr/bin/env python
import sys
import os
import requests
import multiprocessing
def list_files(directory):
for root, subdirs, files in os.walk(directory):
print("ROOT: {}".format(root))
for file in files:
yield os.path.abspath(os.path.join(root, file))
for subdir in subdirs:
for subfile in list_files(subdir):
yield subfile
if __name__ == "__main__":
_, base_url, access_token, listing_dir = sys.argv
endpoint = "{}/services".format(base_url)
print("Base URL: {}".format(base_url))
print("Access token: {}".format(access_token))
print("Listing dir: {}".format(listing_dir))
def post_file(file_path):
with open(file_path) as f:
response = requests.post(
endpoint,
data='{"services":%s}' % f.read(),
headers={
"content-type": "application/json",
"authorization": "Bearer {}".format(access_token),
})
return response
pool = multiprocessing.Pool(10)
for result in pool.imap(post_file, list_files(listing_dir)):
print(result)
| #!/usr/bin/env python
import sys
import json
import os
import requests
import multiprocessing
def list_files(directory):
for root, subdirs, files in os.walk(directory):
print("ROOT: {}".format(root))
for file in files:
yield os.path.abspath(os.path.join(root, file))
for subdir in subdirs:
for subfile in list_files(subdir):
yield subfile
if __name__ == "__main__":
_, base_url, access_token, listing_dir = sys.argv
endpoint = "{}/services".format(base_url)
print("Base URL: {}".format(base_url))
print("Access token: {}".format(access_token))
print("Listing dir: {}".format(listing_dir))
def put_file(file_path):
with open(file_path) as f:
data = json.load(f)
data = {'services': data}
url = '{}/{}'.format(endpoint, data['services']['id'])
response = requests.put(
url,
data=json.dumps(data),
headers={
"content-type": "application/json",
"authorization": "Bearer {}".format(access_token),
})
return response
pool = multiprocessing.Pool(10)
for result in pool.imap(put_file, list_files(listing_dir)):
print(result)
| PUT services rather than POSTing them | PUT services rather than POSTing them
| Python | mit | RichardKnop/digitalmarketplace-api,RichardKnop/digitalmarketplace-api,RichardKnop/digitalmarketplace-api,mtekel/digitalmarketplace-api,mtekel/digitalmarketplace-api,RichardKnop/digitalmarketplace-api,alphagov/digitalmarketplace-api,mtekel/digitalmarketplace-api,alphagov/digitalmarketplace-api,mtekel/digitalmarketplace-api,alphagov/digitalmarketplace-api | ---
+++
@@ -1,5 +1,6 @@
#!/usr/bin/env python
import sys
+import json
import os
import requests
import multiprocessing
@@ -24,11 +25,14 @@
print("Access token: {}".format(access_token))
print("Listing dir: {}".format(listing_dir))
- def post_file(file_path):
+ def put_file(file_path):
with open(file_path) as f:
- response = requests.post(
- endpoint,
- data='{"services":%s}' % f.read(),
+ data = json.load(f)
+ data = {'services': data}
+ url = '{}/{}'.format(endpoint, data['services']['id'])
+ response = requests.put(
+ url,
+ data=json.dumps(data),
headers={
"content-type": "application/json",
"authorization": "Bearer {}".format(access_token),
@@ -37,5 +41,5 @@
pool = multiprocessing.Pool(10)
- for result in pool.imap(post_file, list_files(listing_dir)):
+ for result in pool.imap(put_file, list_files(listing_dir)):
print(result) |
95b08a7cb2d473c25c1d326b0394336955b47af4 | appy/models.py | appy/models.py | from django.db import models
from django.contrib.auth.models import User
class Tag(models.Model):
description = models.TextField()
class Position(models.Model):
company = models.TextField()
job_title = models.TextField()
description = models.TextField()
tags = models.ManyToManyField(Tag)
class Application(models.Model):
user = models.ForeignKey(User)
position = models.ForeignKey(Position)
APPLIED = 'APP'
REJECTED = 'REJ'
INTERVIEWING = 'INT'
NEGOTIATING = 'NEG'
STATUS_CHOICES = (
(APPLIED, 'Applied'),
(REJECTED, 'Rejected'),
(INTERVIEWING, 'Interviewing'),
(NEGOTIATING, 'Negotiating'),
)
status = models.CharField(max_length=3, choices=STATUS_CHOICES, default=APPLIED)
created_at = models.DateTimeField(auto_now_add=True)
updated_at = models.DateTimeField(auto_now=True)
| from django.db import models
from django.contrib.auth.models import User
class Tag(models.Model):
description = models.TextField()
def __unicode__(self):
return self.description
class Position(models.Model):
company = models.TextField()
job_title = models.TextField()
description = models.TextField()
tags = models.ManyToManyField(Tag)
def __unicode__(self):
return u'%s at %s' % (self.job_title, self.company)
class Application(models.Model):
user = models.ForeignKey(User)
position = models.ForeignKey(Position)
APPLIED = 'APP'
REJECTED = 'REJ'
INTERVIEWING = 'INT'
NEGOTIATING = 'NEG'
STATUS_CHOICES = (
(APPLIED, 'Applied'),
(REJECTED, 'Rejected'),
(INTERVIEWING, 'Interviewing'),
(NEGOTIATING, 'Negotiating'),
)
status = models.CharField(max_length=3, choices=STATUS_CHOICES, default=APPLIED)
created_at = models.DateTimeField(auto_now_add=True)
updated_at = models.DateTimeField(auto_now=True)
| Add unicode representations for tags/positions | Add unicode representations for tags/positions
| Python | mit | merdey/ApPy,merdey/ApPy | ---
+++
@@ -5,12 +5,18 @@
class Tag(models.Model):
description = models.TextField()
+ def __unicode__(self):
+ return self.description
+
class Position(models.Model):
company = models.TextField()
job_title = models.TextField()
description = models.TextField()
tags = models.ManyToManyField(Tag)
+
+ def __unicode__(self):
+ return u'%s at %s' % (self.job_title, self.company)
class Application(models.Model): |
2b43ab4eb41e305c5bdadf5c338e134e5569249d | tests/conftest.py | tests/conftest.py | # pylint: disable=C0111
import pytest
import os
import tarfile
BASEDIR = os.path.dirname(__file__)
@pytest.fixture(autouse=False)
def set_up(tmpdir):
# print BASEDIR
tmpdir.chdir()
tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz"))
tar.extractall()
tar.close()
os.chdir('MockRepos')
print('In directory ' + os.getcwd())
# does not need teardown, since tmpdir directories get autodeleted
| # pylint: disable=C0111
import pytest
import os
import tarfile
BASEDIR = os.path.dirname(__file__)
@pytest.fixture(autouse=False)
def set_up(tmpdir):
# print BASEDIR
tmpdir.chdir()
tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz"))
tar.extractall()
tar.close()
os.chdir('MockRepos')
print('In directory ' + os.getcwd())
# does not need teardown, since tmpdir directories get autodeleted
def session_teardown():
"""Tear down testing session"""
print('Tearing down testing session')
os.chdir(BASEDIR) # so html coverage report lands in correct directory
@pytest.fixture(scope='session', autouse=True)
def session_setup(request):
"""Set up testing session"""
print('Setting up testing session')
request.addfinalizer(session_teardown)
| Add session setup and teardown fixtures. | Add session setup and teardown fixtures. | Python | mit | bilderbuchi/ofStateManager | ---
+++
@@ -16,3 +16,17 @@
os.chdir('MockRepos')
print('In directory ' + os.getcwd())
# does not need teardown, since tmpdir directories get autodeleted
+
+
+def session_teardown():
+ """Tear down testing session"""
+ print('Tearing down testing session')
+ os.chdir(BASEDIR) # so html coverage report lands in correct directory
+
+
+@pytest.fixture(scope='session', autouse=True)
+def session_setup(request):
+ """Set up testing session"""
+ print('Setting up testing session')
+
+ request.addfinalizer(session_teardown) |
98bded02b1b5116db640f5e58f73920108af0b0c | tests/test_set.py | tests/test_set.py | import matplotlib
import fishbowl
original = []
updated = [r'\usepackage{mathspec}',
r'\setallmainfonts(Digits,Latin,Greek){Arbitrary}']
def test_context_set():
fishbowl.reset_style()
with fishbowl.style(axes='minimal', palette='gourami', font='Arbitrary'):
assert matplotlib.rcParams['pgf.preamble'] == updated
def test_context_reset():
fishbowl.reset_style()
with fishbowl.style(axes='minimal', palette='gourami', font='Arbitrary'):
pass
assert matplotlib.rcParams['pgf.preamble'] == original
def test_set():
fishbowl.reset_style()
fishbowl.set_style(font='Arbitrary')
assert matplotlib.rcParams['pgf.preamble'] == updated
def test_reset():
fishbowl.set_style(font='Arbitrary')
fishbowl.reset_style()
assert matplotlib.rcParams['pgf.preamble'] == original
| import matplotlib
import fishbowl
original = True
updated = False
def test_context_set():
fishbowl.reset_style()
with fishbowl.style(axes='minimal', palette='gourami', font='Arbitrary'):
assert matplotlib.rcParams['axes.spines.left'] == updated
def test_context_reset():
fishbowl.reset_style()
with fishbowl.style(axes='minimal', palette='gourami', font='Arbitrary'):
pass
assert matplotlib.rcParams['axes.spines.left'] == original
def test_set():
fishbowl.reset_style()
fishbowl.set_style(font='Arbitrary')
assert matplotlib.rcParams['axes.spines.left'] == updated
def test_reset():
fishbowl.set_style(font='Arbitrary')
fishbowl.reset_style()
assert matplotlib.rcParams['axes.spines.left'] == original
| Update tests to remove pgf.preamble | Update tests to remove pgf.preamble
| Python | mit | baxen/fishbowl | ---
+++
@@ -2,31 +2,30 @@
import fishbowl
-original = []
-updated = [r'\usepackage{mathspec}',
- r'\setallmainfonts(Digits,Latin,Greek){Arbitrary}']
+original = True
+updated = False
def test_context_set():
fishbowl.reset_style()
with fishbowl.style(axes='minimal', palette='gourami', font='Arbitrary'):
- assert matplotlib.rcParams['pgf.preamble'] == updated
+ assert matplotlib.rcParams['axes.spines.left'] == updated
def test_context_reset():
fishbowl.reset_style()
with fishbowl.style(axes='minimal', palette='gourami', font='Arbitrary'):
pass
- assert matplotlib.rcParams['pgf.preamble'] == original
+ assert matplotlib.rcParams['axes.spines.left'] == original
def test_set():
fishbowl.reset_style()
fishbowl.set_style(font='Arbitrary')
- assert matplotlib.rcParams['pgf.preamble'] == updated
+ assert matplotlib.rcParams['axes.spines.left'] == updated
def test_reset():
fishbowl.set_style(font='Arbitrary')
fishbowl.reset_style()
- assert matplotlib.rcParams['pgf.preamble'] == original
+ assert matplotlib.rcParams['axes.spines.left'] == original |
e6d965cc36d92ee8f138d487614244c6e3deda69 | run_tests.py | run_tests.py | from __future__ import division
import libtbx.load_env
def discover_pytests(module):
try:
import os
import pytest
except ImportError:
return []
if 'LIBTBX_SKIP_PYTEST' in os.environ:
return []
test_list = []
dist_dir = libtbx.env.dist_path(module)
class TestDiscoveryPlugin:
def pytest_itemcollected(self, item):
test_list.append([ "libtbx.python", "-m", "pytest", '--noconftest',
os.path.join(dist_dir, item.nodeid) ])
print "Discovering pytest tests:"
pytest.main(['-qq', '--collect-only', '--noconftest', dist_dir], plugins=[TestDiscoveryPlugin()])
return test_list
if (__name__ == "__main__"):
import unittest
test_suite = unittest.defaultTestLoader.discover(libtbx.env.dist_path("i19"), pattern="tst_*.py")
result = unittest.TextTestRunner().run(test_suite)
import sys
sys.exit(0 if result.wasSuccessful() else 1)
tst_list = [
# "$D/tests/tst_legacy.py",
["$D/tests/tst_legacy_mult.py", "1"]
# ["$D/tests/tst_legacy_mult.py", "2"]
] + discover_pytests("i19")
| from __future__ import division
import libtbx.load_env
def discover_pytests(module):
try:
import os
import pytest
except ImportError:
def pytest_warning():
print "=" * 60
print "WARNING: Skipping some tests\n"
print "To run all available tests you need to install pytest"
print "eg. with libtbx.python -m pip install pytest"
print "=" * 60
pytest_warning()
import atexit
atexit.register(pytest_warning)
return []
if 'LIBTBX_SKIP_PYTEST' in os.environ:
return []
test_list = []
dist_dir = libtbx.env.dist_path(module)
class TestDiscoveryPlugin:
def pytest_itemcollected(self, item):
test_list.append([ "libtbx.python", "-m", "pytest", '--noconftest',
os.path.join(dist_dir, item.nodeid) ])
print "Discovering pytest tests:"
pytest.main(['-qq', '--collect-only', '--noconftest', dist_dir], plugins=[TestDiscoveryPlugin()])
return test_list
if (__name__ == "__main__"):
import unittest
test_suite = unittest.defaultTestLoader.discover(libtbx.env.dist_path("i19"), pattern="tst_*.py")
result = unittest.TextTestRunner().run(test_suite)
import sys
sys.exit(0 if result.wasSuccessful() else 1)
tst_list = [
# "$D/tests/tst_legacy.py",
["$D/tests/tst_legacy_mult.py", "1"]
# ["$D/tests/tst_legacy_mult.py", "2"]
] + discover_pytests("i19")
| Add warning for skipped tests if pytest not available | Add warning for skipped tests if pytest not available
| Python | bsd-3-clause | xia2/i19 | ---
+++
@@ -6,6 +6,15 @@
import os
import pytest
except ImportError:
+ def pytest_warning():
+ print "=" * 60
+ print "WARNING: Skipping some tests\n"
+ print "To run all available tests you need to install pytest"
+ print "eg. with libtbx.python -m pip install pytest"
+ print "=" * 60
+ pytest_warning()
+ import atexit
+ atexit.register(pytest_warning)
return []
if 'LIBTBX_SKIP_PYTEST' in os.environ:
return [] |
baedff75f2b86f09368e3bd72b72e27bf887cc88 | rotational-cipher/rotational_cipher.py | rotational-cipher/rotational_cipher.py | import string
UPPER = string.ascii_uppercase
LOWER = string.ascii_lowercase
def rotate(s, n):
return "".join(rot_gen(s,n))
def shift_rules(n):
shifted = UPPER[n:] + UPPER[:n] + LOWER[n:] + LOWER[:n]
return {k:v for k,v in zip(UPPER+LOWER, shifted)}
def rot_gen(s, n):
rules = shift_rules(n)
for ch in s:
try:
yield rules[ch]
except KeyError:
yield ch
| import string
UPPER = string.ascii_uppercase
LOWER = string.ascii_lowercase
def rotate(s, n):
rules = shift_rules(n)
return "".join(map(lambda k: rules.get(k, k), s))
def shift_rules(n):
shifted = UPPER[n:] + UPPER[:n] + LOWER[n:] + LOWER[:n]
return {k:v for k,v in zip(UPPER+LOWER, shifted)}
| Use lambda function with method | Use lambda function with method
| Python | agpl-3.0 | CubicComet/exercism-python-solutions | ---
+++
@@ -6,18 +6,10 @@
def rotate(s, n):
- return "".join(rot_gen(s,n))
+ rules = shift_rules(n)
+ return "".join(map(lambda k: rules.get(k, k), s))
def shift_rules(n):
shifted = UPPER[n:] + UPPER[:n] + LOWER[n:] + LOWER[:n]
return {k:v for k,v in zip(UPPER+LOWER, shifted)}
-
-
-def rot_gen(s, n):
- rules = shift_rules(n)
- for ch in s:
- try:
- yield rules[ch]
- except KeyError:
- yield ch |
939ebf2eb4536fd5a6318d6cc4b55a9dc4c8def2 | documentation/compile_documentation.py | documentation/compile_documentation.py | import sys
import os
def find_code(text):
START_TAG = '```lpg'
END_TAG = '```'
first_index = text.find(START_TAG)
if first_index == -1:
return None, None
last_index = text.find(END_TAG, first_index + 1)
return first_index + len(START_TAG), last_index
def process_file(path):
content = open(path, 'r').read()
left, right = find_code(content)
code_pieces = []
while left is not None:
code_pieces.append(content[left:right].strip())
content = content[right:]
left, right = find_code(content)
return code_pieces
def write_to_file(file_name, code):
for i, code_block in enumerate(code):
print_block = '// %s : %i\n%s\n' % (file_name, i, code_block)
destination.write(print_block)
# Main Program
root_dir = '.'
destination = open('documentation.lpg', 'w')
if len(sys.argv) == 2:
root_dir = sys.argv[1]
for root, dirs, files in os.walk(root_dir):
for fi in files:
if fi.endswith('md'):
code = process_file(os.path.join(root, fi))
write_to_file(fi, code)
destination.close()
| import sys
import os
def find_code(text):
START_TAG = '```lpg'
END_TAG = '```'
first_index = text.find(START_TAG)
if first_index == -1:
return None, None
last_index = text.find(END_TAG, first_index + 1)
return first_index + len(START_TAG), last_index
def process_file(path):
content = open(path, 'r').read()
left, right = find_code(content)
code_pieces = []
while left is not None:
code_pieces.append(content[left:right].strip())
content = content[right:]
left, right = find_code(content)
return code_pieces
def write_to_file(file_name, code):
with open(file_name, 'w') as f:
f.write(code)
def write_codeblocks_to_different_files(code_blocks):
for i, code_block in enumerate(code_blocks):
write_to_file('codeblock%i.lpg' % i, code_block)
# Main Program
root_dir = '.'
if len(sys.argv) == 2:
root_dir = sys.argv[1]
if not os.path.exists(root_dir):
raise Exception('Directory does not exist')
for root, dirs, files in os.walk(root_dir):
markdown_files = [f for f in files if f.endswith('md')]
for fi in markdown_files:
code = process_file(os.path.join(root, fi))
write_codeblocks_to_different_files(code)
| Split the documentation compilation into different files | Split the documentation compilation into different files
| Python | mit | TyRoXx/Lpg,TyRoXx/Lpg,TyRoXx/Lpg,mamazu/Lpg,mamazu/Lpg,mamazu/Lpg,TyRoXx/Lpg,mamazu/Lpg,TyRoXx/Lpg | ---
+++
@@ -24,20 +24,24 @@
def write_to_file(file_name, code):
- for i, code_block in enumerate(code):
- print_block = '// %s : %i\n%s\n' % (file_name, i, code_block)
- destination.write(print_block)
+ with open(file_name, 'w') as f:
+ f.write(code)
+
+def write_codeblocks_to_different_files(code_blocks):
+ for i, code_block in enumerate(code_blocks):
+ write_to_file('codeblock%i.lpg' % i, code_block)
# Main Program
root_dir = '.'
-destination = open('documentation.lpg', 'w')
if len(sys.argv) == 2:
root_dir = sys.argv[1]
+if not os.path.exists(root_dir):
+ raise Exception('Directory does not exist')
+
for root, dirs, files in os.walk(root_dir):
- for fi in files:
- if fi.endswith('md'):
- code = process_file(os.path.join(root, fi))
- write_to_file(fi, code)
-destination.close()
+ markdown_files = [f for f in files if f.endswith('md')]
+ for fi in markdown_files:
+ code = process_file(os.path.join(root, fi))
+ write_codeblocks_to_different_files(code) |
c7fd327623dfa84a91931771265932d4da95d766 | 001-xoxoxo-obj/harness.py | 001-xoxoxo-obj/harness.py | from game import Game
from input_con import InputCon
from output_con import OutputCon
class Harness():
def __init__(self, output, inputs):
self._game = Game()
self._output = output
self._inputs = inputs
def Start(self):
self._output.show_welcome()
while True:
self._output.show_board(self._game.get_board())
player_id = self._game.get_turn_no()
player = self._game.get_turn()
self._output.show_player_turn(player)
while True:
move = self._inputs[player_id].get_move()
if move is None:
self._output.show_move_error(player)
continue
if self._game.make_move(move) is False:
self._output.show_move_error(player)
continue
break
if not self._game.get_end():
continue
# End of game.
self._output.show_board(self._game.get_board())
w = self._game.get_winner()
if w is None:
# Draw.
self._output.show_draw()
else:
self._output.show_winner(w)
break
def main():
inputcon1 = InputCon()
inputcon2 = InputCon()
outputcon = OutputCon()
player_inputs = [ inputcon1, inputcon2 ]
player_output = outputcon
h = Harness(player_output, player_inputs)
h.Start()
if __name__ == "__main__":
main()
| from game import Game
from input_con import InputCon
from output_con import OutputCon
class Harness():
def __init__(self, output, inputs):
self._game = Game()
self._output = output
self._inputs = inputs
def Start(self):
self._output.show_welcome()
while True:
self._output.show_board(self._game.get_board())
player_id = self._game.get_turn_no()
player = self._game.get_turn()
self._output.show_player_turn(player)
while True:
move = self._inputs[player_id].get_move()
if move is False:
self._output.show_move_error(player)
continue
if self._game.make_move(move) is False:
self._output.show_move_error(player)
break
break
if not self._game.get_end():
continue
# End of game.
self._output.show_board(self._game.get_board())
w = self._game.get_winner()
if w is None:
# Draw.
self._output.show_draw()
else:
self._output.show_winner(w)
break
def main():
inputcon1 = InputCon()
inputcon2 = InputCon()
outputcon = OutputCon()
player_inputs = [ inputcon1, inputcon2 ]
player_output = outputcon
h = Harness(player_output, player_inputs)
h.Start()
if __name__ == "__main__":
main()
| Fix nierozpoznawania bledu gracza w pierwszym ruchu | Fix nierozpoznawania bledu gracza w pierwszym ruchu
| Python | mit | gynvael/stream,gynvael/stream,gynvael/stream,gynvael/stream,gynvael/stream,gynvael/stream | ---
+++
@@ -22,13 +22,13 @@
while True:
move = self._inputs[player_id].get_move()
- if move is None:
+ if move is False:
self._output.show_move_error(player)
- continue
+ continue
if self._game.make_move(move) is False:
self._output.show_move_error(player)
- continue
+ break
break
@@ -49,7 +49,7 @@
inputcon1 = InputCon()
inputcon2 = InputCon()
outputcon = OutputCon()
-
+
player_inputs = [ inputcon1, inputcon2 ]
player_output = outputcon
|
9a34e663f9ef65cdac91c42dcf198ae28d4385be | txircd/modbase.py | txircd/modbase.py | # The purpose of this file is to provide base classes with the needed functions
# already defined; this allows us to guarantee that any exceptions raised
# during function calls are a problem with the module and not just that the
# particular function isn't defined.
class Module(object):
def hook(self, base):
self.ircd = base
return self
class Mode(object):
def hook(self, base):
self.ircd = base
return self
def prefixSymbol(self):
return None
def checkSet(self, channel, param):
return True
def checkUnset(self, channel, param):
return True
def onJoin(self, channel, user, params):
return "pass"
def onMessage(self, sender, target, message):
return ["pass"]
def onPart(self, channel, user, reason):
pass
def onTopicChange(self, channel, user, topic):
pass
def commandData(self, command, *args):
pass
def Command(object):
def hook(self, base):
self.ircd = base
return self
def onUse(self, user, params):
pass | # The purpose of this file is to provide base classes with the needed functions
# already defined; this allows us to guarantee that any exceptions raised
# during function calls are a problem with the module and not just that the
# particular function isn't defined.
class Module(object):
def hook(self, base):
self.ircd = base
return self
class Mode(object):
def hook(self, base):
self.ircd = base
return self
def prefixSymbol(self):
return None
def checkSet(self, channel, param):
return True
def checkUnset(self, channel, param):
return True
def onJoin(self, channel, user, params):
return "pass"
def onMessage(self, sender, target, message):
return ["pass"]
def onPart(self, channel, user, reason):
pass
def onTopicChange(self, channel, user, topic):
pass
def commandData(self, command, *args):
pass
def Command(object):
def hook(self, base):
self.ircd = base
return self
def onUse(self, user, params):
pass
def processParams(self, user, params):
return {
"user": user,
"params": params
} | Add a function for commands to process parameters | Add a function for commands to process parameters
| Python | bsd-3-clause | ElementalAlchemist/txircd,Heufneutje/txircd,DesertBus/txircd | ---
+++
@@ -35,3 +35,8 @@
return self
def onUse(self, user, params):
pass
+ def processParams(self, user, params):
+ return {
+ "user": user,
+ "params": params
+ } |
a1cc1c08e9ff36ff0293f0224ca0acb41f065073 | alg_selection_sort.py | alg_selection_sort.py | def selection_sort(a_list):
"""Selection Sort algortihm.
Concept: Find out the maximun item's original slot first,
then switch it to the max slot. Iterate the procedure.
"""
for max_slot in reversed(range(len(a_list))):
select_slot = 0
for slot in range(1, max_slot + 1):
if a_list[slot] > a_list[select_slot]:
select_slot = slot
temp = a_list[max_slot]
a_list[max_slot] = a_list[select_slot]
a_list[select_slot] = temp
def main():
a_list = [54, 26, 93, 17, 77, 31, 44, 55, 20]
print('a_list: {}'.format(a_list))
print('By selection sort: ')
selection_sort(a_list)
print(a_list)
if __name__ == '__main__':
main()
| def selection_sort(a_list):
"""Selection Sort algortihm.
Concept:
- Find out the maximun item's original slot first,
- then swap it and the item at the max slot.
- Iterate the procedure.
"""
for max_slot in reversed(range(len(a_list))):
select_slot = 0
for slot in range(1, max_slot + 1):
if a_list[slot] > a_list[select_slot]:
select_slot = slot
temp = a_list[max_slot]
a_list[max_slot] = a_list[select_slot]
a_list[select_slot] = temp
def main():
a_list = [54, 26, 93, 17, 77, 31, 44, 55, 20]
print('a_list: {}'.format(a_list))
print('By selection sort: ')
selection_sort(a_list)
print(a_list)
if __name__ == '__main__':
main()
| Revise doc string for selection sort’s concept | Revise doc string for selection sort’s concept
| Python | bsd-2-clause | bowen0701/algorithms_data_structures | ---
+++
@@ -1,8 +1,10 @@
def selection_sort(a_list):
"""Selection Sort algortihm.
- Concept: Find out the maximun item's original slot first,
- then switch it to the max slot. Iterate the procedure.
+ Concept:
+ - Find out the maximun item's original slot first,
+ - then swap it and the item at the max slot.
+ - Iterate the procedure.
"""
for max_slot in reversed(range(len(a_list))):
select_slot = 0 |
1be7237169f0c0920d4891d7fb20b03327037cdb | osgtest/tests/test_40_proxy.py | osgtest/tests/test_40_proxy.py | import osgtest.library.core as core
import osgtest.library.osgunittest as osgunittest
import unittest
class TestGridProxyInit(osgunittest.OSGTestCase):
def test_01_grid_proxy_init(self):
core.skip_ok_unless_installed('globus-proxy-utils')
command = ('grid-proxy-init', '-debug')
password = core.options.password + '\n'
core.check_system(command, 'Run grid-proxy-init', user=True,
stdin=password)
| import osgtest.library.core as core
import osgtest.library.osgunittest as osgunittest
import pprint
import errno
class TestGridProxyInit(osgunittest.OSGTestCase):
def test_00_list_proxies(self):
command = ('ls', '-lF', '/tmp/x509up_u*')
status, stdout, _ = core.system(command, 'List proxies')
core.log_message(stdout)
def test_01_grid_proxy_init(self):
core.skip_ok_unless_installed('globus-proxy-utils')
command = ('grid-proxy-init', '-debug')
password = core.options.password + '\n'
try:
core.check_system(command, 'Run grid-proxy-init', user=True,
stdin=password)
except OSError, e:
attributes = {}
for x in dir(e):
attributes[x] = getattr(e,x, None)
pprint.pprint(attributes)
raise
def test_02_grid_proxy_info(self):
core.skip_ok_unless_installed('globus-proxy-utils')
command = ('grid-proxy-info', '-debug')
core.check_system(command, 'Run grid-proxy-info', user=True)
| Add debugging code to proxy generation | Add debugging code to proxy generation
git-svn-id: 884a03e47e2adb735d896e55bb5ad6bc3421ba19@18806 4e558342-562e-0410-864c-e07659590f8c
| Python | apache-2.0 | efajardo/osg-test,efajardo/osg-test | ---
+++
@@ -1,12 +1,30 @@
import osgtest.library.core as core
import osgtest.library.osgunittest as osgunittest
-import unittest
+import pprint
+import errno
class TestGridProxyInit(osgunittest.OSGTestCase):
+
+ def test_00_list_proxies(self):
+ command = ('ls', '-lF', '/tmp/x509up_u*')
+ status, stdout, _ = core.system(command, 'List proxies')
+ core.log_message(stdout)
def test_01_grid_proxy_init(self):
core.skip_ok_unless_installed('globus-proxy-utils')
command = ('grid-proxy-init', '-debug')
password = core.options.password + '\n'
- core.check_system(command, 'Run grid-proxy-init', user=True,
- stdin=password)
+ try:
+ core.check_system(command, 'Run grid-proxy-init', user=True,
+ stdin=password)
+ except OSError, e:
+ attributes = {}
+ for x in dir(e):
+ attributes[x] = getattr(e,x, None)
+ pprint.pprint(attributes)
+ raise
+
+ def test_02_grid_proxy_info(self):
+ core.skip_ok_unless_installed('globus-proxy-utils')
+ command = ('grid-proxy-info', '-debug')
+ core.check_system(command, 'Run grid-proxy-info', user=True) |
a30e51ccb74bc55924be6f7f79dc4b6038c9b457 | altair/examples/bar_chart_with_highlighted_segment.py | altair/examples/bar_chart_with_highlighted_segment.py | """
Bar Chart with Highlighted Segment
----------------------------------
This example shows a bar chart that highlights values beyond a threshold.
"""
import altair as alt
import pandas as pd
from vega_datasets import data
source = data.wheat()
threshold = pd.DataFrame([{"threshold": 90}])
bars = alt.Chart(source).mark_bar().encode(
x="year:O",
y="wheat:Q",
)
highlight = alt.Chart(source).mark_bar(color="#e45755").encode(
x='year:O',
y='baseline:Q',
y2='wheat:Q'
).transform_filter(
alt.datum.wheat > 90
).transform_calculate("baseline", "90")
rule = alt.Chart(threshold).mark_rule().encode(
y='threshold:Q'
)
(bars + highlight + rule).properties(width=600)
| """
Bar Chart with Highlighted Segment
----------------------------------
This example shows a bar chart that highlights values beyond a threshold.
"""
# category: bar charts
import altair as alt
import pandas as pd
from vega_datasets import data
source = data.wheat()
threshold = pd.DataFrame([{"threshold": 90}])
bars = alt.Chart(source).mark_bar().encode(
x="year:O",
y="wheat:Q",
)
highlight = alt.Chart(source).mark_bar(color="#e45755").encode(
x='year:O',
y='baseline:Q',
y2='wheat:Q'
).transform_filter(
alt.datum.wheat > 90
).transform_calculate("baseline", "90")
rule = alt.Chart(threshold).mark_rule().encode(
y='threshold:Q'
)
(bars + highlight + rule).properties(width=600)
| Move bar chart with highlighted segment chart into the bar charts section | Move bar chart with highlighted segment chart into the bar charts section | Python | bsd-3-clause | altair-viz/altair | ---
+++
@@ -3,6 +3,7 @@
----------------------------------
This example shows a bar chart that highlights values beyond a threshold.
"""
+# category: bar charts
import altair as alt
import pandas as pd
from vega_datasets import data |
09340916e7db6ba8ccb5697b9444fbccc0512103 | example/example/views.py | example/example/views.py | from django import forms
from django.forms.formsets import formset_factory
from django.shortcuts import render
from djangoformsetjs.utils import formset_media_js
class MyForm(forms.Form):
foo = forms.CharField()
class Media(object):
# The form must have `formset_media_js` in its Media
js = formset_media_js + (
# Other form javascript...
)
MyFormSet = formset_factory(MyForm)
def formset_view(request):
formset = MyFormSet(request.POST or None)
if formset.is_valid():
pass
return render(request, 'formset.html', {'formset': formset})
| from django import forms
from django.forms.formsets import formset_factory
from django.shortcuts import render
from djangoformsetjs.utils import formset_media_js
class MyForm(forms.Form):
foo = forms.CharField()
class Media(object):
# The form must have `formset_media_js` in its Media
js = formset_media_js + (
# Other form javascript...
)
MyFormSet = formset_factory(MyForm, can_delete=True)
def formset_view(request):
formset = MyFormSet(request.POST or None)
if formset.is_valid():
pass
return render(request, 'formset.html', {'formset': formset})
| Add `can_delete=True` to the example formset | Add `can_delete=True` to the example formset
| Python | bsd-2-clause | pretix/django-formset-js,pretix/django-formset-js,pretix/django-formset-js | ---
+++
@@ -14,7 +14,7 @@
)
-MyFormSet = formset_factory(MyForm)
+MyFormSet = formset_factory(MyForm, can_delete=True)
def formset_view(request): |
de11b473a6134ed1403e91d55d30c23e6683a926 | test/runner/versions.py | test/runner/versions.py | #!/usr/bin/env python
"""Show python and pip versions."""
import os
import sys
try:
import pip
except ImportError:
pip = None
print('.'.join(u'%s' % i for i in sys.version_info))
if pip:
print('pip %s from %s' % (pip.__version__, os.path.dirname(pip.__file__)))
| #!/usr/bin/env python
"""Show python and pip versions."""
import os
import sys
try:
import pip
except ImportError:
pip = None
print(sys.version)
if pip:
print('pip %s from %s' % (pip.__version__, os.path.dirname(pip.__file__)))
| Revert "Relax ansible-test python version checking." | Revert "Relax ansible-test python version checking."
This reverts commit d6cc3c41874b64e346639549fd18d8c41be0db8b.
| Python | mit | thaim/ansible,thaim/ansible | ---
+++
@@ -9,7 +9,7 @@
except ImportError:
pip = None
-print('.'.join(u'%s' % i for i in sys.version_info))
+print(sys.version)
if pip:
print('pip %s from %s' % (pip.__version__, os.path.dirname(pip.__file__))) |
c3ead540e7008ba1a3d01df695620bc952564805 | sphinx/fabfile.py | sphinx/fabfile.py | from fabric.api import run, env, roles
from fabric.contrib.project import rsync_project
import sys
sys.path.append("source")
import conf
env.roledefs = {
'web': ['bokeh.pydata.org']
}
env.user = "bokeh"
@roles('web')
def deploy(v=None):
if v is None:
v = conf.version
# make a backup of the old directory
run("rm -rf /www/bokeh/en/%s.bak" % v)
run("mkdir -p /www/bokeh/en/%s" % v)
run("cp -ar /www/bokeh/en/%s /www/bokeh/en/%s.bak" % (v, v))
rsync_project(
local_dir="_build/html/",
remote_dir="/www/bokeh/en/%s" % v,
delete=True
)
# set permissions
run("chmod -R g+w /www/bokeh/en/%s" % v)
@roles('web')
def latest(v=None):
if v is None:
raise RuntimeError("You need to specify version number: fab latest:x.x.x")
# switch the current symlink to new docs
run("rm /www/bokeh/en/latest")
run("ln -s /www/bokeh/en/%s /www/bokeh/en/latest" % v)
| from fabric.api import run, env, roles
from fabric.contrib.files import exists
from fabric.contrib.project import rsync_project
import sys
sys.path.append("source")
import conf
env.roledefs = {
'web': ['bokeh.pydata.org']
}
env.user = "bokeh"
@roles('web')
def deploy(v=None):
if v is None:
v = conf.version
elif v == "latest":
raise RuntimeError("You can not pass 'latest' as fab argument. Use "
"fab latest:x.x.x instead.")
# make a backup of the old directory
run("rm -rf /www/bokeh/en/%s.bak" % v)
run("mkdir -p /www/bokeh/en/%s" % v)
run("cp -ar /www/bokeh/en/%s /www/bokeh/en/%s.bak" % (v, v))
rsync_project(
local_dir="_build/html/",
remote_dir="/www/bokeh/en/%s" % v,
delete=True
)
# set permissions
run("chmod -R g+w /www/bokeh/en/%s" % v)
@roles('web')
def latest(v=None):
if v is None:
raise RuntimeError("You need to specify a version number: fab latest:x.x.x")
if exists("/www/bokeh/en/%s" % v):
# switch the current symlink to new docs
run("rm /www/bokeh/en/latest")
run("ln -s /www/bokeh/en/%s /www/bokeh/en/latest" % v)
else:
raise RuntimeError("We did not detect a %s docs version, please use "
"fab deploy:%s first." % v)
| Add error if latest is passed to deploy. Also check if the path exist before meaking the symlink. | Add error if latest is passed to deploy. Also check if the path exist before meaking the symlink.
| Python | bsd-3-clause | Karel-van-de-Plassche/bokeh,aavanian/bokeh,clairetang6/bokeh,matbra/bokeh,quasiben/bokeh,justacec/bokeh,schoolie/bokeh,rothnic/bokeh,evidation-health/bokeh,ptitjano/bokeh,muku42/bokeh,percyfal/bokeh,daodaoliang/bokeh,KasperPRasmussen/bokeh,aiguofer/bokeh,CrazyGuo/bokeh,draperjames/bokeh,mindriot101/bokeh,percyfal/bokeh,caseyclements/bokeh,muku42/bokeh,rothnic/bokeh,phobson/bokeh,aiguofer/bokeh,daodaoliang/bokeh,timsnyder/bokeh,draperjames/bokeh,paultcochrane/bokeh,saifrahmed/bokeh,evidation-health/bokeh,stonebig/bokeh,paultcochrane/bokeh,ericmjl/bokeh,roxyboy/bokeh,phobson/bokeh,azjps/bokeh,timsnyder/bokeh,dennisobrien/bokeh,maxalbert/bokeh,timsnyder/bokeh,phobson/bokeh,khkaminska/bokeh,deeplook/bokeh,bokeh/bokeh,ericdill/bokeh,maxalbert/bokeh,schoolie/bokeh,xguse/bokeh,aiguofer/bokeh,ptitjano/bokeh,deeplook/bokeh,KasperPRasmussen/bokeh,htygithub/bokeh,Karel-van-de-Plassche/bokeh,carlvlewis/bokeh,justacec/bokeh,aavanian/bokeh,msarahan/bokeh,dennisobrien/bokeh,philippjfr/bokeh,muku42/bokeh,stonebig/bokeh,roxyboy/bokeh,srinathv/bokeh,caseyclements/bokeh,tacaswell/bokeh,ChinaQuants/bokeh,mindriot101/bokeh,josherick/bokeh,paultcochrane/bokeh,ericdill/bokeh,xguse/bokeh,percyfal/bokeh,DuCorey/bokeh,azjps/bokeh,bokeh/bokeh,jakirkham/bokeh,rs2/bokeh,xguse/bokeh,aavanian/bokeh,gpfreitas/bokeh,timsnyder/bokeh,philippjfr/bokeh,rs2/bokeh,phobson/bokeh,schoolie/bokeh,maxalbert/bokeh,draperjames/bokeh,tacaswell/bokeh,aiguofer/bokeh,msarahan/bokeh,aavanian/bokeh,aiguofer/bokeh,ericdill/bokeh,jakirkham/bokeh,maxalbert/bokeh,srinathv/bokeh,htygithub/bokeh,dennisobrien/bokeh,daodaoliang/bokeh,deeplook/bokeh,roxyboy/bokeh,josherick/bokeh,gpfreitas/bokeh,ptitjano/bokeh,clairetang6/bokeh,quasiben/bokeh,jplourenco/bokeh,evidation-health/bokeh,rs2/bokeh,clairetang6/bokeh,CrazyGuo/bokeh,KasperPRasmussen/bokeh,bokeh/bokeh,KasperPRasmussen/bokeh,gpfreitas/bokeh,rothnic/bokeh,khkaminska/bokeh,muku42/bokeh,ericdill/bokeh,rothnic/bokeh,msarahan/bokeh,jplourenco/bokeh,ChinaQuants/bokeh,draperjames/bokeh,caseyclements/bokeh,daodaoliang/bokeh,jplourenco/bokeh,rs2/bokeh,philippjfr/bokeh,ptitjano/bokeh,saifrahmed/bokeh,stonebig/bokeh,evidation-health/bokeh,justacec/bokeh,Karel-van-de-Plassche/bokeh,ericmjl/bokeh,bokeh/bokeh,CrazyGuo/bokeh,paultcochrane/bokeh,Karel-van-de-Plassche/bokeh,DuCorey/bokeh,philippjfr/bokeh,matbra/bokeh,schoolie/bokeh,htygithub/bokeh,DuCorey/bokeh,ChinaQuants/bokeh,percyfal/bokeh,KasperPRasmussen/bokeh,schoolie/bokeh,Karel-van-de-Plassche/bokeh,timsnyder/bokeh,khkaminska/bokeh,ericmjl/bokeh,jakirkham/bokeh,tacaswell/bokeh,saifrahmed/bokeh,mindriot101/bokeh,draperjames/bokeh,gpfreitas/bokeh,DuCorey/bokeh,stonebig/bokeh,jplourenco/bokeh,azjps/bokeh,justacec/bokeh,josherick/bokeh,roxyboy/bokeh,matbra/bokeh,mindriot101/bokeh,dennisobrien/bokeh,jakirkham/bokeh,DuCorey/bokeh,deeplook/bokeh,ChinaQuants/bokeh,philippjfr/bokeh,aavanian/bokeh,carlvlewis/bokeh,saifrahmed/bokeh,rs2/bokeh,carlvlewis/bokeh,xguse/bokeh,matbra/bokeh,ptitjano/bokeh,srinathv/bokeh,bokeh/bokeh,percyfal/bokeh,tacaswell/bokeh,azjps/bokeh,jakirkham/bokeh,khkaminska/bokeh,caseyclements/bokeh,phobson/bokeh,quasiben/bokeh,srinathv/bokeh,carlvlewis/bokeh,CrazyGuo/bokeh,ericmjl/bokeh,msarahan/bokeh,clairetang6/bokeh,dennisobrien/bokeh,htygithub/bokeh,josherick/bokeh,ericmjl/bokeh,azjps/bokeh | ---
+++
@@ -1,4 +1,5 @@
from fabric.api import run, env, roles
+from fabric.contrib.files import exists
from fabric.contrib.project import rsync_project
import sys
@@ -15,6 +16,9 @@
if v is None:
v = conf.version
+ elif v == "latest":
+ raise RuntimeError("You can not pass 'latest' as fab argument. Use "
+ "fab latest:x.x.x instead.")
# make a backup of the old directory
run("rm -rf /www/bokeh/en/%s.bak" % v)
@@ -34,8 +38,12 @@
def latest(v=None):
if v is None:
- raise RuntimeError("You need to specify version number: fab latest:x.x.x")
+ raise RuntimeError("You need to specify a version number: fab latest:x.x.x")
- # switch the current symlink to new docs
- run("rm /www/bokeh/en/latest")
- run("ln -s /www/bokeh/en/%s /www/bokeh/en/latest" % v)
+ if exists("/www/bokeh/en/%s" % v):
+ # switch the current symlink to new docs
+ run("rm /www/bokeh/en/latest")
+ run("ln -s /www/bokeh/en/%s /www/bokeh/en/latest" % v)
+ else:
+ raise RuntimeError("We did not detect a %s docs version, please use "
+ "fab deploy:%s first." % v) |
cb28bba6ee642828df473383ea469a6aa46ca59c | skimage/util/unique.py | skimage/util/unique.py | import numpy as np
def unique_rows(ar):
"""Remove repeated rows from a 2D array.
Parameters
----------
ar : 2D np.ndarray
The input array.
Returns
-------
ar_out : 2D np.ndarray
A copy of the input array with repeated rows removed.
Raises
------
ValueError : if `ar` is not two-dimensional.
Examples
--------
>>> ar = np.array([[1, 0, 1],
[0, 1, 0],
[1, 0, 1]], np.uint8)
>>> aru = unique_rows(ar)
array([[0, 1, 0],
[1, 0, 1]], dtype=uint8)
"""
if ar.ndim != 2:
raise ValueError("unique_rows() only makes sense for 2D arrays, "
"got %dd" % ar.ndim)
# the view in the next line only works if the array is C-contiguous
ar = np.ascontiguousarray(ar)
# np.unique() finds identical items in a raveled array. To make it
# see each row as a single item, we create a view of each row as a
# byte string of length itemsize times number of columns in `ar`
ar_row_view = ar.view('|S%d' % (ar.itemsize * ar.shape[1]))
_, unique_row_indices = np.unique(ar_row_view, return_index=True)
ar_out = ar[unique_row_indices]
return ar_out
| import numpy as np
def unique_rows(ar):
"""Remove repeated rows from a 2D array.
Parameters
----------
ar : 2D np.ndarray
The input array.
Returns
-------
ar_out : 2D np.ndarray
A copy of the input array with repeated rows removed.
Raises
------
ValueError : if `ar` is not two-dimensional.
Notes
-----
The function will generate a copy of `ar` if it is not
C-contiguous, which will negatively affect performance for large
input arrays.
Examples
--------
>>> ar = np.array([[1, 0, 1],
[0, 1, 0],
[1, 0, 1]], np.uint8)
>>> aru = unique_rows(ar)
array([[0, 1, 0],
[1, 0, 1]], dtype=uint8)
"""
if ar.ndim != 2:
raise ValueError("unique_rows() only makes sense for 2D arrays, "
"got %dd" % ar.ndim)
# the view in the next line only works if the array is C-contiguous
ar = np.ascontiguousarray(ar)
# np.unique() finds identical items in a raveled array. To make it
# see each row as a single item, we create a view of each row as a
# byte string of length itemsize times number of columns in `ar`
ar_row_view = ar.view('|S%d' % (ar.itemsize * ar.shape[1]))
_, unique_row_indices = np.unique(ar_row_view, return_index=True)
ar_out = ar[unique_row_indices]
return ar_out
| Add note describing array copy if discontiguous | Add note describing array copy if discontiguous
| Python | bsd-3-clause | paalge/scikit-image,Midafi/scikit-image,michaelpacer/scikit-image,Britefury/scikit-image,SamHames/scikit-image,oew1v07/scikit-image,chintak/scikit-image,vighneshbirodkar/scikit-image,jwiggins/scikit-image,warmspringwinds/scikit-image,Britefury/scikit-image,rjeli/scikit-image,ofgulban/scikit-image,emon10005/scikit-image,juliusbierk/scikit-image,chriscrosscutler/scikit-image,almarklein/scikit-image,keflavich/scikit-image,michaelaye/scikit-image,bennlich/scikit-image,chintak/scikit-image,ClinicalGraphics/scikit-image,warmspringwinds/scikit-image,juliusbierk/scikit-image,blink1073/scikit-image,ofgulban/scikit-image,newville/scikit-image,michaelpacer/scikit-image,bsipocz/scikit-image,GaZ3ll3/scikit-image,ajaybhat/scikit-image,emon10005/scikit-image,Hiyorimi/scikit-image,Midafi/scikit-image,rjeli/scikit-image,newville/scikit-image,ClinicalGraphics/scikit-image,SamHames/scikit-image,paalge/scikit-image,michaelaye/scikit-image,dpshelio/scikit-image,WarrenWeckesser/scikits-image,paalge/scikit-image,oew1v07/scikit-image,youprofit/scikit-image,dpshelio/scikit-image,ofgulban/scikit-image,almarklein/scikit-image,keflavich/scikit-image,SamHames/scikit-image,almarklein/scikit-image,youprofit/scikit-image,WarrenWeckesser/scikits-image,vighneshbirodkar/scikit-image,blink1073/scikit-image,almarklein/scikit-image,chintak/scikit-image,chriscrosscutler/scikit-image,SamHames/scikit-image,robintw/scikit-image,ajaybhat/scikit-image,bennlich/scikit-image,pratapvardhan/scikit-image,vighneshbirodkar/scikit-image,robintw/scikit-image,pratapvardhan/scikit-image,rjeli/scikit-image,GaZ3ll3/scikit-image,chintak/scikit-image,jwiggins/scikit-image,Hiyorimi/scikit-image,bsipocz/scikit-image | ---
+++
@@ -17,6 +17,12 @@
Raises
------
ValueError : if `ar` is not two-dimensional.
+
+ Notes
+ -----
+ The function will generate a copy of `ar` if it is not
+ C-contiguous, which will negatively affect performance for large
+ input arrays.
Examples
-------- |
88ec3066a191f22b1faa3429ae89cbd76d45ac9b | aore/config/dev.py | aore/config/dev.py | # -*- coding: utf-8 -*-
from __future__ import absolute_import
from .common import *
sphinx_conf.listen = "127.0.0.1:9312"
sphinx_conf.var_dir = "C:\\Sphinx"
db_conf.database = "postgres"
db_conf.host = "localhost"
db_conf.port = 5432
db_conf.user = "postgres"
db_conf.password = "intercon"
unrar_config.path = "C:\\Program Files (x86)\\WinRAR\\unrar.exe"
folders.temp = "E:\\!TEMP"
basic.logging = True
| # -*- coding: utf-8 -*-
from __future__ import absolute_import
from .common import *
sphinx_conf.listen = "192.168.0.37:9312"
sphinx_conf.var_dir = "C:\\Sphinx"
db_conf.database = "postgres"
db_conf.host = "192.168.0.37"
db_conf.port = 5432
db_conf.user = "postgres"
db_conf.password = "intercon"
unrar_config.path = "C:\\Program Files (x86)\\WinRAR\\unrar.exe"
folders.temp = "E:\\!TEMP"
basic.logging = True
| Test global ip with Sphinx | Test global ip with Sphinx
| Python | bsd-3-clause | jar3b/py-phias,jar3b/py-phias,jar3b/py-phias,jar3b/py-phias | ---
+++
@@ -4,11 +4,11 @@
from .common import *
-sphinx_conf.listen = "127.0.0.1:9312"
+sphinx_conf.listen = "192.168.0.37:9312"
sphinx_conf.var_dir = "C:\\Sphinx"
db_conf.database = "postgres"
-db_conf.host = "localhost"
+db_conf.host = "192.168.0.37"
db_conf.port = 5432
db_conf.user = "postgres"
db_conf.password = "intercon" |
f60b940205a5e1011ce1c9c5672cb262c4649c0b | app/mod_auth/forms.py | app/mod_auth/forms.py | from flask_wtf import FlaskForm
from wtforms import StringField, BooleanField, PasswordField
from wtforms.validators import DataRequired
class LoginForm(FlaskForm):
username = StringField('username', validators=[DataRequired()])
password = PasswordField('password', validators=[DataRequired()])
remember_me = BooleanField('remember_me', default=False)
class SignupForm(FlaskForm):
username = StringField('username', validators=[DataRequired()])
password = PasswordField('password', validators=[DataRequired()])
password2 = PasswordField('password2', validators=[DataRequired()])
| from flask_wtf import FlaskForm
from wtforms import StringField, BooleanField, PasswordField
from wtforms.validators import DataRequired, ValidationError, EqualTo
from .models import User
class LoginForm(FlaskForm):
username = StringField('username', validators=[DataRequired()])
password = PasswordField('password', validators=[DataRequired()])
remember_me = BooleanField('remember_me', default=False)
class SignupForm(FlaskForm):
username = StringField('username', validators=[DataRequired()])
password = PasswordField('password', validators=[DataRequired()])
password2 = PasswordField('password2', validators=[DataRequired(), EqualTo('password')])
def validate_username(self, username):
user = User.query.filter_by(username=username.data).first()
if user is not None:
raise ValidationError('Username already in use. Please use a different username.')
| Validate username and password on signup | Validate username and password on signup
| Python | mit | ziel980/website,ziel980/website | ---
+++
@@ -1,6 +1,7 @@
from flask_wtf import FlaskForm
from wtforms import StringField, BooleanField, PasswordField
-from wtforms.validators import DataRequired
+from wtforms.validators import DataRequired, ValidationError, EqualTo
+from .models import User
class LoginForm(FlaskForm):
@@ -8,7 +9,13 @@
password = PasswordField('password', validators=[DataRequired()])
remember_me = BooleanField('remember_me', default=False)
+
class SignupForm(FlaskForm):
username = StringField('username', validators=[DataRequired()])
password = PasswordField('password', validators=[DataRequired()])
- password2 = PasswordField('password2', validators=[DataRequired()])
+ password2 = PasswordField('password2', validators=[DataRequired(), EqualTo('password')])
+
+ def validate_username(self, username):
+ user = User.query.filter_by(username=username.data).first()
+ if user is not None:
+ raise ValidationError('Username already in use. Please use a different username.') |
3d38af257f55a0252cb41408a404faa66b30d512 | pyconde/speakers/models.py | pyconde/speakers/models.py | from django.db import models
from django.contrib.auth.models import User
from django.core.urlresolvers import reverse
class Speaker(models.Model):
"""
The speaker model acts as user-abstraction for various session and proposal
related objects.
"""
user = models.OneToOneField(User, related_name='speaker_profile')
def __unicode__(self):
if self.user.first_name and self.user.last_name:
return u"{0} {1}".format(self.user.first_name, self.user.last_name)
return self.user.username
def get_absolute_url(self):
return reverse('account_profile', kwargs={'uid': self.user.id})
def create_speaker_profile(sender, instance, created, raw, **kwargs):
"""
Every user also is a potential speaker in the current implemention so we
also have to create a new speaker object for every newly created user
instance.
"""
if created:
Speaker(user=instance).save()
| # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib.auth.models import User
from django.core.urlresolvers import reverse
from django.db import models
from django.db.models.signals import post_save
from django.dispatch.dispatcher import receiver
class Speaker(models.Model):
"""
The speaker model acts as user-abstraction for various session and proposal
related objects.
"""
user = models.OneToOneField(User, related_name='speaker_profile')
def __unicode__(self):
if self.user.first_name and self.user.last_name:
return u"{0} {1}".format(self.user.first_name, self.user.last_name)
return self.user.username
def get_absolute_url(self):
return reverse('account_profile', kwargs={'uid': self.user.id})
@receiver(post_save, sender=User)
def create_speaker_profile(sender, instance, created, raw, **kwargs):
"""
Every user also is a potential speaker in the current implemention so we
also have to create a new speaker object for every newly created user
instance.
"""
if created:
Speaker(user=instance).save()
| Create a speaker profile for each user that is registered. | Create a speaker profile for each user that is registered.
| Python | bsd-3-clause | pysv/djep,EuroPython/djep,pysv/djep,EuroPython/djep,pysv/djep,EuroPython/djep,pysv/djep,EuroPython/djep,pysv/djep | ---
+++
@@ -1,6 +1,11 @@
-from django.db import models
+# -*- coding: utf-8 -*-
+from __future__ import unicode_literals
+
from django.contrib.auth.models import User
from django.core.urlresolvers import reverse
+from django.db import models
+from django.db.models.signals import post_save
+from django.dispatch.dispatcher import receiver
class Speaker(models.Model):
@@ -19,6 +24,7 @@
return reverse('account_profile', kwargs={'uid': self.user.id})
+@receiver(post_save, sender=User)
def create_speaker_profile(sender, instance, created, raw, **kwargs):
"""
Every user also is a potential speaker in the current implemention so we |
579286426cf20c5cd5d4c94d97fd0a55eb571f8f | talks_keeper/templatetags/tk_filters.py | talks_keeper/templatetags/tk_filters.py | from django import template
from django.forms import CheckboxInput
register = template.Library()
@register.filter(name='addclass')
def addclass(value, arg):
return value.as_widget(attrs={'class': arg})
@register.filter(name='is_checkbox')
def is_checkbox(field):
return field.field.widget.__class__.__name__ == \
CheckboxInput().__class__.__name__
| from django import template
from django.forms import CheckboxInput
from django.contrib.auth.models import Group
register = template.Library()
@register.filter(name='addclass')
def addclass(value, arg):
return value.as_widget(attrs={'class': arg})
@register.filter(name='is_checkbox')
def is_checkbox(field):
return field.field.widget.__class__.__name__ == \
CheckboxInput().__class__.__name__
@register.filter(name='in_group')
def in_group(user, group_name):
group = Group.objects.get(name=group_name)
return True if group in user.groups.all() else False
| Add templatetag to check if user included in group | Add templatetag to check if user included in group
| Python | mit | samitnuk/talks_keeper,samitnuk/talks_keeper,samitnuk/talks_keeper | ---
+++
@@ -1,5 +1,6 @@
from django import template
from django.forms import CheckboxInput
+from django.contrib.auth.models import Group
register = template.Library()
@@ -14,3 +15,9 @@
def is_checkbox(field):
return field.field.widget.__class__.__name__ == \
CheckboxInput().__class__.__name__
+
+
+@register.filter(name='in_group')
+def in_group(user, group_name):
+ group = Group.objects.get(name=group_name)
+ return True if group in user.groups.all() else False |
d52c9731b0c6494e9f4181fc33f00cdf39adb3ca | tests/unit/test_util.py | tests/unit/test_util.py | import pytest
from pmxbot import util
@pytest.has_wordnik
def test_lookup():
assert util.lookup('dachshund') is not None
| import pytest
from pmxbot import util
@pytest.has_wordnik
def test_lookup():
assert util.lookup('dachshund') is not None
@pytest.has_internet
def test_emergency_compliment():
assert util.load_emergency_compliments()
| Add test for emergency compliments | Add test for emergency compliments
| Python | mit | yougov/pmxbot,yougov/pmxbot,yougov/pmxbot | ---
+++
@@ -5,3 +5,7 @@
@pytest.has_wordnik
def test_lookup():
assert util.lookup('dachshund') is not None
+
+@pytest.has_internet
+def test_emergency_compliment():
+ assert util.load_emergency_compliments() |
5b8bcdd802858baae3854fbfb8758dc65bdd8d34 | hardware/unicorn_hat_hd/demo_lights.py | hardware/unicorn_hat_hd/demo_lights.py | #!/usr/bin/env python
import unicornhathd
import os
try:
unicornhathd.set_pixel(0, 0, 255, 255, 255)
unicornhathd.set_pixel(15, 0, 255, 255, 255)
unicornhathd.set_pixel(0, 15, 255, 255, 255)
unicornhathd.set_pixel(15, 15, 255, 255, 255)
unicornhathd.show()
raw_input("Press the <ENTER> key or <CTRL+C> to exit...")
except KeyboardInterrupt:
pass
unicornhathd.off()
| #!/usr/bin/env python
import unicornhathd
import os
from time import sleep
class Point:
def __init__(self, x, y, dx, dy):
self.x = x
self.y = y
self.dx = dx
self.dy = dy
def turn_on(self):
unicornhathd.set_pixel(self.x, self.y, 255, 255, 255)
def turn_off(self):
unicornhathd.set_pixel(self.x, self.y, 0, 0, 0)
def move(self):
self.turn_off()
self.x, self.dx = self.move_one_axis(self.x, self.dx)
self.y, self.dy = self.move_one_axis(self.y, self.dy)
self.turn_on()
def move_one_axis(self, x_or_y, dx_or_dy):
x_or_y += dx_or_dy
if x_or_y < 0 or x_or_y > 15:
dx_or_dy = dx_or_dy * -1
x_or_y += dx_or_dy
return x_or_y, dx_or_dy
print("Press <CTRL+C> to exit...")
unicornhathd.off()
# Bounce backwards and forwards along each edge:
p1 = Point(0, 0, 0, 1)
p2 = Point(0, 15, 1, 0)
p3 = Point(15, 0, -1, 0)
p4 = Point(15, 15, 0, -1)
p1.turn_on()
p2.turn_on()
p3.turn_on()
p4.turn_on()
try:
while True:
p1.move()
p2.move()
p3.move()
p4.move()
unicornhathd.show()
sleep(0.1)
except KeyboardInterrupt:
pass
unicornhathd.off()
| Make the 4 lights bounce backwards and forwards along each edge. | Make the 4 lights bounce backwards and forwards along each edge.
| Python | mit | claremacrae/raspi_code,claremacrae/raspi_code,claremacrae/raspi_code | ---
+++
@@ -1,14 +1,57 @@
#!/usr/bin/env python
import unicornhathd
import os
+from time import sleep
+
+class Point:
+ def __init__(self, x, y, dx, dy):
+ self.x = x
+ self.y = y
+ self.dx = dx
+ self.dy = dy
+
+ def turn_on(self):
+ unicornhathd.set_pixel(self.x, self.y, 255, 255, 255)
+
+ def turn_off(self):
+ unicornhathd.set_pixel(self.x, self.y, 0, 0, 0)
+
+ def move(self):
+ self.turn_off()
+ self.x, self.dx = self.move_one_axis(self.x, self.dx)
+ self.y, self.dy = self.move_one_axis(self.y, self.dy)
+ self.turn_on()
+
+ def move_one_axis(self, x_or_y, dx_or_dy):
+ x_or_y += dx_or_dy
+ if x_or_y < 0 or x_or_y > 15:
+ dx_or_dy = dx_or_dy * -1
+ x_or_y += dx_or_dy
+ return x_or_y, dx_or_dy
+
+print("Press <CTRL+C> to exit...")
+
+unicornhathd.off()
+
+# Bounce backwards and forwards along each edge:
+p1 = Point(0, 0, 0, 1)
+p2 = Point(0, 15, 1, 0)
+p3 = Point(15, 0, -1, 0)
+p4 = Point(15, 15, 0, -1)
+p1.turn_on()
+p2.turn_on()
+p3.turn_on()
+p4.turn_on()
try:
- unicornhathd.set_pixel(0, 0, 255, 255, 255)
- unicornhathd.set_pixel(15, 0, 255, 255, 255)
- unicornhathd.set_pixel(0, 15, 255, 255, 255)
- unicornhathd.set_pixel(15, 15, 255, 255, 255)
- unicornhathd.show()
- raw_input("Press the <ENTER> key or <CTRL+C> to exit...")
+ while True:
+ p1.move()
+ p2.move()
+ p3.move()
+ p4.move()
+ unicornhathd.show()
+ sleep(0.1)
except KeyboardInterrupt:
pass
+
unicornhathd.off() |
4aada26d0de09836f3b67b0fce136805cf11fa37 | thinc/extra/load_nlp.py | thinc/extra/load_nlp.py | import numpy
SPACY_MODELS = {}
VECTORS = {}
def get_spacy(lang, **kwargs):
global SPACY_MODELS
import spacy
if lang not in SPACY_MODELS:
SPACY_MODELS[lang] = spacy.load(lang, **kwargs)
return SPACY_MODELS[lang]
def get_vectors(ops, lang):
global VECTORS
key = (ops.device, lang)
if key not in VECTORS:
nlp = get_spacy(lang)
nV = max(lex.rank for lex in nlp.vocab)+1
nM = nlp.vocab.vectors_length
vectors = numpy.zeros((nV, nM), dtype='float32')
for lex in nlp.vocab:
if lex.has_vector:
vectors[lex.rank] = lex.vector / lex.vector_norm
VECTORS[key] = ops.asarray(vectors)
return VECTORS[key]
| import numpy
SPACY_MODELS = {}
VECTORS = {}
def get_spacy(lang, **kwargs):
global SPACY_MODELS
import spacy
if lang not in SPACY_MODELS:
SPACY_MODELS[lang] = spacy.load(lang, **kwargs)
return SPACY_MODELS[lang]
def get_vectors(ops, lang):
global VECTORS
key = (ops.device, lang)
if key not in VECTORS:
nlp = get_spacy(lang)
nV = max(lex.rank for lex in nlp.vocab)+1
nM = nlp.vocab.vectors_length
vectors = numpy.zeros((nV, nM), dtype='float32')
for lex in nlp.vocab:
if lex.has_vector:
vectors[lex.rank] = lex.vector / (lex.vector_norm+1e-8)
VECTORS[key] = ops.asarray(vectors)
return VECTORS[key]
| Fix divide by zero error in vectors loading | Fix divide by zero error in vectors loading
| Python | mit | explosion/thinc,explosion/thinc,spacy-io/thinc,spacy-io/thinc,explosion/thinc,spacy-io/thinc,explosion/thinc | ---
+++
@@ -22,6 +22,6 @@
vectors = numpy.zeros((nV, nM), dtype='float32')
for lex in nlp.vocab:
if lex.has_vector:
- vectors[lex.rank] = lex.vector / lex.vector_norm
+ vectors[lex.rank] = lex.vector / (lex.vector_norm+1e-8)
VECTORS[key] = ops.asarray(vectors)
return VECTORS[key] |
59c4dd56e427e29eb26e81512c3066fe3f8b13b8 | tools/gdb/gdb_chrome.py | tools/gdb/gdb_chrome.py | #!/usr/bin/python
# Copyright (c) 2011 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""GDB support for Chrome types.
Add this to your gdb by amending your ~/.gdbinit as follows:
python
import sys
sys.path.insert(0, "/path/to/tools/gdb/")
import gdb_chrome
This module relies on the WebKit gdb module already existing in
your Python path.
"""
import gdb
import webkit
class String16Printer(webkit.StringPrinter):
def to_string(self):
return webkit.ustring_to_string(self.val['_M_dataplus']['_M_p'])
class GURLPrinter(webkit.StringPrinter):
def to_string(self):
return self.val['spec_']
def lookup_function(val):
typ = str(val.type)
if typ == 'string16':
return String16Printer(val)
elif typ == 'GURL':
return GURLPrinter(val)
return None
gdb.pretty_printers.append(lookup_function)
| #!/usr/bin/python
# Copyright (c) 2011 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""GDB support for Chrome types.
Add this to your gdb by amending your ~/.gdbinit as follows:
python
import sys
sys.path.insert(0, "/path/to/tools/gdb/")
import gdb_chrome
This module relies on the WebKit gdb module already existing in
your Python path.
"""
import gdb
import webkit
class String16Printer(webkit.StringPrinter):
def to_string(self):
return webkit.ustring_to_string(self.val['_M_dataplus']['_M_p'])
class GURLPrinter(webkit.StringPrinter):
def to_string(self):
return self.val['spec_']
class FilePathPrinter(object):
def __init__(self, val):
self.val = val
def to_string(self):
return self.val['path_']['_M_dataplus']['_M_p']
def lookup_function(val):
type_to_printer = {
'string16': String16Printer,
'GURL': GURLPrinter,
'FilePath': FilePathPrinter,
}
printer = type_to_printer.get(str(val.type), None)
if printer:
return printer(val)
return None
gdb.pretty_printers.append(lookup_function)
| Add FilePath to the gdb pretty printers. | Add FilePath to the gdb pretty printers.
Review URL: http://codereview.chromium.org/6621017
git-svn-id: de016e52bd170d2d4f2344f9bf92d50478b649e0@76956 0039d316-1c4b-4281-b951-d872f2087c98
| Python | bsd-3-clause | zcbenz/cefode-chromium,junmin-zhu/chromium-rivertrail,krieger-od/nwjs_chromium.src,pozdnyakov/chromium-crosswalk,hgl888/chromium-crosswalk-efl,rogerwang/chromium,krieger-od/nwjs_chromium.src,TheTypoMaster/chromium-crosswalk,hujiajie/pa-chromium,hujiajie/pa-chromium,nacl-webkit/chrome_deps,Jonekee/chromium.src,nacl-webkit/chrome_deps,mogoweb/chromium-crosswalk,ChromiumWebApps/chromium,Just-D/chromium-1,rogerwang/chromium,krieger-od/nwjs_chromium.src,timopulkkinen/BubbleFish,timopulkkinen/BubbleFish,rogerwang/chromium,ChromiumWebApps/chromium,keishi/chromium,hujiajie/pa-chromium,anirudhSK/chromium,mohamed--abdel-maksoud/chromium.src,timopulkkinen/BubbleFish,timopulkkinen/BubbleFish,TheTypoMaster/chromium-crosswalk,dushu1203/chromium.src,ondra-novak/chromium.src,markYoungH/chromium.src,crosswalk-project/chromium-crosswalk-efl,axinging/chromium-crosswalk,pozdnyakov/chromium-crosswalk,anirudhSK/chromium,keishi/chromium,Pluto-tv/chromium-crosswalk,M4sse/chromium.src,krieger-od/nwjs_chromium.src,bright-sparks/chromium-spacewalk,Pluto-tv/chromium-crosswalk,Pluto-tv/chromium-crosswalk,ondra-novak/chromium.src,hgl888/chromium-crosswalk-efl,nacl-webkit/chrome_deps,PeterWangIntel/chromium-crosswalk,ondra-novak/chromium.src,TheTypoMaster/chromium-crosswalk,dushu1203/chromium.src,junmin-zhu/chromium-rivertrail,axinging/chromium-crosswalk,hgl888/chromium-crosswalk,dednal/chromium.src,hgl888/chromium-crosswalk,fujunwei/chromium-crosswalk,hujiajie/pa-chromium,patrickm/chromium.src,dednal/chromium.src,Jonekee/chromium.src,littlstar/chromium.src,ondra-novak/chromium.src,junmin-zhu/chromium-rivertrail,Fireblend/chromium-crosswalk,anirudhSK/chromium,mogoweb/chromium-crosswalk,hujiajie/pa-chromium,dushu1203/chromium.src,ondra-novak/chromium.src,robclark/chromium,robclark/chromium,nacl-webkit/chrome_deps,fujunwei/chromium-crosswalk,Pluto-tv/chromium-crosswalk,ChromiumWebApps/chromium,mohamed--abdel-maksoud/chromium.src,Just-D/chromium-1,M4sse/chromium.src,dushu1203/chromium.src,patrickm/chromium.src,nacl-webkit/chrome_deps,axinging/chromium-crosswalk,M4sse/chromium.src,robclark/chromium,jaruba/chromium.src,markYoungH/chromium.src,Pluto-tv/chromium-crosswalk,patrickm/chromium.src,Jonekee/chromium.src,pozdnyakov/chromium-crosswalk,markYoungH/chromium.src,rogerwang/chromium,mohamed--abdel-maksoud/chromium.src,littlstar/chromium.src,hgl888/chromium-crosswalk-efl,zcbenz/cefode-chromium,zcbenz/cefode-chromium,Fireblend/chromium-crosswalk,nacl-webkit/chrome_deps,pozdnyakov/chromium-crosswalk,robclark/chromium,Pluto-tv/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,Fireblend/chromium-crosswalk,chuan9/chromium-crosswalk,markYoungH/chromium.src,pozdnyakov/chromium-crosswalk,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk-efl,dednal/chromium.src,M4sse/chromium.src,Just-D/chromium-1,bright-sparks/chromium-spacewalk,Jonekee/chromium.src,chuan9/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,axinging/chromium-crosswalk,rogerwang/chromium,ChromiumWebApps/chromium,dushu1203/chromium.src,hgl888/chromium-crosswalk,markYoungH/chromium.src,pozdnyakov/chromium-crosswalk,Fireblend/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,patrickm/chromium.src,robclark/chromium,mogoweb/chromium-crosswalk,Fireblend/chromium-crosswalk,dushu1203/chromium.src,axinging/chromium-crosswalk,junmin-zhu/chromium-rivertrail,keishi/chromium,hgl888/chromium-crosswalk-efl,ltilve/chromium,zcbenz/cefode-chromium,ChromiumWebApps/chromium,chuan9/chromium-crosswalk,chuan9/chromium-crosswalk,Pluto-tv/chromium-crosswalk,pozdnyakov/chromium-crosswalk,Pluto-tv/chromium-crosswalk,patrickm/chromium.src,ChromiumWebApps/chromium,hgl888/chromium-crosswalk,M4sse/chromium.src,crosswalk-project/chromium-crosswalk-efl,mohamed--abdel-maksoud/chromium.src,Chilledheart/chromium,mogoweb/chromium-crosswalk,dednal/chromium.src,timopulkkinen/BubbleFish,krieger-od/nwjs_chromium.src,littlstar/chromium.src,axinging/chromium-crosswalk,robclark/chromium,anirudhSK/chromium,pozdnyakov/chromium-crosswalk,M4sse/chromium.src,mohamed--abdel-maksoud/chromium.src,jaruba/chromium.src,pozdnyakov/chromium-crosswalk,hujiajie/pa-chromium,ChromiumWebApps/chromium,crosswalk-project/chromium-crosswalk-efl,krieger-od/nwjs_chromium.src,keishi/chromium,robclark/chromium,rogerwang/chromium,mohamed--abdel-maksoud/chromium.src,Jonekee/chromium.src,anirudhSK/chromium,dednal/chromium.src,mogoweb/chromium-crosswalk,Chilledheart/chromium,mohamed--abdel-maksoud/chromium.src,rogerwang/chromium,nacl-webkit/chrome_deps,M4sse/chromium.src,ltilve/chromium,dushu1203/chromium.src,fujunwei/chromium-crosswalk,jaruba/chromium.src,dednal/chromium.src,patrickm/chromium.src,PeterWangIntel/chromium-crosswalk,markYoungH/chromium.src,mogoweb/chromium-crosswalk,ondra-novak/chromium.src,hgl888/chromium-crosswalk,jaruba/chromium.src,ltilve/chromium,Just-D/chromium-1,anirudhSK/chromium,bright-sparks/chromium-spacewalk,krieger-od/nwjs_chromium.src,dednal/chromium.src,zcbenz/cefode-chromium,dednal/chromium.src,nacl-webkit/chrome_deps,robclark/chromium,robclark/chromium,mogoweb/chromium-crosswalk,keishi/chromium,PeterWangIntel/chromium-crosswalk,hujiajie/pa-chromium,hujiajie/pa-chromium,ltilve/chromium,fujunwei/chromium-crosswalk,axinging/chromium-crosswalk,junmin-zhu/chromium-rivertrail,hujiajie/pa-chromium,markYoungH/chromium.src,bright-sparks/chromium-spacewalk,crosswalk-project/chromium-crosswalk-efl,bright-sparks/chromium-spacewalk,ondra-novak/chromium.src,littlstar/chromium.src,timopulkkinen/BubbleFish,PeterWangIntel/chromium-crosswalk,hgl888/chromium-crosswalk-efl,krieger-od/nwjs_chromium.src,jaruba/chromium.src,junmin-zhu/chromium-rivertrail,chuan9/chromium-crosswalk,timopulkkinen/BubbleFish,littlstar/chromium.src,fujunwei/chromium-crosswalk,littlstar/chromium.src,Just-D/chromium-1,ChromiumWebApps/chromium,ChromiumWebApps/chromium,Jonekee/chromium.src,ltilve/chromium,M4sse/chromium.src,krieger-od/nwjs_chromium.src,Jonekee/chromium.src,hgl888/chromium-crosswalk-efl,pozdnyakov/chromium-crosswalk,markYoungH/chromium.src,zcbenz/cefode-chromium,Pluto-tv/chromium-crosswalk,zcbenz/cefode-chromium,chuan9/chromium-crosswalk,timopulkkinen/BubbleFish,Chilledheart/chromium,TheTypoMaster/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,Chilledheart/chromium,keishi/chromium,PeterWangIntel/chromium-crosswalk,Chilledheart/chromium,timopulkkinen/BubbleFish,jaruba/chromium.src,mohamed--abdel-maksoud/chromium.src,chuan9/chromium-crosswalk,jaruba/chromium.src,ChromiumWebApps/chromium,krieger-od/nwjs_chromium.src,fujunwei/chromium-crosswalk,dushu1203/chromium.src,mogoweb/chromium-crosswalk,rogerwang/chromium,ltilve/chromium,timopulkkinen/BubbleFish,bright-sparks/chromium-spacewalk,keishi/chromium,rogerwang/chromium,ltilve/chromium,Chilledheart/chromium,hgl888/chromium-crosswalk-efl,crosswalk-project/chromium-crosswalk-efl,keishi/chromium,littlstar/chromium.src,Fireblend/chromium-crosswalk,bright-sparks/chromium-spacewalk,littlstar/chromium.src,zcbenz/cefode-chromium,bright-sparks/chromium-spacewalk,junmin-zhu/chromium-rivertrail,ondra-novak/chromium.src,nacl-webkit/chrome_deps,TheTypoMaster/chromium-crosswalk,hgl888/chromium-crosswalk,hgl888/chromium-crosswalk-efl,crosswalk-project/chromium-crosswalk-efl,mohamed--abdel-maksoud/chromium.src,Jonekee/chromium.src,Just-D/chromium-1,hgl888/chromium-crosswalk,dushu1203/chromium.src,keishi/chromium,hujiajie/pa-chromium,junmin-zhu/chromium-rivertrail,dednal/chromium.src,hujiajie/pa-chromium,ChromiumWebApps/chromium,dushu1203/chromium.src,zcbenz/cefode-chromium,Chilledheart/chromium,dednal/chromium.src,PeterWangIntel/chromium-crosswalk,anirudhSK/chromium,hgl888/chromium-crosswalk-efl,axinging/chromium-crosswalk,rogerwang/chromium,M4sse/chromium.src,bright-sparks/chromium-spacewalk,mogoweb/chromium-crosswalk,ltilve/chromium,chuan9/chromium-crosswalk,pozdnyakov/chromium-crosswalk,keishi/chromium,fujunwei/chromium-crosswalk,Jonekee/chromium.src,markYoungH/chromium.src,crosswalk-project/chromium-crosswalk-efl,jaruba/chromium.src,jaruba/chromium.src,mohamed--abdel-maksoud/chromium.src,Just-D/chromium-1,patrickm/chromium.src,TheTypoMaster/chromium-crosswalk,ltilve/chromium,anirudhSK/chromium,crosswalk-project/chromium-crosswalk-efl,patrickm/chromium.src,axinging/chromium-crosswalk,Jonekee/chromium.src,jaruba/chromium.src,nacl-webkit/chrome_deps,Fireblend/chromium-crosswalk,anirudhSK/chromium,hgl888/chromium-crosswalk,chuan9/chromium-crosswalk,Fireblend/chromium-crosswalk,zcbenz/cefode-chromium,axinging/chromium-crosswalk,Chilledheart/chromium,mohamed--abdel-maksoud/chromium.src,anirudhSK/chromium,markYoungH/chromium.src,dednal/chromium.src,fujunwei/chromium-crosswalk,Just-D/chromium-1,markYoungH/chromium.src,anirudhSK/chromium,TheTypoMaster/chromium-crosswalk,junmin-zhu/chromium-rivertrail,Just-D/chromium-1,nacl-webkit/chrome_deps,junmin-zhu/chromium-rivertrail,jaruba/chromium.src,ondra-novak/chromium.src,M4sse/chromium.src,junmin-zhu/chromium-rivertrail,robclark/chromium,keishi/chromium,fujunwei/chromium-crosswalk,axinging/chromium-crosswalk,anirudhSK/chromium,hgl888/chromium-crosswalk,dushu1203/chromium.src,patrickm/chromium.src,timopulkkinen/BubbleFish,Jonekee/chromium.src,M4sse/chromium.src,zcbenz/cefode-chromium,mogoweb/chromium-crosswalk,ChromiumWebApps/chromium,Fireblend/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,PeterWangIntel/chromium-crosswalk,Chilledheart/chromium | ---
+++
@@ -26,12 +26,24 @@
def to_string(self):
return self.val['spec_']
+class FilePathPrinter(object):
+ def __init__(self, val):
+ self.val = val
+
+ def to_string(self):
+ return self.val['path_']['_M_dataplus']['_M_p']
+
+
def lookup_function(val):
- typ = str(val.type)
- if typ == 'string16':
- return String16Printer(val)
- elif typ == 'GURL':
- return GURLPrinter(val)
+ type_to_printer = {
+ 'string16': String16Printer,
+ 'GURL': GURLPrinter,
+ 'FilePath': FilePathPrinter,
+ }
+
+ printer = type_to_printer.get(str(val.type), None)
+ if printer:
+ return printer(val)
return None
gdb.pretty_printers.append(lookup_function) |
97401a56e59d06acdd455f111dbe993265f2a39d | setup.py | setup.py | from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='urbansim',
version='0.2dev',
description='Tool for modeling metropolitan real estate markets',
author='Synthicity',
author_email='ffoti@berkeley.edu',
license='AGPL',
url='https://github.com/synthicity/urbansim',
classifiers=[
'Development Status :: 4 - Beta',
'Programming Language :: Python :: 2.7',
'License :: OSI Approved :: GNU Affero General Public License v3'
],
packages=find_packages(exclude=['urbansimd', '*.tests']),
package_data={'urbansim.urbansim': ['templates/*.template']},
install_requires=[
'Django>=1.6.2',
'jinja2>=2.7.2',
'numpy>=1.8.0',
'pandas>=0.13.1',
'patsy>=0.2.1',
'pyyaml>=3.10',
'scipy>=0.13.3',
'shapely>=1.3.0',
'simplejson>=3.3.3',
'statsmodels>=0.5.0',
'tables>=3.1.0'
],
entry_points={
'console_scripts': [
'urbansim_compile = urbansim.urbansim.compilecli:main',
'urbansim_serve = urbansim.server.servecli:main'
]
}
)
| from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='urbansim',
version='0.2dev',
description='Tool for modeling metropolitan real estate markets',
author='Synthicity',
author_email='ffoti@berkeley.edu',
license='AGPL',
url='https://github.com/synthicity/urbansim',
classifiers=[
'Development Status :: 4 - Beta',
'Programming Language :: Python :: 2.7',
'License :: OSI Approved :: GNU Affero General Public License v3'
],
packages=find_packages(exclude=['*.tests']),
package_data={'urbansim.urbansim': ['templates/*.template']},
install_requires=[
'Django>=1.6.2',
'jinja2>=2.7.2',
'numpy>=1.8.0',
'pandas>=0.13.1',
'patsy>=0.2.1',
'pyyaml>=3.10',
'scipy>=0.13.3',
'shapely>=1.3.0',
'simplejson>=3.3.3',
'statsmodels>=0.5.0',
'tables>=3.1.0'
],
entry_points={
'console_scripts': [
'urbansim_compile = urbansim.urbansim.compilecli:main',
'urbansim_serve = urbansim.server.servecli:main'
]
}
)
| Remove urbansimd from excluded packages. | Remove urbansimd from excluded packages.
| Python | bsd-3-clause | SANDAG/urbansim,apdjustino/urbansim,synthicity/urbansim,UDST/urbansim,UDST/urbansim,bricegnichols/urbansim,AZMAG/urbansim,ual/urbansim,UDST/urbansim,VladimirTyrin/urbansim,waddell/urbansim,VladimirTyrin/urbansim,bricegnichols/urbansim,SANDAG/urbansim,waddell/urbansim,waddell/urbansim,apdjustino/urbansim,apdjustino/urbansim,SANDAG/urbansim,synthicity/urbansim,ual/urbansim,VladimirTyrin/urbansim,synthicity/urbansim,apdjustino/urbansim,AZMAG/urbansim,ual/urbansim,waddell/urbansim,bricegnichols/urbansim,bricegnichols/urbansim,ual/urbansim,VladimirTyrin/urbansim,synthicity/urbansim,AZMAG/urbansim,UDST/urbansim,AZMAG/urbansim,SANDAG/urbansim | ---
+++
@@ -16,7 +16,7 @@
'Programming Language :: Python :: 2.7',
'License :: OSI Approved :: GNU Affero General Public License v3'
],
- packages=find_packages(exclude=['urbansimd', '*.tests']),
+ packages=find_packages(exclude=['*.tests']),
package_data={'urbansim.urbansim': ['templates/*.template']},
install_requires=[
'Django>=1.6.2', |
267b5392adaf3e3f93a22f95e5f1f161225a7f3a | setup.py | setup.py | import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name='django-pgallery',
version=__import__('pgallery').__version__,
description='Photo gallery app for PostgreSQL and Django.',
long_description=read('README.rst'),
author='Zbigniew Siciarz',
author_email='zbigniew@siciarz.net',
url='https://github.com/zsiciarz/django-pgallery',
download_url='https://pypi.python.org/pypi/django-pgallery',
license='MIT',
install_requires=[
'Django>=1.9',
'Pillow',
'psycopg2>=2.5',
'django-markitup>=2.0',
'django-model-utils>=2.0',
],
packages=find_packages(exclude=['tests']),
include_package_data=True,
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Utilities',
],
)
| import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name='django-pgallery',
version=__import__('pgallery').__version__,
description='Photo gallery app for PostgreSQL and Django.',
long_description=read('README.rst'),
author='Zbigniew Siciarz',
author_email='zbigniew@siciarz.net',
url='https://github.com/zsiciarz/django-pgallery',
download_url='https://pypi.python.org/pypi/django-pgallery',
license='MIT',
install_requires=[
'Django>=1.9',
'Pillow',
'psycopg2>=2.5',
'django-markitup>=2.0',
'django-model-utils>=2.0',
],
packages=find_packages(exclude=['tests']),
include_package_data=True,
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
'Topic :: Utilities',
],
)
| Add classifier for Python 3.6 | Add classifier for Python 3.6
| Python | mit | zsiciarz/django-pgallery,zsiciarz/django-pgallery | ---
+++
@@ -35,6 +35,7 @@
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
+ 'Programming Language :: Python :: 3.6',
'Topic :: Utilities',
],
) |
ad5cb91fa011e067a96835e59e05581af3ea3a53 | acctwatch/configcheck.py | acctwatch/configcheck.py | import httplib2
import os
import sys
import time
from apiclient import discovery
from oauth2client import file
from oauth2client import client
from oauth2client import tools
from config import Configuration
try:
import geoip2.database as geoipdb
except ImportError:
geoipdb = None
def main():
config = Configuration()
if config.WITH_GEOIP and not geoipdb:
print ("GeoIP is enabled, but unable to import module, please check installation. Disabling.")
config.WITH_GEOIP = False
credentials = config.get_credentials()
# Create an httplib2.Http object to handle our HTTP requests and authorize it
# with our good Credentials.
http = httplib2.Http()
http = credentials.authorize(http)
# Construct the service object for the interacting with the Admin Reports API.
service = discovery.build('admin', 'reports_v1', http=http)
activities = service.activities()
try:
login_list = activities.list(userKey='all', applicationName='login', maxResults=1000).execute()
print("Success!")
except client.AccessTokenRefreshError:
print("Failure. Access token is invalid.")
if __name__ == '__main__':
main()
| import httplib2
import os
import sys
import time
from apiclient import discovery
from oauth2client import file
from oauth2client import client
from oauth2client import tools
from config import Configuration
try:
import geoip2.database as geoipdb
except ImportError:
print ("GeoIP is missing, please install dependency")
def main():
config = Configuration()
credentials = config.get_credentials()
# Create an httplib2.Http object to handle our HTTP requests and authorize it
# with our good Credentials.
http = httplib2.Http()
http = credentials.authorize(http)
# Construct the service object for the interacting with the Admin Reports API.
service = discovery.build('admin', 'reports_v1', http=http)
activities = service.activities()
try:
login_list = activities.list(userKey='all', applicationName='login', maxResults=1000).execute()
print("Success!")
except client.AccessTokenRefreshError:
print("Failure. Access token is invalid. Please re-run the tool to get a new access token.")
if __name__ == '__main__':
main()
| Clean up configuration check utility | Clean up configuration check utility
| Python | isc | GuardedRisk/Google-Apps-Auditing | ---
+++
@@ -13,14 +13,10 @@
try:
import geoip2.database as geoipdb
except ImportError:
- geoipdb = None
+ print ("GeoIP is missing, please install dependency")
def main():
config = Configuration()
- if config.WITH_GEOIP and not geoipdb:
- print ("GeoIP is enabled, but unable to import module, please check installation. Disabling.")
- config.WITH_GEOIP = False
-
credentials = config.get_credentials()
# Create an httplib2.Http object to handle our HTTP requests and authorize it
@@ -36,7 +32,7 @@
login_list = activities.list(userKey='all', applicationName='login', maxResults=1000).execute()
print("Success!")
except client.AccessTokenRefreshError:
- print("Failure. Access token is invalid.")
+ print("Failure. Access token is invalid. Please re-run the tool to get a new access token.")
if __name__ == '__main__':
main() |
fae3f6aaba91167c5da2f3d5d9b6b1a66068f9f7 | setup.py | setup.py | from distutils.core import setup
import os, glob, string, shutil
# Packages
packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap']
def main():
setup (name = 'neuroimaging',
version = '0.01a',
description = 'This is a neuroimaging python package',
author = 'Various, one of whom is Jonathan Taylor',
author_email = 'jonathan.taylor@stanford.edu',
ext_package = 'neuroimaging',
packages=packages,
package_dir = {'neuroimaging': 'lib'},
url = 'http://neuroimaging.scipy.org',
long_description =
'''
''')
if __name__ == "__main__":
main()
| import os, glob, string, shutil
from distutils.core import setup
# Packages
packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap']
def main():
setup (name = 'neuroimaging',
version = '0.01a',
description = 'This is a neuroimaging python package',
author = 'Various, one of whom is Jonathan Taylor',
author_email = 'jonathan.taylor@stanford.edu',
ext_package = 'neuroimaging',
packages=packages,
package_dir = {'neuroimaging': 'lib'},
url = 'http://neuroimaging.scipy.org',
long_description =
'''
''')
if __name__ == "__main__":
main()
| Test edit - to check svn email hook | Test edit - to check svn email hook | Python | bsd-3-clause | yarikoptic/NiPy-OLD,yarikoptic/NiPy-OLD | ---
+++
@@ -1,5 +1,5 @@
+import os, glob, string, shutil
from distutils.core import setup
-import os, glob, string, shutil
# Packages
|
50efc8ccccecc3f48fcf0e82ebc3390da48e0967 | setup.py | setup.py | from setuptools import setup
setup(
name="balrog",
version="1.0",
description="Mozilla's Update Server",
author="Ben Hearsum",
author_email="ben@hearsum.ca",
packages=[
"auslib",
"auslib.admin",
"auslib.admin.views",
"auslib.blobs",
"auslib.migrate",
"auslib.migrate.versions",
"auslib.util",
"auslib.web",
"auslib.web.views",
],
include_package_data=True,
install_requires=[
"flask==0.10.1",
"Werkzeug==0.9.6",
"wtforms==2.0.1",
"flask-wtf==0.10.2",
"sqlalchemy-migrate==0.7.2",
"tempita==0.5.1",
"decorator==3.3.3",
"blinker==1.2",
"cef==0.5",
"flask-compress==1.0.2",
"itsdangerous==0.24",
"repoze.lru==0.6",
],
url="https://github.com/mozilla/balrog",
)
| from setuptools import find_packages, setup
setup(
name="balrog",
version="1.0",
description="Mozilla's Update Server",
author="Ben Hearsum",
author_email="ben@hearsum.ca",
packages=find_packages(exclude=["vendor"]),
include_package_data=True,
install_requires=[
"flask==0.10.1",
"Werkzeug==0.9.6",
"wtforms==2.0.1",
"flask-wtf==0.10.2",
"sqlalchemy-migrate==0.7.2",
"tempita==0.5.1",
"decorator==3.3.3",
"blinker==1.2",
"cef==0.5",
"flask-compress==1.0.2",
"itsdangerous==0.24",
"repoze.lru==0.6",
],
url="https://github.com/mozilla/balrog",
)
| Use find_packages instead of an explicit list of packages for easier maintenance. | Use find_packages instead of an explicit list of packages for easier maintenance.
| Python | mpl-2.0 | nurav/balrog,testbhearsum/balrog,nurav/balrog,testbhearsum/balrog,mozbhearsum/balrog,aksareen/balrog,nurav/balrog,testbhearsum/balrog,nurav/balrog,mozbhearsum/balrog,tieu/balrog,mozbhearsum/balrog,tieu/balrog,aksareen/balrog,aksareen/balrog,tieu/balrog,mozbhearsum/balrog,tieu/balrog,aksareen/balrog,testbhearsum/balrog | ---
+++
@@ -1,4 +1,4 @@
-from setuptools import setup
+from setuptools import find_packages, setup
setup(
@@ -7,17 +7,7 @@
description="Mozilla's Update Server",
author="Ben Hearsum",
author_email="ben@hearsum.ca",
- packages=[
- "auslib",
- "auslib.admin",
- "auslib.admin.views",
- "auslib.blobs",
- "auslib.migrate",
- "auslib.migrate.versions",
- "auslib.util",
- "auslib.web",
- "auslib.web.views",
- ],
+ packages=find_packages(exclude=["vendor"]),
include_package_data=True,
install_requires=[
"flask==0.10.1", |
2a7877c1ed3e1dc9a5bcc27220847a5a75cf65ab | spiff/membership/management/commands/bill_members.py | spiff/membership/management/commands/bill_members.py | from django.core.management import BaseCommand
from spiff.membership.utils import monthRange
from spiff.membership.models import Member, RankLineItem
from spiff.payment.models import Invoice
class Command(BaseCommand):
help = 'Bills active members for the month'
def handle(self, *args, **options):
for member in Member.objects.all():
if not member.billedForMonth():
if member.highestRank is not None and member.highestRank.monthlyDues > 0:
print "Billing", member, "for the month"
endOfMonth, startOfMonth = monthRange()
invoice = Invoice.objects.create(
user=member.user,
dueDate=endOfMonth,
)
for group in member.user.groups.all():
if group.rank.monthlyDues > 0:
lineItem = RankLineItem.objects.create(
rank = group.rank,
member = member,
activeFromDate=startOfMonth,
activeToDate=endOfMonth,
invoice=invoice
)
print "\tCreated", lineItem
invoice.draft = False
invoice.open = True
invoice.save()
print "\tInvoice saved!"
else:
print "%s has outstanding balance of $%s"%(
member,
member.outstandingBalance
)
| from django.core.management import BaseCommand
from spiff.membership.utils import monthRange
from spiff.membership.models import Member, RankLineItem
from spiff.payment.models import Invoice
class Command(BaseCommand):
help = 'Bills active members for the month'
def handle(self, *args, **options):
for member in Member.objects.all():
if not member.billedForMonth():
if member.highestRank is not None and member.highestRank.monthlyDues > 0:
print "Billing", member, "for the month"
startOfMonth, endOfMonth = monthRange()
invoice = Invoice.objects.create(
user=member.user,
dueDate=endOfMonth,
)
for group in member.user.groups.all():
if group.rank.monthlyDues > 0:
lineItem = RankLineItem.objects.create(
rank = group.rank,
member = member,
activeFromDate=startOfMonth,
activeToDate=endOfMonth,
invoice=invoice
)
print "\tCreated", lineItem
invoice.draft = False
invoice.open = True
invoice.save()
print "\tInvoice saved!"
else:
print "%s has outstanding balance of $%s"%(
member,
member.outstandingBalance
)
| Fix swapped dates on invoices | Fix swapped dates on invoices
| Python | agpl-3.0 | SYNHAK/spiff,SYNHAK/spiff,SYNHAK/spiff | ---
+++
@@ -11,7 +11,7 @@
if not member.billedForMonth():
if member.highestRank is not None and member.highestRank.monthlyDues > 0:
print "Billing", member, "for the month"
- endOfMonth, startOfMonth = monthRange()
+ startOfMonth, endOfMonth = monthRange()
invoice = Invoice.objects.create(
user=member.user,
dueDate=endOfMonth, |
e3409c94b64deac85deada28f57a30ae08d0083d | api/caching/listeners.py | api/caching/listeners.py | from functools import partial
from api.caching.tasks import ban_url
from framework.postcommit_tasks.handlers import enqueue_postcommit_task
from modularodm import signals
@signals.save.connect
def ban_object_from_cache(sender, instance, fields_changed, cached_data):
abs_url = None
if hasattr(instance, 'absolute_api_v2_url'):
abs_url = instance.absolute_api_v2_url
if abs_url is not None:
enqueue_postcommit_task(partial(ban_url, abs_url))
| from functools import partial
from api.caching.tasks import ban_url
from framework.postcommit_tasks.handlers import enqueue_postcommit_task
from modularodm import signals
@signals.save.connect
def ban_object_from_cache(sender, instance, fields_changed, cached_data):
abs_url = None
if hasattr(instance, 'absolute_api_v2_url'):
abs_url = instance.absolute_api_v2_url
if abs_url is not None:
enqueue_postcommit_task(partial(ban_url, instance, fields_changed))
| Fix passing arguments to ban_url | Fix passing arguments to ban_url
h/t @cwisecarver
| Python | apache-2.0 | chrisseto/osf.io,cslzchen/osf.io,Johnetordoff/osf.io,SSJohns/osf.io,alexschiller/osf.io,mluke93/osf.io,rdhyee/osf.io,zachjanicki/osf.io,wearpants/osf.io,leb2dg/osf.io,caneruguz/osf.io,zamattiac/osf.io,TomHeatwole/osf.io,mattclark/osf.io,aaxelb/osf.io,felliott/osf.io,caseyrollins/osf.io,mluke93/osf.io,erinspace/osf.io,RomanZWang/osf.io,cwisecarver/osf.io,brianjgeiger/osf.io,mluo613/osf.io,binoculars/osf.io,mluo613/osf.io,cslzchen/osf.io,adlius/osf.io,cwisecarver/osf.io,alexschiller/osf.io,chrisseto/osf.io,mattclark/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,brianjgeiger/osf.io,brianjgeiger/osf.io,acshi/osf.io,TomHeatwole/osf.io,abought/osf.io,laurenrevere/osf.io,Johnetordoff/osf.io,saradbowman/osf.io,rdhyee/osf.io,Nesiehr/osf.io,chrisseto/osf.io,CenterForOpenScience/osf.io,kwierman/osf.io,rdhyee/osf.io,SSJohns/osf.io,RomanZWang/osf.io,emetsger/osf.io,cslzchen/osf.io,crcresearch/osf.io,jnayak1/osf.io,cslzchen/osf.io,aaxelb/osf.io,mfraezz/osf.io,jnayak1/osf.io,icereval/osf.io,Nesiehr/osf.io,abought/osf.io,abought/osf.io,mluo613/osf.io,alexschiller/osf.io,SSJohns/osf.io,zamattiac/osf.io,sloria/osf.io,binoculars/osf.io,RomanZWang/osf.io,cwisecarver/osf.io,doublebits/osf.io,emetsger/osf.io,SSJohns/osf.io,zachjanicki/osf.io,adlius/osf.io,zamattiac/osf.io,binoculars/osf.io,amyshi188/osf.io,baylee-d/osf.io,sloria/osf.io,emetsger/osf.io,baylee-d/osf.io,TomBaxter/osf.io,laurenrevere/osf.io,aaxelb/osf.io,mfraezz/osf.io,cwisecarver/osf.io,rdhyee/osf.io,jnayak1/osf.io,felliott/osf.io,kch8qx/osf.io,mluke93/osf.io,mluo613/osf.io,felliott/osf.io,samchrisinger/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,kch8qx/osf.io,acshi/osf.io,monikagrabowska/osf.io,jnayak1/osf.io,pattisdr/osf.io,icereval/osf.io,asanfilippo7/osf.io,crcresearch/osf.io,samchrisinger/osf.io,asanfilippo7/osf.io,HalcyonChimera/osf.io,mfraezz/osf.io,icereval/osf.io,caneruguz/osf.io,TomBaxter/osf.io,erinspace/osf.io,hmoco/osf.io,Nesiehr/osf.io,adlius/osf.io,CenterForOpenScience/osf.io,caseyrollins/osf.io,zamattiac/osf.io,hmoco/osf.io,amyshi188/osf.io,kwierman/osf.io,mluke93/osf.io,Johnetordoff/osf.io,samchrisinger/osf.io,DanielSBrown/osf.io,hmoco/osf.io,doublebits/osf.io,doublebits/osf.io,caneruguz/osf.io,acshi/osf.io,crcresearch/osf.io,erinspace/osf.io,mluo613/osf.io,kwierman/osf.io,Nesiehr/osf.io,doublebits/osf.io,RomanZWang/osf.io,baylee-d/osf.io,CenterForOpenScience/osf.io,alexschiller/osf.io,HalcyonChimera/osf.io,RomanZWang/osf.io,wearpants/osf.io,CenterForOpenScience/osf.io,zachjanicki/osf.io,chrisseto/osf.io,alexschiller/osf.io,TomHeatwole/osf.io,TomBaxter/osf.io,chennan47/osf.io,saradbowman/osf.io,DanielSBrown/osf.io,kch8qx/osf.io,hmoco/osf.io,acshi/osf.io,kwierman/osf.io,monikagrabowska/osf.io,caneruguz/osf.io,monikagrabowska/osf.io,TomHeatwole/osf.io,monikagrabowska/osf.io,mfraezz/osf.io,abought/osf.io,sloria/osf.io,asanfilippo7/osf.io,chennan47/osf.io,kch8qx/osf.io,zachjanicki/osf.io,adlius/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,emetsger/osf.io,chennan47/osf.io,amyshi188/osf.io,caseyrollins/osf.io,samchrisinger/osf.io,pattisdr/osf.io,pattisdr/osf.io,HalcyonChimera/osf.io,leb2dg/osf.io,doublebits/osf.io,kch8qx/osf.io,wearpants/osf.io,laurenrevere/osf.io,mattclark/osf.io,acshi/osf.io,aaxelb/osf.io,asanfilippo7/osf.io,monikagrabowska/osf.io,amyshi188/osf.io,felliott/osf.io,HalcyonChimera/osf.io,wearpants/osf.io | ---
+++
@@ -3,7 +3,6 @@
from api.caching.tasks import ban_url
from framework.postcommit_tasks.handlers import enqueue_postcommit_task
from modularodm import signals
-
@signals.save.connect
def ban_object_from_cache(sender, instance, fields_changed, cached_data):
@@ -12,4 +11,4 @@
abs_url = instance.absolute_api_v2_url
if abs_url is not None:
- enqueue_postcommit_task(partial(ban_url, abs_url))
+ enqueue_postcommit_task(partial(ban_url, instance, fields_changed)) |
dae054de92d6d864f77a337d269ba9b0c5ddeec4 | examples/charts/file/stacked_bar.py | examples/charts/file/stacked_bar.py | from bokeh.charts import Bar, output_file, show
from bokeh.charts.operations import blend
from bokeh.charts._attributes import cat, color
from bokeh.charts.utils import df_from_json
from bokeh.sampledata.olympics2014 import data
df = df_from_json(data)
# filter by countries with at least one medal and sort
df = df[df['medals.total'] > 0]
df = df.sort("medals.total", ascending=False)
df = df.rename(columns={'medals.gold': 'gold', 'medals.silver': 'silver',
'medals.bronze': 'bronze', 'medals.total': 'total'})
bar = Bar(df,
values=blend('bronze', 'silver', 'gold', name='medals', labels_name='medal'),
label=cat(columns='abbr', sort=False),
stack=cat(columns='medal', sort=False),
color=color(columns='medal', palette=['SaddleBrown', 'Silver', 'Goldenrod'],
sort=False),
legend='top_right',
title="Stacked bars")
output_file("stacked_bar.html")
show(bar)
| from bokeh.charts import Bar, output_file, show
from bokeh.charts.operations import blend
from bokeh.charts.attributes import cat, color
from bokeh.charts.utils import df_from_json
from bokeh.sampledata.olympics2014 import data
df = df_from_json(data)
# filter by countries with at least one medal and sort
df = df[df['medals.total'] > 0]
df = df.sort("medals.total", ascending=False)
df = df.rename(columns={'medals.gold': 'gold', 'medals.silver': 'silver',
'medals.bronze': 'bronze', 'medals.total': 'total'})
bar = Bar(df,
values=blend('bronze', 'silver', 'gold', name='medals', labels_name='medal'),
label=cat(columns='abbr', sort=False),
stack=cat(columns='medal', sort=False),
color=color(columns='medal', palette=['SaddleBrown', 'Silver', 'Goldenrod'],
sort=False),
legend='top_right',
title="Stacked bars")
output_file("stacked_bar.html")
show(bar)
| Fix import on stacked bar example. | Fix import on stacked bar example.
| Python | bsd-3-clause | percyfal/bokeh,maxalbert/bokeh,phobson/bokeh,ChinaQuants/bokeh,quasiben/bokeh,DuCorey/bokeh,schoolie/bokeh,phobson/bokeh,KasperPRasmussen/bokeh,bokeh/bokeh,Karel-van-de-Plassche/bokeh,rs2/bokeh,DuCorey/bokeh,aavanian/bokeh,gpfreitas/bokeh,aavanian/bokeh,KasperPRasmussen/bokeh,draperjames/bokeh,dennisobrien/bokeh,maxalbert/bokeh,ericmjl/bokeh,jakirkham/bokeh,htygithub/bokeh,mindriot101/bokeh,phobson/bokeh,schoolie/bokeh,ericmjl/bokeh,timsnyder/bokeh,phobson/bokeh,azjps/bokeh,philippjfr/bokeh,ChinaQuants/bokeh,ericmjl/bokeh,schoolie/bokeh,stonebig/bokeh,gpfreitas/bokeh,clairetang6/bokeh,justacec/bokeh,philippjfr/bokeh,azjps/bokeh,ptitjano/bokeh,timsnyder/bokeh,msarahan/bokeh,bokeh/bokeh,jakirkham/bokeh,Karel-van-de-Plassche/bokeh,ptitjano/bokeh,ericmjl/bokeh,msarahan/bokeh,schoolie/bokeh,dennisobrien/bokeh,DuCorey/bokeh,philippjfr/bokeh,DuCorey/bokeh,schoolie/bokeh,mindriot101/bokeh,jakirkham/bokeh,DuCorey/bokeh,bokeh/bokeh,justacec/bokeh,philippjfr/bokeh,dennisobrien/bokeh,justacec/bokeh,ericmjl/bokeh,stonebig/bokeh,stonebig/bokeh,maxalbert/bokeh,draperjames/bokeh,Karel-van-de-Plassche/bokeh,msarahan/bokeh,justacec/bokeh,ptitjano/bokeh,aiguofer/bokeh,mindriot101/bokeh,bokeh/bokeh,bokeh/bokeh,rs2/bokeh,mindriot101/bokeh,rs2/bokeh,htygithub/bokeh,KasperPRasmussen/bokeh,clairetang6/bokeh,azjps/bokeh,aiguofer/bokeh,ChinaQuants/bokeh,clairetang6/bokeh,quasiben/bokeh,aavanian/bokeh,rs2/bokeh,stonebig/bokeh,KasperPRasmussen/bokeh,aiguofer/bokeh,quasiben/bokeh,maxalbert/bokeh,azjps/bokeh,aiguofer/bokeh,philippjfr/bokeh,draperjames/bokeh,timsnyder/bokeh,ptitjano/bokeh,gpfreitas/bokeh,gpfreitas/bokeh,aavanian/bokeh,Karel-van-de-Plassche/bokeh,timsnyder/bokeh,dennisobrien/bokeh,Karel-van-de-Plassche/bokeh,aavanian/bokeh,dennisobrien/bokeh,percyfal/bokeh,htygithub/bokeh,draperjames/bokeh,percyfal/bokeh,azjps/bokeh,clairetang6/bokeh,jakirkham/bokeh,percyfal/bokeh,ptitjano/bokeh,htygithub/bokeh,msarahan/bokeh,ChinaQuants/bokeh,KasperPRasmussen/bokeh,phobson/bokeh,timsnyder/bokeh,aiguofer/bokeh,jakirkham/bokeh,percyfal/bokeh,draperjames/bokeh,rs2/bokeh | ---
+++
@@ -1,6 +1,6 @@
from bokeh.charts import Bar, output_file, show
from bokeh.charts.operations import blend
-from bokeh.charts._attributes import cat, color
+from bokeh.charts.attributes import cat, color
from bokeh.charts.utils import df_from_json
from bokeh.sampledata.olympics2014 import data
|
d3e2a11f72f6de811f39ac10aa0abde74b99d269 | hcibench/pipeline/__init__.py | hcibench/pipeline/__init__.py | """
The :mod:`hcibench.pipeline` module provides a flexible infrastructure for
data processing and implements some common types of processing blocks.
"""
from .base import PipelineBlock, Pipeline, PassthroughPipeline
from .common import Windower, Filter, FeatureExtractor
__all__ = ['PipelineBlock',
'Pipeline',
'PassthroughPipeline',
'Windower',
'Filter',
'FeatureExtractor',
'Estimator']
| """
The :mod:`hcibench.pipeline` module provides a flexible infrastructure for
data processing and implements some common types of processing blocks.
"""
from .base import PipelineBlock, Pipeline, PassthroughPipeline
from .common import Windower, Filter, FeatureExtractor, Estimator
__all__ = ['PipelineBlock',
'Pipeline',
'PassthroughPipeline',
'Windower',
'Filter',
'FeatureExtractor',
'Estimator']
| Make Estimator importable from pipeline. | Make Estimator importable from pipeline.
| Python | mit | ucdrascal/axopy,ucdrascal/hcibench | ---
+++
@@ -4,7 +4,7 @@
"""
from .base import PipelineBlock, Pipeline, PassthroughPipeline
-from .common import Windower, Filter, FeatureExtractor
+from .common import Windower, Filter, FeatureExtractor, Estimator
__all__ = ['PipelineBlock',
'Pipeline', |
dc0f82fb424bce899493be9ba483a1fe16ea4f64 | setup.py | setup.py |
from setuptools import setup, find_packages
setup(
name='jawa',
packages=find_packages(),
version='2.1.1',
python_requires='>=3.6',
description='Doing fun stuff with JVM ClassFiles.',
long_description=open('README.md', 'r').read(),
long_description_content_type='text/markdown',
author='Tyler Kennedy',
author_email='tk@tkte.ch',
url='http://github.com/TkTech/Jawa',
keywords=[
'java',
'disassembly',
'disassembler',
'assembly'
],
include_package_data=True,
classifiers=[
'Programming Language :: Python',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Topic :: Software Development :: Disassemblers',
'Topic :: Software Development :: Assemblers'
],
install_requires=[
'click>=5.0'
],
tests_require=[
'pytest>=2.10',
],
extras_require={
'dev': [
'pytest',
'sphinx',
'sphinxcontrib-googleanalytics',
'sphinx_rtd_theme',
'sphinx-click',
'ghp-import',
'pyyaml',
'ipython',
'twine',
'wheel'
]
},
entry_points='''
[console_scripts]
jawa=jawa.cli:cli
'''
)
|
from setuptools import setup, find_packages
setup(
name='jawa',
packages=find_packages(),
version='2.1.1',
python_requires='>=3.6',
description='Doing fun stuff with JVM ClassFiles.',
long_description=open('README.md', 'r').read(),
long_description_content_type='text/markdown',
author='Tyler Kennedy',
author_email='tk@tkte.ch',
url='http://github.com/TkTech/Jawa',
keywords=[
'java',
'disassembly',
'disassembler',
'assembly'
],
include_package_data=True,
classifiers=[
'Programming Language :: Python',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Topic :: Software Development :: Disassemblers',
'Topic :: Software Development :: Assemblers'
],
install_requires=[
'click>=5.0'
],
tests_require=[
'pytest>=2.10',
],
extras_require={
'dev': [
'pytest',
'sphinx',
'sphinxcontrib-googleanalytics',
'sphinx_rtd_theme',
'sphinx-click',
'ghp-import',
'pyyaml',
'ipython',
'twine',
'wheel',
'bumpversion'
]
},
entry_points='''
[console_scripts]
jawa=jawa.cli:cli
'''
)
| Add bumpversion as a dev requirement. | Add bumpversion as a dev requirement.
| Python | mit | TkTech/Jawa,TkTech/Jawa | ---
+++
@@ -45,7 +45,8 @@
'pyyaml',
'ipython',
'twine',
- 'wheel'
+ 'wheel',
+ 'bumpversion'
]
},
entry_points=''' |
e91b691ba2e9a83d8cc94f42bdc41c9a7350c790 | setup.py | setup.py | #!/usr/bin/env python
"""Distutils installer for extras."""
from setuptools import setup
import os.path
import extras
testtools_cmd = extras.try_import('testtools.TestCommand')
def get_version():
"""Return the version of extras that we are building."""
version = '.'.join(
str(component) for component in extras.__version__[0:3])
return version
def get_long_description():
readme_path = os.path.join(
os.path.dirname(__file__), 'README.rst')
return open(readme_path).read()
cmdclass = {}
if testtools_cmd is not None:
cmdclass['test'] = testtools_cmd
setup(name='extras',
author='Testing cabal',
author_email='testtools-dev@lists.launchpad.net',
url='https://github.com/testing-cabal/extras',
description=('Useful extra bits for Python - things that shold be '
'in the standard library'),
long_description=get_long_description(),
version=get_version(),
classifiers=["License :: OSI Approved :: MIT License"],
packages=[
'extras',
'extras.tests',
],
cmdclass=cmdclass)
| #!/usr/bin/env python
"""Distutils installer for extras."""
from setuptools import setup
import os.path
import extras
testtools_cmd = extras.try_import('testtools.TestCommand')
def get_version():
"""Return the version of extras that we are building."""
version = '.'.join(
str(component) for component in extras.__version__[0:3])
return version
def get_long_description():
readme_path = os.path.join(
os.path.dirname(__file__), 'README.rst')
return open(readme_path).read()
cmdclass = {}
if testtools_cmd is not None:
cmdclass['test'] = testtools_cmd
setup(name='extras',
author='Testing cabal',
author_email='testtools-dev@lists.launchpad.net',
url='https://github.com/testing-cabal/extras',
description=('Useful extra bits for Python - things that shold be '
'in the standard library'),
long_description=get_long_description(),
version=get_version(),
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 3",
],
packages=[
'extras',
'extras.tests',
],
cmdclass=cmdclass)
| Add trove classifiers specifying Python 3 support. | Add trove classifiers specifying Python 3 support.
| Python | mit | testing-cabal/extras | ---
+++
@@ -35,7 +35,12 @@
'in the standard library'),
long_description=get_long_description(),
version=get_version(),
- classifiers=["License :: OSI Approved :: MIT License"],
+ classifiers=[
+ "Intended Audience :: Developers",
+ "License :: OSI Approved :: MIT License",
+ "Programming Language :: Python",
+ "Programming Language :: Python :: 3",
+ ],
packages=[
'extras',
'extras.tests', |
0ed714d6982d8d4cec628a7549a0a348526f0cf2 | setup.py | setup.py | import os
import setuptools
setuptools.setup(
name='factory_djoy',
version='0.5',
description='Factories for Django, creating valid instances every time',
url='http://github.com/jamescooke/factory_djoy',
author='James Cooke',
author_email='github@jamescooke.info',
license='MIT',
packages=setuptools.find_packages(exclude=["*.tests", "*.tests.*", "tests.*", "tests"]),
install_requires=[
'Django>=1.8',
'factory_boy>=2.7',
],
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Framework :: Django',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Software Development :: Testing',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| import os
import setuptools
setuptools.setup(
name='factory_djoy',
version='0.5.0',
description='Factories for Django, creating valid instances every time',
url='http://github.com/jamescooke/factory_djoy',
author='James Cooke',
author_email='github@jamescooke.info',
license='MIT',
packages=setuptools.find_packages(exclude=["*.tests", "*.tests.*", "tests.*", "tests"]),
install_requires=[
'Django>=1.8',
'factory_boy>=2.7',
],
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Framework :: Django',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Software Development :: Testing',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| Add z Patch number as per symver.org | Add z Patch number as per symver.org
| Python | mit | jamescooke/factory_djoy | ---
+++
@@ -4,7 +4,7 @@
setuptools.setup(
name='factory_djoy',
- version='0.5',
+ version='0.5.0',
description='Factories for Django, creating valid instances every time',
url='http://github.com/jamescooke/factory_djoy', |
fdb7c400adb777cdc60cf034569d81e95797cc10 | infupy/backends/common.py | infupy/backends/common.py | import sys
from abc import ABCMeta, abstractmethod
def printerr(msg, e=''):
msg = "Backend: " + str(msg)
print(msg.format(e), file=sys.stderr)
class CommandError(Exception):
def __str__(self):
return "Command error: {}".format(self.args)
class Syringe(metaclass=ABCMeta):
_events = set()
@abstractmethod
def execCommand(self, msg):
"""
Send command and read reply.
"""
pass
# Read Perfusion related values
@abstractmethod
def readRate(self):
return 0
@abstractmethod
def readVolume(self):
return 0
# Infusion control
def setRate(self, rate):
pass
# Events
def registerEvent(self, event):
self._events |= set([event])
def unregisterEvent(self, event):
self._events -= set([event])
def clearEvents(self):
self._events = set()
| import sys
from abc import ABCMeta, abstractmethod
def printerr(msg, e=''):
msg = "Backend: " + str(msg)
print(msg.format(e), file=sys.stderr)
class CommandError(Exception):
def __str__(self):
return "Command error: {}".format(self.args)
class Syringe(metaclass=ABCMeta):
_events = set()
@abstractmethod
def execCommand(self, msg):
raise NotImplementedError
# Read Perfusion related values
@abstractmethod
def readRate(self):
raise NotImplementedError
@abstractmethod
def readVolume(self):
raise NotImplementedError
# Infusion control
def setRate(self, rate):
raise NotImplementedError
# Events
def registerEvent(self, event):
self._events |= set([event])
def unregisterEvent(self, event):
self._events -= set([event])
def clearEvents(self):
self._events = set()
| Make abstract methods raise not implemented | Make abstract methods raise not implemented
| Python | isc | jaj42/infupy | ---
+++
@@ -14,23 +14,20 @@
@abstractmethod
def execCommand(self, msg):
- """
- Send command and read reply.
- """
- pass
+ raise NotImplementedError
# Read Perfusion related values
@abstractmethod
def readRate(self):
- return 0
+ raise NotImplementedError
@abstractmethod
def readVolume(self):
- return 0
+ raise NotImplementedError
# Infusion control
def setRate(self, rate):
- pass
+ raise NotImplementedError
# Events
def registerEvent(self, event): |
e635af49a1f72475980ff91406942707a369935d | setup.py | setup.py |
__author__ = 'matth'
from distutils.core import setup
from setuptools import find_packages
setup(
name='py-exe-builder',
version='0.1',
packages = find_packages(),
license='Apache License, Version 2.0',
description='Uses py2exe to create small exe stubs that leverage a full python installation, rather than packing the required pyc files in to the executable.',
long_description=open('README.md').read(),
url='http://www.j5int.com/',
author='j5 International',
author_email='support@j5int.com',
classifiers=[
'Development Status :: 3 - Alpha',
'License :: OSI Approved :: Apache Software License',
'Environment :: Win32 (MS Windows)',
'Intended Audience :: Developers',
'Operating System :: Microsoft :: Windows',
'Programming Language :: Python',
'Programming Language :: Python :: 2 :: Only',
'Topic :: Software Development :: Libraries :: Python Modules',
],
install_requires = ["py2exe"],
) |
__author__ = 'matth'
from distutils.core import setup
from setuptools import find_packages
setup(
name='py-exe-builder',
version='0.2.dev1',
packages = find_packages(),
license='Apache License, Version 2.0',
description='Uses py2exe to create small exe stubs that leverage a full python installation, rather than packing the required pyc files in to the executable.',
long_description=open('README.md').read(),
url='http://www.j5int.com/',
author='j5 International',
author_email='support@j5int.com',
classifiers=[
'Development Status :: 3 - Alpha',
'License :: OSI Approved :: Apache Software License',
'Environment :: Win32 (MS Windows)',
'Intended Audience :: Developers',
'Operating System :: Microsoft :: Windows',
'Programming Language :: Python',
'Programming Language :: Python :: 2 :: Only',
'Topic :: Software Development :: Libraries :: Python Modules',
],
install_requires = ["py2exe"],
) | Change version to dev version | Change version to dev version
| Python | apache-2.0 | j5int/py-exe-builder | ---
+++
@@ -6,7 +6,7 @@
setup(
name='py-exe-builder',
- version='0.1',
+ version='0.2.dev1',
packages = find_packages(),
license='Apache License, Version 2.0',
description='Uses py2exe to create small exe stubs that leverage a full python installation, rather than packing the required pyc files in to the executable.', |
d93b1c1feadb3e8c2dd0643a19b025faec305a7a | setup.py | setup.py | """Setuptools packaging configuration for pyperf."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst') as f:
README = f.read()
setup(
name='py-perf',
version='0.0.1',
url='https://github.com/kevinconway/PyPerf',
license="Apache2",
description='A service for profiling Python snippets.',
author='Kevin Conway',
author_email='kevinjacobconway@gmail.com',
long_description=README,
classifiers=[],
packages=find_packages(exclude=['tests', 'build', 'dist', 'docs']),
install_requires=[
'psutil',
'memory_profiler',
'daemons',
'six',
'amqp',
'stevedore',
],
entry_points={
"pyperf_messages": [
"profile_request = pyperf.messages:ProfileRequest",
"profile_response = pyperf.messages:ProfileResponse",
"profile_failure = pyperf.messages:ProfileFailure",
],
"pyperf_transports": [
"amqp = pyperf.transports.amqp:AmqpTransport",
],
},
include_package_data=True,
zip_safe=False,
)
| """Setuptools packaging configuration for pyperf."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst') as f:
README = f.read()
setup(
name='py-perf',
version='0.0.1',
url='https://github.com/kevinconway/PyPerf',
license="Apache2",
description='A service for profiling Python snippets.',
author='Kevin Conway',
author_email='kevinjacobconway@gmail.com',
long_description=README,
classifiers=[],
packages=find_packages(exclude=['tests', 'build', 'dist', 'docs']),
install_requires=[
'psutil',
'memory_profiler',
'daemons',
'six',
'amqp',
'stevedore',
],
entry_points={
"pyperf_messages": [
"profile_request = pyperf.messages:ProfileRequest",
"profile_response = pyperf.messages:ProfileResponse",
"profile_failure = pyperf.messages:ProfileFailure",
],
"pyperf_transports": [
"amqp = pyperf.transports.amqp:AmqpTransport",
],
"pyperf_profilers": [
"runtime = pyperf.profilers.runtime:RuntimeProfiler",
"memory_max = pyperf.profilers.memory:MaxMemoryProfiler",
"memory_min = pyperf.profilers.memory:MinMemoryProfiler",
"memory_avg = pyperf.profilers.memory:AvgMemoryProfiler",
],
},
include_package_data=True,
zip_safe=False,
)
| Add entry_point for profiler implementations | Add entry_point for profiler implementations
| Python | apache-2.0 | kevinconway/PyPerf,kevinconway/PyPerf,kevinconway/PyPerf | ---
+++
@@ -36,6 +36,12 @@
"pyperf_transports": [
"amqp = pyperf.transports.amqp:AmqpTransport",
],
+ "pyperf_profilers": [
+ "runtime = pyperf.profilers.runtime:RuntimeProfiler",
+ "memory_max = pyperf.profilers.memory:MaxMemoryProfiler",
+ "memory_min = pyperf.profilers.memory:MinMemoryProfiler",
+ "memory_avg = pyperf.profilers.memory:AvgMemoryProfiler",
+ ],
},
include_package_data=True,
zip_safe=False, |
c656bca77d13afcfd8d6df0286b632433e08def1 | setup.py | setup.py | import os
from setuptools import setup
from sys import platform
REQUIREMENTS = []
if platform.startswith('darwin'):
REQUIREMENTS.append('pyobjc-core >= 2.5')
with open(os.path.join(os.path.dirname(__file__), 'nativeconfig', 'version.py')) as f:
version = None
code = compile(f.read(), 'version.py', 'exec')
exec(code)
assert version
setup(
name='nativeconfig',
version=version,
packages=['nativeconfig'],
url='https://github.com/GreatFruitOmsk/nativeconfig',
license='MIT License',
author='Ilya Kulakov',
author_email='kulakov.ilya@gmail.com',
description="Cross-platform python module to store application config via native subsystems such as Windows Registry or NSUserDefaults.",
platforms=["Mac OS X 10.6+", "Windows XP+", "Linux 2.6+"],
keywords='config',
classifiers=[
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Operating System :: MacOS :: MacOS X',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
],
install_requires=REQUIREMENTS,
test_suite='test'
)
| import os
from setuptools import setup
from sys import platform
REQUIREMENTS = []
if platform.startswith('darwin'):
REQUIREMENTS.append('pyobjc-core >= 2.5')
with open(os.path.join(os.path.dirname(__file__), 'nativeconfig', 'version.py')) as f:
version = None
code = compile(f.read(), 'version.py', 'exec')
exec(code)
assert version
setup(
name='nativeconfig',
version=version,
packages=['nativeconfig', 'nativeconfig.options', 'nativeconfig.config'],
url='https://github.com/GreatFruitOmsk/nativeconfig',
license='MIT License',
author='Ilya Kulakov',
author_email='kulakov.ilya@gmail.com',
description="Cross-platform python module to store application config via native subsystems such as Windows Registry or NSUserDefaults.",
platforms=["Mac OS X 10.6+", "Windows XP+", "Linux 2.6+"],
keywords='config',
classifiers=[
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Operating System :: MacOS :: MacOS X',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
],
install_requires=REQUIREMENTS,
test_suite='test'
)
| Add subdirs of nativeconfig package to build. | Add subdirs of nativeconfig package to build.
| Python | mit | GreatFruitOmsk/nativeconfig | ---
+++
@@ -19,7 +19,7 @@
setup(
name='nativeconfig',
version=version,
- packages=['nativeconfig'],
+ packages=['nativeconfig', 'nativeconfig.options', 'nativeconfig.config'],
url='https://github.com/GreatFruitOmsk/nativeconfig',
license='MIT License',
author='Ilya Kulakov', |
ff921ad5b1d85dc9554dfcd9d94d96f9f80b0d2b | setup.py | setup.py | #!/usr/bin/env python3
# -*- coding: utf-8 -*-
# © 2017 qsuscs, TobiX
import os
import sys
from glob import glob
os.chdir(os.path.dirname(__file__))
exit = 0
for f in glob('dot.*'):
dst = os.path.expanduser('~/' + f[3:].replace(u'\u2571', '/'))
src = os.path.join(os.getcwd(), f)
src_rel = os.path.relpath(src, os.path.dirname(dst))
try:
os.makedirs(os.path.dirname(dst))
except OSError:
pass
try:
os.symlink(src_rel, dst)
except FileExistsError:
# Broken symbolic links do not "exist"
if not os.path.exists(dst) or not os.path.samefile(src, dst):
print(dst + " exists and does not link do " + src)
exit = 1
sys.exit(exit)
| #!/usr/bin/env python3
# -*- coding: utf-8 -*-
# © 2017 qsuscs, TobiX
from __future__ import print_function, unicode_literals
import os
import sys
from glob import glob
os.chdir(os.path.dirname(os.path.abspath(__file__)))
exit = 0
for f in glob('dot.*'):
dst = os.path.expanduser('~/' + f[3:].replace(u'\u2571', '/'))
src = os.path.join(os.getcwd(), f)
src_rel = os.path.relpath(src, os.path.dirname(dst))
try:
os.makedirs(os.path.dirname(dst))
except OSError:
pass
try:
os.symlink(src_rel, dst)
except OSError:
# Broken symbolic links do not "exist"
if not os.path.exists(dst) or not os.path.samefile(src, dst):
print(dst + " exists and does not link do " + src)
exit = 1
sys.exit(exit)
| Fix compatibility with Python 2 ;) | Fix compatibility with Python 2 ;)
| Python | isc | TobiX/dotfiles,TobiX/dotfiles | ---
+++
@@ -1,12 +1,14 @@
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
# © 2017 qsuscs, TobiX
+
+from __future__ import print_function, unicode_literals
import os
import sys
from glob import glob
-os.chdir(os.path.dirname(__file__))
+os.chdir(os.path.dirname(os.path.abspath(__file__)))
exit = 0
for f in glob('dot.*'):
@@ -21,7 +23,7 @@
try:
os.symlink(src_rel, dst)
- except FileExistsError:
+ except OSError:
# Broken symbolic links do not "exist"
if not os.path.exists(dst) or not os.path.samefile(src, dst):
print(dst + " exists and does not link do " + src) |
c638dbf619030c8d207e3bfd2e711da7c6c5cdf4 | passman.py | passman.py | #!/usr/bin/python3
# -*- coding: utf-8 -*-
from splash import showSplash
from functions import quit, getServiceFromUser, getPasswordFromUser, writeToFile
def main():
while True:
service = getServiceFromUser()
pw = getPasswordFromUser()
writeToFile(service, pw)
# run the program
showSplash()
main()
| #!/usr/bin/python3
# -*- coding: utf-8 -*-
import hashlib
from splash import showSplash
from functions import quit, getServiceFromUser, getPasswordFromUser, \
getUserInput, handleLogin, welcomeMessage, showMenu
from database import addUser, getAllServices, checkIfServiceExists, \
addService, removeService, updateService, getServiceByName
def main():
welcomeMessage()
handleLogin()
while True:
showMenu()
# run the program
#showSplash()
main()
| Clean up main a bit | Clean up main a bit
| Python | mit | regexpressyourself/passman | ---
+++
@@ -2,19 +2,27 @@
# -*- coding: utf-8 -*-
+import hashlib
+
from splash import showSplash
-from functions import quit, getServiceFromUser, getPasswordFromUser, writeToFile
+from functions import quit, getServiceFromUser, getPasswordFromUser, \
+ getUserInput, handleLogin, welcomeMessage, showMenu
+
+from database import addUser, getAllServices, checkIfServiceExists, \
+ addService, removeService, updateService, getServiceByName
def main():
+ welcomeMessage()
+ handleLogin()
+
+
while True:
- service = getServiceFromUser()
- pw = getPasswordFromUser()
- writeToFile(service, pw)
+ showMenu()
# run the program
-showSplash()
+#showSplash()
main()
|
8111060bae0818a44dc6669bf8ae011a1e612857 | hunter/reviewsapi.py | hunter/reviewsapi.py | import requests
import os
from .endpoints import *
class UnauthorizedToken(Exception):
pass
class ReviewsAPI:
def __init__(self):
token = os.environ['UDACITY_AUTH_TOKEN']
self.headers = {'Authorization': token, 'Content-Length': '0'}
def execute(self, request):
try:
raw_response = request()
response = raw_response.json()
raw_response.raise_for_status()
return response
except requests.exceptions.HTTPError:
raise UnauthorizedToken('Maybe it\'s time to change your token!')
def certifications(self):
return self.execute(lambda : requests.get(CERTIFICATIONS_URL, headers=self.headers))
def certified_languages(self):
response = self.execute(lambda : requests.get(REVIEWER_URL, headers=self.headers))
return [language for language in response['application']['languages']]
def request_reviews(self, projects):
return self.execute(lambda : requests.post(SUBMISSION_REQUESTS_URL, json=projects, headers=self.headers))
| import requests
import os
from .endpoints import *
class UnauthorizedToken(Exception):
pass
class ReviewsAPI:
def __init__(self):
token = os.environ['UDACITY_AUTH_TOKEN']
self.headers = {'Authorization': token, 'Content-Length': '0'}
def execute(self, request):
try:
raw_response = request()
response = raw_response.json()
raw_response.raise_for_status()
return response
except requests.exceptions.HTTPError:
raise UnauthorizedToken('Maybe it\'s time to change your token!')
def certifications(self):
return self.execute(lambda:requests.get(CERTIFICATIONS_URL, headers=self.headers))
def certified_languages(self):
response = self.execute(lambda:requests.get(REVIEWER_URL, headers=self.headers))
return [language for language in response['application']['languages']]
def request_reviews(self, projects):
return self.execute(lambda:requests.post(SUBMISSION_REQUESTS_URL, json=projects, headers=self.headers))
| Remove whitespaces from lambda expression | Remove whitespaces from lambda expression
| Python | mit | anapaulagomes/reviews-assigner | ---
+++
@@ -25,11 +25,11 @@
raise UnauthorizedToken('Maybe it\'s time to change your token!')
def certifications(self):
- return self.execute(lambda : requests.get(CERTIFICATIONS_URL, headers=self.headers))
+ return self.execute(lambda:requests.get(CERTIFICATIONS_URL, headers=self.headers))
def certified_languages(self):
- response = self.execute(lambda : requests.get(REVIEWER_URL, headers=self.headers))
+ response = self.execute(lambda:requests.get(REVIEWER_URL, headers=self.headers))
return [language for language in response['application']['languages']]
def request_reviews(self, projects):
- return self.execute(lambda : requests.post(SUBMISSION_REQUESTS_URL, json=projects, headers=self.headers))
+ return self.execute(lambda:requests.post(SUBMISSION_REQUESTS_URL, json=projects, headers=self.headers)) |
1dbc30202bddfd4f03bdc9a8005de3c363d2ac1d | blazar/plugins/dummy_vm_plugin.py | blazar/plugins/dummy_vm_plugin.py | # Copyright (c) 2013 Mirantis Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from blazar.plugins import base
class DummyVMPlugin(base.BasePlugin):
"""Plugin for VM resource that does nothing."""
resource_type = 'virtual:instance'
title = 'Dummy VM Plugin'
description = 'This plugin does nothing.'
def reserve_resource(self, reservation_id, values):
return None
def on_start(self, resource_id):
"""Dummy VM plugin does nothing."""
return 'VM %s should be waked up this moment.' % resource_id
def on_end(self, resource_id):
"""Dummy VM plugin does nothing."""
return 'VM %s should be deleted this moment.' % resource_id
| # Copyright (c) 2013 Mirantis Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from blazar.plugins import base
class DummyVMPlugin(base.BasePlugin):
"""Plugin for VM resource that does nothing."""
resource_type = 'virtual:instance'
title = 'Dummy VM Plugin'
description = 'This plugin does nothing.'
def reserve_resource(self, reservation_id, values):
return None
def update_reservation(self, reservation_id, values):
return None
def on_start(self, resource_id):
"""Dummy VM plugin does nothing."""
return 'VM %s should be waked up this moment.' % resource_id
def on_end(self, resource_id):
"""Dummy VM plugin does nothing."""
return 'VM %s should be deleted this moment.' % resource_id
| Add update_reservation to dummy plugin | Add update_reservation to dummy plugin
update_reservation is now an abstract method. It needs to be added to
all plugins.
Change-Id: I921878bd5233613b804b17813af1aac5bdfed9e7
| Python | apache-2.0 | ChameleonCloud/blazar,ChameleonCloud/blazar,openstack/blazar,stackforge/blazar,stackforge/blazar,openstack/blazar | ---
+++
@@ -25,6 +25,9 @@
def reserve_resource(self, reservation_id, values):
return None
+ def update_reservation(self, reservation_id, values):
+ return None
+
def on_start(self, resource_id):
"""Dummy VM plugin does nothing."""
return 'VM %s should be waked up this moment.' % resource_id |
b7627255c04e51ebb28f31365cff28ee9abcd05c | openspending/ui/test/functional/test_home.py | openspending/ui/test/functional/test_home.py | from .. import ControllerTestCase, url
class TestHomeController(ControllerTestCase):
def test_index(self):
response = self.app.get(url(controller='home', action='index'))
assert 'OpenSpending' in response
def test_locale(self):
response = self.app.get(url(controller='home', action='locale'))
| from .. import ControllerTestCase, url
class TestHomeController(ControllerTestCase):
def test_index(self):
response = self.app.get(url(controller='home', action='index'))
assert 'OpenSpending' in response
def test_locale(self):
response = self.app.get(url(controller='home', action='set_locale'))
| Fix test for locale route generation. | Fix test for locale route generation. | Python | agpl-3.0 | CivicVision/datahub,spendb/spendb,johnjohndoe/spendb,pudo/spendb,johnjohndoe/spendb,nathanhilbert/FPA_Core,openspending/spendb,CivicVision/datahub,spendb/spendb,johnjohndoe/spendb,USStateDept/FPA_Core,USStateDept/FPA_Core,openspending/spendb,pudo/spendb,CivicVision/datahub,nathanhilbert/FPA_Core,spendb/spendb,openspending/spendb,nathanhilbert/FPA_Core,pudo/spendb,USStateDept/FPA_Core | ---
+++
@@ -7,4 +7,4 @@
assert 'OpenSpending' in response
def test_locale(self):
- response = self.app.get(url(controller='home', action='locale'))
+ response = self.app.get(url(controller='home', action='set_locale')) |
16d77ff6a8d20773070630e1c7abb23c66345d72 | setup.py | setup.py | from setuptools import setup
setup(name='q', version='2.4', py_modules=['q'],
description='Quick-and-dirty debugging output for tired programmers',
author='Ka-Ping Yee', author_email='ping@zesty.ca',
license='Apache License 2.0',
url='http://github.com/zestyping/q', classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License'
])
| from setuptools import setup
setup(name='q', version='2.5', py_modules=['q'],
description='Quick-and-dirty debugging output for tired programmers',
author='Ka-Ping Yee', author_email='ping@zesty.ca',
license='Apache License 2.0',
url='http://github.com/zestyping/q', classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License'
])
| Advance PyPI version to 2.5. | Advance PyPI version to 2.5.
| Python | apache-2.0 | zestyping/q | ---
+++
@@ -1,5 +1,5 @@
from setuptools import setup
-setup(name='q', version='2.4', py_modules=['q'],
+setup(name='q', version='2.5', py_modules=['q'],
description='Quick-and-dirty debugging output for tired programmers',
author='Ka-Ping Yee', author_email='ping@zesty.ca',
license='Apache License 2.0', |
7fa490cb598aca2848ce886dfc45bb8606f07e58 | backend/geonature/core/gn_profiles/models.py | backend/geonature/core/gn_profiles/models.py | from geonature.utils.env import DB
from utils_flask_sqla.serializers import serializable
@serializable
class VmCorTaxonPhenology(DB.Model):
__tablename__ = "vm_cor_taxon_phenology"
__table_args__ = {"schema": "gn_profiles"}
cd_ref = DB.Column(DB.Integer)
period = DB.Column(DB.Integer)
id_nomenclature_life_stage = DB.Column(DB.Integer)
id_altitude_range = DB.Column(DB.Integer)
count_valid_data = DB.Column(DB.Integer)
| from flask import current_app
from geoalchemy2 import Geometry
from utils_flask_sqla.serializers import serializable
from utils_flask_sqla_geo.serializers import geoserializable
from geonature.utils.env import DB
@serializable
class VmCorTaxonPhenology(DB.Model):
__tablename__ = "vm_cor_taxon_phenology"
__table_args__ = {"schema": "gn_profiles"}
cd_ref = DB.Column(DB.Integer)
period = DB.Column(DB.Integer)
id_nomenclature_life_stage = DB.Column(DB.Integer)
id_altitude_range = DB.Column(DB.Integer)
count_valid_data = DB.Column(DB.Integer)
@serializable
@geoserializable
class VmValidProfiles(DB.Model):
__tablename__ = "vm_valid_profiles"
__table_args__ = {"schema": "gn_profiles"}
cd_ref = DB.Column(DB.Integer)
valid_distribution = DB.Column(Geometry("GEOMETRY", current_app.config["LOCAL_SRID"]))
altitude_min = DB.Column(DB.Integer)
altitude_max = DB.Column(DB.Integer)
first_valid_data = DB.Column(DB.DateTime)
last_valid_data = DB.Column(DB.DateTime)
count_valid_data = DB.Column(DB.Integer) | Add VM valid profile model | Add VM valid profile model
| Python | bsd-2-clause | PnEcrins/GeoNature,PnEcrins/GeoNature,PnEcrins/GeoNature,PnEcrins/GeoNature | ---
+++
@@ -1,5 +1,10 @@
+from flask import current_app
+from geoalchemy2 import Geometry
+
+from utils_flask_sqla.serializers import serializable
+from utils_flask_sqla_geo.serializers import geoserializable
+
from geonature.utils.env import DB
-from utils_flask_sqla.serializers import serializable
@serializable
class VmCorTaxonPhenology(DB.Model):
@@ -10,3 +15,17 @@
id_nomenclature_life_stage = DB.Column(DB.Integer)
id_altitude_range = DB.Column(DB.Integer)
count_valid_data = DB.Column(DB.Integer)
+
+
+@serializable
+@geoserializable
+class VmValidProfiles(DB.Model):
+ __tablename__ = "vm_valid_profiles"
+ __table_args__ = {"schema": "gn_profiles"}
+ cd_ref = DB.Column(DB.Integer)
+ valid_distribution = DB.Column(Geometry("GEOMETRY", current_app.config["LOCAL_SRID"]))
+ altitude_min = DB.Column(DB.Integer)
+ altitude_max = DB.Column(DB.Integer)
+ first_valid_data = DB.Column(DB.DateTime)
+ last_valid_data = DB.Column(DB.DateTime)
+ count_valid_data = DB.Column(DB.Integer) |
52864307e692f40ddfe170a4b8607c4b2b96bff5 | setup.py | setup.py | # -*- coding: utf-8 -*-
import os
from setuptools import setup
def read(fname):
try:
return open(os.path.join(os.path.dirname(__file__), fname)).read()
except:
return ''
setup(
name='todoist-python',
version='0.2.26',
packages=['todoist', 'todoist.managers'],
author='Doist Team',
author_email='info@todoist.com',
license='BSD',
description='todoist-python - The official Todoist Python API library',
long_description = read('README.md'),
install_requires=[
'requests',
],
# see here for complete list of classifiers
# http://pypi.python.org/pypi?%3Aaction=list_classifiers
classifiers=(
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
),
)
| # -*- coding: utf-8 -*-
import os
from setuptools import setup
def read(fname):
try:
return open(os.path.join(os.path.dirname(__file__), fname)).read()
except:
return ''
setup(
name='todoist-python',
version='7.0',
packages=['todoist', 'todoist.managers'],
author='Doist Team',
author_email='info@todoist.com',
license='BSD',
description='todoist-python - The official Todoist Python API library',
long_description = read('README.md'),
install_requires=[
'requests',
],
# see here for complete list of classifiers
# http://pypi.python.org/pypi?%3Aaction=list_classifiers
classifiers=(
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
),
)
| Update the PyPI version to 7.0. | Update the PyPI version to 7.0.
| Python | mit | Doist/todoist-python | ---
+++
@@ -10,7 +10,7 @@
setup(
name='todoist-python',
- version='0.2.26',
+ version='7.0',
packages=['todoist', 'todoist.managers'],
author='Doist Team',
author_email='info@todoist.com', |
823179066c9a7c101a18e9c2d7ccb2a9ccf7f1cb | setup.py | setup.py | import os
from setuptools import setup, find_packages
from hamcrest import __version__
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'PyHamcrest',
version = __version__,
author = 'Jon Reid',
author_email = 'jon.reid@mac.com',
description = 'Hamcrest framework for matcher objects',
license = 'New BSD',
platforms=['All'],
keywords = 'hamcrest matchers pyunit unit test testing unittest unittesting',
url = 'http://code.google.com/p/hamcrest/',
download_url = 'http://pypi.python.org/packages/source/P/PyHamcrest/PyHamcrest-%s.tar.gz' % __version__,
packages = find_packages(),
test_suite = 'hamcrest-unit-test.alltests',
provides = ['hamcrest'],
long_description=read('README.md'),
classifiers = [
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2.6',
'Topic :: Software Development',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
| import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
import re
matched = re.match('__version__.*', read(os.path.join('hamcrest', '__init__.py')))
if matched:
exec(matched.group())
setup(
name = 'PyHamcrest',
version = __version__,
author = 'Jon Reid',
author_email = 'jon.reid@mac.com',
description = 'Hamcrest framework for matcher objects',
license = 'New BSD',
platforms=['All'],
keywords = 'hamcrest matchers pyunit unit test testing unittest unittesting',
url = 'http://code.google.com/p/hamcrest/',
download_url = 'http://pypi.python.org/packages/source/P/PyHamcrest/PyHamcrest-%s.tar.gz' % __version__,
packages = find_packages(),
test_suite = 'hamcrest-unit-test.alltests',
provides = ['hamcrest'],
long_description=read('README.md'),
classifiers = [
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2.6',
'Topic :: Software Development',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
use_2to3 = True,
)
| Support python 3 installtion; need distribute | Support python 3 installtion; need distribute
| Python | bsd-3-clause | msabramo/PyHamcrest,nitishr/PyHamcrest,nitishr/PyHamcrest,msabramo/PyHamcrest | ---
+++
@@ -1,9 +1,13 @@
import os
from setuptools import setup, find_packages
-from hamcrest import __version__
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
+
+import re
+matched = re.match('__version__.*', read(os.path.join('hamcrest', '__init__.py')))
+if matched:
+ exec(matched.group())
setup(
name = 'PyHamcrest',
@@ -32,4 +36,5 @@
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
+ use_2to3 = True,
) |
c5b73be1bf0f0edd05c4743c2449bee568d01c76 | setup.py | setup.py | from distutils.core import setup
from turbasen import VERSION
name = 'turbasen'
setup(
name=name,
packages=[name],
version=VERSION,
description='Client for Nasjonal Turbase REST API',
author='Ali Kaafarani',
author_email='ali.kaafarani@turistforeningen.no',
url='https://github.com/Turbasen/turbasen.py',
download_url='https://github.com/Turbasen/turbasen.py/tarball/v%s' % (VERSION),
keywords=['turbasen', 'nasjonalturbase', 'turistforening', 'rest-api'],
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: Norwegian',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
install_requires=['requests'],
)
| from distutils.core import setup
from os import path
from turbasen import VERSION
name = 'turbasen'
here = path.abspath(path.dirname(__file__))
with open(path.join(here, 'README.md'), encoding='utf-8') as f:
long_description = f.read()
setup(
name=name,
packages=[name],
version=VERSION,
description='Client for Nasjonal Turbase REST API',
long_description=long_description,
author='Ali Kaafarani',
author_email='ali.kaafarani@turistforeningen.no',
url='https://github.com/Turbasen/turbasen.py',
download_url='https://github.com/Turbasen/turbasen.py/tarball/v%s' % (VERSION),
keywords=['turbasen', 'nasjonalturbase', 'turistforening', 'rest-api'],
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: Norwegian',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
install_requires=['requests'],
)
| Add long description from README | Add long description from README
| Python | mit | Turbasen/turbasen.py | ---
+++
@@ -1,14 +1,20 @@
from distutils.core import setup
+from os import path
from turbasen import VERSION
name = 'turbasen'
+
+here = path.abspath(path.dirname(__file__))
+with open(path.join(here, 'README.md'), encoding='utf-8') as f:
+ long_description = f.read()
setup(
name=name,
packages=[name],
version=VERSION,
description='Client for Nasjonal Turbase REST API',
+ long_description=long_description,
author='Ali Kaafarani',
author_email='ali.kaafarani@turistforeningen.no',
url='https://github.com/Turbasen/turbasen.py', |
7a80fd081c6d8ece2b199f4a7915dc59d1805437 | setup.py | setup.py | #!/usr/bin/env python
from os import path
from setuptools import setup
with open(path.join(path.abspath(path.dirname(__file__)), 'README.rst')) as f:
long_description = f.read()
setup(name='django_uncertainty',
version='1.5',
description='A Django middleware to generate predictable errors on sites',
long_description=long_description,
author='Agustin Barto',
author_email='abarto@gmail.com',
url='https://github.com/abarto/django_uncertainty',
license='BSD',
install_requires=[],
tests_require=['Django>=1.10'],
test_suite='uncertainty.tests.runtests.runtests',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
'Topic :: Utilities',
],
packages=['uncertainty'])
| #!/usr/bin/env python
from os import path
from setuptools import setup
with open(path.join(path.abspath(path.dirname(__file__)), 'README.rst')) as f:
long_description = f.read()
setup(name='django_uncertainty',
version='1.5',
description='A Django middleware to generate predictable errors on sites',
long_description=long_description,
author='Agustin Barto',
author_email='abarto@gmail.com',
url='https://github.com/abarto/django_uncertainty',
license='BSD',
install_requires=[],
tests_require=['Django>=1.10'],
test_suite='uncertainty.tests.runtests.runtests',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Utilities',
],
packages=['uncertainty'])
| Remove support for Python 2 | Remove support for Python 2
| Python | bsd-3-clause | abarto/django_uncertainty | ---
+++
@@ -27,7 +27,6 @@
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
- 'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
'Topic :: Utilities',
], |
89dfdc7ebe2fe483d2b306dac83a666aa7c013d7 | setup.py | setup.py | #!/usr/bin/env python
from os.path import join
from setuptools import setup, find_packages
name = 'geokey-sapelli'
version = __import__(name.replace('-', '_')).__version__
repository = join('https://github.com/ExCiteS', name)
setup(
name=name,
version=version,
description='Read Sapelli project and load data from CSVs to GeoKey',
url=repository,
download_url=join(repository, 'tarball', version),
author='ExCiteS',
author_email='excites@ucl.ac.uk',
packages=find_packages(exclude=['*.tests', '*.tests.*', 'tests.*']),
include_package_data=True,
install_requires=['qrcode==5.2.2'],
)
| #!/usr/bin/env python
from os.path import dirname, join
from setuptools import setup, find_packages
def read(file_name):
with open(join(dirname(__file__), file_name)) as file_object:
return file_object.read()
name = 'geokey-sapelli'
version = __import__(name.replace('-', '_')).__version__
repository = join('https://github.com/ExCiteS', name)
setup(
name=name,
version=version,
description='Read Sapelli project and load data from CSVs to GeoKey',
long_description=read('README.rst'),
url=repository,
download_url=join(repository, 'tarball', version),
author='ExCiteS',
author_email='excites@ucl.ac.uk',
packages=find_packages(exclude=['*.tests', '*.tests.*', 'tests.*']),
include_package_data=True,
install_requires=['qrcode==5.2.2'],
)
| Include README for PyPI release | Include README for PyPI release
| Python | mit | ExCiteS/geokey-sapelli,ExCiteS/geokey-sapelli | ---
+++
@@ -1,7 +1,12 @@
#!/usr/bin/env python
-from os.path import join
+from os.path import dirname, join
from setuptools import setup, find_packages
+
+
+def read(file_name):
+ with open(join(dirname(__file__), file_name)) as file_object:
+ return file_object.read()
name = 'geokey-sapelli'
@@ -12,6 +17,7 @@
name=name,
version=version,
description='Read Sapelli project and load data from CSVs to GeoKey',
+ long_description=read('README.rst'),
url=repository,
download_url=join(repository, 'tarball', version),
author='ExCiteS', |
2c7924b879be6536f0a6f0f7f78e5813156734af | setup.py | setup.py | from setuptools import setup, find_packages
setup(
name='neomodel',
version='0.4.0',
description='An object mapper for the neo4j graph database.',
long_description=open('README.rst').read(),
author='Robin Edwards',
author_email='robin.ge@gmail.com',
zip_safe=True,
url='http://github.com/robinedwards/neomodel',
license='MIT',
packages=find_packages(),
keywords='graph neo4j py2neo ORM',
tests_require=['nose==1.1.2'],
test_suite='nose.collector',
install_requires=['py2neo==1.6.1', 'pytz==2013.8', 'lucene-querybuilder==0.2'],
classifiers=[
"Development Status :: 5 - Production/Stable",
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Topic :: Software Development :: Libraries :: Python Modules',
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
])
| from setuptools import setup, find_packages
setup(
name='neomodel',
version='1.0.0',
description='An object mapper for the neo4j graph database.',
long_description=open('README.rst').read(),
author='Robin Edwards',
author_email='robin.ge@gmail.com',
zip_safe=True,
url='http://github.com/robinedwards/neomodel',
license='MIT',
packages=find_packages(),
keywords='graph neo4j py2neo ORM',
tests_require=['nose==1.1.2'],
test_suite='nose.collector',
install_requires=['py2neo==1.6.1', 'pytz==2013.8'],
classifiers=[
"Development Status :: 5 - Production/Stable",
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Topic :: Software Development :: Libraries :: Python Modules',
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
])
| Remove lucene query builder and bumper version | Remove lucene query builder and bumper version
| Python | mit | bleib1dj/neomodel,fpieper/neomodel,robinedwards/neomodel,cristigociu/neomodel_dh,bleib1dj/neomodel,pombredanne/neomodel,robinedwards/neomodel,andrefsp/neomodel,wcooley/neomodel | ---
+++
@@ -2,7 +2,7 @@
setup(
name='neomodel',
- version='0.4.0',
+ version='1.0.0',
description='An object mapper for the neo4j graph database.',
long_description=open('README.rst').read(),
author='Robin Edwards',
@@ -14,7 +14,7 @@
keywords='graph neo4j py2neo ORM',
tests_require=['nose==1.1.2'],
test_suite='nose.collector',
- install_requires=['py2neo==1.6.1', 'pytz==2013.8', 'lucene-querybuilder==0.2'],
+ install_requires=['py2neo==1.6.1', 'pytz==2013.8'],
classifiers=[
"Development Status :: 5 - Production/Stable",
'Intended Audience :: Developers', |
69281da6f69bbdc5cfb832efa0b0c1b7810eb262 | setup.py | setup.py | # -*- coding: utf-8 -*-
from distutils.core import setup
readme_file = open('README.rst')
setup(
name='django-db-file-storage',
version='0.3.1',
author='Victor Oliveira da Silva',
author_email='victor_o_silva@hotmail.com',
packages=['db_file_storage'],
url='https://github.com/victor-o-silva/db_file_storage',
download_url='https://github.com/victor-o-silva/db_file_storage'
'/tarball/0.3.1',
license='GNU GPL v3',
description="Custom FILE_STORAGE for Django. Saves files "
"in your database instead of your file system.",
long_description=readme_file.read(),
install_requires=[
"Django",
],
)
readme_file.close()
| # -*- coding: utf-8 -*-
from distutils.core import setup
readme_file = open('README.rst')
setup(
name='django-db-file-storage',
version='0.3.1',
author='Victor Oliveira da Silva',
author_email='victor_o_silva@hotmail.com',
packages=['db_file_storage'],
url='https://github.com/victor-o-silva/db_file_storage',
download_url='https://github.com/victor-o-silva/db_file_storage'
'/tarball/0.3.1',
license='GNU GPL v3',
description="Custom FILE_STORAGE for Django. Saves files "
"in your database instead of your file system.",
long_description=readme_file.read(),
install_requires=[
"Django",
],
classifiers=[
'Framework :: Django',
'Intended Audience :: Developers',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
readme_file.close()
| Mark as compatible for python 2.7, 3.3 and 3.4 | Mark as compatible for python 2.7, 3.3 and 3.4
Add `classifiers` parameter to `setup` function call in `setup.py` file. | Python | mit | victor-o-silva/db_file_storage,victor-o-silva/db_file_storage | ---
+++
@@ -20,6 +20,16 @@
install_requires=[
"Django",
],
+ classifiers=[
+ 'Framework :: Django',
+ 'Intended Audience :: Developers',
+ 'Programming Language :: Python',
+ 'Programming Language :: Python :: 2',
+ 'Programming Language :: Python :: 2.7',
+ 'Programming Language :: Python :: 3',
+ 'Programming Language :: Python :: 3.3',
+ 'Programming Language :: Python :: 3.4',
+ ],
)
readme_file.close() |
b04ef68d079bf4ef172c6fa7f84946369b080cb7 | setup.py | setup.py | """setup.py"""
from codecs import open as codecs_open
from setuptools import setup
with codecs_open('README.rst', 'r', 'utf-8') as f:
README = f.read()
with codecs_open('HISTORY.rst', 'r', 'utf-8') as f:
HISTORY = f.read()
setup(
name='jsonrpcserver',
version='3.4.0',
description='Process JSON-RPC requests',
long_description=README + '\n\n' + HISTORY,
author='Beau Barker',
author_email='beauinmelbourne@gmail.com',
url='https://jsonrpcserver.readthedocs.io/',
license='MIT',
packages=['jsonrpcserver'],
package_data={'jsonrpcserver': ['request-schema.json']},
include_package_data=True,
install_requires=['jsonschema', 'six', 'funcsigs'],
extras_require={
'tox': ['tox','pylint'],
'examples': [
'aiohttp',
'flask',
'flask-socketio',
'tornado',
'werkzeug',
'pyzmq'
]
},
classifiers=[
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6'
],
)
| """setup.py"""
from codecs import open as codecs_open
from setuptools import setup
with codecs_open('README.rst', 'r', 'utf-8') as f:
README = f.read()
with codecs_open('HISTORY.rst', 'r', 'utf-8') as f:
HISTORY = f.read()
setup(
name='jsonrpcserver',
version='3.4.0',
description='Process JSON-RPC requests',
long_description=README + '\n\n' + HISTORY,
author='Beau Barker',
author_email='beauinmelbourne@gmail.com',
url='https://jsonrpcserver.readthedocs.io/',
license='MIT',
packages=['jsonrpcserver'],
package_data={'jsonrpcserver': ['request-schema.json']},
include_package_data=True,
install_requires=['jsonschema', 'six', 'funcsigs'],
extras_require={
'tox': ['tox','pylint'],
'examples': [
'aiohttp',
'aiozmq',
'flask',
'flask-socketio',
'pyzmq',
'tornado',
'websockets',
'werkzeug',
]
},
classifiers=[
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6'
],
)
| Add to extras_require examples section | Add to extras_require examples section
| Python | mit | bcb/jsonrpcserver | ---
+++
@@ -26,11 +26,13 @@
'tox': ['tox','pylint'],
'examples': [
'aiohttp',
+ 'aiozmq',
'flask',
'flask-socketio',
+ 'pyzmq',
'tornado',
+ 'websockets',
'werkzeug',
- 'pyzmq'
]
},
classifiers=[ |
ad889be15a374ff07492b549ba454bafd3a76fd7 | setup.py | setup.py | import os
from setuptools import find_packages, setup
with open(os.path.join(os.path.dirname(__file__), 'README.rst'), encoding='utf-8') as readme:
README = readme.read()
setup(
name='django-postgres-extra',
version='1.21a2',
packages=find_packages(),
include_package_data=True,
license='MIT License',
description='Bringing all of PostgreSQL\'s awesomeness to Django.',
long_description=README,
url='https://github.com/SectorLabs/django-postgres-extra',
author='Sector Labs',
author_email='open-source@sectorlabs.ro',
keywords=['django', 'postgres', 'extra', 'hstore', 'ltree'],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3.5',
'Topic :: Internet :: WWW/HTTP',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
]
)
| import os
from setuptools import find_packages, setup
with open(os.path.join(os.path.dirname(__file__), 'README.rst'), encoding='utf-8') as readme:
README = readme.read().split('h1>', 2)[1]
setup(
name='django-postgres-extra',
version='1.21a2',
packages=find_packages(),
include_package_data=True,
license='MIT License',
description='Bringing all of PostgreSQL\'s awesomeness to Django.',
long_description=README,
url='https://github.com/SectorLabs/django-postgres-extra',
author='Sector Labs',
author_email='open-source@sectorlabs.ro',
keywords=['django', 'postgres', 'extra', 'hstore', 'ltree'],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3.5',
'Topic :: Internet :: WWW/HTTP',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
]
)
| Cut out logo in README uploaded to PyPi | Cut out logo in README uploaded to PyPi
| Python | mit | SectorLabs/django-postgres-extra | ---
+++
@@ -3,7 +3,7 @@
from setuptools import find_packages, setup
with open(os.path.join(os.path.dirname(__file__), 'README.rst'), encoding='utf-8') as readme:
- README = readme.read()
+ README = readme.read().split('h1>', 2)[1]
setup(
name='django-postgres-extra', |
3ab6dbf87053634ce48627787151a24b33b546af | setup.py | setup.py | # -*- coding: utf-8 -*-
import os
from setuptools import setup
def read(fname):
try:
return open(os.path.join(os.path.dirname(__file__), fname)).read()
except:
return ''
setup(
name='todoist-python',
version='0.2.5',
packages=['todoist', 'todoist.managers'],
author='Doist Team',
author_email='info@todoist.com',
license='BSD',
description='todoist-python - The official Todoist Python API library',
long_description = read('README.md'),
install_requires=[
'requests',
],
# see here for complete list of classifiers
# http://pypi.python.org/pypi?%3Aaction=list_classifiers
classifiers=(
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
),
)
| # -*- coding: utf-8 -*-
import os
from setuptools import setup
def read(fname):
try:
return open(os.path.join(os.path.dirname(__file__), fname)).read()
except:
return ''
setup(
name='todoist-python',
version='0.2.6',
packages=['todoist', 'todoist.managers'],
author='Doist Team',
author_email='info@todoist.com',
license='BSD',
description='todoist-python - The official Todoist Python API library',
long_description = read('README.md'),
install_requires=[
'requests',
],
# see here for complete list of classifiers
# http://pypi.python.org/pypi?%3Aaction=list_classifiers
classifiers=(
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
),
)
| Update the PyPI version to 0.2.6. | Update the PyPI version to 0.2.6.
| Python | mit | Doist/todoist-python,electronick1/todoist-python | ---
+++
@@ -10,7 +10,7 @@
setup(
name='todoist-python',
- version='0.2.5',
+ version='0.2.6',
packages=['todoist', 'todoist.managers'],
author='Doist Team',
author_email='info@todoist.com', |
b98b804ea83723d476f50a46e9ba4dbe8fc6e4a4 | setup.py | setup.py | from setuptools import find_packages
from setuptools import setup
setup(
name='kms-encryption-toolbox',
version='0.0.5',
url='https://github.com/ApplauseOSS/kms-encryption-toolbox',
license='Applause',
description='Encryption toolbox to be used with the Amazon Key Management Service for securing your deployment secrets. It encapsulates the aws-encryption-sdk package to expose cmdline actions.',
author='Applause',
author_email='architecture@applause.com',
zip_safe=False,
packages=find_packages(),
install_requires=[
'aws-encryption-sdk>=1.2.0',
'click>=6.6',
'attrs>=16.3.0,<17.0.0'
],
entry_points={
"console_scripts": [
"kms-encryption = kmsencryption.__main__:main",
]
},
scripts=["kmsencryption/scripts/decrypt-and-start.sh"]
)
| from setuptools import find_packages
from setuptools import setup
setup(
name='kms-encryption-toolbox',
version='0.0.6',
url='https://github.com/ApplauseOSS/kms-encryption-toolbox',
license='Applause',
description='Encryption toolbox to be used with the Amazon Key Management Service for securing your deployment secrets. It encapsulates the aws-encryption-sdk package to expose cmdline actions.',
author='Applause',
author_email='architecture@applause.com',
zip_safe=False,
packages=find_packages(),
install_requires=[
'cffi>=1.10.0',
'aws-encryption-sdk>=1.2.0',
'click>=6.6',
'attrs>=16.3.0,<17.0.0'
],
entry_points={
"console_scripts": [
"kms-encryption = kmsencryption.__main__:main",
]
},
scripts=["kmsencryption/scripts/decrypt-and-start.sh"]
)
| Add cffi to the package requirements. | Add cffi to the package requirements.
| Python | mit | ApplauseOSS/kms-encryption-toolbox | ---
+++
@@ -3,7 +3,7 @@
setup(
name='kms-encryption-toolbox',
- version='0.0.5',
+ version='0.0.6',
url='https://github.com/ApplauseOSS/kms-encryption-toolbox',
license='Applause',
description='Encryption toolbox to be used with the Amazon Key Management Service for securing your deployment secrets. It encapsulates the aws-encryption-sdk package to expose cmdline actions.',
@@ -12,6 +12,7 @@
zip_safe=False,
packages=find_packages(),
install_requires=[
+ 'cffi>=1.10.0',
'aws-encryption-sdk>=1.2.0',
'click>=6.6',
'attrs>=16.3.0,<17.0.0' |
548d5b46607d1023e7b0a95c6a2995419fa06b50 | fickle/api.py | fickle/api.py | import flask
from flask import request, json
def Response(data, status = 200):
body = json.dumps(data)
return flask.Response(body, status = status, mimetype = 'application/json')
def SuccessResponse(dataset_id = None):
return Response({ 'success': True, 'id': dataset_id })
def ErrorResponse(status = 400):
return Response({ 'success': False }, status = status)
def API(name, backend):
app = flask.Flask(name)
app.config.from_object(name)
@app.route('/')
def api_root():
return SuccessResponse(backend.dataset_id)
@app.route('/load', methods=['POST'])
def api_load():
backend.load(request.json)
return SuccessResponse(backend.dataset_id)
@app.route('/fit', methods=['POST'])
def api_fit():
if not backend.loaded():
return ErrorResponse()
backend.fit()
return SuccessResponse(backend.dataset_id)
@app.route('/validate', methods=['PUT'])
def api_validate():
if not backend.loaded():
return ErrorResponse()
data = backend.validate()
return Response(data)
@app.route('/predict', methods=['POST'])
def api_predict():
if not backend.trained():
return ErrorResponse()
data = backend.predict(request.json).tolist()
return Response(data)
return app
| import flask
from flask import request, json
def Response(data, status = 200):
body = json.dumps(data)
return flask.Response(body, status = status, mimetype = 'application/json')
def SuccessResponse(dataset_id = None):
return Response({ 'success': True, 'id': dataset_id })
def ErrorResponse(status = 400):
return Response({ 'success': False }, status = status)
def API(name, backend):
app = flask.Flask(name)
app.config.from_object(name)
@app.route('/')
def api_root():
return SuccessResponse(backend.dataset_id)
@app.route('/load', methods=['POST'])
def api_load():
backend.load(request.json)
return SuccessResponse(backend.dataset_id)
@app.route('/fit', methods=['POST'])
def api_fit():
if not backend.loaded():
return ErrorResponse()
backend.fit()
return SuccessResponse(backend.dataset_id)
@app.route('/validate', methods=['POST'])
def api_validate():
if not backend.loaded():
return ErrorResponse()
data = backend.validate()
return Response(data)
@app.route('/predict', methods=['POST'])
def api_predict():
if not backend.trained():
return ErrorResponse()
data = backend.predict(request.json).tolist()
return Response(data)
return app
| Change HTTP method for validate endpoint | Change HTTP method for validate endpoint
| Python | mit | norbert/fickle | ---
+++
@@ -31,7 +31,7 @@
backend.fit()
return SuccessResponse(backend.dataset_id)
- @app.route('/validate', methods=['PUT'])
+ @app.route('/validate', methods=['POST'])
def api_validate():
if not backend.loaded():
return ErrorResponse() |
6007c5800ea1f59ac26417b5081323c0b40446ad | almanacbot/almanac-bot.py | almanacbot/almanac-bot.py | import json
import logging
import logging.config
import os
import sys
import config
import constants
def setup_logging(
path='logging.json',
log_level=logging.DEBUG,
env_key='LOG_CFG'
):
env_path = os.getenv(env_key, None)
if env_path:
path = env_path
if os.path.exists(path):
with open(path, 'rt') as f:
log_conf = json.load(f)
logging.config.dictConfig(log_conf)
else:
logging.basicConfig(level=log_level)
if __name__ == '__main__':
# configure logger
setup_logging()
# read configuration
try:
configuration = config.Configuration(constants.CONFIG_FILE_NAME)
except Exception as exc:
logging.error("Error getting configuration.", exc)
sys.exit(1)
| import json
import logging
import logging.config
import os
import sys
import config
import constants
import twitter
conf = None
twitter_api = None
def setup_logging(
path='logging.json',
log_level=logging.DEBUG,
env_key='LOG_CFG'
):
env_path = os.getenv(env_key, None)
if env_path:
path = env_path
if os.path.exists(path):
with open(path, 'rt') as f:
log_conf = json.load(f)
logging.config.dictConfig(log_conf)
else:
logging.basicConfig(level=log_level)
def setup_twitter():
logging.info("Setting up Twitter API client...")
twitter_api = twitter.Api(
consumer_key=conf.config["twitter"]["consumer_key"],
consumer_secret=conf.config["twitter"]["consumer_secret"],
access_token_key=conf.config["twitter"]["access_token_key"],
access_token_secret=conf.config["twitter"]["access_token_secret"])
logging.info("Verifying Twitter API client credentials...")
twitter_api.VerifyCredentials()
logging.info("Twitter API client credentials verified.")
logging.info("Twitter API client set up.")
if __name__ == '__main__':
# configure logger
setup_logging()
# read configuration
try:
conf = config.Configuration(constants.CONFIG_FILE_NAME)
except Exception as exc:
logging.error("Error getting configuration.", exc)
sys.exit(1)
try:
setup_twitter()
except Exception as exc:
logging.error("Error setting up Twitter API client.", exc)
sys.exit(1)
| Add Twitter API client initialization. | Add Twitter API client initialization.
| Python | mit | logoff/almanac-bot,logoff/almanac-bot | ---
+++
@@ -6,6 +6,10 @@
import config
import constants
+import twitter
+
+conf = None
+twitter_api = None
def setup_logging(
@@ -24,13 +28,34 @@
logging.basicConfig(level=log_level)
+def setup_twitter():
+ logging.info("Setting up Twitter API client...")
+
+ twitter_api = twitter.Api(
+ consumer_key=conf.config["twitter"]["consumer_key"],
+ consumer_secret=conf.config["twitter"]["consumer_secret"],
+ access_token_key=conf.config["twitter"]["access_token_key"],
+ access_token_secret=conf.config["twitter"]["access_token_secret"])
+
+ logging.info("Verifying Twitter API client credentials...")
+ twitter_api.VerifyCredentials()
+ logging.info("Twitter API client credentials verified.")
+
+ logging.info("Twitter API client set up.")
+
if __name__ == '__main__':
# configure logger
setup_logging()
# read configuration
try:
- configuration = config.Configuration(constants.CONFIG_FILE_NAME)
+ conf = config.Configuration(constants.CONFIG_FILE_NAME)
except Exception as exc:
logging.error("Error getting configuration.", exc)
sys.exit(1)
+
+ try:
+ setup_twitter()
+ except Exception as exc:
+ logging.error("Error setting up Twitter API client.", exc)
+ sys.exit(1) |
ae966a3cb7f99e5604c8302680f125e12087003a | blackhole/state.py | blackhole/state.py |
class MailState():
_reading_data = False
def set_reading(self, val):
self._reading_data = val
def get_reading(self):
return self._reading_data
| class MailState():
_reading_data = False
def set_reading(self, val):
self._reading_data = val
@property
def reading(self):
return self._reading_data
| Change getter in to property | Change getter in to property | Python | mit | kura/blackhole,kura/blackhole | ---
+++
@@ -1,9 +1,9 @@
-
class MailState():
_reading_data = False
def set_reading(self, val):
self._reading_data = val
- def get_reading(self):
+ @property
+ def reading(self):
return self._reading_data |
e50fc12459e6ff77864fe499b512a57e89f7ead2 | pi_control_service/gpio_service.py | pi_control_service/gpio_service.py | from rpc import RPCService
from pi_pin_manager import PinManager
ALLOWED_ACTIONS = ('on', 'off', 'read')
class GPIOService(RPCService):
def __init__(self, rabbit_url, device_key, pin_config):
self.pins = PinManager(config_file=pin_config)
super(GPIOService, self).__init__(
rabbit_url=rabbit_url,
queue_name='gpio_service',
device_key=device_key,
request_action=self._perform_gpio_action)
def _perform_gpio_action(self, instruction):
result = {'error': 1, 'pin': instruction['pin'], 'response': "An error occurred"}
if instruction['action'] not in ALLOWED_ACTIONS:
result['response'] = "'action' must be one of: {0}".format(', '.join(ALLOWED_ACTIONS))
return result
try:
result['response'] = getattr(self.pins, instruction['action'])(int(instruction['pin']))
result['error'] = 0
except ValueError:
result['response'] = "'pin' value must be an integer"
except:
pass
return result
def stop(self):
self.pins.cleanup()
super(GPIOService, self).stop()
| from rpc import RPCService
from pi_pin_manager import PinManager
ALLOWED_ACTIONS = ('on', 'off', 'read')
class GPIOService(RPCService):
def __init__(self, rabbit_url, device_key, pin_config):
self.pins = PinManager(config_file=pin_config)
super(GPIOService, self).__init__(
rabbit_url=rabbit_url,
queue_name='gpio_service',
device_key=device_key,
request_action=self._perform_gpio_action)
def _perform_gpio_action(self, instruction):
result = {'error': 1, 'pin': instruction['pin'], 'response': "An error occurred"}
if instruction['action'] not in ALLOWED_ACTIONS:
result['response'] = "'action' must be one of: {0}".format(', '.join(ALLOWED_ACTIONS))
return result
try:
result['response'] = getattr(self.pins, instruction['action'])(int(instruction['pin']))
result['error'] = 0
except ValueError:
result['response'] = "'pin' value must be an integer"
except Exception as e:
result['response'] = e.message
return result
def stop(self):
self.pins.cleanup()
super(GPIOService, self).stop()
| Send exception message in response | Send exception message in response
| Python | mit | projectweekend/Pi-Control-Service,HydAu/ProjectWeekds_Pi-Control-Service | ---
+++
@@ -27,8 +27,8 @@
result['error'] = 0
except ValueError:
result['response'] = "'pin' value must be an integer"
- except:
- pass
+ except Exception as e:
+ result['response'] = e.message
return result
def stop(self): |
fe4fec66cbf4100752c4b7414090019ab8ddb8ce | ideascube/conf/idb_bdi.py | ideascube/conf/idb_bdi.py | """Generic config for Ideasbox of Burundi"""
from .idb import * # noqa
from django.utils.translation import ugettext_lazy as _
USER_FORM_FIELDS = (
('Ideasbox', ['serial', 'box_awareness']),
(_('Personal informations'), ['refugee_id', 'short_name', 'full_name', 'birth_year', 'gender', 'phone']), # noqa
(_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa
(_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa
(_('Origin'), ['country', 'city', 'country_of_origin_occupation', 'school_level', 'is_sent_to_school']), # noqa
(_('Language skills'), ['rn_level', 'sw_level', 'fr_level']),
(_('National residents'), ['id_card_number']),
)
| """Generic config for Ideasbox of Burundi"""
from .idb import * # noqa
from django.utils.translation import ugettext_lazy as _
USER_FORM_FIELDS = (
('Ideasbox', ['serial', 'box_awareness']),
(_('Personal informations'), ['refugee_id', 'short_name', 'full_name', 'birth_year', 'gender', 'phone']), # noqa
(_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa
(_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa
(_('Origin'), ['country', 'city', 'country_of_origin_occupation', 'school_level', 'is_sent_to_school']), # noqa
(_('Language skills'), ['rn_level', 'sw_level', 'fr_level']),
(_('National residents'), ['id_card_number']),
)
STAFF_HOME_CARDS = [c for c in STAFF_HOME_CARDS
if c['url'] in ['user_list', 'server:power',
'server:backup']]
HOME_CARDS = STAFF_HOME_CARDS + [
{
'id': 'blog',
},
{
'id': 'mediacenter',
},
{
'id': 'wikipedia',
},
{
'id': 'khanacademy',
},
{
'id': 'vikidia',
},
{
'id': 'gutenberg',
},
{
'id': 'cpassorcier',
},
{
'id': 'ted',
},
]
| Add cards for Ideasbox in Burundi | Add cards for Ideasbox in Burundi
| Python | agpl-3.0 | ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube | ---
+++
@@ -11,3 +11,33 @@
(_('Language skills'), ['rn_level', 'sw_level', 'fr_level']),
(_('National residents'), ['id_card_number']),
)
+STAFF_HOME_CARDS = [c for c in STAFF_HOME_CARDS
+ if c['url'] in ['user_list', 'server:power',
+ 'server:backup']]
+
+HOME_CARDS = STAFF_HOME_CARDS + [
+ {
+ 'id': 'blog',
+ },
+ {
+ 'id': 'mediacenter',
+ },
+ {
+ 'id': 'wikipedia',
+ },
+ {
+ 'id': 'khanacademy',
+ },
+ {
+ 'id': 'vikidia',
+ },
+ {
+ 'id': 'gutenberg',
+ },
+ {
+ 'id': 'cpassorcier',
+ },
+ {
+ 'id': 'ted',
+ },
+] |
d75a79d10658ad32a9b1d71e472372d8335c7bb6 | ml/test_amaranth_lib.py | ml/test_amaranth_lib.py | # Lint as: python3
"""These tests ensure correctness for the helper functions in amaranth_lib."""
import unittest
class TestAmaranthHelpers(unittest.TestCase):
def test_load_calorie_data(self):
raise NotImplementedError
def test_clean_data(self):
raise NotImplementedError
def test_add_calorie_labels(self):
raise NotImplementedError
def test_num_unique_words(self):
raise NotImplementedError
def test_max_sequence_length(self):
raise NotImplementedError
def test_add_input_labels(self):
raise NotImplementedError
if __name__ == '__main__':
unittest.main()
| # Lint as: python3
"""These tests ensure correctness for the helper functions in amaranth_lib."""
import unittest
class TestAmaranthHelpers(unittest.TestCase):
def test_combine_dataframes(self):
raise NotImplementedError
def test_get_calorie_data(self):
raise NotImplementedError
def test_clean_data(self):
raise NotImplementedError
def test_add_calorie_labels(self):
raise NotImplementedError
def test_num_unique_words(self):
raise NotImplementedError
def test_max_sequence_length(self):
raise NotImplementedError
def test_add_input_labels(self):
raise NotImplementedError
if __name__ == '__main__':
unittest.main()
| Update testing stubs with helper lib changes | Update testing stubs with helper lib changes
| Python | apache-2.0 | googleinterns/amaranth,googleinterns/amaranth | ---
+++
@@ -6,7 +6,10 @@
class TestAmaranthHelpers(unittest.TestCase):
- def test_load_calorie_data(self):
+ def test_combine_dataframes(self):
+ raise NotImplementedError
+
+ def test_get_calorie_data(self):
raise NotImplementedError
def test_clean_data(self): |
3bc52b94479e3b0e147ff6ef4bcc8379a5b57249 | trunk/mobile_portal/mobile_portal/core/middleware.py | trunk/mobile_portal/mobile_portal/core/middleware.py | from django.conf import settings
import geolocation
from mobile_portal.wurfl.wurfl_data import devices
from mobile_portal.wurfl import device_parents
from pywurfl.algorithms import DeviceNotFound
from mobile_portal.wurfl.vsm import VectorSpaceAlgorithm
class LocationMiddleware(object):
vsa = VectorSpaceAlgorithm(devices)
def process_request(self, request):
ua = request.META.get('HTTP_USER_AGENT', '')
try:
request.browser = devices.select_ua(
request.META['HTTP_USER_AGENT'],
search=LocationMiddleware.vsa
)
except (KeyError, DeviceNotFound):
request.browser = devices.select_id('generic_xhtml')
if 'opera_mini_ver1' in device_parents[request.browser.devid]:
opera_device = request.META.get('HTTP_X_OPERAMINI_PHONE')
request.device = devices.select_ua(
opera_device,
search=LocationMiddleware.vsa
)
else:
request.device = request.browser
from django.db import connection
class PrintQueriesMiddleware(object):
def process_response(self, request, response):
for query in connection.queries:
print '-'*80
print query['sql']
return response | from django.conf import settings
import geolocation
from mobile_portal.wurfl.wurfl_data import devices
from mobile_portal.wurfl import device_parents
from pywurfl.algorithms import DeviceNotFound
from mobile_portal.wurfl.vsm import VectorSpaceAlgorithm
class LocationMiddleware(object):
vsa = VectorSpaceAlgorithm(devices)
def process_request(self, request):
ua = request.META.get('HTTP_USER_AGENT', '')
try:
request.browser = devices.select_ua(
request.META['HTTP_USER_AGENT'],
search=LocationMiddleware.vsa
)
except (KeyError, DeviceNotFound):
request.browser = devices.select_id('generic_xhtml')
if 'HTTP_X_OPERAMINI_PHONE' in request.META:
opera_device = request.META['HTTP_X_OPERAMINI_PHONE']
request.device = devices.select_ua(
opera_device,
search=LocationMiddleware.vsa
)
else:
request.device = request.browser
from django.db import connection
class PrintQueriesMiddleware(object):
def process_response(self, request, response):
for query in connection.queries:
print '-'*80
print query['sql']
return response | Fix bug when Opera Mini (and possibly others) present no X-OperaMini-Phone header. | Fix bug when Opera Mini (and possibly others) present no X-OperaMini-Phone header.
| Python | apache-2.0 | mollyproject/mollyproject,mollyproject/mollyproject,mollyproject/mollyproject | ---
+++
@@ -20,8 +20,8 @@
except (KeyError, DeviceNotFound):
request.browser = devices.select_id('generic_xhtml')
- if 'opera_mini_ver1' in device_parents[request.browser.devid]:
- opera_device = request.META.get('HTTP_X_OPERAMINI_PHONE')
+ if 'HTTP_X_OPERAMINI_PHONE' in request.META:
+ opera_device = request.META['HTTP_X_OPERAMINI_PHONE']
request.device = devices.select_ua(
opera_device,
search=LocationMiddleware.vsa |
ef81886e4ecf08c12783e0cc2b934ed812accb97 | Zika_vdb_upload.py | Zika_vdb_upload.py | import os, re, time, datetime, csv, sys
import rethinkdb as r
from Bio import SeqIO
from vdb_upload import vdb_upload
from vdb_upload import parser
class Zika_vdb_upload(vdb_upload):
def __init__(self, fasta_fields, fasta_fname, database, virus, source, locus=None, vsubtype=None, authors=None, path=None, auth_key=None):
'''
:param fasta_fields: Dictionary defining position in fasta field to be included in database
'''
vdb_upload.__init__(self, fasta_fields, fasta_fname, database, virus, source, locus, vsubtype, authors, path, auth_key)
if __name__=="__main__":
args = parser.parse_args()
fasta_fields = {0:'accession', 1:'strain', 2:'date', 4:'country', 5:'division', 6:'location'}
run = Zika_vdb_upload(fasta_fields, fasta_fname=args.fname, database=args.database, virus=args.virus, source=args.source, locus=args.locus, vsubtype=args.subtype, authors=args.authors, path=args.path)
run.upload() | import os, re, time, datetime, csv, sys
import rethinkdb as r
from Bio import SeqIO
from vdb_upload import vdb_upload
from vdb_upload import parser
class Zika_vdb_upload(vdb_upload):
def __init__(self, fasta_fields, fasta_fname, database, virus, source, locus=None, vsubtype=None, authors=None, path=None, auth_key=None):
'''
:param fasta_fields: Dictionary defining position in fasta field to be included in database
'''
vdb_upload.__init__(self, fasta_fields, fasta_fname, database, virus, source, locus, vsubtype, authors, path, auth_key)
if __name__=="__main__":
args = parser.parse_args()
fasta_fields = {0:'accession', 2:'strain', 4:'date', 6:'country'}
# 0 1 2 3 4 5 6
# >KU647676|Zika_virus|MRS_OPY_Martinique_PaRi_2015|NA|2015_12|Human|Martinique
run = Zika_vdb_upload(fasta_fields, fasta_fname=args.fname, database=args.database, virus=args.virus, source=args.source, locus=args.locus, vsubtype=args.subtype, authors=args.authors, path=args.path)
run.upload() | Modify Zika fasta fields to match default VIPRBRC ordering. | Modify Zika fasta fields to match default VIPRBRC ordering.
| Python | agpl-3.0 | blab/nextstrain-db,nextstrain/fauna,blab/nextstrain-db,nextstrain/fauna | ---
+++
@@ -14,6 +14,8 @@
if __name__=="__main__":
args = parser.parse_args()
- fasta_fields = {0:'accession', 1:'strain', 2:'date', 4:'country', 5:'division', 6:'location'}
+ fasta_fields = {0:'accession', 2:'strain', 4:'date', 6:'country'}
+ # 0 1 2 3 4 5 6
+ # >KU647676|Zika_virus|MRS_OPY_Martinique_PaRi_2015|NA|2015_12|Human|Martinique
run = Zika_vdb_upload(fasta_fields, fasta_fname=args.fname, database=args.database, virus=args.virus, source=args.source, locus=args.locus, vsubtype=args.subtype, authors=args.authors, path=args.path)
run.upload() |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.