commit stringlengths 40 40 | old_file stringlengths 4 150 | new_file stringlengths 4 150 | old_contents stringlengths 0 3.26k | new_contents stringlengths 1 4.43k | subject stringlengths 15 501 | message stringlengths 15 4.06k | lang stringclasses 4 values | license stringclasses 13 values | repos stringlengths 5 91.5k | diff stringlengths 0 4.35k |
|---|---|---|---|---|---|---|---|---|---|---|
489ac862bf5efc70d1af3933d1c518489a0e95a6 | tests/test_pubtator.py | tests/test_pubtator.py |
import kindred
def test_pubtator():
data = kindred.pubtator.load([19894120,19894121])
assert isinstance(data,list)
for d in data:
assert isinstance(d,kindred.RelationData)
fileCount = len(data)
entityCount = sum([ len(d.getEntities()) for d in data ])
relationCount = sum([ len(d.getRelations()) for d in data ])
assert fileCount == 2
assert relationCount == 0
assert entityCount == 17
if __name__ == '__main__':
test_pubtator()
|
import kindred
def test_pubtator():
data = kindred.pubtator.load([19894120,19894121])
assert isinstance(data,list)
for d in data:
assert isinstance(d,kindred.RelationData)
fileCount = len(data)
entityCount = sum([ len(d.getEntities()) for d in data ])
relationCount = sum([ len(d.getRelations()) for d in data ])
assert fileCount == 2
assert relationCount == 0
assert entityCount == 39
if __name__ == '__main__':
test_pubtator()
| Update pubtator test given new pubtator output | Update pubtator test given new pubtator output
| Python | mit | jakelever/kindred,jakelever/kindred | ---
+++
@@ -14,7 +14,7 @@
assert fileCount == 2
assert relationCount == 0
- assert entityCount == 17
+ assert entityCount == 39
if __name__ == '__main__':
test_pubtator() |
11103afa4a46cc1835f1479651bcd7c808d6a33c | sdks/python/apache_beam/runners/api/__init__.py | sdks/python/apache_beam/runners/api/__init__.py | #
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
| #
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
"""Checked in to avoid protoc dependency for Python development.
Regenerate files with::
protoc -I../common/runner-api/src/main/proto/ \
--python_out=apache_beam/runners/api/ \
../common/runner-api/src/main/proto/*.proto
protoc -I../common/{fn,runner}-api/src/main/proto/ \
--python_out=apache_beam/runners/api/ \
--grpc_python_out=apache_beam/runners/api/ \
../common/fn-api/src/main/proto/*.proto
"""
| Add instructions to regenerate Python proto wrappers. | Add instructions to regenerate Python proto wrappers.
| Python | apache-2.0 | apache/beam,staslev/incubator-beam,jbonofre/beam,iemejia/incubator-beam,lukecwik/incubator-beam,rangadi/beam,wtanaka/beam,rangadi/beam,markflyhigh/incubator-beam,chamikaramj/beam,manuzhang/beam,tgroh/incubator-beam,wangyum/beam,charlesccychen/beam,RyanSkraba/beam,manuzhang/beam,eljefe6a/incubator-beam,apache/beam,charlesccychen/beam,amarouni/incubator-beam,rangadi/incubator-beam,charlesccychen/incubator-beam,vikkyrk/incubator-beam,staslev/incubator-beam,robertwb/incubator-beam,chamikaramj/beam,dhalperi/beam,sammcveety/incubator-beam,lukecwik/incubator-beam,apache/beam,chamikaramj/beam,rangadi/beam,apache/beam,markflyhigh/incubator-beam,peihe/incubator-beam,charlesccychen/beam,wangyum/beam,lukecwik/incubator-beam,wangyum/beam,robertwb/incubator-beam,wtanaka/beam,mxm/incubator-beam,markflyhigh/incubator-beam,robertwb/incubator-beam,mxm/incubator-beam,lukecwik/incubator-beam,RyanSkraba/beam,robertwb/incubator-beam,apache/beam,robertwb/incubator-beam,manuzhang/beam,lukecwik/incubator-beam,jbonofre/incubator-beam,eljefe6a/incubator-beam,lukecwik/incubator-beam,dhalperi/beam,markflyhigh/incubator-beam,RyanSkraba/beam,robertwb/incubator-beam,wangyum/beam,chamikaramj/beam,chamikaramj/beam,jbonofre/incubator-beam,vikkyrk/incubator-beam,vikkyrk/incubator-beam,staslev/beam,jbonofre/beam,staslev/beam,tgroh/beam,RyanSkraba/beam,rangadi/incubator-beam,markflyhigh/incubator-beam,jbonofre/beam,sammcveety/incubator-beam,lukecwik/incubator-beam,iemejia/incubator-beam,markflyhigh/incubator-beam,amarouni/incubator-beam,tgroh/beam,eljefe6a/incubator-beam,staslev/beam,apache/beam,rangadi/beam,RyanSkraba/beam,markflyhigh/incubator-beam,apache/beam,manuzhang/incubator-beam,lukecwik/incubator-beam,dhalperi/beam,lukecwik/incubator-beam,tgroh/beam,rangadi/beam,dhalperi/incubator-beam,charlesccychen/beam,peihe/incubator-beam,dhalperi/incubator-beam,charlesccychen/beam,sammcveety/incubator-beam,chamikaramj/beam,charlesccychen/incubator-beam,robertwb/incubator-beam,peihe/incubator-beam,manuzhang/incubator-beam,yk5/beam,tgroh/beam,apache/beam,rangadi/incubator-beam,apache/beam,tgroh/incubator-beam,chamikaramj/beam,chamikaramj/beam,RyanSkraba/beam,lukecwik/incubator-beam,yk5/beam,rangadi/beam,wtanaka/beam,chamikaramj/beam,charlesccychen/beam,charlesccychen/beam,robertwb/incubator-beam,rangadi/beam,jbonofre/beam,chamikaramj/beam,apache/beam,charlesccychen/incubator-beam,robertwb/incubator-beam,apache/beam,RyanSkraba/beam,robertwb/incubator-beam,yk5/beam | ---
+++
@@ -14,3 +14,17 @@
# See the License for the specific language governing permissions and
# limitations under the License.
#
+
+"""Checked in to avoid protoc dependency for Python development.
+
+Regenerate files with::
+
+ protoc -I../common/runner-api/src/main/proto/ \
+ --python_out=apache_beam/runners/api/ \
+ ../common/runner-api/src/main/proto/*.proto
+
+ protoc -I../common/{fn,runner}-api/src/main/proto/ \
+ --python_out=apache_beam/runners/api/ \
+ --grpc_python_out=apache_beam/runners/api/ \
+ ../common/fn-api/src/main/proto/*.proto
+""" |
7176ec5d4abe678d8f0d01baeacf4dc78204b18f | tests/integration/modules/grains.py | tests/integration/modules/grains.py | '''
Test the grains module
'''
import integration
class TestModulesGrains(integration.ModuleCase):
'''
Test the grains module
'''
def test_items(self):
'''
grains.items
'''
opts = self.minion_opts
self.assertEqual(
self.run_function('grains.items')['test_grain'],
opts['grains']['test_grain']
)
def test_item(self):
'''
grains.item
'''
opts = self.minion_opts
self.assertEqual(
self.run_function('grains.item', ['test_grain']),
opts['grains']['test_grain']
)
def test_ls(self):
'''
grains.ls
'''
check_for = (
'cpuarch',
'cpu_flags',
'cpu_model',
'domain',
'fqdn',
'host',
'kernel',
'kernelrelease',
'localhost',
'mem_total',
'num_cpus',
'os',
'path',
'ps',
'pythonpath',
'pythonversion',
'saltpath',
'saltversion',
'virtual',
)
lsgrains = self.run_function('grains.ls')
for grain_name in check_for:
self.assertTrue(grain_name in lsgrains)
if __name__ == '__main__':
from integration import run_tests
run_tests(TestModulesGrains)
| '''
Test the grains module
'''
import integration
class TestModulesGrains(integration.ModuleCase):
'''
Test the grains module
'''
def test_items(self):
'''
grains.items
'''
opts = self.minion_opts
self.assertEqual(
self.run_function('grains.items')['test_grain'],
opts['grains']['test_grain']
)
def test_item(self):
'''
grains.item
'''
opts = self.minion_opts
self.assertEqual(
self.run_function('grains.item', ['test_grain']),
opts['grains']['test_grain']
)
def test_ls(self):
'''
grains.ls
'''
check_for = (
'cpuarch',
'cpu_flags',
'cpu_model',
'domain',
'fqdn',
'host',
'kernel',
'kernelrelease',
'localhost',
'mem_total',
'num_cpus',
'os',
'os_family',
'path',
'ps',
'pythonpath',
'pythonversion',
'saltpath',
'saltversion',
'virtual',
)
lsgrains = self.run_function('grains.ls')
for grain_name in check_for:
self.assertTrue(grain_name in lsgrains)
if __name__ == '__main__':
from integration import run_tests
run_tests(TestModulesGrains)
| Add test to test if os_family grain is provided. | Add test to test if os_family grain is provided.
Corey Quinn reported a issue where __grains__['os_family'] returned a
KeyError. This commits adds a check to the grains module test to ensure
os_family is present.
| Python | apache-2.0 | saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt | ---
+++
@@ -45,6 +45,7 @@
'mem_total',
'num_cpus',
'os',
+ 'os_family',
'path',
'ps',
'pythonpath', |
f3b960da24f2f9486eca3936a5be0d5001c24ab3 | package_name/module.py | package_name/module.py | """
Module provides a simple cubic_rectification function.
"""
import numpy as np
def cubic_rectification(x):
"""
Rectified cube of an array.
Parameters
----------
X : numpy.ndarray
Input array.
Returns
-------
numpy.ndarray
Elementwise, the cube of `X` where it is positive and `0` otherwise.
Note
----
This is a sample function, using a Google docstring format.
Note
----
The use of intersphinx will cause numpy.ndarray above to link to its
documentation, but not inside this Note.
"""
return np.maximum(0, x ** 3)
| """
Module provides a simple cubic_rectification function.
"""
import numpy as np
def cubic_rectification(x):
"""
Rectified cube of an array.
Parameters
----------
x : numpy.ndarray
Input array.
Returns
-------
numpy.ndarray
Elementwise, the cube of `x` where it is positive and `0` otherwise.
Note
----
This is a sample function, using a Google docstring format.
Note
----
The use of intersphinx will cause numpy.ndarray above to link to its
documentation, but not inside this Note.
"""
return np.maximum(0, x ** 3)
| Fix X -> x in cubic_rectification | DOC: Fix X -> x in cubic_rectification
| Python | mit | scottclowe/python-continuous-integration,scottclowe/python-ci,scottclowe/python-continuous-integration,scottclowe/python-ci | ---
+++
@@ -11,13 +11,13 @@
Parameters
----------
- X : numpy.ndarray
+ x : numpy.ndarray
Input array.
Returns
-------
numpy.ndarray
- Elementwise, the cube of `X` where it is positive and `0` otherwise.
+ Elementwise, the cube of `x` where it is positive and `0` otherwise.
Note
---- |
6f199c9f2fb54931fa852700598b3c0fe24e40ad | viewflow/exceptions.py | viewflow/exceptions.py | class FlowRuntimeError(Exception):
"""Unrecovable flow runtime error."""
class FlowLockFailed(Exception):
"""Flow lock failed."""
| class FlowRuntimeError(Exception):
"""Unrecoverable flow runtime error."""
class FlowLockFailed(Exception):
"""Flow lock failed."""
| Fix typo in exception docstring | Fix typo in exception docstring | Python | agpl-3.0 | viewflow/viewflow,viewflow/viewflow,viewflow/viewflow | ---
+++
@@ -1,5 +1,5 @@
class FlowRuntimeError(Exception):
- """Unrecovable flow runtime error."""
+ """Unrecoverable flow runtime error."""
class FlowLockFailed(Exception): |
fb8f77553d75cdcf2c50423cb2e4159350a13e5a | tests/test_person.py | tests/test_person.py | import unittest
from classes.person import Person
class PersonClassTest(unittest.TestCase):
def test_full_name_only_returns_strings(self):
with self.assertRaises(ValueError, msg='Only strings are allowed as names'):
my_class_instance = Person("staff", "Peter", "Musonye")
my_class_instance.full_name()
| import unittest
from classes.person import Person
class PersonClassTest(unittest.TestCase):
def test_full_name_only_returns_strings(self):
with self.assertRaises(ValueError, msg='Only strings are allowed as names'):
my_class_instance = Person("staff", "Peter", 1234)
my_class_instance.full_name()
| Fix test case for class Person | Fix test case for class Person
| Python | mit | peterpaints/room-allocator | ---
+++
@@ -5,5 +5,5 @@
class PersonClassTest(unittest.TestCase):
def test_full_name_only_returns_strings(self):
with self.assertRaises(ValueError, msg='Only strings are allowed as names'):
- my_class_instance = Person("staff", "Peter", "Musonye")
+ my_class_instance = Person("staff", "Peter", 1234)
my_class_instance.full_name() |
892393458612ea78319cceeb98957c34ccb91d2d | django_react_templatetags/encoders.py | django_react_templatetags/encoders.py | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.core.serializers.json import DjangoJSONEncoder
from django_react_templatetags.mixins import RepresentationMixin
def json_encoder_cls_factory(context):
class ReqReactRepresentationJSONEncoder(ReactRepresentationJSONEncoder):
context = None
ReqReactRepresentationJSONEncoder.context = context
return ReqReactRepresentationJSONEncoder
class ReactRepresentationJSONEncoder(DjangoJSONEncoder):
'''
Custom json encoder that adds support for RepresentationMixin
'''
def default(self, o):
if isinstance(o, RepresentationMixin):
# Allow backwards compability with react_representation prop
if not hasattr(o, 'to_react_representation'):
return o.react_representation
args = [self.context if hasattr(self, 'context') else None]
args = [x for x in args if x is not None]
return o.to_react_representation(*args)
return super(ReactRepresentationJSONEncoder, self).default(o)
| # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.core.serializers.json import DjangoJSONEncoder
from django_react_templatetags.mixins import RepresentationMixin
def json_encoder_cls_factory(context):
class ReqReactRepresentationJSONEncoder(ReactRepresentationJSONEncoder):
context = None
ReqReactRepresentationJSONEncoder.context = context
return ReqReactRepresentationJSONEncoder
class ReactRepresentationJSONEncoder(DjangoJSONEncoder):
'''
Custom json encoder that adds support for RepresentationMixin
'''
def default(self, o):
if isinstance(o, RepresentationMixin):
args = [self.context if hasattr(self, 'context') else None]
args = [x for x in args if x is not None]
return o.to_react_representation(*args)
return super(ReactRepresentationJSONEncoder, self).default(o)
| Drop support for reacct_representation property | Drop support for reacct_representation property
| Python | mit | Frojd/django-react-templatetags,Frojd/django-react-templatetags,Frojd/django-react-templatetags | ---
+++
@@ -20,10 +20,6 @@
def default(self, o):
if isinstance(o, RepresentationMixin):
- # Allow backwards compability with react_representation prop
- if not hasattr(o, 'to_react_representation'):
- return o.react_representation
-
args = [self.context if hasattr(self, 'context') else None]
args = [x for x in args if x is not None]
|
cfe78dabea226e24928d26183f4b135c52b64663 | feder/cases/forms.py | feder/cases/forms.py | # -*- coding: utf-8 -*-
from atom.ext.crispy_forms.forms import SingleButtonMixin
from braces.forms import UserKwargModelFormMixin
from django import forms
from .models import Case
class CaseForm(SingleButtonMixin, UserKwargModelFormMixin, forms.ModelForm):
def __init__(self, *args, **kwargs):
self.monitoring = kwargs.pop('monitoring', None)
super(CaseForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
if self.monitoring:
self.instance.monitoring = self.monitoring
super(CaseForm, self).save(*args, **kwargs)
class Meta:
model = Case
fields = ['name', 'institution']
| # -*- coding: utf-8 -*-
from atom.ext.crispy_forms.forms import SingleButtonMixin
from braces.forms import UserKwargModelFormMixin
from django import forms
from .models import Case
class CaseForm(SingleButtonMixin, UserKwargModelFormMixin, forms.ModelForm):
def __init__(self, *args, **kwargs):
self.monitoring = kwargs.pop('monitoring', None)
super(CaseForm, self).__init__(*args, **kwargs)
if self.monitoring:
self.instance.monitoring = self.monitoring
class Meta:
model = Case
fields = ['name', 'institution']
| Clean up form in CaseForm | Clean up form in CaseForm
| Python | mit | watchdogpolska/feder,watchdogpolska/feder,watchdogpolska/feder,watchdogpolska/feder | ---
+++
@@ -11,11 +11,8 @@
def __init__(self, *args, **kwargs):
self.monitoring = kwargs.pop('monitoring', None)
super(CaseForm, self).__init__(*args, **kwargs)
-
- def save(self, *args, **kwargs):
if self.monitoring:
self.instance.monitoring = self.monitoring
- super(CaseForm, self).save(*args, **kwargs)
class Meta:
model = Case |
06ec0a7f0a6a53fddfb2038b0ae8cc1bad2c8511 | blankspot/node_registration/models.py | blankspot/node_registration/models.py | from django.db import models
class Contact(models.Model):
first_name = models.CharField(max_length=50, blank=True, null=True)
last_name = models.CharField(max_length=50, blank=True, null=True)
nick = models.CharField(max_length=128)
email = models.EmailField(max_length=254)
def __unicode__(self):
return (self.nick)
def get_absolute_url(self):
return reverse('contact-detail', kwargs={'pk': self.pk})
class Position(models.Model):
contact = models.ForeignKey('Contact')
street = models.CharField(max_length=200, blank=True, null=True)
city = models.CharField(max_length=100, blank=True, null=True)
address_description = models.TextField(blank=True, null=True)
altitude = models.FloatField(blank=True, null=True)
latitude = models.FloatField(blank=True, null=True)
def __unicode__(self):
return (self.street)
def get_absolute_url(self):
return reverse('position-detail', kwargs={'pk': self.pk})
| from django.db import models
class Position(models.Model):
first_name = models.CharField(max_length=50, blank=True, null=True)
last_name = models.CharField(max_length=50, blank=True, null=True)
nick = models.CharField(max_length=128)
email = models.EmailField(max_length=254)
street = models.CharField(max_length=200, blank=True, null=True)
city = models.CharField(max_length=100, blank=True, null=True)
address_description = models.TextField(blank=True, null=True)
altitude = models.FloatField(blank=True, null=True)
latitude = models.FloatField(blank=True, null=True)
def __unicode__(self):
return (self.street)
def get_absolute_url(self):
return reverse('position-detail', kwargs={'pk': self.pk})
| Revert splitting of model as its adding to much complexitiy for the timebeing to later logics IIt's just not adding enought value for having a more complicated implementation. | Revert splitting of model as its adding to much complexitiy for the timebeing to later logics
IIt's just not adding enought value for having a more complicated implementation.
| Python | agpl-3.0 | frlan/blankspot | ---
+++
@@ -1,19 +1,10 @@
from django.db import models
-class Contact(models.Model):
+class Position(models.Model):
first_name = models.CharField(max_length=50, blank=True, null=True)
last_name = models.CharField(max_length=50, blank=True, null=True)
nick = models.CharField(max_length=128)
email = models.EmailField(max_length=254)
-
- def __unicode__(self):
- return (self.nick)
-
- def get_absolute_url(self):
- return reverse('contact-detail', kwargs={'pk': self.pk})
-
-class Position(models.Model):
- contact = models.ForeignKey('Contact')
street = models.CharField(max_length=200, blank=True, null=True)
city = models.CharField(max_length=100, blank=True, null=True)
address_description = models.TextField(blank=True, null=True) |
b86d88a10839ba642f992dcaf3e69de3a244f984 | golingo/urls.py | golingo/urls.py | """golingo URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
]
| """golingo URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
from quiz.views import QuestionTemplateView
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^question/$', QuestionTemplateView.as_view(), name='question'),
]
| Add initial url to question | Add initial url to question
| Python | bsd-3-clause | jesuejunior/golingo,jesuejunior/golingo,jesuejunior/golingo | ---
+++
@@ -15,7 +15,9 @@
"""
from django.conf.urls import include, url
from django.contrib import admin
+from quiz.views import QuestionTemplateView
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
+ url(r'^question/$', QuestionTemplateView.as_view(), name='question'),
] |
4a330e190dcb727cb7483b826f2927b94b081e8a | yardcam.py | yardcam.py | import capture
from picamera import PiCamera
import time
import delay
def image_cap_loop(camera, status=None):
"""Set image parameters, capture image, set wait time, repeat"""
resolution = (1640, 1232)
wait = delay.next_capture() # Delay time in seconds from delay.py
waithours = wait / 60 / 60 # Convert seconds to hours
print('Next capture begins in {} hours.'.format(waithours))
time.sleep(wait)
images = 18
for i in range(images):
latest = capture.cap(camera, resolution, status)
status = latest[0]
capture.copy_latest(latest[1])
time.sleep(300)
status = camera.shutdown(camera)
image_cap_loop(camera, status)
return latest
def main():
camera = PiCamera()
image_cap_loop(camera)
print("Images captured")
if __name__ == '__main__':
main()
| import capture
from picamera import PiCamera
import time
import delay
def image_cap_loop(camera, status=None):
"""Set image parameters, capture image, set wait time, repeat"""
resolution = (1640, 1232)
# wait = delay.next_capture() # Delay time in seconds from delay.py
wait = 60
waithours = wait / 60 / 60 # Convert seconds to hours
print('Next capture begins in {} hours.'.format(waithours))
time.sleep(wait)
images = 18
for i in range(images):
latest = capture.cap(camera, resolution, status)
status = latest[0]
capture.copy_latest(latest[1])
time.sleep(300)
status = camera.shutdown(camera)
image_cap_loop(camera, status)
return latest
def main():
camera = PiCamera()
image_cap_loop(camera)
print("Images captured")
if __name__ == '__main__':
main()
| Remove delay from loop for testing | Remove delay from loop for testing
| Python | mit | gnfrazier/YardCam | ---
+++
@@ -8,7 +8,8 @@
"""Set image parameters, capture image, set wait time, repeat"""
resolution = (1640, 1232)
- wait = delay.next_capture() # Delay time in seconds from delay.py
+ # wait = delay.next_capture() # Delay time in seconds from delay.py
+ wait = 60
waithours = wait / 60 / 60 # Convert seconds to hours
print('Next capture begins in {} hours.'.format(waithours))
time.sleep(wait) |
f2d3d52c4118b9ff0161a8db5d53d83c199918e3 | gunicorn_cfg.py | gunicorn_cfg.py | """
This file contains gunicorn settings.
To run sqmpy with gunicorn run the following command:
gunicorn -c gunicorn_cfg.py run:app
In order to daemonize gunicorn add -D flag:
gunicorn -c gunicorn_cfg.py run:app -D
"""
import multiprocessing
# Gunicorn will listen on the given host:port
bind = '0.0.0.0:3000'
# The only tested worker class is gevent
worker_class = 'gevent'
# Set number of workers based on CPU count
workers = multiprocessing.cpu_count() * 2 + 1
# Uncomment for development
# reload = True
# Daemonize the application
daemon = False
# Comment only for development. Use your own certificates here.
keyfile = 'server.key'
certfile = 'server.crt'
# Application loglevel
loglevel = 'debug'
| """
This file contains gunicorn settings.
To run sqmpy with gunicorn run the following command:
gunicorn -c gunicorn_cfg.py run:app
In order to daemonize gunicorn add -D flag:
gunicorn -c gunicorn_cfg.py run:app -D
"""
import multiprocessing
# Gunicorn will listen on the given host:port
bind = '0.0.0.0:3000'
# The only tested worker class is gevent
#worker_class = 'gevent'
# Set number of workers based on CPU count
workers = multiprocessing.cpu_count() * 2 + 1
# Uncomment for development
# reload = True
# Daemonize the application
daemon = False
# Comment only for development. Use your own certificates here.
keyfile = 'server.key'
certfile = 'server.crt'
# Application log level
loglevel = 'debug'
| Update gunicorn config with docs | Update gunicorn config with docs
| Python | bsd-3-clause | mehdisadeghi/sqmpy,simphony/sqmpy,mehdisadeghi/sqmpy,simphony/sqmpy,simphony/sqmpy,mehdisadeghi/sqmpy | ---
+++
@@ -16,7 +16,7 @@
bind = '0.0.0.0:3000'
# The only tested worker class is gevent
-worker_class = 'gevent'
+#worker_class = 'gevent'
# Set number of workers based on CPU count
workers = multiprocessing.cpu_count() * 2 + 1
@@ -31,5 +31,5 @@
keyfile = 'server.key'
certfile = 'server.crt'
-# Application loglevel
+# Application log level
loglevel = 'debug' |
3f6b18304a3f947cc165201a507a672a56af851f | warehouse/cli.py | warehouse/cli.py | # Copyright 2013 Donald Stufft
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import absolute_import, division, print_function
from __future__ import unicode_literals
import warehouse.migrations.cli
def include(module):
return getattr(module, "__commands__", {})
__commands__ = {
"migrate": include(warehouse.migrations.cli),
}
| # Copyright 2013 Donald Stufft
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import absolute_import, division, print_function
from __future__ import unicode_literals
import warehouse.migrations.cli
__commands__ = {
"migrate": warehouse.migrations.cli.__commands__,
}
| Remove a useless function call | Remove a useless function call
| Python | apache-2.0 | robhudson/warehouse,mattrobenolt/warehouse,techtonik/warehouse,mattrobenolt/warehouse,mattrobenolt/warehouse,robhudson/warehouse,techtonik/warehouse | ---
+++
@@ -17,10 +17,6 @@
import warehouse.migrations.cli
-def include(module):
- return getattr(module, "__commands__", {})
-
-
__commands__ = {
- "migrate": include(warehouse.migrations.cli),
+ "migrate": warehouse.migrations.cli.__commands__,
} |
aff8cebfd168493a4a9dff77cf9722507429d570 | contrib/examples/actions/pythonactions/isprime.py | contrib/examples/actions/pythonactions/isprime.py | import math
class PrimeChecker(object):
def run(self, **kwargs):
return self._is_prime(**kwargs)
def _is_prime(self, value=0):
if math.floor(value) != value:
raise ValueError('%s should be an integer.' % value)
if value < 2:
return False
for test in range(2, int(math.floor(math.sqrt(value)))+1):
if value % test == 0:
return False
return True
if __name__ == '__main__':
checker = PrimeChecker()
for i in range(0, 10):
print '%s : %s' % (i, checker.run(**{'value': i}))
| import math
class PrimeChecker(object):
def run(self, value=0):
if math.floor(value) != value:
raise ValueError('%s should be an integer.' % value)
if value < 2:
return False
for test in range(2, int(math.floor(math.sqrt(value)))+1):
if value % test == 0:
return False
return True
if __name__ == '__main__':
checker = PrimeChecker()
for i in range(0, 10):
print '%s : %s' % (i, checker.run(**{'value': i}))
| Update pythonaction sample for simpler run. | Update pythonaction sample for simpler run.
| Python | apache-2.0 | peak6/st2,lakshmi-kannan/st2,pixelrebel/st2,StackStorm/st2,jtopjian/st2,pinterb/st2,Plexxi/st2,punalpatel/st2,armab/st2,grengojbo/st2,grengojbo/st2,punalpatel/st2,pixelrebel/st2,Itxaka/st2,lakshmi-kannan/st2,emedvedev/st2,lakshmi-kannan/st2,pixelrebel/st2,nzlosh/st2,peak6/st2,dennybaa/st2,pinterb/st2,Plexxi/st2,nzlosh/st2,Itxaka/st2,grengojbo/st2,alfasin/st2,nzlosh/st2,pinterb/st2,Plexxi/st2,jtopjian/st2,emedvedev/st2,StackStorm/st2,armab/st2,jtopjian/st2,StackStorm/st2,dennybaa/st2,alfasin/st2,emedvedev/st2,peak6/st2,punalpatel/st2,tonybaloney/st2,Plexxi/st2,tonybaloney/st2,alfasin/st2,nzlosh/st2,StackStorm/st2,tonybaloney/st2,dennybaa/st2,armab/st2,Itxaka/st2 | ---
+++
@@ -3,10 +3,7 @@
class PrimeChecker(object):
- def run(self, **kwargs):
- return self._is_prime(**kwargs)
-
- def _is_prime(self, value=0):
+ def run(self, value=0):
if math.floor(value) != value:
raise ValueError('%s should be an integer.' % value)
if value < 2: |
035ae3b2acf5c29304a1c2ec327feb5cc7160559 | django_vend/core/forms.py | django_vend/core/forms.py | import re
from django import forms
from django.utils.dateparse import parse_datetime
from django.core.exceptions import ValidationError
def valid_date(date):
regex = ("^(?:[1-9]\d{3}-(?:(?:0[1-9]|1[0-2])-(?:0[1-9]|1\d|2[0-8])|(?:0[13"
"-9]|1[0-2])-(?:29|30)|(?:0[13578]|1[02])-31)|(?:[1-9]\d(?:0[48]|["
"2468][048]|[13579][26])|(?:[2468][048]|[13579][26])00)-02-29)T(?:"
"[01]\d|2[0-3]):[0-5]\d:[0-5]\d(?:Z|[+-][01]\d:[0-5]\d)$")
return re.search(regex, date)
class VendDateTimeField(forms.DateTimeField):
def to_python(self, value):
if value not in self.empty_values and valid_date(value):
try:
value = parse_datetime(value)
except ValueError:
pass
return super(VendDateTimeField, self).to_python(value)
| import re
from django import forms
from django.utils.dateparse import parse_datetime
from django.core.exceptions import ValidationError
def valid_date(date):
regex = ("^(?:[1-9]\d{3}-(?:(?:0[1-9]|1[0-2])-(?:0[1-9]|1\d|2[0-8])|(?:0[13"
"-9]|1[0-2])-(?:29|30)|(?:0[13578]|1[02])-31)|(?:[1-9]\d(?:0[48]|["
"2468][048]|[13579][26])|(?:[2468][048]|[13579][26])00)-02-29)T(?:"
"[01]\d|2[0-3]):[0-5]\d:[0-5]\d(?:Z|[+-][01]\d:[0-5]\d)$")
return re.search(regex, date)
class VendDateTimeField(forms.DateTimeField):
def to_python(self, value):
if value not in self.empty_values and valid_date(value):
try:
value = parse_datetime(value)
except ValueError:
pass
elif value == "null":
value = None
return super(VendDateTimeField, self).to_python(value)
| Allow VendDateTimeField to accept null dates (if required is set to False) | Allow VendDateTimeField to accept null dates (if required is set to False)
| Python | bsd-3-clause | remarkablerocket/django-vend,remarkablerocket/django-vend | ---
+++
@@ -18,4 +18,6 @@
value = parse_datetime(value)
except ValueError:
pass
+ elif value == "null":
+ value = None
return super(VendDateTimeField, self).to_python(value) |
080637c99898082d38b306ef73983552b263e628 | inbox/ignition.py | inbox/ignition.py | from sqlalchemy import create_engine
from inbox.sqlalchemy_ext.util import ForceStrictMode
from inbox.config import db_uri, config
DB_POOL_SIZE = config.get_required('DB_POOL_SIZE')
def main_engine(pool_size=DB_POOL_SIZE, max_overflow=5):
engine = create_engine(db_uri(),
listeners=[ForceStrictMode()],
isolation_level='READ COMMITTED',
echo=False,
pool_size=pool_size,
max_overflow=max_overflow,
connect_args={'charset': 'utf8mb4'})
return engine
def init_db():
""" Make the tables.
This is called only from bin/create-db, which is run during setup.
Previously we allowed this to run everytime on startup, which broke some
alembic revisions by creating new tables before a migration was run.
From now on, we should ony be creating tables+columns via SQLalchemy *once*
and all subscequent changes done via migration scripts.
"""
from inbox.models.base import MailSyncBase
engine = main_engine(pool_size=1)
MailSyncBase.metadata.create_all(engine)
| from sqlalchemy import create_engine
from inbox.sqlalchemy_ext.util import ForceStrictMode
from inbox.config import db_uri, config
DB_POOL_SIZE = config.get_required('DB_POOL_SIZE')
def main_engine(pool_size=DB_POOL_SIZE, max_overflow=5):
engine = create_engine(db_uri(),
listeners=[ForceStrictMode()],
isolation_level='READ COMMITTED',
echo=False,
pool_size=pool_size,
pool_recycle=3600,
max_overflow=max_overflow,
connect_args={'charset': 'utf8mb4'})
return engine
def init_db():
""" Make the tables.
This is called only from bin/create-db, which is run during setup.
Previously we allowed this to run everytime on startup, which broke some
alembic revisions by creating new tables before a migration was run.
From now on, we should ony be creating tables+columns via SQLalchemy *once*
and all subscequent changes done via migration scripts.
"""
from inbox.models.base import MailSyncBase
engine = main_engine(pool_size=1)
MailSyncBase.metadata.create_all(engine)
| Set pool_recycle to deal with MySQL closing idle connections. | Set pool_recycle to deal with MySQL closing idle connections.
See http://docs.sqlalchemy.org/en/latest/dialects/mysql.html#connection-timeouts
Cherry-picking this onto master so it definitely gets deployed.
| Python | agpl-3.0 | Eagles2F/sync-engine,Eagles2F/sync-engine,ErinCall/sync-engine,ErinCall/sync-engine,ErinCall/sync-engine,PriviPK/privipk-sync-engine,closeio/nylas,Eagles2F/sync-engine,wakermahmud/sync-engine,Eagles2F/sync-engine,wakermahmud/sync-engine,jobscore/sync-engine,jobscore/sync-engine,Eagles2F/sync-engine,wakermahmud/sync-engine,PriviPK/privipk-sync-engine,jobscore/sync-engine,jobscore/sync-engine,ErinCall/sync-engine,wakermahmud/sync-engine,PriviPK/privipk-sync-engine,PriviPK/privipk-sync-engine,closeio/nylas,nylas/sync-engine,nylas/sync-engine,wakermahmud/sync-engine,EthanBlackburn/sync-engine,EthanBlackburn/sync-engine,nylas/sync-engine,PriviPK/privipk-sync-engine,EthanBlackburn/sync-engine,gale320/sync-engine,EthanBlackburn/sync-engine,gale320/sync-engine,ErinCall/sync-engine,gale320/sync-engine,closeio/nylas,EthanBlackburn/sync-engine,nylas/sync-engine,gale320/sync-engine,closeio/nylas,gale320/sync-engine | ---
+++
@@ -12,6 +12,7 @@
isolation_level='READ COMMITTED',
echo=False,
pool_size=pool_size,
+ pool_recycle=3600,
max_overflow=max_overflow,
connect_args={'charset': 'utf8mb4'})
return engine |
1513532e473866438ac9dabbfb462e9348a5895e | hug/output_format.py | hug/output_format.py | import json as json_converter
from datetime import date, datetime
from hug.format import content_type
def _json_converter(item):
if isinstance(item, (date, datetime)):
return item.isoformat()
elif isinstance(item, bytes):
return item.decode('utf8')
raise TypeError("Type not serializable")
@content_type('application/json')
def json(content, **kwargs):
"""JSON (Javascript Serialized Object Notation)"""
return json_converter.dumps(content, default=_json_converter, **kwargs).encode('utf8')
@content_type('text/plain')
def text(content):
"""Free form UTF8 text"""
return content.encode('utf8')
def _camelcase(dictionary):
if not isinstance(dictionary, dict):
return dictionary
new_dictionary = {}
for key, value in dictionary.items():
if isinstance(key, str):
key = key[0] + "".join(key.title().split('_'))[1:]
new_dictionary[key] = _camelcase(value)
return new_dictionary
@content_type('application/json')
def json_camelcase(content):
"""JSON (Javascript Serialized Object Notation) with all keys camelCased"""
return json(_camelcase(content))
@content_type('application/json')
def pretty_json(content):
"""JSON (Javascript Serialized Object Notion) pretty printed and indented"""
return json(content, indent=4, separators=(',', ': '))
| import json as json_converter
from datetime import date, datetime
from hug.format import content_type
def _json_converter(item):
if isinstance(item, (date, datetime)):
return item.isoformat()
elif isinstance(item, bytes):
return item.decode('utf8')
elif getattr(item, '__json__', None):
return item.__json__()
raise TypeError("Type not serializable")
@content_type('application/json')
def json(content, **kwargs):
"""JSON (Javascript Serialized Object Notation)"""
return json_converter.dumps(content, default=_json_converter, **kwargs).encode('utf8')
@content_type('text/plain')
def text(content):
"""Free form UTF8 text"""
return content.encode('utf8')
def _camelcase(dictionary):
if not isinstance(dictionary, dict):
return dictionary
new_dictionary = {}
for key, value in dictionary.items():
if isinstance(key, str):
key = key[0] + "".join(key.title().split('_'))[1:]
new_dictionary[key] = _camelcase(value)
return new_dictionary
@content_type('application/json')
def json_camelcase(content):
"""JSON (Javascript Serialized Object Notation) with all keys camelCased"""
return json(_camelcase(content))
@content_type('application/json')
def pretty_json(content):
"""JSON (Javascript Serialized Object Notion) pretty printed and indented"""
return json(content, indent=4, separators=(',', ': '))
| Add the ability for individual objects to define how they would like there data to be outputed for json | Add the ability for individual objects to define how they would like there data to be outputed for json
| Python | mit | janusnic/hug,yasoob/hug,janusnic/hug,shaunstanislaus/hug,timothycrosley/hug,alisaifee/hug,gbn972/hug,MuhammadAlkarouri/hug,philiptzou/hug,giserh/hug,timothycrosley/hug,STANAPO/hug,shaunstanislaus/hug,STANAPO/hug,origingod/hug,MuhammadAlkarouri/hug,MuhammadAlkarouri/hug,alisaifee/hug,giserh/hug,yasoob/hug,gbn972/hug,philiptzou/hug,jean/hug,timothycrosley/hug,jean/hug,origingod/hug | ---
+++
@@ -9,6 +9,8 @@
return item.isoformat()
elif isinstance(item, bytes):
return item.decode('utf8')
+ elif getattr(item, '__json__', None):
+ return item.__json__()
raise TypeError("Type not serializable")
|
e507461dba5020726c9505fef187098ad234a68a | kazoo/tests/__init__.py | kazoo/tests/__init__.py | import os
import unittest
import time
import uuid
from kazoo.client import KazooClient, KazooState
# if this env variable is set, ZK client integration tests are run
# against the specified host list
ENV_TEST_HOSTS = "KAZOO_TEST_HOSTS"
def get_hosts_or_skip():
if ENV_TEST_HOSTS in os.environ:
return os.environ[ENV_TEST_HOSTS]
raise unittest.SkipTest("Skipping ZooKeeper test. To run, set " +
"%s env to a host list. (ex: localhost:2181)" %
ENV_TEST_HOSTS)
def get_client_or_skip(**kwargs):
hosts = get_hosts_or_skip()
return KazooClient(hosts, **kwargs)
def until_timeout(timeout, value=None):
"""Returns an iterator that repeats until a timeout is reached
timeout is in seconds
"""
start = time.time()
while True:
if time.time() - start >= timeout:
raise Exception("timed out before success!")
yield value
class KazooTestCase(unittest.TestCase):
def _get_client(self):
return KazooClient(self.hosts)
def setUp(self):
namespace = "/kazootests" + uuid.uuid4().hex
self.hosts = get_hosts_or_skip() + namespace
self.client = self._get_client()
def tearDown(self):
if self.client.state == KazooState.LOST:
self.client.connect()
self.client.stop()
| import os
import unittest
import time
import uuid
from nose import SkipTest
from kazoo.client import KazooClient, KazooState
# if this env variable is set, ZK client integration tests are run
# against the specified host list
ENV_TEST_HOSTS = "KAZOO_TEST_HOSTS"
def get_hosts_or_skip():
if ENV_TEST_HOSTS in os.environ:
return os.environ[ENV_TEST_HOSTS]
raise SkipTest("Skipping ZooKeeper test. To run, set " +
"%s env to a host list. (ex: localhost:2181)" %
ENV_TEST_HOSTS)
def get_client_or_skip(**kwargs):
hosts = get_hosts_or_skip()
return KazooClient(hosts, **kwargs)
def until_timeout(timeout, value=None):
"""Returns an iterator that repeats until a timeout is reached
timeout is in seconds
"""
start = time.time()
while True:
if time.time() - start >= timeout:
raise Exception("timed out before success!")
yield value
class KazooTestCase(unittest.TestCase):
def _get_client(self):
return KazooClient(self.hosts)
def setUp(self):
namespace = "/kazootests" + uuid.uuid4().hex
self.hosts = get_hosts_or_skip() + namespace
self.client = self._get_client()
def tearDown(self):
if self.client.state == KazooState.LOST:
self.client.connect()
self.client.stop()
| Use SkipTest that works on Py2.6 | Use SkipTest that works on Py2.6
| Python | apache-2.0 | kormat/kazoo,rackerlabs/kazoo,tempbottle/kazoo,max0d41/kazoo,rgs1/kazoo,rockerbox/kazoo,harlowja/kazoo,kormat/kazoo,rgs1/kazoo,harlowja/kazoo,pombredanne/kazoo,python-zk/kazoo,python-zk/kazoo,pombredanne/kazoo,rockerbox/kazoo,tempbottle/kazoo,AlexanderplUs/kazoo,jacksontj/kazoo,max0d41/kazoo,Asana/kazoo,jacksontj/kazoo,rackerlabs/kazoo,bsanders/kazoo,bsanders/kazoo,AlexanderplUs/kazoo | ---
+++
@@ -2,6 +2,8 @@
import unittest
import time
import uuid
+
+from nose import SkipTest
from kazoo.client import KazooClient, KazooState
@@ -13,9 +15,9 @@
def get_hosts_or_skip():
if ENV_TEST_HOSTS in os.environ:
return os.environ[ENV_TEST_HOSTS]
- raise unittest.SkipTest("Skipping ZooKeeper test. To run, set " +
- "%s env to a host list. (ex: localhost:2181)" %
- ENV_TEST_HOSTS)
+ raise SkipTest("Skipping ZooKeeper test. To run, set " +
+ "%s env to a host list. (ex: localhost:2181)" %
+ ENV_TEST_HOSTS)
def get_client_or_skip(**kwargs): |
544f00f73657bbc06f4dd7f5faae2ce2546fe788 | lava_server/__init__.py | lava_server/__init__.py | # Copyright (C) 2010, 2011 Linaro Limited
#
# Author: Zygmunt Krynicki <zygmunt.krynicki@linaro.org>
#
# This file is part of LAVA Server.
#
# LAVA Server is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License version 3
# as published by the Free Software Foundation
#
# LAVA Server is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with LAVA Server. If not, see <http://www.gnu.org/licenses/>.
__version__ = (0, 2, 0, "dev", 0)
| # Copyright (C) 2010, 2011 Linaro Limited
#
# Author: Zygmunt Krynicki <zygmunt.krynicki@linaro.org>
#
# This file is part of LAVA Server.
#
# LAVA Server is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License version 3
# as published by the Free Software Foundation
#
# LAVA Server is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with LAVA Server. If not, see <http://www.gnu.org/licenses/>.
__version__ = (0, 2, 0, "final", 0)
| Mark final for 0.2 release | Mark final for 0.2 release
| Python | agpl-3.0 | Linaro/lava-server,OSSystems/lava-server,OSSystems/lava-server,Linaro/lava-server,OSSystems/lava-server,Linaro/lava-server,Linaro/lava-server | ---
+++
@@ -17,4 +17,4 @@
# along with LAVA Server. If not, see <http://www.gnu.org/licenses/>.
-__version__ = (0, 2, 0, "dev", 0)
+__version__ = (0, 2, 0, "final", 0) |
f55af10f1767d39fdba65fb4c17beee526f96748 | lib/__init__.py | lib/__init__.py | """retriever.lib contains the core EcoData Retriever modules."""
|
"""retriever.lib contains the core EcoData Retriever modules."""
import os
def set_proxy():
proxies = ["https_proxy", "http_proxy", "ftp_proxy", "HTTP_PROXY", "HTTPS_PROXY", "FTP_PROXY"]
for proxy in proxies:
if os.getenv(proxy):
if len(os.environ[proxy]) != 0:
for i in proxies:
os.environ[i] = os.environ[proxy]
break
set_proxy()
| Check for and use system proxies for downloading files | Check for and use system proxies for downloading files
In some cases when the user is using a proxy urlib.urlopen() will fail to successfully open https files. This prevents the retriever from accessing the scripts stored on GitHub and causes the installation to fail (see #268). This change checks for the existence of proxies and makes them available in a way that urllib.urlopen() can find them | Python | mit | embaldridge/retriever,davharris/retriever,davharris/retriever,davharris/retriever,embaldridge/retriever,goelakash/retriever,henrykironde/deletedret,goelakash/retriever,henrykironde/deletedret,embaldridge/retriever | ---
+++
@@ -1 +1,14 @@
+
"""retriever.lib contains the core EcoData Retriever modules."""
+import os
+
+def set_proxy():
+ proxies = ["https_proxy", "http_proxy", "ftp_proxy", "HTTP_PROXY", "HTTPS_PROXY", "FTP_PROXY"]
+ for proxy in proxies:
+ if os.getenv(proxy):
+ if len(os.environ[proxy]) != 0:
+ for i in proxies:
+ os.environ[i] = os.environ[proxy]
+ break
+
+set_proxy() |
a0d32bb36674863a0e9a04aa97e8c1f7e8ca0f37 | lecturer/tests.py | lecturer/tests.py |
from django.test import Client, TestCase
c = Client()
class WebsiteStabilityTestCase(TestCase):
def test_availability(self):
self.assertEqual(c.get('/lecturer/').status_code, 302) # We are getting redirect when not logged in, so 302
# TODO maybe check for something more reliable than 302?
|
from django.test import Client, TestCase
from django.contrib.auth import get_user_model
c = Client()
class WebsiteStabilityTestCase(TestCase):
def test_availability(self):
self.assertEqual(c.get('/lecturer/').status_code, 302) # We are getting redirect when not logged in, so 302
# TODO maybe check for something more reliable than 302?
def test_lecturer_login(self):
user = get_user_model().objects.create_user('test_user', 'test@test.com', 'kNouYH8J3KjJH3')
user.save()
# Test if lecturer is logged in upon login-request
self.assertEqual(c.post('/login/', {'username': 'test_user', 'password': 'kNouYH8J3KjJH3'}).status_code, 200) | Test if lecturer is logged in upon login-request | Test if lecturer is logged in upon login-request
| Python | mit | martinlunde/RealBack,martinlunde/RealBack,martinlunde/RealBack | ---
+++
@@ -1,6 +1,6 @@
from django.test import Client, TestCase
-
+from django.contrib.auth import get_user_model
c = Client()
@@ -9,3 +9,10 @@
def test_availability(self):
self.assertEqual(c.get('/lecturer/').status_code, 302) # We are getting redirect when not logged in, so 302
# TODO maybe check for something more reliable than 302?
+
+ def test_lecturer_login(self):
+ user = get_user_model().objects.create_user('test_user', 'test@test.com', 'kNouYH8J3KjJH3')
+ user.save()
+
+ # Test if lecturer is logged in upon login-request
+ self.assertEqual(c.post('/login/', {'username': 'test_user', 'password': 'kNouYH8J3KjJH3'}).status_code, 200) |
0867054258e231b2ce9b028c5ce2bc3a26bca7be | gamernews/apps/threadedcomments/views.py | gamernews/apps/threadedcomments/views.py | from django.shortcuts import render_to_response, get_object_or_404
from django.template import RequestContext
from django.contrib.auth.decorators import login_required
from django.utils.translation import ugettext as _
from django.views.generic.list import ListView
from core.models import Account as User
from django_comments.models import Comment
from news.models import Blob, BlobInstance
from .models import ThreadedComment
def single_comment(request, id):
comment = get_object_or_404(ThreadedComment, id=id)
variables = RequestContext(request, {'comment': comment})
return render_to_response('comments/single.html', variables)
def comment_posted( request ):
if request.GET['c']:
comment_id, blob_id = request.GET['c'].split( ':' )
blob = Blob.objects.get( pk=blob_id )
if post:
return HttpResponseRedirect( blob.get_absolute_url() )
return HttpResponseRedirect( "/" )
| from django.shortcuts import render_to_response, get_object_or_404
from django.template import RequestContext
from django.contrib.auth.decorators import login_required
from django.utils.translation import ugettext as _
from django.views.generic.list import ListView
from core.models import Account as User
from django_comments.models import Comment
from news.models import Blob, BlobInstance
from .models import ThreadedComment
def single_comment(request, id):
comment = get_object_or_404(ThreadedComment, id=id)
variables = RequestContext(request, {'comment': comment})
return render_to_response('comments/single.html', variables)
def comment_posted(request):
if request.GET['c']:
comment_id, blob_id = request.GET['c']
comment = Comment.objects.get( pk=comment_id )
blob = Blob.objects.get(pk=blob_id)
if blob:
return HttpResponseRedirect( blob.get_absolute_url() )
return HttpResponseRedirect( "/" )
| Remove name, url and email from comment form | Remove name, url and email from comment form
| Python | mit | underlost/GamerNews,underlost/GamerNews | ---
+++
@@ -14,12 +14,13 @@
variables = RequestContext(request, {'comment': comment})
return render_to_response('comments/single.html', variables)
-def comment_posted( request ):
+def comment_posted(request):
if request.GET['c']:
- comment_id, blob_id = request.GET['c'].split( ':' )
- blob = Blob.objects.get( pk=blob_id )
+ comment_id, blob_id = request.GET['c']
+ comment = Comment.objects.get( pk=comment_id )
+ blob = Blob.objects.get(pk=blob_id)
- if post:
+ if blob:
return HttpResponseRedirect( blob.get_absolute_url() )
return HttpResponseRedirect( "/" ) |
8cd11782d4b3558d204f438accdc15b3b702839b | unn/cli.py | unn/cli.py | import sys
commands = {}
args = []
kwargs = {}
def EXIT(msg, code=1):
print(msg)
sys.exit(code)
def command(fn):
commands[fn.__name__] = fn
return fn
def run():
if len(sys.argv) < 2:
EXIT('No command provided')
cmd = sys.argv[1]
if cmd not in commands:
EXIT('Unkown command')
args = [x for x in sys.argv[2:] if '=' not in x]
kwargs = dict([x.split('=') for x in sys.argv[2:] if '=' in x])
kwargs = dict([(k.replace('-', ''),v) for k,v in kwargs.items()])
commands[cmd](*args, **kwargs)
| import sys
commands = {}
args = []
kwargs = {}
def EXIT(msg, code=1):
print(msg)
sys.exit(code)
def command(fn):
commands[fn.__name__] = fn
return fn
def run():
if len(sys.argv) < 2:
EXIT('Valid commands are:\n ' + '\n '.join(commands))
cmd = sys.argv[1]
if cmd not in commands:
EXIT('Unkown command')
args = [x for x in sys.argv[2:] if '=' not in x]
kwargs = dict([x.split('=') for x in sys.argv[2:] if '=' in x])
kwargs = dict([(k.replace('-', ''),v) for k,v in kwargs.items()])
commands[cmd](*args, **kwargs)
| Add a helpful message if no command given | Add a helpful message if no command given
| Python | mit | runningskull/unn | ---
+++
@@ -16,7 +16,7 @@
def run():
if len(sys.argv) < 2:
- EXIT('No command provided')
+ EXIT('Valid commands are:\n ' + '\n '.join(commands))
cmd = sys.argv[1]
|
ab6526b14f5bdc544367bcaa281a861d2314330b | gi2fasta.py | gi2fasta.py | import sys
from Bio import Entrez
from Bio import SeqIO
Entrez.email = "davidsshin@lbl.gov"
infilename = sys.argv[1]
outfilename = sys.argv[2]
with open(infilename) as f:
gi_numbers=', '.join(line.rstrip() for line in f)
handle = Entrez.efetch(db="protein", rettype="fasta", retmode="text", id=gi_numbers)
records = SeqIO.parse(handle, "fasta")
fout = open(outfilename, 'w')
for record in records:
#print ">" + record.seq
#print record.id
print record.description
#print record.seq
fout.write(">" + str(record.description) + "\n")
fout.write(str(record.seq) + "\n")
fout.close()
#for seq_record in SeqIO.parse(record, "fasta"):
# print seq_record.id
#fob2.write("high percent identity cutoff: " + str(high_identity2) + "\n")
| import sys
#from Bio import Entrez
#from Bio import SeqIO
user_email = "" # User must supply email here to access NCBI api
# Add error message in the event no email address is supplied
if user_email == "":
sys.exit("Error: Please supply your email address to line 5 of gi2fasta.py")
Entrez.email = user_email
infilename = sys.argv[1]
outfilename = sys.argv[2]
with open(infilename) as f:
gi_numbers=', '.join(line.rstrip() for line in f)
handle = Entrez.efetch(db="protein", rettype="fasta", retmode="text", id=gi_numbers)
records = SeqIO.parse(handle, "fasta")
fout = open(outfilename, 'w')
for record in records:
#print ">" + record.seq
#print record.id
print record.description
#print record.seq
fout.write(">" + str(record.description) + "\n")
fout.write(str(record.seq) + "\n")
fout.close()
#for seq_record in SeqIO.parse(record, "fasta"):
# print seq_record.id
#fob2.write("high percent identity cutoff: " + str(high_identity2) + "\n")
| Add error message if User does not enter email address | Add error message if User does not enter email address
| Python | bsd-2-clause | datadaveshin/bioinformatics,datadaveshin/bioinformatics | ---
+++
@@ -1,7 +1,13 @@
import sys
-from Bio import Entrez
-from Bio import SeqIO
-Entrez.email = "davidsshin@lbl.gov"
+#from Bio import Entrez
+#from Bio import SeqIO
+
+user_email = "" # User must supply email here to access NCBI api
+# Add error message in the event no email address is supplied
+if user_email == "":
+ sys.exit("Error: Please supply your email address to line 5 of gi2fasta.py")
+
+Entrez.email = user_email
infilename = sys.argv[1]
outfilename = sys.argv[2] |
10c6112dd343901b502c31655a001e612ed6e441 | api/logs/permissions.py | api/logs/permissions.py | # -*- coding: utf-8 -*-
from rest_framework import permissions
from website.models import Node, NodeLog
from api.nodes.permissions import ContributorOrPublic
from api.base.utils import get_object_or_error
class ContributorOrPublicForLogs(permissions.BasePermission):
def has_object_permission(self, request, view, obj):
assert isinstance(obj, (NodeLog)), 'obj must be a NodeLog, got {}'.format(obj)
for node_id in obj._backrefs['logged']['node']['logs']:
node = get_object_or_error(Node, node_id, display_name='node')
if ContributorOrPublic().has_object_permission(request, view, node):
return True
return False
| # -*- coding: utf-8 -*-
from rest_framework import permissions
from website.models import Node, NodeLog
from api.nodes.permissions import ContributorOrPublic
from api.base.utils import get_object_or_error
class ContributorOrPublicForLogs(permissions.BasePermission):
def has_object_permission(self, request, view, obj):
assert isinstance(obj, (NodeLog)), 'obj must be a NodeLog, got {}'.format(obj)
if obj._backrefs.get('logged'):
for node_id in obj._backrefs['logged']['node']['logs']:
node = get_object_or_error(Node, node_id, display_name='node')
if ContributorOrPublic().has_object_permission(request, view, node):
return True
if getattr(obj, 'node'):
if ContributorOrPublic().has_object_permission(request, view, obj.node):
return True
return False
| Add case for when there are no node backrefs on logs. Again, this whole method will change when eliminating backrefs from nodelogs is merged. | Add case for when there are no node backrefs on logs. Again, this whole method will change when eliminating backrefs from nodelogs is merged.
| Python | apache-2.0 | doublebits/osf.io,mluo613/osf.io,cwisecarver/osf.io,billyhunt/osf.io,baylee-d/osf.io,caneruguz/osf.io,mattclark/osf.io,Johnetordoff/osf.io,kwierman/osf.io,kwierman/osf.io,amyshi188/osf.io,acshi/osf.io,mfraezz/osf.io,zamattiac/osf.io,pattisdr/osf.io,samchrisinger/osf.io,RomanZWang/osf.io,hmoco/osf.io,alexschiller/osf.io,chrisseto/osf.io,felliott/osf.io,laurenrevere/osf.io,chrisseto/osf.io,alexschiller/osf.io,abought/osf.io,felliott/osf.io,mluo613/osf.io,TomBaxter/osf.io,abought/osf.io,TomHeatwole/osf.io,RomanZWang/osf.io,doublebits/osf.io,emetsger/osf.io,mluo613/osf.io,kwierman/osf.io,aaxelb/osf.io,mluke93/osf.io,caneruguz/osf.io,Nesiehr/osf.io,leb2dg/osf.io,brianjgeiger/osf.io,sloria/osf.io,CenterForOpenScience/osf.io,billyhunt/osf.io,kwierman/osf.io,binoculars/osf.io,mluke93/osf.io,billyhunt/osf.io,doublebits/osf.io,chennan47/osf.io,Nesiehr/osf.io,saradbowman/osf.io,hmoco/osf.io,chennan47/osf.io,DanielSBrown/osf.io,baylee-d/osf.io,CenterForOpenScience/osf.io,mluo613/osf.io,cslzchen/osf.io,wearpants/osf.io,zachjanicki/osf.io,monikagrabowska/osf.io,adlius/osf.io,SSJohns/osf.io,icereval/osf.io,alexschiller/osf.io,mattclark/osf.io,wearpants/osf.io,zachjanicki/osf.io,jnayak1/osf.io,jnayak1/osf.io,chrisseto/osf.io,baylee-d/osf.io,amyshi188/osf.io,billyhunt/osf.io,samchrisinger/osf.io,erinspace/osf.io,rdhyee/osf.io,crcresearch/osf.io,hmoco/osf.io,pattisdr/osf.io,brianjgeiger/osf.io,doublebits/osf.io,SSJohns/osf.io,cwisecarver/osf.io,RomanZWang/osf.io,kch8qx/osf.io,erinspace/osf.io,kch8qx/osf.io,wearpants/osf.io,rdhyee/osf.io,HalcyonChimera/osf.io,caneruguz/osf.io,HalcyonChimera/osf.io,sloria/osf.io,cslzchen/osf.io,doublebits/osf.io,Johnetordoff/osf.io,icereval/osf.io,samchrisinger/osf.io,DanielSBrown/osf.io,crcresearch/osf.io,aaxelb/osf.io,emetsger/osf.io,Johnetordoff/osf.io,jnayak1/osf.io,kch8qx/osf.io,laurenrevere/osf.io,icereval/osf.io,emetsger/osf.io,leb2dg/osf.io,abought/osf.io,CenterForOpenScience/osf.io,rdhyee/osf.io,billyhunt/osf.io,monikagrabowska/osf.io,zamattiac/osf.io,adlius/osf.io,acshi/osf.io,alexschiller/osf.io,TomBaxter/osf.io,RomanZWang/osf.io,caneruguz/osf.io,adlius/osf.io,jnayak1/osf.io,amyshi188/osf.io,asanfilippo7/osf.io,cwisecarver/osf.io,acshi/osf.io,DanielSBrown/osf.io,monikagrabowska/osf.io,mfraezz/osf.io,saradbowman/osf.io,CenterForOpenScience/osf.io,SSJohns/osf.io,zamattiac/osf.io,mfraezz/osf.io,TomBaxter/osf.io,caseyrollins/osf.io,kch8qx/osf.io,leb2dg/osf.io,cwisecarver/osf.io,brianjgeiger/osf.io,Nesiehr/osf.io,caseyrollins/osf.io,asanfilippo7/osf.io,TomHeatwole/osf.io,brianjgeiger/osf.io,mluke93/osf.io,abought/osf.io,acshi/osf.io,alexschiller/osf.io,aaxelb/osf.io,monikagrabowska/osf.io,mfraezz/osf.io,asanfilippo7/osf.io,mluke93/osf.io,adlius/osf.io,laurenrevere/osf.io,TomHeatwole/osf.io,HalcyonChimera/osf.io,zachjanicki/osf.io,leb2dg/osf.io,felliott/osf.io,chennan47/osf.io,pattisdr/osf.io,TomHeatwole/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,acshi/osf.io,caseyrollins/osf.io,emetsger/osf.io,cslzchen/osf.io,wearpants/osf.io,mluo613/osf.io,DanielSBrown/osf.io,binoculars/osf.io,erinspace/osf.io,mattclark/osf.io,asanfilippo7/osf.io,hmoco/osf.io,zamattiac/osf.io,Johnetordoff/osf.io,zachjanicki/osf.io,amyshi188/osf.io,felliott/osf.io,SSJohns/osf.io,HalcyonChimera/osf.io,Nesiehr/osf.io,kch8qx/osf.io,samchrisinger/osf.io,chrisseto/osf.io,crcresearch/osf.io,sloria/osf.io,RomanZWang/osf.io,binoculars/osf.io,aaxelb/osf.io,rdhyee/osf.io | ---
+++
@@ -12,8 +12,14 @@
def has_object_permission(self, request, view, obj):
assert isinstance(obj, (NodeLog)), 'obj must be a NodeLog, got {}'.format(obj)
- for node_id in obj._backrefs['logged']['node']['logs']:
- node = get_object_or_error(Node, node_id, display_name='node')
- if ContributorOrPublic().has_object_permission(request, view, node):
+ if obj._backrefs.get('logged'):
+ for node_id in obj._backrefs['logged']['node']['logs']:
+ node = get_object_or_error(Node, node_id, display_name='node')
+ if ContributorOrPublic().has_object_permission(request, view, node):
+ return True
+
+ if getattr(obj, 'node'):
+ if ContributorOrPublic().has_object_permission(request, view, obj.node):
return True
+
return False |
66f6529880da4e8c444c2872820fb6b207c3794e | Server/main.py | Server/main.py | from flask import Flask
from flask_restful import Api
app = Flask(__name__)
api = Api(app)
debug = True
@app.before_first_request
def before_first_request():
pass
@app.before_request
def before_request():
pass
@app.after_request
def after_request(response):
# flask.wrapper.Response 클래스의 인스턴스
return response
@app.teardown_request
def teardown_request(exception):
pass
@app.teardown_appcontext
def teardown_appcontext(exception):
pass
@app.route('/')
def index():
return 'hello'
if __name__ == '__main__':
app.run(debug=debug)
| from flask import Flask
from flask_restful import Api
import logging
from logging.handlers import RotatingFileHandler
app = Flask(__name__)
api = Api(app)
debug = True
@app.before_first_request
def before_first_request():
handler = RotatingFileHandler('server_log.log', maxBytes=100000, backupCount=5)
handler.setLevel(logging.DEBUG if debug else logging.INFO)
formatter = logging.Formatter("[%(asctime)s] %(levelname)s - %(message)s")
handler.setFormatter(formatter)
app.logger.addHandler(handler)
app.logger.info('Logger started')
@app.before_request
def before_request():
pass
@app.after_request
def after_request(response):
# flask.wrapper.Response 클래스의 인스턴스
return response
@app.teardown_request
def teardown_request(exception):
pass
@app.teardown_appcontext
def teardown_appcontext(exception):
pass
@app.route('/')
def index():
return 'hello'
if __name__ == '__main__':
app.run(debug=debug)
| Add logging to before_first_request decorator | Add logging to before_first_request decorator
| Python | mit | DSM-GRAM/Artist,DSM-GRAM/Artist | ---
+++
@@ -1,6 +1,8 @@
from flask import Flask
from flask_restful import Api
+import logging
+from logging.handlers import RotatingFileHandler
app = Flask(__name__)
api = Api(app)
@@ -10,7 +12,14 @@
@app.before_first_request
def before_first_request():
- pass
+ handler = RotatingFileHandler('server_log.log', maxBytes=100000, backupCount=5)
+ handler.setLevel(logging.DEBUG if debug else logging.INFO)
+
+ formatter = logging.Formatter("[%(asctime)s] %(levelname)s - %(message)s")
+ handler.setFormatter(formatter)
+
+ app.logger.addHandler(handler)
+ app.logger.info('Logger started')
@app.before_request |
acec4dd403201dec5d22623c37ce1aff3324bc67 | drivnal/remote_snapshot.py | drivnal/remote_snapshot.py | from constants import *
from core_snapshot import CoreSnapshot
import logging
logger = logging.getLogger(APP_NAME)
class RemoteSnapshot(CoreSnapshot):
def _get_path(self):
return ''
def _get_log_path(self):
return ''
def _setup_snapshot(self, last_snapshot):
pass
def set_state(self, state):
if self.state == state:
return
self.state = state
| from constants import *
from core_snapshot import CoreSnapshot
import logging
logger = logging.getLogger(APP_NAME)
class RemoteSnapshot(CoreSnapshot):
def _get_path(self):
dir_name = str(self.id)
if self.state != COMPLETE:
dir_name = '%s.%s' % (dir_name, self.state)
return '%s@%s%s' % (self.volume.ssh_user, self.volume.ssh_path,
os.sep + os.path.join(SNAPSHOT_DIR, dir_name) + os.sep)
def _get_log_path(self):
return ''
def _setup_snapshot(self, last_snapshot):
pass
def set_state(self, state):
if self.state == state:
return
self.state = state
| Add get path for remote snapshot | Add get path for remote snapshot
| Python | agpl-3.0 | drivnal/drivnal,drivnal/drivnal,drivnal/drivnal | ---
+++
@@ -6,7 +6,11 @@
class RemoteSnapshot(CoreSnapshot):
def _get_path(self):
- return ''
+ dir_name = str(self.id)
+ if self.state != COMPLETE:
+ dir_name = '%s.%s' % (dir_name, self.state)
+ return '%s@%s%s' % (self.volume.ssh_user, self.volume.ssh_path,
+ os.sep + os.path.join(SNAPSHOT_DIR, dir_name) + os.sep)
def _get_log_path(self):
return '' |
5b1ab860a0706831b8abc77a060d6ba89cf8946a | interface/subprocess/001.backticks.py | interface/subprocess/001.backticks.py | import subprocess
# --- replacing shell backticks ---
# https://docs.python.org/2/library/subprocess.html#replacing-bin-sh-shell-backquote
# output=`mycmd myarg`
# output = check_output(["mycmd", "myarg"])
# not true, because mycmd is not passed to shell
try:
output = subprocess.check_output(["mycmd", "myarg"], shell=True)
except OSError as ex:
# command not found.
# it is impossible to catch output here, but shell outputs
# message to stderr, which backticks doesn't catch either
output = ''
except subprocess.CalledProcessError as ex:
output = ex.output
# ^ information about error condition is lost
# ^ output in case of OSError is lost
# ux notes:
# - `mycmd myarg` > ["mycmd", "myarg"]
# - `` is invisible
# subprocess.check_output is hardly rememberable
# - exception checking is excessive and not needed
# (common pattern is to check return code)
def backticks(command):
try:
# this doesn't escape shell patterns, such as:
# ^ (windows cmd.exe shell)
output = subprocess.check_output(command, shell=True)
except OSError as ex:
# command not found.
# it is impossible to catch output here, but shell outputs
# message to stderr, which backticks doesn't catch either
output = ''
except subprocess.CalledProcessError as ex:
output = ex.output
return output
| import subprocess
# --- replacing shell backticks ---
# https://docs.python.org/2/library/subprocess.html#replacing-bin-sh-shell-backquote
# output=`mycmd myarg`
# output = check_output(["mycmd", "myarg"])
# not true, because mycmd is not passed to shell
try:
output = subprocess.check_output(["mycmd", "myarg"], shell=True)
except OSError as ex:
# command not found.
# it is impossible to catch output here, but shell outputs
# message to stderr, which backticks doesn't catch either
output = ''
except subprocess.CalledProcessError as ex:
output = ex.output
# ^ information about error condition is lost
# ^ output in case of OSError is lost
# ux notes:
# - `mycmd myarg` > ["mycmd", "myarg"]
# - `` is invisible
# subprocess.check_output is hardly rememberable
# - exception checking is excessive and not needed
# (common pattern is to check return code)
def backticks(command):
'''
Execute `command and return output.
- no return code
- no stderr capture
- bailed out with MemoryError on Windows with 500Mb of output
'''
try:
# this doesn't escape shell patterns, such as:
# ^ (windows cmd.exe shell)
output = subprocess.check_output(command, shell=True)
except OSError as ex:
# command not found.
# it is impossible to catch output here, but shell outputs
# message to stderr, which backticks doesn't catch either
output = ''
except subprocess.CalledProcessError as ex:
output = ex.output
return output
| Add docs to backtics function | interface.subprocess: Add docs to backtics function
| Python | unlicense | techtonik/discovery,techtonik/discovery,techtonik/discovery | ---
+++
@@ -26,6 +26,12 @@
def backticks(command):
+ '''
+ Execute `command and return output.
+ - no return code
+ - no stderr capture
+ - bailed out with MemoryError on Windows with 500Mb of output
+ '''
try:
# this doesn't escape shell patterns, such as:
# ^ (windows cmd.exe shell) |
91bb9574ec760efd8aba2d9ae8fe67fe2e69d0a2 | jacquard/buckets/tests/test_bucket.py | jacquard/buckets/tests/test_bucket.py | import pytest
from jacquard.buckets.constants import NUM_BUCKETS
@pytest.mark.parametrize('divisor', (
2,
3,
4,
5,
6,
10,
100,
))
def test_divisible(divisor):
assert NUM_BUCKETS % divisor == 0
def test_at_least_three_buckets_per_percent():
assert NUM_BUCKETS / 100 >= 3
| import pytest
from jacquard.odm import Session
from jacquard.buckets import Bucket
from jacquard.buckets.constants import NUM_BUCKETS
@pytest.mark.parametrize('divisor', (
2,
3,
4,
5,
6,
10,
100,
))
def test_divisible(divisor):
assert NUM_BUCKETS % divisor == 0
def test_at_least_three_buckets_per_percent():
assert NUM_BUCKETS / 100 >= 3
def test_can_get_empty_bucket_from_old_format():
session = Session({'buckets/1': []})
bucket = session.get(Bucket, 1)
assert not bucket.needs_constraints()
| Add a test for getting an empty bucket | Add a test for getting an empty bucket
| Python | mit | prophile/jacquard,prophile/jacquard | ---
+++
@@ -1,5 +1,7 @@
import pytest
+from jacquard.odm import Session
+from jacquard.buckets import Bucket
from jacquard.buckets.constants import NUM_BUCKETS
@@ -18,3 +20,9 @@
def test_at_least_three_buckets_per_percent():
assert NUM_BUCKETS / 100 >= 3
+
+
+def test_can_get_empty_bucket_from_old_format():
+ session = Session({'buckets/1': []})
+ bucket = session.get(Bucket, 1)
+ assert not bucket.needs_constraints() |
6c4e94f1133c9c9cd18b97a386f04f56b229f9a8 | las_reader/las2excel.py | las_reader/las2excel.py | try:
import argparse
except ImportError:
argparse = None
import sys
import core
def main():
if argparse:
args = get_parser().parse_args(sys.argv[1:])
lasfn = args.las_filename
xlsfn = args.xls_filename
else:
if len(sys.argv >= 3):
lasfn = sys.argv[1]
xlsfn = sys.argv[2]
else:
print('Convert LAS file to Excel.\n\n'
'Usage:\n\n'
'las2excel.py example.las output.xls')
sys.exit(1)
las = core.LASFile(lasfn)
converter = core.ExcelConverter(las)
converter.write_excel(xlsfn)
def get_parser():
parser = argparse.ArgumentParser('Convert LAS file to Excel')
parser.add_argument('las-filename')
parser.add_argument('xls-filename')
return parser
if __name__ == '__main__':
main() | try:
import argparse
except ImportError:
argparse = None
import sys
import core
def main():
if argparse:
args = get_parser().parse_args(sys.argv[1:])
print args.__dict__.keys()
lasfn = args.las_filename
xlsfn = args.xls_filename
else:
if len(sys.argv >= 3):
lasfn = sys.argv[1]
xlsfn = sys.argv[2]
else:
print('Convert LAS file to Excel.\n\n'
'Usage:\n\n'
'las2excel.py example.las output.xls')
sys.exit(1)
las = core.LASFile(lasfn)
converter = core.ExcelConverter(las)
converter.write_excel(xlsfn)
def get_parser():
parser = argparse.ArgumentParser('Convert LAS file to Excel')
parser.add_argument('las_filename')
parser.add_argument('xls_filename')
return parser
if __name__ == '__main__':
main() | Fix Namespace for cmd line args | Fix Namespace for cmd line args
| Python | mit | kinverarity1/las-reader,Kramer477/lasio,kinverarity1/lasio,VelizarVESSELINOV/las-reader,kwinkunks/lasio | ---
+++
@@ -10,6 +10,7 @@
def main():
if argparse:
args = get_parser().parse_args(sys.argv[1:])
+ print args.__dict__.keys()
lasfn = args.las_filename
xlsfn = args.xls_filename
else:
@@ -29,8 +30,8 @@
def get_parser():
parser = argparse.ArgumentParser('Convert LAS file to Excel')
- parser.add_argument('las-filename')
- parser.add_argument('xls-filename')
+ parser.add_argument('las_filename')
+ parser.add_argument('xls_filename')
return parser
if __name__ == '__main__': |
ccb774b58ab7dbe704abfb7df3fa29915fad8f8f | examples/memnn/download.py | examples/memnn/download.py | #!/usr/bin/env python
from six.moves.urllib import request
def main():
opener = request.FancyURLopener()
opener.addheaders = [('User-Agent', '')]
opener.retrieve(
'http://www.thespermwhale.com/jaseweston/babi/tasks_1-20_v1-2.tar.gz',
'tasks_1-20_v1-2.tar.gz')
if __name__ == '__main__':
main()
| #!/usr/bin/env python
from six.moves.urllib import request
def main():
request.urlretrieve(
'http://www.thespermwhale.com/jaseweston/babi/tasks_1-20_v1-2.tar.gz',
'tasks_1-20_v1-2.tar.gz')
if __name__ == '__main__':
main()
| Replace deprecated URLopener in `donwload.py` | Replace deprecated URLopener in `donwload.py`
| Python | mit | niboshi/chainer,keisuke-umezawa/chainer,wkentaro/chainer,wkentaro/chainer,pfnet/chainer,keisuke-umezawa/chainer,wkentaro/chainer,niboshi/chainer,niboshi/chainer,okuta/chainer,okuta/chainer,chainer/chainer,hvy/chainer,chainer/chainer,keisuke-umezawa/chainer,wkentaro/chainer,okuta/chainer,keisuke-umezawa/chainer,hvy/chainer,hvy/chainer,tkerola/chainer,hvy/chainer,chainer/chainer,chainer/chainer,okuta/chainer,niboshi/chainer | ---
+++
@@ -4,9 +4,7 @@
def main():
- opener = request.FancyURLopener()
- opener.addheaders = [('User-Agent', '')]
- opener.retrieve(
+ request.urlretrieve(
'http://www.thespermwhale.com/jaseweston/babi/tasks_1-20_v1-2.tar.gz',
'tasks_1-20_v1-2.tar.gz')
|
1a1600b0cd27d5e004be344574901c64cdd6f7a2 | scripts/imgtool/__init__.py | scripts/imgtool/__init__.py | # Copyright 2017 Linaro Limited
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
imgtool_version = "1.6.0"
| # Copyright 2017-2020 Linaro Limited
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
imgtool_version = "1.7.0a1"
| Change imgtool version to 1.7.0a1 | Change imgtool version to 1.7.0a1
Signed-off-by: Ihor Slabkyy <5b878c9a28a92b9cb7e9988086921fcb7ae33592@cypress.com>
| Python | apache-2.0 | utzig/mcuboot,tamban01/mcuboot,utzig/mcuboot,runtimeco/mcuboot,runtimeco/mcuboot,runtimeco/mcuboot,tamban01/mcuboot,runtimeco/mcuboot,ATmobica/mcuboot,ATmobica/mcuboot,utzig/mcuboot,tamban01/mcuboot,tamban01/mcuboot,ATmobica/mcuboot,runtimeco/mcuboot,tamban01/mcuboot,utzig/mcuboot,utzig/mcuboot,ATmobica/mcuboot,ATmobica/mcuboot | ---
+++
@@ -1,4 +1,4 @@
-# Copyright 2017 Linaro Limited
+# Copyright 2017-2020 Linaro Limited
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
@@ -12,4 +12,4 @@
# See the License for the specific language governing permissions and
# limitations under the License.
-imgtool_version = "1.6.0"
+imgtool_version = "1.7.0a1" |
361ebc774fba5489c1911ac40dde4828f6cbd374 | flysight_manager/report.py | flysight_manager/report.py | #!/usr/bin/env python
import log
from jinja2 import Template
import traceback
class Report(object):
def __init__(self):
self.logs = log.LogAggregator.new()
def format_exception_as_reason(exc):
return traceback.format_exc(exc)
@log.make_loggable
class UploadReport(Report):
TEMPLATE_FILENAME = 'templates/uploader_report.jinja2'
def __init__(self, mailer, mail_cfg):
self.files = []
self.mailer = mailer
self.mail_cfg = mail_cfg
self.reason = None
super(UploadReport, self).__init__()
def add_uploaded_file(self, filename):
self.files.append(filename)
def finish_with_exception(self, exc):
reason = format_exception_as_reason(exc)
self.finish
def finish(self, reason):
self.reason = reason
def render(self):
tpl = Template(open(self.TEMPLATE_FILENAME).read())
return tpl.render(
reason=self.reason,
files=self.files,
logs=self.logs
)
def send(self):
content = self.render()
self.mailer.mail(
self.mail_cfg['to'],
self.mail_cfg['from'],
self.mail_cfg['subject'],
content)
| #!/usr/bin/env python
import log
import time
from jinja2 import Template
import traceback
class Report(object):
TIME_FMT = ": %y/%m/%d %H:%M %z (%Z)"
def __init__(self):
self.logs = log.LogAggregator.new()
self.started = time.strftime(TIME_FMT)
def format_exception_as_reason(exc):
return traceback.format_exc(exc)
@log.make_loggable
class UploadReport(Report):
TEMPLATE_FILENAME = 'templates/uploader_report.jinja2'
def __init__(self, mailer, mail_cfg):
self.files = []
self.mailer = mailer
self.mail_cfg = mail_cfg
self.reason = None
super(UploadReport, self).__init__()
def add_uploaded_file(self, filename):
self.files.append(filename)
def finish_with_exception(self, exc):
reason = format_exception_as_reason(exc)
self.finish
def finish(self, reason):
self.reason = reason
def render(self):
tpl = Template(open(self.TEMPLATE_FILENAME).read())
return tpl.render(
reason=self.reason,
files=self.files,
logs=self.logs
)
def send(self):
content = self.render()
self.mailer.mail(
self.mail_cfg['to'],
self.mail_cfg['from'],
self.mail_cfg['subject'] + self.started,
content)
| Include the time in the email | Include the time in the email
| Python | mit | richo/flysight-manager,richo/flysight-manager | ---
+++
@@ -1,12 +1,15 @@
#!/usr/bin/env python
import log
+import time
from jinja2 import Template
import traceback
class Report(object):
+ TIME_FMT = ": %y/%m/%d %H:%M %z (%Z)"
def __init__(self):
self.logs = log.LogAggregator.new()
+ self.started = time.strftime(TIME_FMT)
def format_exception_as_reason(exc):
return traceback.format_exc(exc)
@@ -48,5 +51,5 @@
self.mailer.mail(
self.mail_cfg['to'],
self.mail_cfg['from'],
- self.mail_cfg['subject'],
+ self.mail_cfg['subject'] + self.started,
content) |
e2919039a20a255232fefe3b78e173587710baf0 | cla_backend/apps/core/middleware.py | cla_backend/apps/core/middleware.py | from django.http import Http404
from django_statsd.clients import statsd
class GraphiteMiddleware(object):
def process_response(self, request, response):
statsd.incr("response.%s" % response.status_code)
return response
def process_exception(self, request, exception):
if not isinstance(exception, Http404):
statsd.incr("response.500")
| class GraphiteMiddleware(object):
def process_response(self, request, response):
return response
| Remove statsd code and resultant redundant code | Remove statsd code and resultant redundant code | Python | mit | ministryofjustice/cla_backend,ministryofjustice/cla_backend,ministryofjustice/cla_backend,ministryofjustice/cla_backend | ---
+++
@@ -1,12 +1,3 @@
-from django.http import Http404
-from django_statsd.clients import statsd
-
-
class GraphiteMiddleware(object):
def process_response(self, request, response):
- statsd.incr("response.%s" % response.status_code)
return response
-
- def process_exception(self, request, exception):
- if not isinstance(exception, Http404):
- statsd.incr("response.500") |
6784c455cf93c16237661d6d9fed6af06726a880 | conveyor/processor.py | conveyor/processor.py | from __future__ import absolute_import
from __future__ import division
import collections
from xmlrpc2 import client as xmlrpc2
class BaseProcessor(object):
def __init__(self, index, *args, **kwargs):
super(BaseProcessor, self).__init__(*args, **kwargs)
self.index = index
self.client = xmlrpc2.Client(self.index)
def process(self):
raise NotImplementedError
def get_releases(self, name, version=None):
if version is None:
versions = self.client.package_releases(name, True)
else:
versions = [version]
for version in versions:
item = self.client.release_data(name, version)
url = self.client.release_urls(item["name"], item["version"])
if isinstance(url, collections.Mapping):
urls = [url]
elif isinstance(url, collections.Iterable):
urls = url
else:
raise RuntimeError("Do not understand the type returned by release_urls")
item.update({"files": urls})
yield item
class BulkProcessor(BaseProcessor):
def process(self):
pass
| from __future__ import absolute_import
from __future__ import division
import collections
import slumber
import slumber.exceptions
import xmlrpc2.client
class BaseProcessor(object):
def __init__(self, index, warehouse, *args, **kwargs):
super(BaseProcessor, self).__init__(*args, **kwargs)
wargs, wkwargs = warehouse
self.client = xmlrpc2.client.Client(index)
self.warehouse = slumber.API(*wargs, **wkwargs)
def process(self):
raise NotImplementedError
def get_releases(self, name, version=None):
if version is None:
versions = self.client.package_releases(name, True)
else:
versions = [version]
for version in versions:
item = self.client.release_data(name, version)
url = self.client.release_urls(item["name"], item["version"])
if isinstance(url, collections.Mapping):
urls = [url]
elif isinstance(url, collections.Iterable):
urls = url
else:
raise RuntimeError("Do not understand the type returned by release_urls")
item.update({"files": urls})
yield item
class BulkProcessor(BaseProcessor):
def process(self):
pass
| Switch to more obvious imports | Switch to more obvious imports
| Python | bsd-2-clause | crateio/carrier | ---
+++
@@ -3,16 +3,20 @@
import collections
-from xmlrpc2 import client as xmlrpc2
+import slumber
+import slumber.exceptions
+import xmlrpc2.client
class BaseProcessor(object):
- def __init__(self, index, *args, **kwargs):
+ def __init__(self, index, warehouse, *args, **kwargs):
super(BaseProcessor, self).__init__(*args, **kwargs)
- self.index = index
- self.client = xmlrpc2.Client(self.index)
+ wargs, wkwargs = warehouse
+
+ self.client = xmlrpc2.client.Client(index)
+ self.warehouse = slumber.API(*wargs, **wkwargs)
def process(self):
raise NotImplementedError |
59becaccfc93e1d737be26483a18e9de1dfe1db1 | openedx/core/release.py | openedx/core/release.py | """
Information about the release line of this Open edX code.
"""
# The release line: an Open edX release name ("ficus"), or "master".
# This should always be "master" on the master branch, and will be changed
# manually when we start release-line branches, like open-release/ficus.master.
RELEASE_LINE = "hawthorn"
def doc_version():
"""The readthedocs.org version name used in documentation references.
Returns a short string like "latest" or "open-release-ficus.master".
"""
if RELEASE_LINE == "master":
return "latest"
else:
return "open-release-{}.master".format(RELEASE_LINE)
| """
Information about the release line of this Open edX code.
"""
# The release line: an Open edX release name ("ficus"), or "master".
# This should always be "master" on the master branch, and will be changed
# manually when we start release-line branches, like open-release/ficus.master.
RELEASE_LINE = "master"
def doc_version():
"""The readthedocs.org version name used in documentation references.
Returns a short string like "latest" or "open-release-ficus.master".
"""
if RELEASE_LINE == "master":
return "latest"
else:
return "open-release-{}.master".format(RELEASE_LINE)
| Revert "Set the RELEASE_LINE to hawthorn" | Revert "Set the RELEASE_LINE to hawthorn"
This reverts commit f44ac32bc9060cfa5ea6ce4284ce7f15b466be28.
| Python | agpl-3.0 | appsembler/edx-platform,appsembler/edx-platform,appsembler/edx-platform,appsembler/edx-platform | ---
+++
@@ -5,7 +5,7 @@
# The release line: an Open edX release name ("ficus"), or "master".
# This should always be "master" on the master branch, and will be changed
# manually when we start release-line branches, like open-release/ficus.master.
-RELEASE_LINE = "hawthorn"
+RELEASE_LINE = "master"
def doc_version(): |
2f56f7dccbc3c9fc416200160bd8616a5e4ab954 | barf/barf/__init__.py | barf/barf/__init__.py | # Copyright (c) 2014, Fundacion Dr. Manuel Sadosky
# All rights reserved.
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
# 1. Redistributions of source code must retain the above copyright notice, this
# list of conditions and the following disclaimer.
# 2. Redistributions in binary form must reproduce the above copyright notice,
# this list of conditions and the following disclaimer in the documentation
# and/or other materials provided with the distribution.
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
import logging
from barf import BARF
import analysis
import core
import arch
# Setup logging module.
logging.basicConfig(
filename="barf.log",
format="%(asctime)s: %(name)s:%(levelname)s: %(message)s",
level=logging.DEBUG
)
| # Copyright (c) 2014, Fundacion Dr. Manuel Sadosky
# All rights reserved.
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
# 1. Redistributions of source code must retain the above copyright notice, this
# list of conditions and the following disclaimer.
# 2. Redistributions in binary form must reproduce the above copyright notice,
# this list of conditions and the following disclaimer in the documentation
# and/or other materials provided with the distribution.
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
import logging
from barf import BARF
import analysis
import core
import arch
# Setup logging module.
logging.basicConfig(
filename="barf.log",
format="%(asctime)s: %(name)s:%(levelname)s: %(message)s",
filemode='w',
level=logging.DEBUG
)
| Set logging in overwrite mode | Set logging in overwrite mode
| Python | bsd-2-clause | cnheitman/barf-project,chubbymaggie/barf-project,programa-stic/barf-project,programa-stic/barf-project,cnheitman/barf-project,chubbymaggie/barf-project,chubbymaggie/barf-project,cnheitman/barf-project | ---
+++
@@ -34,5 +34,6 @@
logging.basicConfig(
filename="barf.log",
format="%(asctime)s: %(name)s:%(levelname)s: %(message)s",
+ filemode='w',
level=logging.DEBUG
) |
1d63f615ac58cc8c548cdd8e359694355e5b1843 | portal/forms.py | portal/forms.py | from django.contrib.auth.models import User
from django import forms
# Create your forms here.
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput)
class Meta:
model = User
fields = ['username', 'email', 'password']
| from django.contrib.auth.models import User
from django import forms
# Create your forms here.
class BootstrapForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
for visible in self.visible_fields():
visible.field.widget.attrs['class'] = 'form-control'
visible.field.widget.attrs['placeholder'] = 'Please enter value'
class UserForm(BootstrapForm):
password = forms.CharField(widget=forms.PasswordInput)
class Meta:
model = User
fields = ['username', 'email', 'password']
| Add BootstrapForm to beautify form_template | Add BootstrapForm to beautify form_template
| Python | mit | huangsam/chowist,huangsam/chowist,huangsam/chowist | ---
+++
@@ -3,7 +3,16 @@
# Create your forms here.
-class UserForm(forms.ModelForm):
+class BootstrapForm(forms.ModelForm):
+
+ def __init__(self, *args, **kwargs):
+ super().__init__(*args, **kwargs)
+ for visible in self.visible_fields():
+ visible.field.widget.attrs['class'] = 'form-control'
+ visible.field.widget.attrs['placeholder'] = 'Please enter value'
+
+
+class UserForm(BootstrapForm):
password = forms.CharField(widget=forms.PasswordInput)
class Meta: |
d0b9824fe29e9c5772c0fc2838e2c6e373013819 | portal/forms.py | portal/forms.py | from django import forms
from django.contrib.auth.models import User
from portal.models import Profile
# Create your forms here.
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput)
class Meta:
model = User
fields = ['username', 'email', 'password']
class ProfileForm(forms.ModelForm):
class Meta:
model = Profile
fields = ['user', 'bio', 'address', 'birth_date']
| from django import forms
from django.contrib.auth.models import User
from portal.models import Profile
# Create your forms here.
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput)
class Meta:
model = User
fields = ['username', 'email', 'password']
class ProfileForm(forms.ModelForm):
class Meta:
model = Profile
fields = ['user', 'bio', 'address', 'birth_date']
| Fix indents on form module | Fix indents on form module
| Python | mit | huangsam/chowist,huangsam/chowist,huangsam/chowist | ---
+++
@@ -12,6 +12,7 @@
model = User
fields = ['username', 'email', 'password']
+
class ProfileForm(forms.ModelForm):
class Meta:
model = Profile |
6b880f3c783e6a278906b8da2aabea29bb106252 | thinc/neural/_classes/resnet.py | thinc/neural/_classes/resnet.py | from .model import Model
from ...api import layerize
from .affine import Affine
import cytoolz as toolz
def Residual(layer):
def residual_fwd(X, drop=0.):
y, bp_y = layer.begin_update(X, drop=drop)
output = X+y
def residual_bwd(d_output, sgd=None):
return d_output + bp_y(d_output, sgd)
return output, residual_bwd
model = layerize(residual_fwd)
model._layers.append(layer)
def on_data(self, X, y=None):
for layer in self._layers:
for hook in layer.on_data_hooks:
hook(layer, X, y)
model.on_data_hooks.append(on_data)
return model
| from .model import Model
from ...api import layerize
from .affine import Affine
import cytoolz as toolz
class Residual(Model):
def __init__(self, layer):
Model.__init__(self)
self._layers.append(layer)
self.on_data_hooks.append(on_data)
def __call__(self, X):
return X + self._layers[0](X)
def begin_update(self, X, drop=0.):
y, bp_y = self._layer[0].begin_update(X, drop=drop)
output = X+y
def residual_bwd(d_output, sgd=None):
return d_output + bp_y(d_output, sgd)
return output, residual_bwd
def on_data(self, X, y=None):
for layer in self._layers:
for hook in layer.on_data_hooks:
hook(layer, X, y)
if hasattr(layer, 'W'):
layer.W.fill(0)
| Add predict path for Residual | Add predict path for Residual
| Python | mit | spacy-io/thinc,spacy-io/thinc,spacy-io/thinc,explosion/thinc,explosion/thinc,explosion/thinc,explosion/thinc | ---
+++
@@ -5,18 +5,25 @@
import cytoolz as toolz
-def Residual(layer):
- def residual_fwd(X, drop=0.):
- y, bp_y = layer.begin_update(X, drop=drop)
+class Residual(Model):
+ def __init__(self, layer):
+ Model.__init__(self)
+ self._layers.append(layer)
+ self.on_data_hooks.append(on_data)
+
+ def __call__(self, X):
+ return X + self._layers[0](X)
+
+ def begin_update(self, X, drop=0.):
+ y, bp_y = self._layer[0].begin_update(X, drop=drop)
output = X+y
def residual_bwd(d_output, sgd=None):
return d_output + bp_y(d_output, sgd)
return output, residual_bwd
- model = layerize(residual_fwd)
- model._layers.append(layer)
- def on_data(self, X, y=None):
- for layer in self._layers:
- for hook in layer.on_data_hooks:
- hook(layer, X, y)
- model.on_data_hooks.append(on_data)
- return model
+
+def on_data(self, X, y=None):
+ for layer in self._layers:
+ for hook in layer.on_data_hooks:
+ hook(layer, X, y)
+ if hasattr(layer, 'W'):
+ layer.W.fill(0) |
105ac0020dbc60fe57da7db75fb82cf872a0834d | crm_switzerland/models/res_partner.py | crm_switzerland/models/res_partner.py | # -*- coding: utf-8 -*-
##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Emanuel Cino <ecino@compassion.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import api, models
class ResPartner(models.Model):
_inherit = 'res.partner'
@api.multi
def schedule_meeting(self):
old_action = super(ResPartner, self).schedule_meeting()
new_action = self.env.ref(
'crm_switzerland.action_calendar_event_partner').read()[0]
new_action['domain'] = [('partner_ids', 'in', self.ids)]
new_action['context'] = {
'default_partner_ids': old_action['context'][
'default_partner_ids']
}
return new_action
@api.model
def _notify_prepare_template_context(self, message):
# modification of context for lang
message = message.with_context(lang=self.lang)
return super(ResPartner, self).\
_notify_prepare_template_context(message)
| # -*- coding: utf-8 -*-
##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Emanuel Cino <ecino@compassion.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import api, models
class ResPartner(models.Model):
_inherit = 'res.partner'
@api.multi
def schedule_meeting(self):
old_action = super(ResPartner, self).schedule_meeting()
new_action = self.env.ref(
'crm_switzerland.action_calendar_event_partner').read()[0]
new_action['domain'] = [('partner_ids', 'in', self.ids)]
new_action['context'] = {
'default_partner_ids': old_action['context'][
'default_partner_ids']
}
return new_action
@api.model
def _notify_prepare_template_context(self, message):
# modification of context for lang
message = message.with_context(lang=self[:1].lang or self.env.lang)
return super(ResPartner, self).\
_notify_prepare_template_context(message)
| FIX bug when sending notification to multiple partners | FIX bug when sending notification to multiple partners
| Python | agpl-3.0 | ecino/compassion-switzerland,eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,ecino/compassion-switzerland,eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,ecino/compassion-switzerland | ---
+++
@@ -30,6 +30,6 @@
@api.model
def _notify_prepare_template_context(self, message):
# modification of context for lang
- message = message.with_context(lang=self.lang)
+ message = message.with_context(lang=self[:1].lang or self.env.lang)
return super(ResPartner, self).\
_notify_prepare_template_context(message) |
bb8d2fa458dd565b88db4e2185062f641864e990 | tornado/test/httpserver_test.py | tornado/test/httpserver_test.py | #!/usr/bin/env python
from tornado.testing import AsyncHTTPTestCase, LogTrapTestCase
from tornado.web import Application, RequestHandler
import os
import pycurl
import re
import unittest
import urllib
class HelloWorldRequestHandler(RequestHandler):
def get(self):
self.finish("Hello world")
class SSLTest(AsyncHTTPTestCase, LogTrapTestCase):
def get_app(self):
return Application([('/', HelloWorldRequestHandler)])
def get_httpserver_options(self):
# Testing keys were generated with:
# openssl req -new -keyout tornado/test/test.key -out tornado/test/test.crt -nodes -days 3650 -x509
test_dir = os.path.dirname(__file__)
return dict(ssl_options=dict(
certfile=os.path.join(test_dir, 'test.crt'),
keyfile=os.path.join(test_dir, 'test.key')))
def test_ssl(self):
def disable_cert_check(curl):
# Our certificate was not signed by a CA, so don't check it
curl.setopt(pycurl.SSL_VERIFYPEER, 0)
self.http_client.fetch(self.get_url('/').replace('http', 'https'),
self.stop,
prepare_curl_callback=disable_cert_check)
response = self.wait()
self.assertEqual(response.body, "Hello world")
| #!/usr/bin/env python
from tornado.testing import AsyncHTTPTestCase, LogTrapTestCase
from tornado.web import Application, RequestHandler
import os
import pycurl
import re
import unittest
import urllib
try:
import ssl
except ImportError:
ssl = None
class HelloWorldRequestHandler(RequestHandler):
def get(self):
self.finish("Hello world")
class SSLTest(AsyncHTTPTestCase, LogTrapTestCase):
def get_app(self):
return Application([('/', HelloWorldRequestHandler)])
def get_httpserver_options(self):
# Testing keys were generated with:
# openssl req -new -keyout tornado/test/test.key -out tornado/test/test.crt -nodes -days 3650 -x509
test_dir = os.path.dirname(__file__)
return dict(ssl_options=dict(
certfile=os.path.join(test_dir, 'test.crt'),
keyfile=os.path.join(test_dir, 'test.key')))
def test_ssl(self):
def disable_cert_check(curl):
# Our certificate was not signed by a CA, so don't check it
curl.setopt(pycurl.SSL_VERIFYPEER, 0)
self.http_client.fetch(self.get_url('/').replace('http', 'https'),
self.stop,
prepare_curl_callback=disable_cert_check)
response = self.wait()
self.assertEqual(response.body, "Hello world")
if ssl is None:
# Don't try to run ssl tests if we don't have the ssl module
del SSLTest
| Disable SSL test on python 2.5 | Disable SSL test on python 2.5
| Python | apache-2.0 | bywbilly/tornado,AlphaStaxLLC/tornado,felixonmars/tornado,jarrahwu/tornado,BencoLee/tornado,MjAbuz/tornado,shaohung001/tornado,sunjeammy/tornado,zhuochenKIDD/tornado,VShangxiao/tornado,LTD-Beget/tornado,Snamint/tornado,304471720/tornado,Callwoola/tornado,shashankbassi92/tornado,InverseLina/tornado,Batterfii/tornado,cyrusin/tornado,gwillem/tornado,bywbilly/tornado,erichuang1994/tornado,AlphaStaxLLC/tornado,liqueur/tornado,mlyundin/tornado,zguangyu/tornado,fengsp/tornado,0x73/tornado,ColorFuzzy/tornado,ms7s/tornado,shashankbassi92/tornado,ListFranz/tornado,bywbilly/tornado,BencoLee/tornado,wsyzxcn/tornado,ms7s/tornado,mivade/tornado,eXcomm/tornado,kippandrew/tornado,nbargnesi/tornado,304471720/tornado,eXcomm/tornado,yuyangit/tornado,mehmetkose/tornado,nephics/tornado,leekchan/tornado_test,noxiouz/tornado,jonashagstedt/tornado,Lancher/tornado,pombredanne/tornado,erichuang1994/tornado,304471720/tornado,z-fork/tornado,ubear/tornado,kevinge314gh/tornado,ovidiucp/tornado,wechasing/tornado,hhru/tornado,jarrahwu/tornado,coderhaoxin/tornado,VShangxiao/tornado,304471720/tornado,fengsp/tornado,yuezhonghua/tornado,elelianghh/tornado,kaushik94/tornado,erichuang1994/tornado,hzruandd/tornado,erichuang1994/tornado,MjAbuz/tornado,Fydot/tornado,nephics/tornado,andyaguiar/tornado,Aaron1992/tornado,dongpinglai/my_tornado,legnaleurc/tornado,chenxiaba/tornado,arthurdarcet/tornado,ColorFuzzy/tornado,ms7s/tornado,yuyangit/tornado,elelianghh/tornado,jsjohnst/tornado,futurechallenger/tornado,gwillem/tornado,Drooids/tornado,codeb2cc/tornado,leekchan/tornado_test,mlyundin/tornado,allenl203/tornado,mivade/tornado,akalipetis/tornado,elijah513/tornado,ovidiucp/tornado,chenxiaba/tornado,noxiouz/tornado,shaohung001/tornado,ydaniv/tornado,jsjohnst/tornado,zguangyu/tornado,hzruandd/tornado,dongpinglai/my_tornado,kippandrew/tornado,johan--/tornado,lujinda/tornado,lujinda/tornado,jarrahwu/tornado,Geoion/tornado,hhru/tornado,elijah513/tornado,LTD-Beget/tornado,ajdavis/tornado,xinyu7/tornado,icejoywoo/tornado,Snamint/tornado,whip112/tornado,Acidburn0zzz/tornado,xinyu7/tornado,shashankbassi92/tornado,djt5019/tornado,Windsooon/tornado,kangbiao/tornado,ajdavis/tornado,Lancher/tornado,Snamint/tornado,tianyk/tornado-research,jehiah/tornado,yangkf1985/tornado,hzruandd/tornado,nordaux/tornado,jparise/tornado,BencoLee/tornado,felixonmars/tornado,nbargnesi/tornado,BencoLee/tornado,zguangyu/tornado,ubear/tornado,zhuochenKIDD/tornado,mehmetkose/tornado,fengshao0907/tornado,wujuguang/tornado,Fydot/tornado,jarrahwu/tornado,liqueur/tornado,hzruandd/tornado,ZhuPeng/tornado,jehiah/tornado,codecov/tornado,VShangxiao/tornado,wujuguang/tornado,sevenguin/tornado,z-fork/tornado,eXcomm/tornado,obsh/tornado,Polyconseil/tornado,tianyk/tornado-research,leekchan/tornado_test,ovidiucp/tornado,djt5019/tornado,ymero/tornado,takeshineshiro/tornado,tornadoweb/tornado,johan--/tornado,lsanotes/tornado,mehmetkose/tornado,johan--/tornado,noxiouz/tornado,futurechallenger/tornado,bufferx/tornado,elelianghh/tornado,gwillem/tornado,gitchs/tornado,bdarnell/tornado,lilydjwg/tornado,gitchs/tornado,futurechallenger/tornado,Windsooon/tornado,kaushik94/tornado,mivade/tornado,ajdavis/tornado,liqueur/tornado,AlphaStaxLLC/tornado,jparise/tornado,wsyzxcn/tornado,yuezhonghua/tornado,z-fork/tornado,djt5019/tornado,MjAbuz/tornado,gwillem/tornado,dongpinglai/my_tornado,Aaron1992/tornado,wxhzk/tornado-1,ydaniv/tornado,ColorFuzzy/tornado,sxfmol/tornado,hhru/tornado,Snamint/tornado,LTD-Beget/tornado,sevenguin/tornado,Windsooon/tornado,ydaniv/tornado,zhuochenKIDD/tornado,kevinge314gh/tornado,NoyaInRain/tornado,cyrilMargaria/tornado,ovidiucp/tornado,jonashagstedt/tornado,zhuochenKIDD/tornado,elijah513/tornado,Windsooon/tornado,BencoLee/tornado,Polyconseil/tornado,mr-ping/tornado,Acidburn0zzz/tornado,QuanZag/tornado,frtmelody/tornado,mlyundin/tornado,codeb2cc/tornado,liqueur/tornado,coderhaoxin/tornado,legnaleurc/tornado,kippandrew/tornado,ListFranz/tornado,pombredanne/tornado,fengshao0907/tornado,kevinge314gh/tornado,NoyaInRain/tornado,gitchs/tornado,akalipetis/tornado,sunjeammy/tornado,wxhzk/tornado-1,leekchan/tornado_test,QuanZag/tornado,gitchs/tornado,dsseter/tornado,fengsp/tornado,AlphaStaxLLC/tornado,codecov/tornado,nbargnesi/tornado,futurechallenger/tornado,cyrilMargaria/tornado,NoyaInRain/tornado,fengshao0907/tornado,anandology/tornado,mr-ping/tornado,lujinda/tornado,leekchan/tornado_test,SuminAndrew/tornado,anandology/tornado,0xkag/tornado,mr-ping/tornado,frtmelody/tornado,icejoywoo/tornado,QuanZag/tornado,wsyzxcn/tornado,eklitzke/tornado,elijah513/tornado,insflow/tornado,hzruandd/tornado,anjan-srivastava/tornado,jehiah/tornado,Aaron1992/tornado,dongpinglai/my_tornado,Geoion/tornado,arthurdarcet/tornado,arthurdarcet/tornado,kangbiao/tornado,Batterfii/tornado,noxiouz/tornado,lilydjwg/tornado,nordaux/tornado,codeb2cc/tornado,arthurdarcet/tornado,Lancher/tornado,insflow/tornado,Callwoola/tornado,anjan-srivastava/tornado,jampp/tornado,InverseLina/tornado,Acidburn0zzz/tornado,andyaguiar/tornado,mehmetkose/tornado,mr-ping/tornado,tornadoweb/tornado,kangbiao/tornado,dongpinglai/my_tornado,304471720/tornado,Polyconseil/tornado,shaohung001/tornado,SuminAndrew/tornado,elijah513/tornado,mivade/tornado,wechasing/tornado,yangkf1985/tornado,NoyaInRain/tornado,xinyu7/tornado,jonashagstedt/tornado,hzruandd/tornado,drewmiller/tornado,obsh/tornado,jparise/tornado,nephics/tornado,andyaguiar/tornado,ymero/tornado,lilydjwg/tornado,ymero/tornado,lujinda/tornado,coderhaoxin/tornado,nbargnesi/tornado,whip112/tornado,lsanotes/tornado,jarrahwu/tornado,ifduyue/tornado,NoyaInRain/tornado,sevenguin/tornado,0xkag/tornado,allenl203/tornado,codeb2cc/tornado,jampp/tornado,lsanotes/tornado,anandology/tornado,xinyu7/tornado,bdarnell/tornado,mehmetkose/tornado,allenl203/tornado,noxiouz/tornado,lsanotes/tornado,sxfmol/tornado,frtmelody/tornado,Polyconseil/tornado,wxhzk/tornado-1,drewmiller/tornado,Snamint/tornado,z-fork/tornado,dsseter/tornado,zguangyu/tornado,bywbilly/tornado,whip112/tornado,Batterfii/tornado,cyrusin/tornado,legnaleurc/tornado,wsyzxcn/tornado,pombredanne/tornado,obsh/tornado,Polyconseil/tornado,xinyu7/tornado,0x73/tornado,wxhzk/tornado-1,ifduyue/tornado,QuanZag/tornado,mlyundin/tornado,cyrilMargaria/tornado,liqueur/tornado,eklitzke/tornado,eXcomm/tornado,ListFranz/tornado,tianyk/tornado-research,andyaguiar/tornado,Batterfii/tornado,ovidiucp/tornado,takeshineshiro/tornado,chenxiaba/tornado,Snamint/tornado,felixonmars/tornado,0xkag/tornado,bufferx/tornado,ColorFuzzy/tornado,ms7s/tornado,wxhzk/tornado-1,wujuguang/tornado,jarrahwu/tornado,Batterfii/tornado,chenxiaba/tornado,SuminAndrew/tornado,yuezhonghua/tornado,legnaleurc/tornado,eXcomm/tornado,chenxiaba/tornado,codecov/tornado,whip112/tornado,sxfmol/tornado,whip112/tornado,LTD-Beget/tornado,insflow/tornado,elelianghh/tornado,Drooids/tornado,NoyaInRain/tornado,eklitzke/tornado,takeshineshiro/tornado,Geoion/tornado,ydaniv/tornado,sevenguin/tornado,0xkag/tornado,Drooids/tornado,mlyundin/tornado,zhuochenKIDD/tornado,chenxiaba/tornado,kippandrew/tornado,mehmetkose/tornado,futurechallenger/tornado,hhru/tornado,anjan-srivastava/tornado,djt5019/tornado,ubear/tornado,arthurdarcet/tornado,nbargnesi/tornado,takeshineshiro/tornado,dsseter/tornado,Fydot/tornado,coderhaoxin/tornado,icejoywoo/tornado,cyrilMargaria/tornado,cyrusin/tornado,fengshao0907/tornado,frtmelody/tornado,felixonmars/tornado,VShangxiao/tornado,Drooids/tornado,drewmiller/tornado,akalipetis/tornado,kangbiao/tornado,Fydot/tornado,jonashagstedt/tornado,bdarnell/tornado,coderhaoxin/tornado,SuminAndrew/tornado,ubear/tornado,InverseLina/tornado,bywbilly/tornado,0x73/tornado,jsjohnst/tornado,Geoion/tornado,icejoywoo/tornado,lujinda/tornado,Acidburn0zzz/tornado,anandology/tornado,anjan-srivastava/tornado,shashankbassi92/tornado,kevinge314gh/tornado,anandology/tornado,gitchs/tornado,tornadoweb/tornado,LTD-Beget/tornado,gitchs/tornado,zguangyu/tornado,insflow/tornado,zguangyu/tornado,elelianghh/tornado,Fydot/tornado,zhuochenKIDD/tornado,ColorFuzzy/tornado,Lancher/tornado,dsseter/tornado,z-fork/tornado,obsh/tornado,elelianghh/tornado,sxfmol/tornado,sunjeammy/tornado,icejoywoo/tornado,allenl203/tornado,johan--/tornado,lilydjwg/tornado,sunjeammy/tornado,akalipetis/tornado,Callwoola/tornado,drewmiller/tornado,yangkf1985/tornado,ifduyue/tornado,cyrilMargaria/tornado,akalipetis/tornado,fengsp/tornado,ymero/tornado,codeb2cc/tornado,wujuguang/tornado,SuminAndrew/tornado,jehiah/tornado,anjan-srivastava/tornado,Drooids/tornado,Batterfii/tornado,mivade/tornado,legnaleurc/tornado,ifduyue/tornado,ZhuPeng/tornado,jparise/tornado,304471720/tornado,wechasing/tornado,ymero/tornado,andyaguiar/tornado,ifduyue/tornado,wechasing/tornado,jparise/tornado,bufferx/tornado,erichuang1994/tornado,ms7s/tornado,eklitzke/tornado,bufferx/tornado,InverseLina/tornado,importcjj/tornado,VShangxiao/tornado,MjAbuz/tornado,yuyangit/tornado,bdarnell/tornado,nephics/tornado,anandology/tornado,sunjeammy/tornado,yangkf1985/tornado,nephics/tornado,Callwoola/tornado,importcjj/tornado,nordaux/tornado,InverseLina/tornado,allenl203/tornado,sevenguin/tornado,cyrusin/tornado,z-fork/tornado,Acidburn0zzz/tornado,shashankbassi92/tornado,nbargnesi/tornado,tianyk/tornado-research,eXcomm/tornado,ajdavis/tornado,kaushik94/tornado,insflow/tornado,ymero/tornado,ydaniv/tornado,importcjj/tornado,takeshineshiro/tornado,whip112/tornado,hhru/tornado,obsh/tornado,wechasing/tornado,cyrilMargaria/tornado,fengsp/tornado,MjAbuz/tornado,QuanZag/tornado,Drooids/tornado,futurechallenger/tornado,VShangxiao/tornado,mr-ping/tornado,lujinda/tornado,dsseter/tornado,sxfmol/tornado,codeb2cc/tornado,wujuguang/tornado,bdarnell/tornado,kippandrew/tornado,yuezhonghua/tornado,takeshineshiro/tornado,jsjohnst/tornado,kangbiao/tornado,AlphaStaxLLC/tornado,noxiouz/tornado,mr-ping/tornado,LTD-Beget/tornado,shaohung001/tornado,gwillem/tornado,drewmiller/tornado,jampp/tornado,dongpinglai/my_tornado,kangbiao/tornado,ColorFuzzy/tornado,kippandrew/tornado,jampp/tornado,frtmelody/tornado,fengshao0907/tornado,liqueur/tornado,wsyzxcn/tornado,kevinge314gh/tornado,ZhuPeng/tornado,kaushik94/tornado,ListFranz/tornado,ZhuPeng/tornado,cyrusin/tornado,djt5019/tornado,sevenguin/tornado,Lancher/tornado,jparise/tornado,xinyu7/tornado,pombredanne/tornado,wxhzk/tornado-1,lsanotes/tornado,yangkf1985/tornado,importcjj/tornado,elijah513/tornado,shaohung001/tornado,0x73/tornado,Windsooon/tornado,MjAbuz/tornado,ajdavis/tornado,ZhuPeng/tornado,fengsp/tornado,erichuang1994/tornado,arthurdarcet/tornado,QuanZag/tornado,gwillem/tornado,yuyangit/tornado,lsanotes/tornado,InverseLina/tornado,wsyzxcn/tornado,coderhaoxin/tornado,Callwoola/tornado,johan--/tornado,kaushik94/tornado,sxfmol/tornado,ZhuPeng/tornado,0xkag/tornado,jampp/tornado,obsh/tornado,Geoion/tornado,importcjj/tornado,ListFranz/tornado,jsjohnst/tornado,yuezhonghua/tornado,pombredanne/tornado,jehiah/tornado,Fydot/tornado,BencoLee/tornado,icejoywoo/tornado,importcjj/tornado,fengshao0907/tornado,ubear/tornado,tianyk/tornado-research,Aaron1992/tornado,bywbilly/tornado,shashankbassi92/tornado,drewmiller/tornado,ms7s/tornado,Polyconseil/tornado,0x73/tornado,kevinge314gh/tornado,dsseter/tornado,pombredanne/tornado,akalipetis/tornado,frtmelody/tornado,felixonmars/tornado,jsjohnst/tornado,tornadoweb/tornado,Callwoola/tornado,wsyzxcn/tornado,jampp/tornado,Acidburn0zzz/tornado,nordaux/tornado,djt5019/tornado,Windsooon/tornado,nordaux/tornado,anjan-srivastava/tornado,eklitzke/tornado,johan--/tornado,ListFranz/tornado,yuyangit/tornado,Geoion/tornado,cyrusin/tornado,jonashagstedt/tornado,mlyundin/tornado,shaohung001/tornado,insflow/tornado,ubear/tornado,Aaron1992/tornado,bufferx/tornado,yangkf1985/tornado,yuezhonghua/tornado,ydaniv/tornado,wechasing/tornado,AlphaStaxLLC/tornado,andyaguiar/tornado,codecov/tornado,ovidiucp/tornado | ---
+++
@@ -7,6 +7,11 @@
import re
import unittest
import urllib
+
+try:
+ import ssl
+except ImportError:
+ ssl = None
class HelloWorldRequestHandler(RequestHandler):
def get(self):
@@ -33,3 +38,7 @@
prepare_curl_callback=disable_cert_check)
response = self.wait()
self.assertEqual(response.body, "Hello world")
+
+if ssl is None:
+ # Don't try to run ssl tests if we don't have the ssl module
+ del SSLTest |
39cc30f2f6c74d3a506c5d1a46cf0ccc6377b80f | pylibscrypt/__init__.py | pylibscrypt/__init__.py |
# First, try loading libscrypt
_done = False
try:
from pylibscrypt import *
except ImportError:
pass
else:
_done = True
# If that didn't work, get the inlined Python version
if not _done:
try:
from pypyscrypt_inline import *
except ImportError:
pass
else:
_done = True
# Finally the non-inlined
if not _done:
from pypyscrypt import *
|
# First, try loading libscrypt
_done = False
try:
from pylibscrypt import *
except ImportError:
pass
else:
_done = True
# If that didn't work, try the scrypt module
if not _done:
try:
from pyscrypt import *
except ImportError:
pass
else:
_done = True
# If that didn't work either, the inlined Python version
if not _done:
try:
from pypyscrypt_inline import *
except ImportError:
pass
else:
_done = True
# Finally the non-inlined
if not _done:
from pypyscrypt import *
| Use pyscrypt.py in package import if libscrypt isn't available | Use pyscrypt.py in package import if libscrypt isn't available
| Python | isc | jvarho/pylibscrypt,jvarho/pylibscrypt | ---
+++
@@ -8,7 +8,16 @@
else:
_done = True
-# If that didn't work, get the inlined Python version
+# If that didn't work, try the scrypt module
+if not _done:
+ try:
+ from pyscrypt import *
+ except ImportError:
+ pass
+ else:
+ _done = True
+
+# If that didn't work either, the inlined Python version
if not _done:
try:
from pypyscrypt_inline import * |
00aa59468c4dbfde282891f1396e29bd3f28fb62 | gunny/reveille/service.py | gunny/reveille/service.py | from twisted.application import internet
from twisted.application.service import Service
from twisted.internet import reactor
from autobahn.websocket import connectWS
class ControlService(Service):
pass
class PlayerService(Service):
def __init__(self, factory):
self.factory = factory
self.conn = None
def startService(self):
self.factory.startFactory()
self.conn = connectWS(self.factory)
self.running = 1
def stopService(self):
self.factory.stopFactory()
if self.conn is not None:
self.conn.disconnect()
self.running = 0
| from twisted.application.service import Service
from twisted.internet import stdio
from autobahn.websocket import connectWS
class CoxswainService(Service):
def __init__(self, factory):
self.factory = factory
self.conn = None
def startService(self):
#self.factory(ReveilleCommandProtocol())
self.conn = connectWS(self.factory)
self.running = True
def stopService(self):
self.factory.stopFactory()
if self.conn is not None:
self.conn.disconnect()
self.running = False
| Rename classes to reflect intended use. | Rename classes to reflect intended use.
| Python | bsd-2-clause | davidblewett/gunny,davidblewett/gunny | ---
+++
@@ -1,27 +1,22 @@
-from twisted.application import internet
from twisted.application.service import Service
-from twisted.internet import reactor
+from twisted.internet import stdio
from autobahn.websocket import connectWS
-class ControlService(Service):
- pass
-
-
-class PlayerService(Service):
+class CoxswainService(Service):
def __init__(self, factory):
self.factory = factory
self.conn = None
def startService(self):
- self.factory.startFactory()
+ #self.factory(ReveilleCommandProtocol())
self.conn = connectWS(self.factory)
- self.running = 1
+ self.running = True
def stopService(self):
self.factory.stopFactory()
if self.conn is not None:
self.conn.disconnect()
- self.running = 0
+ self.running = False |
7df1ed120281c82d166fa1c2218def4c84b48a3d | alfred_db/migrations/versions/30c0aec2ca06_improve_repository_o.py | alfred_db/migrations/versions/30c0aec2ca06_improve_repository_o.py | """Improve repository owner information
Revision ID: 30c0aec2ca06
Revises: 4fdf1059c4ba
Create Date: 2012-09-02 14:45:05.241933
"""
# revision identifiers, used by Alembic.
revision = '30c0aec2ca06'
down_revision = '4fdf1059c4ba'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.add_column(
'repositories',
sa.Column('owner_type', sa.Enum('organization', 'user', native_enum=False),
nullable=False)
)
op.add_column(
'repositories',
sa.Column('owner_name', sa.String(), nullable=False)
)
op.add_column(
'repositories',
sa.Column('owner_id', sa.Integer(), nullable=False)
)
op.drop_column('repositories', u'user')
op.create_unique_constraint(
"uq_owner_type_owner_name",
"repositories",
["owner_type", "owner_name"],
)
def downgrade():
op.add_column(
'repositories',
sa.Column(u'user', sa.String(), nullable=False)
)
op.drop_constraint('uq_owner_type_owner_name', 'repositories', 'unique')
op.drop_column('repositories', 'owner_id')
op.drop_column('repositories', 'owner_name')
op.drop_column('repositories', 'owner_type')
| """Improve repository owner information
Revision ID: 30c0aec2ca06
Revises: 4fdf1059c4ba
Create Date: 2012-09-02 14:45:05.241933
"""
# revision identifiers, used by Alembic.
revision = '30c0aec2ca06'
down_revision = '4fdf1059c4ba'
from alembic import op
import sqlalchemy as sa
def upgrade():
op.add_column(
'repositories',
sa.Column('owner_type', sa.Enum('organization', 'user', native_enum=False),
nullable=False)
)
op.add_column(
'repositories',
sa.Column('owner_name', sa.String(), nullable=False)
)
op.add_column(
'repositories',
sa.Column('owner_id', sa.Integer(), nullable=False)
)
op.drop_column('repositories', 'user')
op.create_unique_constraint(
"uq_owner_type_owner_name",
"repositories",
["owner_type", "owner_name"],
)
def downgrade():
op.add_column(
'repositories',
sa.Column('user', sa.String(), nullable=False)
)
op.drop_constraint('uq_owner_type_owner_name', 'repositories', 'unique')
op.drop_column('repositories', 'owner_id')
op.drop_column('repositories', 'owner_name')
op.drop_column('repositories', 'owner_type')
| Remove unicode string markers which are removed in python3 | Remove unicode string markers which are removed in python3
| Python | isc | alfredhq/alfred-db | ---
+++
@@ -28,7 +28,7 @@
'repositories',
sa.Column('owner_id', sa.Integer(), nullable=False)
)
- op.drop_column('repositories', u'user')
+ op.drop_column('repositories', 'user')
op.create_unique_constraint(
"uq_owner_type_owner_name",
"repositories",
@@ -39,7 +39,7 @@
def downgrade():
op.add_column(
'repositories',
- sa.Column(u'user', sa.String(), nullable=False)
+ sa.Column('user', sa.String(), nullable=False)
)
op.drop_constraint('uq_owner_type_owner_name', 'repositories', 'unique')
op.drop_column('repositories', 'owner_id') |
9938678e05270c06d328aeb466ab827bab232e3a | solar_neighbourhood/prepare_data_add_kinematics.py | solar_neighbourhood/prepare_data_add_kinematics.py | """
Add very large RV errors for stars with no known RVs.
Convert to cartesian.
"""
import numpy as np
import sys
sys.path.insert(0, '..')
from chronostar import tabletool
from astropy.table import Table
datafile = Table.read('../data/ScoCen_box_result.fits')
d = Table.read(datafile)
# Set missing radial velocities (nan) to 0
d['radial_velocity'] = np.nan_to_num(d['radial_velocity'])
# Set missing radial velocity errors (nan) to 1e+10
d['radial_velocity_error'][np.isnan(d['radial_velocity_error'])] = 1e+4
print('Convert to cartesian')
tabletool.convert_table_astro2cart(table=d, return_table=True)
d.write('../data/ScoCen_box_result_15M_ready_for_bg_ols.fits')
print('Cartesian written.', len(d)) | """
Add very large RV errors for stars with no known RVs.
Convert to cartesian.
"""
import numpy as np
import sys
sys.path.insert(0, '..')
from chronostar import tabletool
from astropy.table import Table
datafile = '../data/ScoCen_box_result.fits')
d = tabletool.read(datafile)
# Set missing radial velocities (nan) to 0
d['radial_velocity'] = np.nan_to_num(d['radial_velocity'])
# Set missing radial velocity errors (nan) to 1e+10
d['radial_velocity_error'][np.isnan(d['radial_velocity_error'])] = 1e+4
print('Convert to cartesian')
tabletool.convert_table_astro2cart(table=d, return_table=True)
d.write('../data/ScoCen_box_result_15M_ready_for_bg_ols.fits')
print('Cartesian written.', len(d)) | Convert entire table to cartesian | Convert entire table to cartesian
| Python | mit | mikeireland/chronostar,mikeireland/chronostar,mikeireland/chronostar,mikeireland/chronostar | ---
+++
@@ -9,8 +9,8 @@
from chronostar import tabletool
from astropy.table import Table
-datafile = Table.read('../data/ScoCen_box_result.fits')
-d = Table.read(datafile)
+datafile = '../data/ScoCen_box_result.fits')
+d = tabletool.read(datafile)
# Set missing radial velocities (nan) to 0
d['radial_velocity'] = np.nan_to_num(d['radial_velocity']) |
364d83c8add1fdde679aa2823ae94ad7f264cb48 | raco/relation_key.py | raco/relation_key.py | """Representation of a Myria relation key.
Myria relations are identified by a tuple of user, program, relation_name."""
class RelationKey(object):
def __init__(self, user='public', program='adhoc', relation=None):
assert relation
self.user = user
self.program = program
self.relation = relation
def __repr__(self):
return 'RelationKey(%s,%s,%s)' % (self.user, self.program,
self.relation)
def __str__(self):
return '%s:%s:%s' % (self.user, self.program, self.relation)
def __eq__(self, other):
return self.user == other.user and self.program == other.program \
and self.relation == other.relation
@classmethod
def from_string(cls, s):
"""Create a RelationKey from a colon-delimited string."""
toks = s.split(':')
assert len(toks) <= 3
args = {'relation' : toks[-1]}
try:
args['program'] = toks[-2]
args['user'] = toks[-3]
except IndexError:
pass
return cls(**args)
| """Representation of a Myria relation key.
Myria relations are identified by a tuple of user, program, relation_name."""
class RelationKey(object):
def __init__(self, user='public', program='adhoc', relation=None):
assert relation
self.user = user
self.program = program
self.relation = relation
def __repr__(self):
return 'RelationKey(%s,%s,%s)' % (self.user, self.program,
self.relation)
def __str__(self):
return '%s:%s:%s' % (self.user, self.program, self.relation)
def __eq__(self, other):
return self.__dict__ == other.__dict__
def __hash__(self):
return hash(str(self))
@classmethod
def from_string(cls, s):
"""Create a RelationKey from a colon-delimited string."""
toks = s.split(':')
assert len(toks) <= 3
args = {'relation' : toks[-1]}
try:
args['program'] = toks[-2]
args['user'] = toks[-3]
except IndexError:
pass
return cls(**args)
| Add hash function to RelationKey | Add hash function to RelationKey
| Python | bsd-3-clause | uwescience/raco,uwescience/raco,uwescience/raco,uwescience/raco,uwescience/raco | ---
+++
@@ -16,8 +16,10 @@
return '%s:%s:%s' % (self.user, self.program, self.relation)
def __eq__(self, other):
- return self.user == other.user and self.program == other.program \
- and self.relation == other.relation
+ return self.__dict__ == other.__dict__
+
+ def __hash__(self):
+ return hash(str(self))
@classmethod
def from_string(cls, s): |
93a23b3aed48da6953914036ae488c5b3ab891c7 | scikits/audiolab/soundio/alsa.py | scikits/audiolab/soundio/alsa.py | import numpy as np
from _alsa import card_name, card_indexes, asoundlib_version
from _alsa import Device, AlsaException
def play(input, samplerate = 48000):
if input.ndim == 1:
n = input.size
nc = 1
elif input.ndim == 2:
n, nc = input.shape
else:
raise ValueError("Only ndim 1 or 2 supported")
try:
dev = Device(samplerate = samplerate, channels = nc)
assert nc == dev.channels
assert input.dtype == np.float32 or input.dtype == np.float64
dev.play_short((16384 * input).astype(np.int16))
except AlsaException, e:
raise IOError(str(e))
if __name__ == '__main__':
print "Asoundlib version is", asoundlib_version()
for i in card_indexes():
print card_name(i)
dev = Device()
print "Device name:", dev.name
a = 0.2 * np.random.randn(4e4)
play(a, 16000)
play(a, 8000)
play(a, 22050)
| import numpy as np
from _alsa import card_name, card_indexes, asoundlib_version
from _alsa import Device, AlsaException
def play(input, samplerate = 48000):
if input.ndim == 1:
n = input.size
nc = 1
elif input.ndim == 2:
n, nc = input.shape
else:
raise ValueError("Only ndim 1 or 2 supported")
if not input.dtype in (np.float32, np.float64):
raise ValueError("input should be array of float32 or float64 !")
try:
dev = Device(samplerate = samplerate, channels = nc)
dev.play_short((16384 * input).astype(np.int16))
except AlsaException, e:
raise IOError(str(e))
if __name__ == '__main__':
print "Asoundlib version is", asoundlib_version()
for i in card_indexes():
print card_name(i)
dev = Device()
print "Device name:", dev.name
a = 0.2 * np.random.randn(4e4)
play(a, 16000)
play(a, 8000)
play(a, 22050)
| Check input dtype before creating pcm device. | Check input dtype before creating pcm device.
| Python | lgpl-2.1 | cournape/audiolab,cournape/audiolab,cournape/audiolab | ---
+++
@@ -12,12 +12,11 @@
else:
raise ValueError("Only ndim 1 or 2 supported")
+ if not input.dtype in (np.float32, np.float64):
+ raise ValueError("input should be array of float32 or float64 !")
+
try:
dev = Device(samplerate = samplerate, channels = nc)
-
- assert nc == dev.channels
- assert input.dtype == np.float32 or input.dtype == np.float64
-
dev.play_short((16384 * input).astype(np.int16))
except AlsaException, e:
raise IOError(str(e)) |
9c650cb3fb37e8c96ef9642af553ce77a28a1587 | problem-static/Intro-Eval_50/admin/eval.py | problem-static/Intro-Eval_50/admin/eval.py | #!/usr/bin/python2.7
import sys
del __builtins__.__dict__['__import__']
del __builtins__.__dict__['reload']
flag = "eval_is_fun"
class UnbufferedStream(object):
def __init__(self, stream):
self.stream = stream
def write(self, data):
self.stream.write(data)
self.stream.flush()
def __getattr__(self, attr):
return getattr(self.stream, attr)
sys.stdout = UnbufferedStream(sys.stdout)
def main():
while True:
print "Welcome to the flag database! We are currently under construction. Please do not hack the flags."
try:
command = str(raw_input("What would you like to do? "))
result = str(eval(command))
print "This is the result: %s" %(result)
except Exception, e:
print "Invalid command!!!! EXITING!!!!!"
return
main() | #!/usr/bin/python2.7
import sys
del __builtins__.__dict__['__import__']
del __builtins__.__dict__['reload']
flag = "eval_is_fun"
class UnbufferedStream(object):
def __init__(self, stream):
self.stream = stream
def write(self, data):
self.stream.write(data)
self.stream.flush()
def __getattr__(self, attr):
return getattr(self.stream, attr)
sys.stdout = UnbufferedStream(sys.stdout)
def main():
print "Welcome to the flag database! We are currently under construction. Please do not hack the flags."
while True:
try:
command = str(raw_input("What would you like to do? "))
result = str(eval(command))
print "This is the result: %s" %(result)
except Exception, e:
print "Invalid command!!!! EXITING!!!!!"
return
main() | Move welcome message to outside the loop | Move welcome message to outside the loop
| Python | mit | james9909/IntroCTF,james9909/IntroCTF,james9909/IntroCTF,james9909/IntroCTF,james9909/IntroCTF,james9909/IntroCTF | ---
+++
@@ -17,9 +17,9 @@
return getattr(self.stream, attr)
sys.stdout = UnbufferedStream(sys.stdout)
-def main():
+def main():
+ print "Welcome to the flag database! We are currently under construction. Please do not hack the flags."
while True:
- print "Welcome to the flag database! We are currently under construction. Please do not hack the flags."
try:
command = str(raw_input("What would you like to do? "))
result = str(eval(command)) |
1bbffc2152ea1c48b47153005beeb2974b682f3c | bot/actions/action.py | bot/actions/action.py | from bot.api.api import Api
from bot.storage import Config, State, Cache
from bot.utils.dictionaryobject import DictionaryObject
class Event(DictionaryObject):
pass
class Update(Event):
def __init__(self, update, is_pending):
super().__init__()
self.update = update
self.is_pending = is_pending
class Action:
def __init__(self):
pass
def get_name(self):
return self.__class__.__name__
def setup(self, api: Api, config: Config, state: State, cache: Cache):
self.api = api
self.config = config
self.state = state
self.cache = cache
self.post_setup()
def post_setup(self):
pass
def process(self, event):
pass
class ActionGroup(Action):
def __init__(self, *actions):
super().__init__()
self.actions = list(actions)
def add(self, *actions):
self.actions.extend(actions)
def setup(self, *args):
self.for_each(lambda action: action.setup(*args))
super().setup(*args)
def process(self, event):
self.for_each(lambda action: action.process(event._copy()))
def for_each(self, func):
map(func, self.actions)
class IntermediateAction(ActionGroup):
def __init__(self):
super().__init__()
def then(self, *next_actions):
self.add(*next_actions)
return self
def _continue(self, event):
super().process(event)
| from bot.api.api import Api
from bot.storage import Config, State, Cache
from bot.utils.dictionaryobject import DictionaryObject
class Event(DictionaryObject):
pass
class Update(Event):
def __init__(self, update, is_pending):
super().__init__()
self.update = update
self.is_pending = is_pending
class Action:
def __init__(self):
pass
def get_name(self):
return self.__class__.__name__
def setup(self, api: Api, config: Config, state: State, cache: Cache):
self.api = api
self.config = config
self.state = state
self.cache = cache
self.post_setup()
def post_setup(self):
pass
def process(self, event):
pass
class ActionGroup(Action):
def __init__(self, *actions):
super().__init__()
self.actions = list(actions)
def add(self, *actions):
self.actions.extend(actions)
def setup(self, *args):
self.for_each(lambda action: action.setup(*args))
super().setup(*args)
def process(self, event):
self.for_each(lambda action: action.process(event._copy()))
def for_each(self, func):
for action in self.actions:
func(action)
class IntermediateAction(ActionGroup):
def __init__(self):
super().__init__()
def then(self, *next_actions):
self.add(*next_actions)
return self
def _continue(self, event):
super().process(event)
| Fix for_each incorrectly using lazy map operator | Fix for_each incorrectly using lazy map operator
| Python | agpl-3.0 | alvarogzp/telegram-bot,alvarogzp/telegram-bot | ---
+++
@@ -51,7 +51,8 @@
self.for_each(lambda action: action.process(event._copy()))
def for_each(self, func):
- map(func, self.actions)
+ for action in self.actions:
+ func(action)
class IntermediateAction(ActionGroup): |
eb7ff9cec9360af0b5c18915164a54d4755e657b | mistraldashboard/dashboards/mistral/executions/tables.py | mistraldashboard/dashboards/mistral/executions/tables.py | # -*- coding: utf-8 -*-
#
# Copyright 2014 - StackStorm, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from django.utils.translation import ugettext_lazy as _
from horizon import tables
class ExecutionsTable(tables.DataTable):
id = tables.Column("id",
verbose_name=_("ID"),
link=("horizon:mistral:executions:tasks"))
wb_name = tables.Column("workbook_name", verbose_name=_("Workbook"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "executions"
verbose_name = _("Executions")
class TaskTable(tables.DataTable):
id = tables.Column("id", verbose_name=_("ID"))
name = tables.Column("name", verbose_name=_("Name"))
action = tables.Column("action", verbose_name=_("Action"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "tasks"
verbose_name = _("Tasks")
| # -*- coding: utf-8 -*-
#
# Copyright 2014 - StackStorm, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from django.utils.translation import ugettext_lazy as _
from horizon import tables
class ExecutionsTable(tables.DataTable):
id = tables.Column("id",
verbose_name=_("ID"),
link=("horizon:mistral:executions:tasks"))
wb_name = tables.Column("workbook_name", verbose_name=_("Workbook"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "executions"
verbose_name = _("Executions")
class TaskTable(tables.DataTable):
id = tables.Column("id", verbose_name=_("ID"))
name = tables.Column("name", verbose_name=_("Name"))
parameters = tables.Column("parameters", verbose_name=_("Parameters"))
output = tables.Column("output", verbose_name=_("Output"))
state = tables.Column("state", verbose_name=_("State"))
class Meta:
name = "tasks"
verbose_name = _("Tasks")
| Add Task's output and parameters columns | Add Task's output and parameters columns
Change-Id: I98f57a6a0178bb7258d82f3a165127f060f42f7b
Implements: blueprint mistral-ui
| Python | apache-2.0 | openstack/mistral-dashboard,openstack/mistral-dashboard,openstack/mistral-dashboard | ---
+++
@@ -34,7 +34,8 @@
class TaskTable(tables.DataTable):
id = tables.Column("id", verbose_name=_("ID"))
name = tables.Column("name", verbose_name=_("Name"))
- action = tables.Column("action", verbose_name=_("Action"))
+ parameters = tables.Column("parameters", verbose_name=_("Parameters"))
+ output = tables.Column("output", verbose_name=_("Output"))
state = tables.Column("state", verbose_name=_("State"))
class Meta: |
557cf0bc733c49e973a12bd14fb596af6a7fb5ff | refugeedata/admin.py | refugeedata/admin.py | from django.contrib import admin
from refugeedata import models, forms
class NumberAdmin(admin.ModelAdmin):
list_display = ("number", "short_id", "active")
class BatchAdmin(admin.ModelAdmin):
list_display = ("registration_number_format", "data_file")
def get_form(self, request, obj=None, **kwargs):
if not obj:
kwargs["form"] = forms.BatchAdminForm
return super(BatchAdmin, self).get_form(request, obj=obj, **kwargs)
def save_related(self, request, form, formsets, change):
if not change: # create
numbers = form.cleaned_data["registration_numbers"]
models.RegistrationNumber.objects.bulk_create(numbers)
form.cleaned_data["registration_numbers"] = (
models.RegistrationNumber.objects.filter(
id__in=[n.id for n in numbers]))
return super(BatchAdmin, self).save_related(
request, form, formsets, change)
class LanguageAdmin(admin.ModelAdmin):
list_display = ("iso_code", "description", "example_text")
admin.site.register(models.RegistrationNumber, NumberAdmin)
admin.site.register(models.RegistrationCardBatch, BatchAdmin)
admin.site.register(models.Language, LanguageAdmin)
| from django.core.management import call_command
from django.contrib import admin
from refugeedata import models, forms
class NumberAdmin(admin.ModelAdmin):
list_display = ("number", "short_id", "active")
class BatchAdmin(admin.ModelAdmin):
list_display = ("registration_number_format", "data_file")
def get_form(self, request, obj=None, **kwargs):
if not obj:
kwargs["form"] = forms.BatchAdminForm
return super(BatchAdmin, self).get_form(request, obj=obj, **kwargs)
def save_related(self, request, form, formsets, change):
if not change: # create
numbers = form.cleaned_data["registration_numbers"]
models.RegistrationNumber.objects.bulk_create(numbers)
form.cleaned_data["registration_numbers"] = (
models.RegistrationNumber.objects.filter(
id__in=[n.id for n in numbers]))
super(BatchAdmin, self).save_related(request, form, formsets, change)
call_command("export_card_data", str(form.instance.id), "--save")
class LanguageAdmin(admin.ModelAdmin):
list_display = ("iso_code", "description", "example_text")
admin.site.register(models.RegistrationNumber, NumberAdmin)
admin.site.register(models.RegistrationCardBatch, BatchAdmin)
admin.site.register(models.Language, LanguageAdmin)
| Call export_card_data on batch save | Call export_card_data on batch save
| Python | mit | ukch/refugeedata,ukch/refugeedata,ukch/refugeedata,ukch/refugeedata | ---
+++
@@ -1,3 +1,4 @@
+from django.core.management import call_command
from django.contrib import admin
from refugeedata import models, forms
@@ -24,8 +25,8 @@
form.cleaned_data["registration_numbers"] = (
models.RegistrationNumber.objects.filter(
id__in=[n.id for n in numbers]))
- return super(BatchAdmin, self).save_related(
- request, form, formsets, change)
+ super(BatchAdmin, self).save_related(request, form, formsets, change)
+ call_command("export_card_data", str(form.instance.id), "--save")
class LanguageAdmin(admin.ModelAdmin): |
a4fa3b9866ac9712f029c7cabe64121f80875207 | biobox_cli/main.py | biobox_cli/main.py | """
biobox - A command line interface for running biobox Docker containers
Usage:
biobox <command> <biobox_type> <image> [<args>...]
Options:
-h, --help Show this screen.
-v, --version Show version.
Commands:
run Run a biobox Docker image with input parameters
verify Verify that a Docker image matches the given specification type
login Log in to a biobox container with mounted test data
Biobox types:
short_read_assembler Assemble short reads into contigs
"""
import sys, string
from fn import F
import biobox_cli.util.misc as util
import biobox_cli.util.functional as fn
def run():
args = input_args()
opts = util.parse_docopt(__doc__, args, True)
util.select_module("command", opts["<command>"]).run(args)
def input_args():
"""
Get command line args excluding those consisting of only whitespace
"""
return fn.thread([
sys.argv[1:],
F(map, string.strip),
F(filter, fn.is_not_empty)])
| """
biobox - A command line interface for running biobox Docker containers
Usage:
biobox <command> <biobox_type> <image> [<args>...]
Options:
-h, --help Show this screen.
-v, --version Show version.
Commands:
run Run a biobox Docker image with input parameters
verify Verify that a Docker image matches the given specification type
login Log in to a biobox container with mounted test data
Biobox types:
short_read_assembler Assemble short reads into contigs
"""
import sys
from fn import F
import biobox_cli.util.misc as util
import biobox_cli.util.functional as fn
def run():
args = input_args()
opts = util.parse_docopt(__doc__, args, True)
util.select_module("command", opts["<command>"]).run(args)
def input_args():
"""
Get command line args excluding those consisting of only whitespace
"""
return fn.thread([
sys.argv[1:],
F(map, str.strip),
F(filter, fn.is_not_empty)])
| Use str methods instead of string module | Use str methods instead of string module
| Python | mit | bioboxes/command-line-interface,pbelmann/command-line-interface,michaelbarton/command-line-interface,michaelbarton/command-line-interface,bioboxes/command-line-interface,pbelmann/command-line-interface | ---
+++
@@ -17,7 +17,7 @@
short_read_assembler Assemble short reads into contigs
"""
-import sys, string
+import sys
from fn import F
@@ -35,5 +35,5 @@
"""
return fn.thread([
sys.argv[1:],
- F(map, string.strip),
+ F(map, str.strip),
F(filter, fn.is_not_empty)]) |
a6804dd0baefbbd9681edc2f0ba0ec13e84f5cc3 | nimp/utilities/paths.py | nimp/utilities/paths.py | # -*- coding: utf-8 -*-
import os
import os.path
import sys
import fnmatch
import glob
from nimp.utilities.logging import *
from nimp.utilities.system import *
#-------------------------------------------------------------------------------
def split_path(path):
splitted_path = []
while True:
(path, folder) = os.path.split(path)
if folder != "":
splitted_path.insert(0, folder)
else:
if path != "":
splitted_path.insert(0, path)
break
return splitted_path
def sanitize_path(path):
if is_windows() and not is_msys():
if path[0:1] == '/' and path[1:2].isalpha() and path[2:3] == '/':
return '%s:\\%s' % (path[1], path[3:].replace('/', '\\'))
if os.sep is '\\':
return path.replace('/', '\\')
# elif os.sep is '/':
return path.replace('\\', '/')
#-------------------------------------------------------------------------------
# This function is necessary because Python’s makedirs cannot create a
# directory such as "d:\data\foo/bar" because it’ll split it as "d:\data"
# and "foo/bar" then try to create a directory named "foo/bar".
def safe_makedirs(path):
path = sanitize_path(path)
try:
os.makedirs(path)
except FileExistsError:
# Maybe someone else created the directory for us; if so, ignore error
if os.path.exists(path):
return
raise
| # -*- coding: utf-8 -*-
import os
import os.path
import sys
import fnmatch
import glob
from nimp.utilities.logging import *
from nimp.utilities.system import *
#-------------------------------------------------------------------------------
def split_path(path):
splitted_path = []
while True:
(path, folder) = os.path.split(path)
if folder != "":
splitted_path.insert(0, folder)
else:
if path != "":
splitted_path.insert(0, path)
break
return splitted_path
def sanitize_path(path):
if path is None:
return None
if is_windows() and not is_msys():
if path[0:1] == '/' and path[1:2].isalpha() and path[2:3] == '/':
return '%s:\\%s' % (path[1], path[3:].replace('/', '\\'))
if os.sep is '\\':
return path.replace('/', '\\')
# elif os.sep is '/':
return path.replace('\\', '/')
#-------------------------------------------------------------------------------
# This function is necessary because Python’s makedirs cannot create a
# directory such as "d:\data\foo/bar" because it’ll split it as "d:\data"
# and "foo/bar" then try to create a directory named "foo/bar".
def safe_makedirs(path):
path = sanitize_path(path)
try:
os.makedirs(path)
except FileExistsError:
# Maybe someone else created the directory for us; if so, ignore error
if os.path.exists(path):
return
raise
| Fix case where path is None in sanitize_path | Fix case where path is None in sanitize_path
| Python | mit | dontnod/nimp | ---
+++
@@ -26,6 +26,9 @@
def sanitize_path(path):
+ if path is None:
+ return None
+
if is_windows() and not is_msys():
if path[0:1] == '/' and path[1:2].isalpha() and path[2:3] == '/':
return '%s:\\%s' % (path[1], path[3:].replace('/', '\\')) |
65c22394fad7929a7de1e78be7569a2895915dc9 | protocols/admin.py | protocols/admin.py | from django.contrib import admin
from .models import Protocol, Topic, Institution
class ProtocolAdmin(admin.ModelAdmin):
list_display = ['number', 'start_time', 'get_topics', 'information', 'majority', 'current_majority', 'institution']
list_display_links = ['number']
list_filter = ['institution__name', 'topics']
search_fields =['number', 'institution__name', 'topics__name', 'information']
admin.site.register(Institution)
admin.site.register(Topic)
admin.site.register(Protocol, ProtocolAdmin)
| from django.contrib import admin
from .models import Protocol, Topic, Institution
class ProtocolAdminIndex(admin.ModelAdmin):
list_display = ['number', 'start_time', 'get_topics', 'information', 'majority', 'current_majority', 'institution']
list_display_links = ['number']
list_filter = ['institution__name', 'topics']
search_fields =['number', 'institution__name', 'topics__name', 'information']
class TopicAdminIndex(admin.ModelAdmin):
list_display = ['name', 'voted_for', 'voted_against', 'voted_abstain', 'protocol']
list_filter = ['protocol__number']
search_fields =['name', 'protocol__number']
admin.site.register(Institution)
admin.site.register(Topic, TopicAdminIndex)
admin.site.register(Protocol, ProtocolAdminIndex)
| Add Topics index page customization | Add Topics index page customization
| Python | mit | Hackfmi/Diaphanum,Hackfmi/Diaphanum | ---
+++
@@ -2,7 +2,7 @@
from .models import Protocol, Topic, Institution
-class ProtocolAdmin(admin.ModelAdmin):
+class ProtocolAdminIndex(admin.ModelAdmin):
list_display = ['number', 'start_time', 'get_topics', 'information', 'majority', 'current_majority', 'institution']
list_display_links = ['number']
@@ -11,6 +11,15 @@
search_fields =['number', 'institution__name', 'topics__name', 'information']
+
+class TopicAdminIndex(admin.ModelAdmin):
+
+ list_display = ['name', 'voted_for', 'voted_against', 'voted_abstain', 'protocol']
+
+ list_filter = ['protocol__number']
+
+ search_fields =['name', 'protocol__number']
+
admin.site.register(Institution)
-admin.site.register(Topic)
-admin.site.register(Protocol, ProtocolAdmin)
+admin.site.register(Topic, TopicAdminIndex)
+admin.site.register(Protocol, ProtocolAdminIndex) |
22326bdd9265d8ae97055cbcc1f64939dd6bfcda | reviewboard/notifications/templatetags/markdown_email.py | reviewboard/notifications/templatetags/markdown_email.py | from __future__ import unicode_literals
import markdown
from django import template
from django.utils.safestring import mark_safe
from djblets.markdown import markdown_unescape
register = template.Library()
@register.filter
def markdown_email_html(text, is_rich_text):
if not is_rich_text:
return text
# We use XHTML1 instead of HTML5 to ensure the results can be parsed by
# an XML parser. This is actually needed for the main Markdown renderer
# for the web UI, but consistency is good here.
return mark_safe(markdown.markdown(
text,
output_format='xhtml1',
extensions=[
'markdown.extensions.fenced_code',
'markdown.extensions.codehilite',
'markdown.extensions.tables',
'markdown.extensions.sane_lists',
'markdown.extensions.smart_strong',
'pymdownx.tilde',
'djblets.markdown.extensions.escape_html',
'djblets.markdown.extensions.wysiwyg_email',
],
extension_configs={
'codehilite': {
'noclasses': True,
},
}))
@register.filter
def markdown_email_text(text, is_rich_text):
if not is_rich_text:
return text
return markdown_unescape(text)
| from __future__ import unicode_literals
import markdown
from django import template
from django.utils.safestring import mark_safe
from djblets.markdown import markdown_unescape
register = template.Library()
@register.filter
def markdown_email_html(text, is_rich_text):
if not is_rich_text:
return text
# We use XHTML1 instead of HTML5 to ensure the results can be parsed by
# an XML parser. This is actually needed for the main Markdown renderer
# for the web UI, but consistency is good here.
return mark_safe(markdown.markdown(
text,
output_format='xhtml1',
extensions=[
'markdown.extensions.fenced_code',
'markdown.extensions.codehilite',
'markdown.extensions.tables',
'markdown.extensions.sane_lists',
'pymdownx.tilde',
'djblets.markdown.extensions.escape_html',
'djblets.markdown.extensions.wysiwyg_email',
],
extension_configs={
'codehilite': {
'noclasses': True,
},
}))
@register.filter
def markdown_email_text(text, is_rich_text):
if not is_rich_text:
return text
return markdown_unescape(text)
| Remove a legacy Markdown extension when generating e-mails. | Remove a legacy Markdown extension when generating e-mails.
The recent updates for using Python-Markdown 3.x removed the
`smart_strong` extension from the main Markdown procssing, but failed to
remove it for the list of extensions used in e-mails. This is a trivial
change that simply removes that entry.
| Python | mit | reviewboard/reviewboard,chipx86/reviewboard,reviewboard/reviewboard,reviewboard/reviewboard,chipx86/reviewboard,reviewboard/reviewboard,chipx86/reviewboard,chipx86/reviewboard | ---
+++
@@ -25,7 +25,6 @@
'markdown.extensions.codehilite',
'markdown.extensions.tables',
'markdown.extensions.sane_lists',
- 'markdown.extensions.smart_strong',
'pymdownx.tilde',
'djblets.markdown.extensions.escape_html',
'djblets.markdown.extensions.wysiwyg_email', |
681cc1dc53851a2d127b4c00fc4e7d9e54bd8fba | cms/envs/devstack_docker.py | cms/envs/devstack_docker.py | """ Overrides for Docker-based devstack. """
from .devstack import * # pylint: disable=wildcard-import, unused-wildcard-import
# Docker does not support the syslog socket at /dev/log. Rely on the console.
LOGGING['handlers']['local'] = LOGGING['handlers']['tracking'] = {
'class': 'logging.NullHandler',
}
LOGGING['loggers']['tracking']['handlers'] = ['console']
HOST = 'edx.devstack.edxapp:18000'
SITE_NAME = HOST
LMS_ROOT_URL = 'http://{}:18000'.format(HOST)
OAUTH_OIDC_ISSUER = '{}/oauth2'.format(LMS_ROOT_URL)
JWT_AUTH.update({
'JWT_SECRET_KEY': 'lms-secret',
'JWT_ISSUER': OAUTH_OIDC_ISSUER,
'JWT_AUDIENCE': 'lms-key',
})
| """ Overrides for Docker-based devstack. """
from .devstack import * # pylint: disable=wildcard-import, unused-wildcard-import
# Docker does not support the syslog socket at /dev/log. Rely on the console.
LOGGING['handlers']['local'] = LOGGING['handlers']['tracking'] = {
'class': 'logging.NullHandler',
}
LOGGING['loggers']['tracking']['handlers'] = ['console']
HOST = 'edx.devstack.edxapp:18000'
SITE_NAME = HOST
LMS_ROOT_URL = 'http://{}:18000'.format(HOST)
# This is the public-facing host used for previews
LMS_BASE = 'localhost:18000'
OAUTH_OIDC_ISSUER = '{}/oauth2'.format(LMS_ROOT_URL)
JWT_AUTH.update({
'JWT_SECRET_KEY': 'lms-secret',
'JWT_ISSUER': OAUTH_OIDC_ISSUER,
'JWT_AUDIENCE': 'lms-key',
})
| Set LMS_BASE setting for Studio | Set LMS_BASE setting for Studio
This allows previews in LMS to work properly.
ECOM-6634
| Python | agpl-3.0 | jolyonb/edx-platform,ahmedaljazzar/edx-platform,proversity-org/edx-platform,raccoongang/edx-platform,edx/edx-platform,hastexo/edx-platform,fintech-circle/edx-platform,Stanford-Online/edx-platform,eduNEXT/edunext-platform,jolyonb/edx-platform,eduNEXT/edx-platform,ESOedX/edx-platform,a-parhom/edx-platform,prarthitm/edxplatform,gymnasium/edx-platform,appsembler/edx-platform,kmoocdev2/edx-platform,EDUlib/edx-platform,CredoReference/edx-platform,philanthropy-u/edx-platform,fintech-circle/edx-platform,miptliot/edx-platform,TeachAtTUM/edx-platform,cpennington/edx-platform,arbrandes/edx-platform,miptliot/edx-platform,romain-li/edx-platform,ESOedX/edx-platform,Lektorium-LLC/edx-platform,msegado/edx-platform,gsehub/edx-platform,TeachAtTUM/edx-platform,procangroup/edx-platform,gsehub/edx-platform,mitocw/edx-platform,edx-solutions/edx-platform,procangroup/edx-platform,mitocw/edx-platform,teltek/edx-platform,cpennington/edx-platform,CredoReference/edx-platform,msegado/edx-platform,angelapper/edx-platform,lduarte1991/edx-platform,proversity-org/edx-platform,gsehub/edx-platform,eduNEXT/edx-platform,arbrandes/edx-platform,angelapper/edx-platform,msegado/edx-platform,Edraak/edraak-platform,Edraak/edraak-platform,fintech-circle/edx-platform,BehavioralInsightsTeam/edx-platform,philanthropy-u/edx-platform,pepeportela/edx-platform,ahmedaljazzar/edx-platform,edx/edx-platform,msegado/edx-platform,gymnasium/edx-platform,ahmedaljazzar/edx-platform,ESOedX/edx-platform,stvstnfrd/edx-platform,msegado/edx-platform,Stanford-Online/edx-platform,teltek/edx-platform,Lektorium-LLC/edx-platform,romain-li/edx-platform,philanthropy-u/edx-platform,CredoReference/edx-platform,a-parhom/edx-platform,cpennington/edx-platform,appsembler/edx-platform,romain-li/edx-platform,edx-solutions/edx-platform,arbrandes/edx-platform,kmoocdev2/edx-platform,pepeportela/edx-platform,BehavioralInsightsTeam/edx-platform,hastexo/edx-platform,teltek/edx-platform,hastexo/edx-platform,gymnasium/edx-platform,prarthitm/edxplatform,appsembler/edx-platform,romain-li/edx-platform,appsembler/edx-platform,TeachAtTUM/edx-platform,proversity-org/edx-platform,pabloborrego93/edx-platform,eduNEXT/edx-platform,Lektorium-LLC/edx-platform,romain-li/edx-platform,Edraak/edraak-platform,edx/edx-platform,kmoocdev2/edx-platform,pabloborrego93/edx-platform,philanthropy-u/edx-platform,eduNEXT/edunext-platform,EDUlib/edx-platform,gsehub/edx-platform,arbrandes/edx-platform,pepeportela/edx-platform,stvstnfrd/edx-platform,procangroup/edx-platform,Stanford-Online/edx-platform,Lektorium-LLC/edx-platform,kmoocdev2/edx-platform,pepeportela/edx-platform,raccoongang/edx-platform,BehavioralInsightsTeam/edx-platform,prarthitm/edxplatform,edx-solutions/edx-platform,lduarte1991/edx-platform,angelapper/edx-platform,TeachAtTUM/edx-platform,stvstnfrd/edx-platform,lduarte1991/edx-platform,stvstnfrd/edx-platform,ESOedX/edx-platform,raccoongang/edx-platform,Stanford-Online/edx-platform,BehavioralInsightsTeam/edx-platform,a-parhom/edx-platform,hastexo/edx-platform,CredoReference/edx-platform,raccoongang/edx-platform,kmoocdev2/edx-platform,edx/edx-platform,proversity-org/edx-platform,miptliot/edx-platform,mitocw/edx-platform,angelapper/edx-platform,Edraak/edraak-platform,cpennington/edx-platform,teltek/edx-platform,eduNEXT/edunext-platform,fintech-circle/edx-platform,procangroup/edx-platform,edx-solutions/edx-platform,lduarte1991/edx-platform,miptliot/edx-platform,ahmedaljazzar/edx-platform,a-parhom/edx-platform,eduNEXT/edx-platform,jolyonb/edx-platform,eduNEXT/edunext-platform,EDUlib/edx-platform,jolyonb/edx-platform,mitocw/edx-platform,prarthitm/edxplatform,pabloborrego93/edx-platform,gymnasium/edx-platform,EDUlib/edx-platform,pabloborrego93/edx-platform | ---
+++
@@ -13,6 +13,9 @@
SITE_NAME = HOST
LMS_ROOT_URL = 'http://{}:18000'.format(HOST)
+# This is the public-facing host used for previews
+LMS_BASE = 'localhost:18000'
+
OAUTH_OIDC_ISSUER = '{}/oauth2'.format(LMS_ROOT_URL)
JWT_AUTH.update({ |
594cd5d490786bbbdcf877d8c155530c36acd2c1 | src/services/TemperatureMonitor/src/temperature.py | src/services/TemperatureMonitor/src/temperature.py | import smbus
class TemperatureSensor:
temp_history = []
last_temp = 0
def __init__(self, address):
self.bus = smbus.SMBus(1)
self.address = address
def get_temp(self):
MSB = self.bus.read_byte_data(self.address, 0)
LSB = self.bus.read_byte_data(self.address, 1)
temp = ((MSB << 8 | LSB) >> 4) * 0.0625
result = temp
# smooth the data slightly
history_length = 3
for t in self.temp_history:
if abs(t - temp) > 0.2:
result = self.last_temp
break
self.temp_history.append(temp)
self.temp_history = self.temp_history[0:history_length]
self.last_temp = result
return result | import smbus
class TemperatureSensor:
temp_history = []
last_temp = 0
def __init__(self, address):
self.bus = smbus.SMBus(1)
self.address = address
def get_temp(self):
MSB = self.bus.read_byte_data(self.address, 0)
LSB = self.bus.read_byte_data(self.address, 1)
temp = ((MSB << 8 | LSB) >> 4) * 0.0625
result = temp
return result | Remove Smoothing From Temp Sensor | Remove Smoothing From Temp Sensor
| Python | mit | IAPark/PITherm | ---
+++
@@ -16,16 +16,4 @@
temp = ((MSB << 8 | LSB) >> 4) * 0.0625
result = temp
- # smooth the data slightly
- history_length = 3
-
- for t in self.temp_history:
- if abs(t - temp) > 0.2:
- result = self.last_temp
- break
-
- self.temp_history.append(temp)
- self.temp_history = self.temp_history[0:history_length]
- self.last_temp = result
-
return result |
70f5a3fd7e28c574912a0318eba83f11789c2c7b | ankieta/contact/urls.py | ankieta/contact/urls.py | from django.conf.urls import patterns, url
from django.views.generic import TemplateView
from . import views
urlpatterns = patterns('',
url(r'^$', views.ContactView.as_view(), name="form"),
url(r'^success$', TemplateView.as_view(template_name="contact/success.html"), name="success"),
)
| from django.conf.urls import patterns, url
from django.views.generic import TemplateView
from . import views
urlpatterns = patterns('',
url(r'^$', views.ContactView.as_view(), name="form"),
url(r'^/success$', TemplateView.as_view(template_name="contact/success.html"), name="success"),
)
| Fix URLS in contact form | Fix URLS in contact form
| Python | bsd-3-clause | watchdogpolska/prezydent.siecobywatelska.pl,watchdogpolska/prezydent.siecobywatelska.pl,watchdogpolska/prezydent.siecobywatelska.pl | ---
+++
@@ -4,6 +4,6 @@
urlpatterns = patterns('',
url(r'^$', views.ContactView.as_view(), name="form"),
- url(r'^success$', TemplateView.as_view(template_name="contact/success.html"), name="success"),
+ url(r'^/success$', TemplateView.as_view(template_name="contact/success.html"), name="success"),
) |
72d33ea47458cace13dac920ce2a82e55f83caba | statsmodels/stats/tests/test_outliers_influence.py | statsmodels/stats/tests/test_outliers_influence.py | from numpy.testing import assert_almost_equal
from statsmodels.datasets import statecrime, get_rdataset
from statsmodels.regression.linear_model import OLS
from statsmodels.stats.outliers_influence import reset_ramsey
from statsmodels.stats.outliers_influence import variance_inflation_factor
from statsmodels.tools import add_constant
import numpy as np
data = statecrime.load_pandas().data
def test_reset_stata():
mod = OLS(data.violent, add_constant(data[['murder', 'hs_grad']]))
res = mod.fit()
stat = reset_ramsey(res, degree=4)
assert_almost_equal(stat.fvalue[0, 0], 1.52, decimal=2)
assert_almost_equal(stat.pvalue, 0.2221, decimal=4)
exog_idx = list(data.columns).index('urban')
X_arr = np.asarray(data)
vif = variance_inflation_factor(X_arr, exog_idx)
assert_almost_equal(vif, 16.4394, decimal=4)
| from numpy.testing import assert_almost_equal
from statsmodels.datasets import statecrime
from statsmodels.regression.linear_model import OLS
from statsmodels.stats.outliers_influence import reset_ramsey
from statsmodels.stats.outliers_influence import variance_inflation_factor
from statsmodels.tools import add_constant
import numpy as np
data = statecrime.load_pandas().data
def test_reset_stata():
mod = OLS(data.violent, add_constant(data[['murder', 'hs_grad']]))
res = mod.fit()
stat = reset_ramsey(res, degree=4)
assert_almost_equal(stat.fvalue[0, 0], 1.52, decimal=2)
assert_almost_equal(stat.pvalue, 0.2221, decimal=4)
exog_idx = list(data.columns).index('urban')
X_arr = np.asarray(data)
vif = variance_inflation_factor(X_arr, exog_idx)
assert_almost_equal(vif, 16.4394, decimal=4)
| Add pandas dataframe capability in variance_inflation_factor | ENH: Add pandas dataframe capability in variance_inflation_factor
| Python | bsd-3-clause | bashtage/statsmodels,josef-pkt/statsmodels,josef-pkt/statsmodels,bashtage/statsmodels,bashtage/statsmodels,statsmodels/statsmodels,josef-pkt/statsmodels,josef-pkt/statsmodels,statsmodels/statsmodels,statsmodels/statsmodels,statsmodels/statsmodels,josef-pkt/statsmodels,josef-pkt/statsmodels,statsmodels/statsmodels,bashtage/statsmodels,bashtage/statsmodels,bashtage/statsmodels,statsmodels/statsmodels | ---
+++
@@ -1,6 +1,6 @@
from numpy.testing import assert_almost_equal
-from statsmodels.datasets import statecrime, get_rdataset
+from statsmodels.datasets import statecrime
from statsmodels.regression.linear_model import OLS
from statsmodels.stats.outliers_influence import reset_ramsey
from statsmodels.stats.outliers_influence import variance_inflation_factor |
e787e4981441198e2b015b1b4f4971fbc112c78b | cyder/base/eav/utils.py | cyder/base/eav/utils.py | import re
from django.core.exceptions import ValidationError
default_validator = lambda x: x != '' # FIXME: Do we need this?
def validate_list(value, validator=default_validator, separator=',',
strip_whitespace=True, min_length=0, die=False):
"""Validate a "list" of things
separator: the char that separates list items (None means whitespace)
allow_whitespace: whether to strip whitespace around separators before
validating (unnecessary if separator is None)
Returns whether validator returned True for every item in value. Note that
this is not terribly useful.
"""
items = value.split(separator)
length = len(items)
all_valid = all([validator(x.strip() if strip_whitespace else x)
for x in items])
if not all_valid:
if die:
raise ValidationError("One or more list items are invalid")
else:
return False
elif length < min_length:
if die:
raise ValidationError("List must contain at least {0} items"
.format(length))
else:
return False
else:
return True
def is_hex_byte(value):
return bool(re.match(r'^[0-9a-fA-F]{2}$', value))
def is_hex_byte_sequence(value):
return validate_list(value, _hex_byte, separator=':',
strip_whitespace=False)
def strip_and_get_base(value):
if value.startswith('0x'):
value = value[len('0x'):]
base = 16
else:
base = 10
return (value, base)
| import re
from django.core.exceptions import ValidationError
default_validator = lambda x: x != '' # FIXME: Do we need this?
def validate_list(value, validator=default_validator, separator=',',
strip_whitespace=True, min_length=0, die=False):
"""Validate a "list" of things
separator: the char that separates list items (None means whitespace)
allow_whitespace: whether to strip whitespace around separators before
validating (unnecessary if separator is None)
Returns whether validator returned True for every item in value. Note that
this is not terribly useful.
"""
items = value.split(separator)
length = len(items)
all_valid = all([validator(x.strip() if strip_whitespace else x)
for x in items])
if not all_valid:
if die:
raise ValidationError("One or more list items are invalid")
else:
return False
elif length < min_length:
if die:
raise ValidationError("List must contain at least {0} items"
.format(length))
else:
return False
else:
return True
def is_hex_byte(value):
return bool(re.match(r'^[0-9a-fA-F]{2}$', value))
def is_hex_byte_sequence(value):
return validate_list(value, is_hex_byte, separator=':',
strip_whitespace=False)
def strip_and_get_base(value):
if value.startswith('0x'):
value = value[len('0x'):]
base = 16
else:
base = 10
return (value, base)
| Fix yet another stupid mistake | Fix yet another stupid mistake
| Python | bsd-3-clause | akeym/cyder,murrown/cyder,murrown/cyder,murrown/cyder,zeeman/cyder,drkitty/cyder,drkitty/cyder,murrown/cyder,akeym/cyder,zeeman/cyder,drkitty/cyder,OSU-Net/cyder,akeym/cyder,akeym/cyder,OSU-Net/cyder,OSU-Net/cyder,drkitty/cyder,zeeman/cyder,OSU-Net/cyder,zeeman/cyder | ---
+++
@@ -41,7 +41,7 @@
def is_hex_byte_sequence(value):
- return validate_list(value, _hex_byte, separator=':',
+ return validate_list(value, is_hex_byte, separator=':',
strip_whitespace=False)
|
295fc64b8fac9852e92356f61ff4698e011c798e | seam/util.py | seam/util.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
""" util.py
Utility functions/constants across seam
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
import sys
PY2 = sys.version_info[0] == 2
if PY2:
STRING_TYPE = basestring
else:
STRING_TYPE = str
import os
from string import digits, ascii_letters
from random import choice
total = digits + ascii_letters
def get_tmp_filename(ext='out', basename='/tmp', fname_length=32):
fname = ''.join(choice(total) for _ in range(fname_length))
return os.path.join(basename, '{}.{}'.format(fname, ext))
def wrap_with_xvfb(command, wait=5, server_args='-screen 0, 1600x1200x24'):
parts = ['xvfb-run',
'-a', # automatically get a free server number
'-f {}'.format(get_tmp_filename()),
'-e {}'.format(get_tmp_filename()),
'--wait={:d}'.format(wait),
'--server-args="{}"'.format(server_args),
command]
return ' '.join(parts)
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
""" util.py
Utility functions/constants across seam
"""
__author__ = 'Scott Burns <scott.s.burns@vanderbilt.edu>'
__copyright__ = 'Copyright 2014 Vanderbilt University. All Rights Reserved'
import sys
PY2 = sys.version_info[0] == 2
if PY2:
STRING_TYPE = basestring
else:
STRING_TYPE = str
import os
from string import digits, ascii_letters
from random import choice
total = digits + ascii_letters
def get_tmp_filename(ext='out', basename='/tmp', fname_length=32):
fname = ''.join(choice(total) for _ in range(fname_length))
return os.path.join(basename, '{}.{}'.format(fname, ext))
def wrap_with_xvfb(command, wait=5, server_args='-screen 0, 1600x1200x24'):
parts = ['xvfb-run',
'-a', # automatically get a free server number
'-f {}'.format(get_tmp_filename()),
'-e {}'.format(get_tmp_filename(ext='err')),
'--wait={:d}'.format(wait),
'--server-args="{}"'.format(server_args),
command]
return ' '.join(parts)
| Make xvfb e files end with .err | Make xvfb e files end with .err
| Python | mit | VUIIS/seam,VUIIS/seam | ---
+++
@@ -30,7 +30,7 @@
parts = ['xvfb-run',
'-a', # automatically get a free server number
'-f {}'.format(get_tmp_filename()),
- '-e {}'.format(get_tmp_filename()),
+ '-e {}'.format(get_tmp_filename(ext='err')),
'--wait={:d}'.format(wait),
'--server-args="{}"'.format(server_args),
command] |
d18c715f8f0d86d58fcec4df8710f0370ff87308 | channels/worker.py | channels/worker.py | from __future__ import unicode_literals
import logging
import time
from .message import Message
from .utils import name_that_thing
logger = logging.getLogger('django.channels')
class Worker(object):
"""
A "worker" process that continually looks for available messages to run
and runs their consumers.
"""
def __init__(self, channel_layer, callback=None):
self.channel_layer = channel_layer
self.callback = callback
def run(self):
"""
Tries to continually dispatch messages to consumers.
"""
channels = self.channel_layer.registry.all_channel_names()
while True:
channel, content = self.channel_layer.receive_many(channels, block=True)
logger.debug("Worker got message on %s: repl %s", channel, content.get("reply_channel", "none"))
# If no message, stall a little to avoid busy-looping then continue
if channel is None:
time.sleep(0.01)
continue
# Create message wrapper
message = Message(
content=content,
channel_name=channel,
channel_layer=self.channel_layer,
)
# Handle the message
consumer = self.channel_layer.registry.consumer_for_channel(channel)
if self.callback:
self.callback(channel, message)
try:
consumer(message)
except:
logger.exception("Error processing message with consumer %s:", name_that_thing(consumer))
| from __future__ import unicode_literals
import logging
import time
from .message import Message
from .utils import name_that_thing
logger = logging.getLogger('django.channels')
class Worker(object):
"""
A "worker" process that continually looks for available messages to run
and runs their consumers.
"""
def __init__(self, channel_layer, callback=None):
self.channel_layer = channel_layer
self.callback = callback
def run(self):
"""
Tries to continually dispatch messages to consumers.
"""
channels = self.channel_layer.registry.all_channel_names()
while True:
channel, content = self.channel_layer.receive_many(channels, block=True)
# If no message, stall a little to avoid busy-looping then continue
if channel is None:
time.sleep(0.01)
continue
# Create message wrapper
logger.debug("Worker got message on %s: repl %s", channel, content.get("reply_channel", "none"))
message = Message(
content=content,
channel_name=channel,
channel_layer=self.channel_layer,
)
# Handle the message
consumer = self.channel_layer.registry.consumer_for_channel(channel)
if self.callback:
self.callback(channel, message)
try:
consumer(message)
except:
logger.exception("Error processing message with consumer %s:", name_that_thing(consumer))
| Fix core logging when no message on channel | Fix core logging when no message on channel
| Python | bsd-3-clause | Krukov/channels,Coread/channels,Krukov/channels,raiderrobert/channels,Coread/channels,andrewgodwin/django-channels,andrewgodwin/channels,linuxlewis/channels,django/channels,raphael-boucher/channels | ---
+++
@@ -26,12 +26,12 @@
channels = self.channel_layer.registry.all_channel_names()
while True:
channel, content = self.channel_layer.receive_many(channels, block=True)
- logger.debug("Worker got message on %s: repl %s", channel, content.get("reply_channel", "none"))
# If no message, stall a little to avoid busy-looping then continue
if channel is None:
time.sleep(0.01)
continue
# Create message wrapper
+ logger.debug("Worker got message on %s: repl %s", channel, content.get("reply_channel", "none"))
message = Message(
content=content,
channel_name=channel, |
da3c8b01512466e8b0c71a239378b8643d5384ef | tests/state_tests.py | tests/state_tests.py | """
state_tests.py
"""
import os
from os.path import splitext
from minicps.state import SQLiteState
def test_SQLiteState():
# TODO: change to /tmp when install SQLitesutdio in ubuntu
PATH = "temp/state_test_db.sqlite"
# sqlite use text instead of VARCHAR
SCHEMA = """
CREATE TABLE state_test (
name TEXT NOT NULL,
datatype TEXT NOT NULL,
value TEXT,
pid INTEGER NOT NULL,
PRIMARY KEY (name, pid)
);
"""
SCHEMA_INIT = """
INSERT INTO state_test VALUES ('SENSOR1', 'int', '0', 1);
INSERT INTO state_test VALUES ('SENSOR2', 'float', '0.0', 1);
INSERT INTO state_test VALUES ('SENSOR3', 'int', '0.0', 2);
INSERT INTO state_test VALUES ('ACTUATOR1', 'int', '1', 1);
INSERT INTO state_test VALUES ('ACTUATOR2', 'int', '0', 1);
"""
path, extension = splitext(PATH)
sqlite_state = SQLiteState(path, extension)
os.remove(PATH)
sqlite_state._create(PATH, SCHEMA)
sqlite_state._init(PATH, SCHEMA_INIT)
# sqlite_state._delete()
| """
state_tests.py
"""
import os
from os.path import splitext
from minicps.state import SQLiteState
def test_SQLiteState():
# TODO: change to /tmp when install SQLitesutdio in ubuntu
PATH = "temp/state_test_db.sqlite"
# sqlite use text instead of VARCHAR
SCHEMA = """
CREATE TABLE state_test (
name TEXT NOT NULL,
datatype TEXT NOT NULL,
value TEXT,
pid INTEGER NOT NULL,
PRIMARY KEY (name, pid)
);
"""
SCHEMA_INIT = """
INSERT INTO state_test VALUES ('SENSOR1', 'int', '0', 1);
INSERT INTO state_test VALUES ('SENSOR2', 'float', '0.0', 1);
INSERT INTO state_test VALUES ('SENSOR3', 'int', '0.0', 1);
INSERT INTO state_test VALUES ('SENSOR3', 'int', '0.0', 2);
INSERT INTO state_test VALUES ('ACTUATOR1', 'int', '1', 1);
INSERT INTO state_test VALUES ('ACTUATOR2', 'int', '0', 1);
"""
path, extension = splitext(PATH)
sqlite_state = SQLiteState(path, extension)
os.remove(PATH)
sqlite_state._create(PATH, SCHEMA)
sqlite_state._init(PATH, SCHEMA_INIT)
# sqlite_state._delete()
| Use same tagname but different scope | Use same tagname but different scope
| Python | mit | remmihsorp/minicps,scy-phy/minicps,remmihsorp/minicps,scy-phy/minicps | ---
+++
@@ -26,6 +26,7 @@
SCHEMA_INIT = """
INSERT INTO state_test VALUES ('SENSOR1', 'int', '0', 1);
INSERT INTO state_test VALUES ('SENSOR2', 'float', '0.0', 1);
+ INSERT INTO state_test VALUES ('SENSOR3', 'int', '0.0', 1);
INSERT INTO state_test VALUES ('SENSOR3', 'int', '0.0', 2);
INSERT INTO state_test VALUES ('ACTUATOR1', 'int', '1', 1);
INSERT INTO state_test VALUES ('ACTUATOR2', 'int', '0', 1); |
c820e3ed4d78b975a6bdff54a2ecae26354ae10e | tests/test_itunes.py | tests/test_itunes.py | """
test_itunes.py
Copyright © 2015 Alex Danoff. All Rights Reserved.
2015-08-02
This file tests the functionality provided by the itunes module.
"""
import unittest
from itunes.itunes import parse_value
class ITunesTests(unittest.TestCase):
"""
Test cases for iTunes functionality.
"""
def test_parse_value(self):
self.assertEquals(parse_value("10"), 10)
self.assertEquals(parse_value("1.0"), 1.0)
self.assertTrue(parse_value("true"))
self.assertFalse(parse_value("false"))
self.assertIsNone(parse_value(""))
self.assertIsNone(parse_value('""'))
self.assertIsNone(parse_value("missing value"))
| """
test_itunes.py
Copyright © 2015 Alex Danoff. All Rights Reserved.
2015-08-02
This file tests the functionality provided by the itunes module.
"""
import unittest
from datetime import datetime
from itunes.itunes import parse_value
class ITunesTests(unittest.TestCase):
"""
Test cases for iTunes functionality.
"""
def test_parse_value(self):
self.assertEquals(parse_value("10"), 10)
self.assertEquals(parse_value("1.0"), 1.0)
self.assertTrue(parse_value("true"))
self.assertFalse(parse_value("false"))
self.assertIsNone(parse_value(""))
self.assertIsNone(parse_value('""'))
self.assertIsNone(parse_value("missing value"))
self.assertEquals(parse_value('date: "Saturday, March 13, 2010 at ' \
'5:02:22 PM"'), datetime.fromtimestamp(1268517742))
| Add `parse_value` test for AppleScript dates | Add `parse_value` test for AppleScript dates
Added a test case to `parse_value` to parse dates returned in
AppleScript responses.
| Python | mit | adanoff/iTunesTUI | ---
+++
@@ -8,6 +8,7 @@
"""
import unittest
+from datetime import datetime
from itunes.itunes import parse_value
@@ -24,3 +25,5 @@
self.assertIsNone(parse_value(""))
self.assertIsNone(parse_value('""'))
self.assertIsNone(parse_value("missing value"))
+ self.assertEquals(parse_value('date: "Saturday, March 13, 2010 at ' \
+ '5:02:22 PM"'), datetime.fromtimestamp(1268517742)) |
d69b137bd19e0363173b120ff4f68becc6be7b3c | mama_cas/tests/backends.py | mama_cas/tests/backends.py | from django.contrib.auth.backends import ModelBackend
from django.contrib.auth.models import User
class ExceptionBackend(ModelBackend):
"""Raise an exception on authentication for testing purposes."""
def authenticate(self, username=None, password=None):
raise Exception
class CaseInsensitiveBackend(ModelBackend):
"""A case-insenstitive authentication backend."""
def authenticate(self, username=None, password=None):
try:
user = User.objects.get(username__iexact=username)
if user.check_password(password):
return user
except User.DoesNotExist:
return None
| from django.contrib.auth.backends import ModelBackend
from mama_cas.compat import get_user_model
class ExceptionBackend(ModelBackend):
"""Raise an exception on authentication for testing purposes."""
def authenticate(self, username=None, password=None):
raise Exception
class CaseInsensitiveBackend(ModelBackend):
"""A case-insenstitive authentication backend."""
def authenticate(self, username=None, password=None):
user_model = get_user_model()
try:
user = user_model.objects.get(username__iexact=username)
if user.check_password(password):
return user
except user_model.DoesNotExist:
return None
| Use get_user_model within test backend | Use get_user_model within test backend
| Python | bsd-3-clause | orbitvu/django-mama-cas,harlov/django-mama-cas,forcityplatform/django-mama-cas,jbittel/django-mama-cas,orbitvu/django-mama-cas,forcityplatform/django-mama-cas,jbittel/django-mama-cas,harlov/django-mama-cas | ---
+++
@@ -1,5 +1,6 @@
from django.contrib.auth.backends import ModelBackend
-from django.contrib.auth.models import User
+
+from mama_cas.compat import get_user_model
class ExceptionBackend(ModelBackend):
@@ -11,9 +12,10 @@
class CaseInsensitiveBackend(ModelBackend):
"""A case-insenstitive authentication backend."""
def authenticate(self, username=None, password=None):
+ user_model = get_user_model()
try:
- user = User.objects.get(username__iexact=username)
+ user = user_model.objects.get(username__iexact=username)
if user.check_password(password):
return user
- except User.DoesNotExist:
+ except user_model.DoesNotExist:
return None |
46077269450f98505308736251b3f08ed3c6827f | scripts/poweron/DRAC.py | scripts/poweron/DRAC.py | import subprocess, sys, os.path
class DRAC_NO_SUPP_PACK(Exception):
"""Base Exception class for all transfer plugin errors."""
def __init__(self, *args):
Exception.__init__(self, *args)
class DRAC_POWERON_FAILED(Exception):
"""Base Exception class for all transfer plugin errors."""
def __init__(self, *args):
Exception.__init__(self, *args)
def run2(command):
run = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
# Wait for the process to return
out, err = [ e.splitlines() for e in run.communicate() ]
return run.returncode, out, err
drac_path='/usr/sbin/racadm'
def DRAC( power_on_ip, user, password):
if( not os.path.exists(drac_path)):
raise DRAC_NO_SUPP_PACK()
cmd='%s -r %s -u %s -p %s serveraction powerup' % (drac_path, power_on_ip, user, password)
retcode,out,err=run2(cmd)
if(len(err)==0):
return str(True)
else:
raise DRAC_POWERON_FAILED()
def main():
if len(sys.argv)<3:
exit(0)
ip=sys.argv[1]
user=sys.argv[2]
password=sys.argv[3]
print DRAC(ip,user,password)
if __name__ == "__main__":
main() | import subprocess, sys, os.path
class DRAC_NO_SUPP_PACK(Exception):
"""Base Exception class for all transfer plugin errors."""
def __init__(self, *args):
Exception.__init__(self, *args)
class DRAC_POWERON_FAILED(Exception):
"""Base Exception class for all transfer plugin errors."""
def __init__(self, *args):
Exception.__init__(self, *args)
def run2(command):
run = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
# Wait for the process to return
out, err = [ e.splitlines() for e in run.communicate() ]
return run.returncode, out, err
drac_path='/opt/dell/srvadmin/sbin/racadm'
def DRAC( power_on_ip, user, password):
if( not os.path.exists(drac_path)):
raise DRAC_NO_SUPP_PACK()
cmd='%s -r %s -u %s -p %s serveraction powerup' % (drac_path, power_on_ip, user, password)
retcode,out,err=run2(cmd)
if(len(err)==0):
return str(True)
else:
raise DRAC_POWERON_FAILED()
def main():
if len(sys.argv)<3:
exit(0)
ip=sys.argv[1]
user=sys.argv[2]
password=sys.argv[3]
print DRAC(ip,user,password)
if __name__ == "__main__":
main() | Change path to the supplemental pack | CA-40618: Change path to the supplemental pack
Signed-off-by: Javier Alvarez-Valle <cf4c8668a0b4c5e013f594a6940d05b3d4d9ddcf@citrix.com>
| Python | lgpl-2.1 | Frezzle/xen-api,vasilenkomike/xen-api,simonjbeaumont/xen-api,euanh/xen-api,cheng-z/xen-api,jjd27/xen-api,vasilenkomike/xen-api,rafalmiel/xen-api,robertbreker/xen-api,koushikcgit/xen-api,huizh/xen-api,agimofcarmen/xen-api,cheng--zhang/xen-api,salvocambria/xen-api,jjd27/xen-api,djs55/xen-api,thomassa/xen-api,huizh/xen-api,rafalmiel/xen-api,salvocambria/xen-api,robertbreker/xen-api,cheng--zhang/xen-api,thomassa/xen-api,srowe/xen-api,koushikcgit/xen-api,koushikcgit/xen-api,robertbreker/xen-api,robertbreker/xen-api,simonjbeaumont/xen-api,anoobs/xen-api,srowe/xen-api,cheng-z/xen-api,jjd27/xen-api,huizh/xen-api,guard163/xen-api,guard163/xen-api,salvocambria/xen-api,jjd27/xen-api,rafalmiel/xen-api,djs55/xen-api,Frezzle/xen-api,jjd27/xen-api,srowe/xen-api,anoobs/xen-api,huizh/xen-api,agimofcarmen/xen-api,robertbreker/xen-api,cheng-z/xen-api,jjd27/xen-api,guard163/xen-api,cheng-z/xen-api,guard163/xen-api,cheng-z/xen-api,simonjbeaumont/xen-api,anoobs/xen-api,djs55/xen-api,huizh/xen-api,thomassa/xen-api,Frezzle/xen-api,djs55/xen-api,simonjbeaumont/xen-api,cheng-z/xen-api,djs55/xen-api,rafalmiel/xen-api,robertbreker/xen-api,euanh/xen-api,cheng--zhang/xen-api,agimofcarmen/xen-api,rafalmiel/xen-api,ravippandey/xen-api,guard163/xen-api,salvocambria/xen-api,cheng--zhang/xen-api,cheng-z/xen-api,euanh/xen-api,cheng--zhang/xen-api,ravippandey/xen-api,vasilenkomike/xen-api,vasilenkomike/xen-api,koushikcgit/xen-api,thomassa/xen-api,ravippandey/xen-api,euanh/xen-api,euanh/xen-api,thomassa/xen-api,srowe/xen-api,ravippandey/xen-api,thomassa/xen-api,djs55/xen-api,Frezzle/xen-api,ravippandey/xen-api,cheng--zhang/xen-api,Frezzle/xen-api,euanh/xen-api,guard163/xen-api,cheng--zhang/xen-api,koushikcgit/xen-api,salvocambria/xen-api,koushikcgit/xen-api,anoobs/xen-api,agimofcarmen/xen-api,srowe/xen-api,agimofcarmen/xen-api,vasilenkomike/xen-api,agimofcarmen/xen-api,simonjbeaumont/xen-api,rafalmiel/xen-api,ravippandey/xen-api,simonjbeaumont/xen-api,salvocambria/xen-api,anoobs/xen-api,vasilenkomike/xen-api,huizh/xen-api,anoobs/xen-api | ---
+++
@@ -18,7 +18,7 @@
return run.returncode, out, err
-drac_path='/usr/sbin/racadm'
+drac_path='/opt/dell/srvadmin/sbin/racadm'
def DRAC( power_on_ip, user, password):
if( not os.path.exists(drac_path)):
raise DRAC_NO_SUPP_PACK() |
1fe22f9750c618ede99f9b0a0d088aa67b7929a1 | stock_available_unreserved/models/quant.py | stock_available_unreserved/models/quant.py | # Copyright 2018 Camptocamp SA
# Copyright 2016-19 ForgeFlow S.L. (https://www.forgeflow.com)
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockQuant(models.Model):
_inherit = "stock.quant"
contains_unreserved = fields.Boolean(
string="Contains unreserved products",
compute="_compute_contains_unreserved",
store=True,
)
@api.depends("product_id", "location_id", "quantity", "reserved_quantity")
def _compute_contains_unreserved(self):
for record in self:
available = record._get_available_quantity(
record.product_id, record.location_id
)
record.contains_unreserved = True if available > 0 else False
| # Copyright 2018 Camptocamp SA
# Copyright 2016-19 ForgeFlow S.L. (https://www.forgeflow.com)
# License LGPL-3.0 or later (https://www.gnu.org/licenses/lgpl.html).
from odoo import api, fields, models
class StockQuant(models.Model):
_inherit = "stock.quant"
contains_unreserved = fields.Boolean(
string="Contains unreserved products",
compute="_compute_contains_unreserved",
store=True,
)
@api.depends("product_id", "location_id", "quantity", "reserved_quantity")
def _compute_contains_unreserved(self):
for record in self:
# Avoid error when adding a new line on manually Update Quantity
if isinstance(record.id, models.NewId):
record.contains_unreserved = False
continue
available = record._get_available_quantity(
record.product_id, record.location_id
)
record.contains_unreserved = True if available > 0 else False
| Fix compute contains_unreserved on NewId records | [FIX] Fix compute contains_unreserved on NewId records
| Python | agpl-3.0 | OCA/stock-logistics-warehouse,OCA/stock-logistics-warehouse,OCA/stock-logistics-warehouse,OCA/stock-logistics-warehouse | ---
+++
@@ -17,6 +17,10 @@
@api.depends("product_id", "location_id", "quantity", "reserved_quantity")
def _compute_contains_unreserved(self):
for record in self:
+ # Avoid error when adding a new line on manually Update Quantity
+ if isinstance(record.id, models.NewId):
+ record.contains_unreserved = False
+ continue
available = record._get_available_quantity(
record.product_id, record.location_id
) |
828b78767c17419513337ca29b5c2dab08995714 | ctypeslib/test/test_dynmodule.py | ctypeslib/test/test_dynmodule.py | # Basic test of dynamic code generation
import unittest
import os, glob
import stdio
from ctypes import POINTER, c_int
class DynModTest(unittest.TestCase):
def tearDown(self):
for fnm in glob.glob(stdio._gen_basename + ".*"):
try:
os.remove(fnm)
except IOError:
pass
def test_fopen(self):
self.failUnlessEqual(stdio.fopen.restype, POINTER(stdio.FILE))
self.failUnlessEqual(stdio.fopen.argtypes, [stdio.STRING, stdio.STRING])
def test_constants(self):
self.failUnlessEqual(stdio.O_RDONLY, 0)
self.failUnlessEqual(stdio.O_WRONLY, 1)
self.failUnlessEqual(stdio.O_RDWR, 2)
def test_compiler_errors(self):
from ctypeslib.codegen.cparser import CompilerError
from ctypeslib.dynamic_module import include
self.failUnlessRaises(CompilerError, lambda: include("#error"))
if __name__ == "__main__":
unittest.main()
| # Basic test of dynamic code generation
import unittest
import os, glob
import stdio
from ctypes import POINTER, c_int
class DynModTest(unittest.TestCase):
def test_fopen(self):
self.failUnlessEqual(stdio.fopen.restype, POINTER(stdio.FILE))
self.failUnlessEqual(stdio.fopen.argtypes, [stdio.STRING, stdio.STRING])
def test_constants(self):
self.failUnlessEqual(stdio.O_RDONLY, 0)
self.failUnlessEqual(stdio.O_WRONLY, 1)
self.failUnlessEqual(stdio.O_RDWR, 2)
def test_compiler_errors(self):
from ctypeslib.codegen.cparser import CompilerError
from ctypeslib.dynamic_module import include
self.failUnlessRaises(CompilerError, lambda: include("#error"))
if __name__ == "__main__":
unittest.main()
| Remove now useless TearDown method. | Remove now useless TearDown method.
git-svn-id: ac2c3632cb6543e7ab5fafd132c7fe15057a1882@53797 6015fed2-1504-0410-9fe1-9d1591cc4771
| Python | mit | trolldbois/ctypeslib,luzfcb/ctypeslib,trolldbois/ctypeslib,luzfcb/ctypeslib,luzfcb/ctypeslib,trolldbois/ctypeslib | ---
+++
@@ -6,12 +6,6 @@
from ctypes import POINTER, c_int
class DynModTest(unittest.TestCase):
- def tearDown(self):
- for fnm in glob.glob(stdio._gen_basename + ".*"):
- try:
- os.remove(fnm)
- except IOError:
- pass
def test_fopen(self):
self.failUnlessEqual(stdio.fopen.restype, POINTER(stdio.FILE)) |
ddb79d01e7ae0c840a3f3181a600aae34613c4e5 | login_token/models.py | login_token/models.py | import random
import re
from django.contrib.auth.models import User
from django.db import models
from instances.models import InstanceMixin
NUMBER_OF_TOKEN_WORDS = 3
def generate_token():
def useful_word(w):
# FIXME: should try to exclude offensive words
if len(w) < 4:
return False
if re.search('^[a-z]*$', w):
return True
words = []
with open('/usr/share/dict/words') as fp:
for line in fp:
word = line.strip()
if useful_word(word):
words.append(word)
return " ".join(random.choice(words)
for i in range(NUMBER_OF_TOKEN_WORDS))
class LoginToken(InstanceMixin, models.Model):
'''Represents a readable login token for mobile devices
To enable logging in to a SayIt instance as a particular user, we
ask the user to type in a three word phrase; this model records
tokens that allow login for a particular instance by a particular
user.'''
user = models.ForeignKey(User)
token = models.TextField(max_length=255,
default=generate_token)
def regenerate_token(self):
token = generate_token()
token.save()
| import random
import re
from django.contrib.auth.models import User
from django.db import models
from instances.models import InstanceMixin
NUMBER_OF_TOKEN_WORDS = 3
def generate_token():
def useful_word(w):
# FIXME: should try to exclude offensive words
if len(w) < 4:
return False
if re.search('^[a-z]*$', w):
return True
words = []
with open('/usr/share/dict/words') as fp:
for line in fp:
word = line.strip()
if useful_word(word):
words.append(word)
return " ".join(random.choice(words)
for i in range(NUMBER_OF_TOKEN_WORDS))
class LoginToken(InstanceMixin, models.Model):
'''Represents a readable login token for mobile devices
To enable logging in to a SayIt instance as a particular user, we
ask the user to type in a three word phrase; this model records
tokens that allow login for a particular instance by a particular
user.'''
user = models.ForeignKey(User)
token = models.TextField(max_length=255,
default=generate_token)
def regenerate_token(self):
token = generate_token()
token.save()
def __repr__(self):
repr_format = '<LoginToken: "%s" user="%s" instance="%s">'
return repr_format % (self.token,
self.user.username,
self.instance.label)
| Add a __repr__ method for LoginToken | Add a __repr__ method for LoginToken
| Python | agpl-3.0 | opencorato/sayit,opencorato/sayit,opencorato/sayit,opencorato/sayit | ---
+++
@@ -40,3 +40,9 @@
def regenerate_token(self):
token = generate_token()
token.save()
+
+ def __repr__(self):
+ repr_format = '<LoginToken: "%s" user="%s" instance="%s">'
+ return repr_format % (self.token,
+ self.user.username,
+ self.instance.label) |
23e3197f15d13445defe6ec7cfb4f08484089068 | tests/test_scripts/test_simulate_data.py | tests/test_scripts/test_simulate_data.py | import json
import numpy as np
from click.testing import CliRunner
from fastimgproto.scripts.simulate_data import cli as sim_cli
def test_simulate_data():
runner = CliRunner()
with runner.isolated_filesystem():
output_filename = 'simdata.npz'
result = runner.invoke(sim_cli,
[output_filename,])
assert result.exit_code == 0
with open(output_filename, 'rb') as f:
output_data = np.load(f)
expected_keys = ('uvw_lambda', 'model', 'vis')
for k in expected_keys:
assert k in output_data | import json
import numpy as np
from click.testing import CliRunner
from fastimgproto.scripts.simulate_data import cli as sim_cli
def test_simulate_data():
runner = CliRunner()
with runner.isolated_filesystem():
output_filename = 'simdata.npz'
result = runner.invoke(sim_cli,
[output_filename,
'--nstep','5'
])
assert result.exit_code == 0
with open(output_filename, 'rb') as f:
output_data = np.load(f)
expected_keys = ('uvw_lambda', 'model', 'vis')
for k in expected_keys:
assert k in output_data | Use few nsteps for testing sim-script | Use few nsteps for testing sim-script
| Python | apache-2.0 | SKA-ScienceDataProcessor/FastImaging-Python,SKA-ScienceDataProcessor/FastImaging-Python | ---
+++
@@ -11,7 +11,9 @@
output_filename = 'simdata.npz'
result = runner.invoke(sim_cli,
- [output_filename,])
+ [output_filename,
+ '--nstep','5'
+ ])
assert result.exit_code == 0
with open(output_filename, 'rb') as f:
output_data = np.load(f) |
5972644fe7d0267849440d8e60509baba6e013a3 | test/test_exception.py | test/test_exception.py | from mock import MagicMock
import pyaem
import unittest
class TestPyAemException(unittest.TestCase):
def test_init(self):
exception = pyaem.PyAemException(123, 'somemessage')
self.assertEqual(exception.code, 123)
self.assertEqual(exception.message, 'somemessage')
if __name__ == '__main__':
unittest.main() | import pyaem
import unittest
class TestException(unittest.TestCase):
def test_init(self):
exception = pyaem.PyAemException(123, 'somemessage')
self.assertEqual(exception.code, 123)
self.assertEqual(exception.message, 'somemessage')
if __name__ == '__main__':
unittest.main() | Rename class name to be consistent with file name. Remove unused import. | Rename class name to be consistent with file name. Remove unused import.
| Python | mit | Sensis/pyaem,wildone/pyaem | ---
+++
@@ -1,8 +1,7 @@
-from mock import MagicMock
import pyaem
import unittest
-class TestPyAemException(unittest.TestCase):
+class TestException(unittest.TestCase):
def test_init(self): |
dfea77df6e6ba27bada1c80da6efab392507736b | forklift/services/satellite.py | forklift/services/satellite.py | #
# Copyright 2014 Infoxchange Australia
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""
Satellite processes started by Forklift itself to provide services.
"""
import os
import threading
from time import sleep
def start_satellite(target, args=(), kwargs=None, stop=None):
"""
Start a process configured to run the target but kill it after the parent
exits.
"""
if kwargs is None:
kwargs = {}
pid = os.fork()
if pid == 0:
# Run target daemonized.
payload = threading.Thread(
target=target,
args=args,
kwargs=kwargs,
daemon=True,
)
payload.start()
# Cannot wait for the process that's not our child
ppid = os.getppid()
try:
while True:
os.kill(ppid, 0)
sleep(1)
except OSError:
if stop:
stop()
os._exit(os.EX_OK) # pylint:disable=protected-access
| #
# Copyright 2014 Infoxchange Australia
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""
Satellite processes started by Forklift itself to provide services.
"""
import os
import threading
from time import sleep
def start_satellite(target, args=(), kwargs=None, stop=None):
"""
Start a process configured to run the target but kill it after the parent
exits.
"""
if kwargs is None:
kwargs = {}
pid = os.fork()
if pid == 0:
# Run target daemonized.
payload = threading.Thread(
target=target,
args=args,
kwargs=kwargs,
)
payload.daemon = True
payload.start()
# Cannot wait for the process that's not our child
ppid = os.getppid()
try:
while True:
os.kill(ppid, 0)
sleep(1)
except OSError:
if stop:
stop()
os._exit(os.EX_OK) # pylint:disable=protected-access
| Fix making threads daemonic on Python 3.2 | Fix making threads daemonic on Python 3.2
| Python | apache-2.0 | infoxchange/docker-forklift,infoxchange/docker-forklift | ---
+++
@@ -38,8 +38,8 @@
target=target,
args=args,
kwargs=kwargs,
- daemon=True,
)
+ payload.daemon = True
payload.start()
# Cannot wait for the process that's not our child |
d41af20b1bdf5b630962a2e474b5d9c7ed62cd5c | nuxeo-drive-client/nxdrive/gui/resources.py | nuxeo-drive-client/nxdrive/gui/resources.py | """Helper to lookup UI resources from package"""
import re
import os
from nxdrive.logging_config import get_logger
log = get_logger(__name__)
def find_icon(icon_filename):
"""Find the FS path of an icon on various OS binary packages"""
import nxdrive
nxdrive_path = os.path.dirname(nxdrive.__file__)
icons_path = os.path.join(nxdrive_path, 'data', 'icons')
cxfreeze_suffix = os.path.join('library.zip', 'nxdrive')
app_resources = '/Contents/Resources/'
if app_resources in nxdrive_path:
# OSX frozen distribution, bundled as an app
icons_path = re.sub(app_resources + ".*", app_resources + 'icons',
nxdrive_path)
elif nxdrive_path.endswith(cxfreeze_suffix):
# Frozen distribution of nxdrive, data is out of the zip
icons_path = nxdrive_path.replace(cxfreeze_suffix, 'icons')
if not os.path.exists(icons_path):
log.warning("Could not find the icons folder at: %s", icons_path)
return None
icon_filepath = os.path.join(icons_path, icon_filename)
if not os.path.exists(icon_filepath):
log.warning("Could not find icon file: %s", icon_filepath)
return None
return icon_filepath
| """Helper to lookup UI resources from package"""
import os
from nxdrive.logging_config import get_logger
from nxdrive.utils import find_resource_dir
log = get_logger(__name__)
def find_icon(icon_filename):
"""Find the FS path of an icon in various OS binary packages"""
import nxdrive
nxdrive_path = os.path.dirname(nxdrive.__file__)
icons_path = os.path.join(nxdrive_path, 'data', 'icons')
icons_dir = find_resource_dir('icons', icons_path)
if icons_dir is None:
log.warning("Could not find icon file %s as icons directory"
" could not be found",
icon_filename)
return None
icon_filepath = os.path.join(icons_dir, icon_filename)
if not os.path.exists(icon_filepath):
log.warning("Could not find icon file: %s", icon_filepath)
return None
return icon_filepath
| Use generic resource directory finder for icon files | NXP-12694: Use generic resource directory finder for icon files
| Python | lgpl-2.1 | arameshkumar/base-nuxeo-drive,IsaacYangSLA/nuxeo-drive,rsoumyassdi/nuxeo-drive,arameshkumar/nuxeo-drive,ssdi-drive/nuxeo-drive,DirkHoffmann/nuxeo-drive,arameshkumar/nuxeo-drive,arameshkumar/base-nuxeo-drive,DirkHoffmann/nuxeo-drive,loopingz/nuxeo-drive,rsoumyassdi/nuxeo-drive,DirkHoffmann/nuxeo-drive,arameshkumar/nuxeo-drive,loopingz/nuxeo-drive,loopingz/nuxeo-drive,ssdi-drive/nuxeo-drive,DirkHoffmann/nuxeo-drive,loopingz/nuxeo-drive,loopingz/nuxeo-drive,arameshkumar/base-nuxeo-drive,rsoumyassdi/nuxeo-drive,IsaacYangSLA/nuxeo-drive,IsaacYangSLA/nuxeo-drive,DirkHoffmann/nuxeo-drive,arameshkumar/base-nuxeo-drive,IsaacYangSLA/nuxeo-drive,rsoumyassdi/nuxeo-drive,ssdi-drive/nuxeo-drive,arameshkumar/nuxeo-drive,IsaacYangSLA/nuxeo-drive | ---
+++
@@ -1,35 +1,26 @@
"""Helper to lookup UI resources from package"""
-import re
import os
from nxdrive.logging_config import get_logger
+from nxdrive.utils import find_resource_dir
log = get_logger(__name__)
def find_icon(icon_filename):
- """Find the FS path of an icon on various OS binary packages"""
+ """Find the FS path of an icon in various OS binary packages"""
import nxdrive
nxdrive_path = os.path.dirname(nxdrive.__file__)
icons_path = os.path.join(nxdrive_path, 'data', 'icons')
+ icons_dir = find_resource_dir('icons', icons_path)
- cxfreeze_suffix = os.path.join('library.zip', 'nxdrive')
- app_resources = '/Contents/Resources/'
-
- if app_resources in nxdrive_path:
- # OSX frozen distribution, bundled as an app
- icons_path = re.sub(app_resources + ".*", app_resources + 'icons',
- nxdrive_path)
-
- elif nxdrive_path.endswith(cxfreeze_suffix):
- # Frozen distribution of nxdrive, data is out of the zip
- icons_path = nxdrive_path.replace(cxfreeze_suffix, 'icons')
-
- if not os.path.exists(icons_path):
- log.warning("Could not find the icons folder at: %s", icons_path)
+ if icons_dir is None:
+ log.warning("Could not find icon file %s as icons directory"
+ " could not be found",
+ icon_filename)
return None
- icon_filepath = os.path.join(icons_path, icon_filename)
+ icon_filepath = os.path.join(icons_dir, icon_filename)
if not os.path.exists(icon_filepath):
log.warning("Could not find icon file: %s", icon_filepath)
return None |
72dea9616a84cefd8424f965060552c84cfd241d | tests/test_luabject.py | tests/test_luabject.py | try:
import unittest2 as unittest
except ImportError:
import unittest
from village import _luabject
class TestDirect(unittest.TestCase):
def test_new(self):
state = _luabject.new()
# PyCObject isn't available to assertIsInstance, so:
self.assertEqual(type(state).__name__, 'PyCObject')
def test_load_script(self):
state = _luabject.new()
_luabject.load_script(state, "")
# Can load multiple scripts in one state.
_luabject.load_script(state, "")
# Can load a syntactically correct script.
state = _luabject.new()
_luabject.load_script(state, "function foo() prant() end")
# Can load multiple syntactically correct scripts in one state.
_luabject.load_script(state, "function bar() prant() end")
# Loading a syntactically incorrect script raises an exception.
state = _luabject.new()
with self.assertRaises(ValueError):
_luabject.load_script(state, "1+1")
# Can load a syntactically correct script even after loading an incorrect script raises an exception.
_luabject.load_script(state, "function foo() prant() end")
| try:
import unittest2 as unittest
except ImportError:
import unittest
from village import _luabject
class TestDirect(unittest.TestCase):
def test_new(self):
state = _luabject.new()
# PyCObject isn't available to assertIsInstance, so:
self.assertEqual(type(state).__name__, 'PyCObject')
def test_load_script(self):
state = _luabject.new()
_luabject.load_script(state, "")
# Can load multiple scripts in one state.
_luabject.load_script(state, "")
# Can load a syntactically correct script.
state = _luabject.new()
_luabject.load_script(state, "function foo() prant() end")
# Can load multiple syntactically correct scripts in one state.
_luabject.load_script(state, "function bar() prant() end")
# Loading a syntactically incorrect script raises an exception.
state = _luabject.new()
with self.assertRaises(ValueError):
_luabject.load_script(state, "1+1")
# Can load a syntactically correct script even after a load_script() exception.
_luabject.load_script(state, "function foo() prant() end")
# Loading a syntactically correct script that causes an error raises an exception.
state = _luabject.new()
with self.assertRaises(ValueError):
_luabject.load_script(state, "hi()")
# Can load a syntactically correct script even after a load_script() exception.
_luabject.load_script(state, "function foo() prant() end")
| Test unrunnable script exceptions too | Test unrunnable script exceptions too
| Python | mit | markpasc/luabject,markpasc/luabject | ---
+++
@@ -33,5 +33,13 @@
with self.assertRaises(ValueError):
_luabject.load_script(state, "1+1")
- # Can load a syntactically correct script even after loading an incorrect script raises an exception.
+ # Can load a syntactically correct script even after a load_script() exception.
_luabject.load_script(state, "function foo() prant() end")
+
+ # Loading a syntactically correct script that causes an error raises an exception.
+ state = _luabject.new()
+ with self.assertRaises(ValueError):
+ _luabject.load_script(state, "hi()")
+
+ # Can load a syntactically correct script even after a load_script() exception.
+ _luabject.load_script(state, "function foo() prant() end") |
9f20f232a9507f0002adc682a87bb792f6fbdd4e | django_plim/template.py | django_plim/template.py | #!/usr/bin/env python
#-*- coding: UTF-8 -*-
from functools import partial
from django.conf import settings
from plim import preprocessor as plim_preprocessor
from mako.template import Template as MakoTemplate
from mako.lookup import TemplateLookup
lookup = TemplateLookup(directories=settings.TEMPLATE_DIRS)
Template = partial(MakoTemplate, lookup=lookup,
preprocessor=plim_preprocessor)
| #!/usr/bin/env python
#-*- coding: UTF-8 -*-
from functools import partial
from django.conf import settings
from plim import preprocessor as plim_preprocessor
from mako.template import Template as MakoTemplate
from mako.lookup import TemplateLookup
from django.template.loaders import app_directories
lookup = TemplateLookup(directories=settings.TEMPLATE_DIRS)
Template = partial(MakoTemplate, lookup=lookup,
preprocessor=plim_preprocessor)
class Template(MakoTemplate):
def render(self, context):
context_dict = {}
for d in context.dicts:
context_dict.update(d)
return super(Template, self).render(context_dict)
class Loader(app_directories.Loader):
is_usable = True
def load_template(self, template_name, template_dirs=None):
source, origin = self.load_template_source(template_name, template_dirs)
template = Template(source)
return template, origin | Add example code copied from django doc | Add example code copied from django doc
| Python | mit | imom0/django-plim | ---
+++
@@ -7,8 +7,26 @@
from plim import preprocessor as plim_preprocessor
from mako.template import Template as MakoTemplate
from mako.lookup import TemplateLookup
+from django.template.loaders import app_directories
lookup = TemplateLookup(directories=settings.TEMPLATE_DIRS)
Template = partial(MakoTemplate, lookup=lookup,
preprocessor=plim_preprocessor)
+
+
+class Template(MakoTemplate):
+ def render(self, context):
+ context_dict = {}
+ for d in context.dicts:
+ context_dict.update(d)
+ return super(Template, self).render(context_dict)
+
+
+class Loader(app_directories.Loader):
+ is_usable = True
+
+ def load_template(self, template_name, template_dirs=None):
+ source, origin = self.load_template_source(template_name, template_dirs)
+ template = Template(source)
+ return template, origin |
c53824a3427235c814cfe35c5c85fd5e1e312b40 | i3/.config/i3/scripts/lock_screen/lock_screen.py | i3/.config/i3/scripts/lock_screen/lock_screen.py | #!/usr/bin/env python
from subprocess import check_call, CalledProcessError
from tempfile import NamedTemporaryFile
from dpms import DPMS
from mss import mss
from PIL import Image, ImageFilter
GAUSSIAN_BLUR_RADIUS = 5
SCREEN_TIMEOUT = (5, 5, 5) # Standby, Suspend, Off
# Get current DPMS settings
dpms = DPMS()
current_timeouts = dpms.GetTimeouts()
with mss() as sct:
# Get the "All-in-one" monitor
monitor = sct.monitors[0]
# Get raw pixels of the screen
sct_img = sct.grab(monitor)
# Create Image object using Pillow
img = Image.frombytes("RGB", sct_img.size, sct_img.rgb)
with NamedTemporaryFile(suffix=".png") as tempfile:
# Apply filters to Image
img = img.filter(ImageFilter.GaussianBlur(radius=GAUSSIAN_BLUR_RADIUS))
# Save temporary file
img.save(tempfile.name, optimize=False, compress_level=1)
# Set monitor timeout to SCREEN_TIMEOUT
dpms.SetTimeouts(*SCREEN_TIMEOUT)
dpms.GetTimeouts()
try:
# Load image in i3lock
check_call(["i3lock", "-nei", tempfile.name])
except CalledProcessError:
# Something went wrong, lock it anyway
check_call(["i3lock", "-ne"])
finally:
# Restore DPMS settings
dpms.SetTimeouts(*current_timeouts)
dpms.GetTimeouts()
| #!/usr/bin/env python
from subprocess import check_call, CalledProcessError
from tempfile import NamedTemporaryFile
from dpms import DPMS
from mss import mss
from PIL import Image, ImageFilter
GAUSSIAN_BLUR_RADIUS = 5
SCREEN_TIMEOUT = (5, 5, 5) # Standby, Suspend, Off
# Get current DPMS settings
dpms = DPMS()
current_timeouts = dpms.GetTimeouts()
with mss() as sct:
# Get the "All-in-one" monitor
monitor = sct.monitors[0]
# Get raw pixels of the screen
sct_img = sct.grab(monitor)
# Create Image object using Pillow
img = Image.frombytes("RGB", sct_img.size, sct_img.rgb)
with NamedTemporaryFile(suffix=".png") as tempfile:
# Apply filters to Image
img = img.filter(ImageFilter.GaussianBlur(radius=GAUSSIAN_BLUR_RADIUS))
# Save temporary file
img.save(tempfile.name, optimize=False, compress_level=1)
# Set monitor timeout to SCREEN_TIMEOUT
dpms.SetTimeouts(*SCREEN_TIMEOUT)
try:
# Load image in i3lock
check_call(["i3lock", "-nei", tempfile.name])
except CalledProcessError:
# Something went wrong, lock it anyway
check_call(["i3lock", "-ne"])
finally:
# Restore DPMS settings
dpms.SetTimeouts(*current_timeouts)
| Remove call to GetTimeouts() after SetTimeouts() | i3: Remove call to GetTimeouts() after SetTimeouts()
Fixed in commit 72e984a54049c77208546b8565cece100e87be48 from
m45t3r/python-dpms.
| Python | mit | m45t3r/dotfiles,m45t3r/dotfiles,m45t3r/dotfiles | ---
+++
@@ -29,7 +29,6 @@
img.save(tempfile.name, optimize=False, compress_level=1)
# Set monitor timeout to SCREEN_TIMEOUT
dpms.SetTimeouts(*SCREEN_TIMEOUT)
- dpms.GetTimeouts()
try:
# Load image in i3lock
check_call(["i3lock", "-nei", tempfile.name])
@@ -39,4 +38,3 @@
finally:
# Restore DPMS settings
dpms.SetTimeouts(*current_timeouts)
- dpms.GetTimeouts() |
55c72a5297244ba51fba5ebc5b71efc3001e0dd4 | otz/__init__.py | otz/__init__.py | from otz.Timestream import CalibrationTimestream, CapturedTimestream
from otz.Calibration import Calibration
| from otz.Timestream import CalibrationTimestream, CapturedTimestream
from otz.Calibration import Calibration
from otz.Beam import Beam, Bead
| Add Beam, Bead to main module | Add Beam, Bead to main module
| Python | unlicense | ghallsimpsons/optical_tweezers | ---
+++
@@ -1,2 +1,3 @@
from otz.Timestream import CalibrationTimestream, CapturedTimestream
from otz.Calibration import Calibration
+from otz.Beam import Beam, Bead |
cd75c139910e8968e5262d0f0f5289119b258f21 | phileo/views.py | phileo/views.py | from django.contrib.auth.decorators import login_required
from django.contrib.contenttypes.models import ContentType
from django.http import HttpResponse
from django.utils import simplejson as json
from django.shortcuts import get_object_or_404, redirect
from django.views.decorators.http import require_POST
from phileo.models import Like
from phileo.signals import object_liked, object_unliked
@require_POST
@login_required
def like_toggle(request, content_type_id, object_id):
content_type = get_object_or_404(ContentType, pk=content_type_id)
like, created = Like.objects.get_or_create(
sender = request.user,
receiver_content_type = content_type,
receiver_object_id = object_id
)
if created:
object_liked.send(sender=Like, like=like)
else:
like.delete()
object_unliked.send(
sender=Like,
object=content_type.get_object_for_this_type(
pk=object_id
)
)
if request.is_ajax():
return HttpResponse(json.dumps({
"likes_count": Like.objects.filter(
sender = request.user,
receiver_content_type = content_type,
receiver_object_id = object_id
).count()
}), mimetype="application/json")
return redirect(request.META["HTTP_REFERER"])
| from django.contrib.auth.decorators import login_required
from django.contrib.contenttypes.models import ContentType
from django.http import HttpResponse
from django.utils import simplejson as json
from django.shortcuts import get_object_or_404, redirect
from django.views.decorators.http import require_POST
from phileo.models import Like
from phileo.signals import object_liked, object_unliked
@require_POST
@login_required
def like_toggle(request, content_type_id, object_id):
content_type = get_object_or_404(ContentType, pk=content_type_id)
like, created = Like.objects.get_or_create(
sender = request.user,
receiver_content_type = content_type,
receiver_object_id = object_id
)
if created:
object_liked.send(sender=Like, like=like)
else:
like.delete()
object_unliked.send(
sender=Like,
object=content_type.get_object_for_this_type(
pk=object_id
)
)
if request.is_ajax():
return HttpResponse(json.dumps({
"likes_count": Like.objects.filter(
receiver_content_type = content_type,
receiver_object_id = object_id
).count()
}), mimetype="application/json")
return redirect(request.META["HTTP_REFERER"])
| Remove user from count query to show likes count for all users for obj | Remove user from count query to show likes count for all users for obj
| Python | mit | pinax/phileo,jacobwegner/phileo,rizumu/pinax-likes,rizumu/pinax-likes,jacobwegner/phileo,pinax/pinax-likes,pinax/phileo | ---
+++
@@ -35,7 +35,6 @@
if request.is_ajax():
return HttpResponse(json.dumps({
"likes_count": Like.objects.filter(
- sender = request.user,
receiver_content_type = content_type,
receiver_object_id = object_id
).count() |
8b0e39eec8a82fd3f5a424ec75678426b2bf523e | cinder/version.py | cinder/version.py | # Copyright 2011 OpenStack Foundation
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from pbr import version as pbr_version
CINDER_VENDOR = "OpenStack Foundation"
CINDER_PRODUCT = "OpenStack Cinder"
CINDER_PACKAGE = None # OS distro package version suffix
loaded = False
version_info = pbr_version.VersionInfo('cinder')
version_string = version_info.version_string
| # Copyright 2011 OpenStack Foundation
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
CINDER_VENDOR = "OpenStack Foundation"
CINDER_PRODUCT = "OpenStack Cinder"
CINDER_PACKAGE = None # OS distro package version suffix
loaded = False
class VersionInfo:
version = "REDHATCINDERVERSION"
release = "REDHATCINDERRELEASE"
def release_string(self):
return '%s-%s' % (self.version, self.release)
def version_string(self):
return self.version
version_info = VersionInfo()
version_string = version_info.version_string
| Remove runtime dep on python-pbr, python-d2to1 | Remove runtime dep on python-pbr, python-d2to1
Requires RPM spec to fill in REDHATCINDERVERSION.
| Python | apache-2.0 | alex8866/cinder,alex8866/cinder | ---
+++
@@ -12,12 +12,22 @@
# License for the specific language governing permissions and limitations
# under the License.
-from pbr import version as pbr_version
-
CINDER_VENDOR = "OpenStack Foundation"
CINDER_PRODUCT = "OpenStack Cinder"
CINDER_PACKAGE = None # OS distro package version suffix
loaded = False
-version_info = pbr_version.VersionInfo('cinder')
+
+
+class VersionInfo:
+ version = "REDHATCINDERVERSION"
+ release = "REDHATCINDERRELEASE"
+
+ def release_string(self):
+ return '%s-%s' % (self.version, self.release)
+
+ def version_string(self):
+ return self.version
+
+version_info = VersionInfo()
version_string = version_info.version_string |
9c6739830ea8ccfbe697bc691de001a42f01f9c6 | serial_protocol/test.py | serial_protocol/test.py | import serial
import time
import binascii
import struct
def establishConnection():
# Define Constants
SERIAL_DEVICE = "/dev/ttyACM0"
# Establish Connection
ser = serial.Serial(SERIAL_DEVICE, 9600)
time.sleep(2)
print("Connection Established")
return ser
# Each motor speed is a float from -1.0 to 1.0
def sendDrive(ser, left, right):
if(left < -1 or left > 1 or right < -1 or right > 1):
print("Incorrectly formated drive command!")
return;
ser.write('1')
#ser.write('0' if left >= 0 else '1')
#ser.write(struct.pack("B", abs(left) * 255))
#ser.write('0' if right >= 0 else '1')
#ser.write(struct.pack("B", abs(right) * 255))
ser.write('0')
ser.write(bytes(255))
ser.write('0')
ser.write(bytes(255))
ser.write('0')
ser.write('0')
ser.write('0')
ser.write('0')
print('test')
if __name__ == '__main__':
ser = establishConnection()
sendDrive(ser, -1.0, -1.0)
time.sleep(5)
sendDrive(ser, 1.0, 1.0)
time.sleep(5)
sendDrive(ser, 0.0, 0.0)
| import serial
import time
import binascii
import struct
def establishConnection():
# Define Constants
SERIAL_DEVICE = "/dev/ttyACM0"
# Establish Connection
ser = serial.Serial(SERIAL_DEVICE, 9600)
time.sleep(2)
print("Connection Established")
return ser
# Each motor speed is a float from -1.0 to 1.0
def sendDrive(ser, left, right):
if(left < -1 or left > 1 or right < -1 or right > 1):
print("Incorrectly formated drive command!")
return;
# Write OpCode
ser.write('1')
# Write Left Motor Direction
if (left >= 0):
ser.write(bytes(0))
else:
ser.write(bytes(1))
# Write Left Motor Speed
ser.write(bytes(abs(left * 255)))
# Write Right Motor Direction
if (right >= 0):
ser.write(bytes(0))
else:
ser.write(bytes(1))
# Write Right Motor Speed
ser.write(bytes(abs(right * 255)))
# Pad message to 9 bytes
ser.write(bytes(0))
ser.write(bytes(0))
ser.write(bytes(0))
ser.write(bytes(0))
print('Test')
if __name__ == '__main__':
ser = establishConnection()
sendDrive(ser, -1.0, -1.0)
time.sleep(5)
sendDrive(ser, 1.0, 1.0)
time.sleep(5)
sendDrive(ser, 0.0, 0.0)
| Write each byte at a time in protocol | Write each byte at a time in protocol
| Python | mit | zacharylawrence/ENEE408I-Team-9,zacharylawrence/ENEE408I-Team-9,zacharylawrence/ENEE408I-Team-9 | ---
+++
@@ -20,23 +20,34 @@
print("Incorrectly formated drive command!")
return;
+ # Write OpCode
ser.write('1')
- #ser.write('0' if left >= 0 else '1')
- #ser.write(struct.pack("B", abs(left) * 255))
- #ser.write('0' if right >= 0 else '1')
- #ser.write(struct.pack("B", abs(right) * 255))
- ser.write('0')
- ser.write(bytes(255))
- ser.write('0')
- ser.write(bytes(255))
+ # Write Left Motor Direction
+ if (left >= 0):
+ ser.write(bytes(0))
+ else:
+ ser.write(bytes(1))
- ser.write('0')
- ser.write('0')
- ser.write('0')
- ser.write('0')
+ # Write Left Motor Speed
+ ser.write(bytes(abs(left * 255)))
- print('test')
+ # Write Right Motor Direction
+ if (right >= 0):
+ ser.write(bytes(0))
+ else:
+ ser.write(bytes(1))
+
+ # Write Right Motor Speed
+ ser.write(bytes(abs(right * 255)))
+
+ # Pad message to 9 bytes
+ ser.write(bytes(0))
+ ser.write(bytes(0))
+ ser.write(bytes(0))
+ ser.write(bytes(0))
+
+ print('Test')
if __name__ == '__main__':
ser = establishConnection() |
1503bf01903f088d812a9cb38a4ce7582e063a58 | gaphor/misc/tests/test_gidlethread.py | gaphor/misc/tests/test_gidlethread.py | import pytest
from gaphor.misc.gidlethread import GIdleThread
def counter(count):
for x in range(count):
yield x
@pytest.fixture
def gidle_counter(request):
# Setup GIdle Thread with 0.02 sec timeout
t = GIdleThread(counter(request.param))
t.start()
assert t.is_alive()
wait_result = t.wait(0.02)
yield wait_result
# Teardown GIdle Thread
t.interrupt()
@pytest.mark.parametrize(argnames="gidle_counter", argvalues=[20000], indirect=True)
def test_wait_with_timeout(gidle_counter):
# GIVEN a long coroutine thread
# WHEN waiting short timeout
# THEN timeout is True
assert gidle_counter
@pytest.mark.parametrize(argnames="gidle_counter", argvalues=[2], indirect=True)
def test_wait_until_finished(gidle_counter):
# GIVEN a short coroutine thread
# WHEN wait for coroutine to finish
# THEN coroutine finished
assert not gidle_counter
| import pytest
from gaphor.misc.gidlethread import GIdleThread
def counter(count):
for x in range(count):
yield x
@pytest.fixture
def gidle_counter(request):
# Setup GIdle Thread with 0.05 sec timeout
t = GIdleThread(counter(request.param))
t.start()
assert t.is_alive()
wait_result = t.wait(0.05)
yield wait_result
# Teardown GIdle Thread
t.interrupt()
@pytest.mark.parametrize(argnames="gidle_counter", argvalues=[20000], indirect=True)
def test_wait_with_timeout(gidle_counter):
# GIVEN a long coroutine thread
# WHEN waiting short timeout
# THEN timeout is True
assert gidle_counter
@pytest.mark.parametrize(argnames="gidle_counter", argvalues=[2], indirect=True)
def test_wait_until_finished(gidle_counter):
# GIVEN a short coroutine thread
# WHEN wait for coroutine to finish
# THEN coroutine finished
assert not gidle_counter
| Fix test flakiness by changing timeout to a longer value | Fix test flakiness by changing timeout to a longer value
Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>
| Python | lgpl-2.1 | amolenaar/gaphor,amolenaar/gaphor | ---
+++
@@ -10,11 +10,11 @@
@pytest.fixture
def gidle_counter(request):
- # Setup GIdle Thread with 0.02 sec timeout
+ # Setup GIdle Thread with 0.05 sec timeout
t = GIdleThread(counter(request.param))
t.start()
assert t.is_alive()
- wait_result = t.wait(0.02)
+ wait_result = t.wait(0.05)
yield wait_result
# Teardown GIdle Thread
t.interrupt() |
072bc480cbc489cd89d03405026f152934893b7e | go/routers/keyword/view_definition.py | go/routers/keyword/view_definition.py | from django import forms
from go.router.view_definition import RouterViewDefinitionBase, EditRouterView
class KeywordForm(forms.Form):
keyword = forms.CharField()
target_endpoint = forms.CharField()
class BaseKeywordFormSet(forms.formsets.BaseFormSet):
@staticmethod
def initial_from_config(data):
return [{'keyword': k, 'target_endpoint': v}
for k, v in sorted(data.items())]
def to_config(self):
keyword_endpoint_mapping = {}
for form in self:
if not form.is_valid():
continue
keyword = form.cleaned_data['keyword']
target_endpoint = form.cleaned_data['target_endpoint']
keyword_endpoint_mapping[keyword] = target_endpoint
return keyword_endpoint_mapping
KeywordFormSet = forms.formsets.formset_factory(
KeywordForm, can_delete=True, extra=1, formset=BaseKeywordFormSet)
class EditKeywordView(EditRouterView):
edit_forms = (
('keyword_endpoint_mapping', KeywordFormSet),
)
class RouterViewDefinition(RouterViewDefinitionBase):
edit_view = EditKeywordView
| from django import forms
from go.router.view_definition import RouterViewDefinitionBase, EditRouterView
class KeywordForm(forms.Form):
keyword = forms.CharField()
target_endpoint = forms.CharField()
class BaseKeywordFormSet(forms.formsets.BaseFormSet):
@staticmethod
def initial_from_config(data):
return [{'keyword': k, 'target_endpoint': v}
for k, v in sorted(data.items())]
def to_config(self):
keyword_endpoint_mapping = {}
for form in self:
if (not form.is_valid()) or form.cleaned_data['DELETE']:
continue
keyword = form.cleaned_data['keyword']
target_endpoint = form.cleaned_data['target_endpoint']
keyword_endpoint_mapping[keyword] = target_endpoint
return keyword_endpoint_mapping
KeywordFormSet = forms.formsets.formset_factory(
KeywordForm, can_delete=True, extra=1, formset=BaseKeywordFormSet)
class EditKeywordView(EditRouterView):
edit_forms = (
('keyword_endpoint_mapping', KeywordFormSet),
)
class RouterViewDefinition(RouterViewDefinitionBase):
edit_view = EditKeywordView
| Revert "Remove unnecessary and broken DELETE check." | Revert "Remove unnecessary and broken DELETE check."
This reverts commit 7906153b4718f34ed31c193a8e80b171e567209c.
Reverting commit accidentally commited straight to develop.
| Python | bsd-3-clause | praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go | ---
+++
@@ -17,7 +17,7 @@
def to_config(self):
keyword_endpoint_mapping = {}
for form in self:
- if not form.is_valid():
+ if (not form.is_valid()) or form.cleaned_data['DELETE']:
continue
keyword = form.cleaned_data['keyword']
target_endpoint = form.cleaned_data['target_endpoint'] |
2867ea119a846e1eb61c64a206058dc0d28d090b | src/foremast/utils/check_task.py | src/foremast/utils/check_task.py | """Check Taskid status."""
import logging
import requests
from tryagain import retries
from ..exceptions import SpinnakerTaskError
HEADERS = {'Content-Type': 'application/json', 'Accept': '*/*'}
GATE_URL = "http://gate-api.build.example.com:8084"
LOG = logging.getLogger(__name__)
@retries(max_attempts=10, wait=10, exceptions=(AssertionError, ValueError))
def check_task(taskid, app_name):
"""Check task status.
Args:
taskid: the task id returned from create_elb.
app_name: application name related to this task.
Returns:
polls for task status.
"""
try:
taskurl = taskid.get('ref', '0000')
except AttributeError:
taskurl = taskid
taskid = taskurl.split('/tasks/')[-1]
LOG.info('Checking taskid %s', taskid)
url = '{0}/applications/{1}/tasks/{2}'.format(GATE_URL, app_name, taskid)
task_response = requests.get(url, headers=HEADERS)
LOG.debug(task_response.json())
assert task_response.ok
task_state = task_response.json()
status = task_state['status']
LOG.info('Current task status: %s', status)
if status == 'SUCCEEDED':
return status
elif status == 'TERMINAL':
raise SpinnakerTaskError(task_state)
else:
raise ValueError
| """Check Taskid status."""
import logging
import requests
from tryagain import retries
from ..consts import API_URL, HEADERS
from ..exceptions import SpinnakerTaskError
LOG = logging.getLogger(__name__)
@retries(max_attempts=10, wait=10, exceptions=(AssertionError, ValueError))
def check_task(taskid, app_name):
"""Check task status.
Args:
taskid: the task id returned from create_elb.
app_name: application name related to this task.
Returns:
polls for task status.
"""
try:
taskurl = taskid.get('ref', '0000')
except AttributeError:
taskurl = taskid
taskid = taskurl.split('/tasks/')[-1]
LOG.info('Checking taskid %s', taskid)
url = '{0}/applications/{1}/tasks/{2}'.format(API_URL, app_name, taskid)
task_response = requests.get(url, headers=HEADERS)
LOG.debug(task_response.json())
assert task_response.ok
task_state = task_response.json()
status = task_state['status']
LOG.info('Current task status: %s', status)
if status == 'SUCCEEDED':
return status
elif status == 'TERMINAL':
raise SpinnakerTaskError(task_state)
else:
raise ValueError
| Use global API_URL and HEADERS | refactor: Use global API_URL and HEADERS
See also: PSOBAT-1197
| Python | apache-2.0 | gogoair/foremast,gogoair/foremast | ---
+++
@@ -4,10 +4,9 @@
import requests
from tryagain import retries
+from ..consts import API_URL, HEADERS
from ..exceptions import SpinnakerTaskError
-HEADERS = {'Content-Type': 'application/json', 'Accept': '*/*'}
-GATE_URL = "http://gate-api.build.example.com:8084"
LOG = logging.getLogger(__name__)
@@ -31,7 +30,7 @@
LOG.info('Checking taskid %s', taskid)
- url = '{0}/applications/{1}/tasks/{2}'.format(GATE_URL, app_name, taskid)
+ url = '{0}/applications/{1}/tasks/{2}'.format(API_URL, app_name, taskid)
task_response = requests.get(url, headers=HEADERS)
LOG.debug(task_response.json()) |
939998db349c364aa0f5ba4705d4feb2da7104d5 | nn/flags.py | nn/flags.py | import functools
import tensorflow as tf
FLAGS = tf.app.flags.FLAGS
tf.app.flags.DEFINE_string("batch-size", 64, "")
tf.app.flags.DEFINE_float("dropout-prob", 0, "")
tf.app.flags.DEFINE_string("word-file", None, "")
tf.app.flags.DEFINE_integer("num-threads-per-queue", 2, "")
tf.app.flags.DEFINE_integer("queue-capacity", 2, "")
tf.app.flags.DEFINE_string("length-boundaries", "", "")
tf.app.flags.DEFINE_string("rnn-cell", "ln_lstm", "Default RNN cell")
tf.app.flags.DEFINE_string("float32", "", "")
@functools.lru_cache()
def words():
with open(tf.app.flags.FLAGS.word_file) as file_:
return sorted([line.strip() for line in file_.readlines()])
@functools.lru_cache()
def word_indices():
# 0 -> null, 1 -> unknown
return { word: index + 2 for index, word in enumerate(flags.words()) }
@functools.lru_cache()
def word_space_size():
return len(words())
def rnn_cell():
from .rnn import cell
return getattr(cell, FLAGS.rnn_cell)
def float_type():
return getattr(tf, FLAGS.float_type)
| import functools
import tensorflow as tf
FLAGS = tf.app.flags.FLAGS
tf.app.flags.DEFINE_string("batch-size", 64, "")
tf.app.flags.DEFINE_float("dropout-prob", 0, "")
tf.app.flags.DEFINE_string("word-file", None, "")
tf.app.flags.DEFINE_integer("num-threads-per-queue", 2, "")
tf.app.flags.DEFINE_integer("queue-capacity", 2, "")
tf.app.flags.DEFINE_string("length-boundaries", "", "")
tf.app.flags.DEFINE_string("rnn-cell", "ln_lstm", "Default RNN cell")
tf.app.flags.DEFINE_string("float-type", "float32", "")
@functools.lru_cache()
def words():
with open(tf.app.flags.FLAGS.word_file) as file_:
return sorted([line.strip() for line in file_.readlines()])
@functools.lru_cache()
def word_indices():
# 0 -> null, 1 -> unknown
return { word: index + 2 for index, word in enumerate(flags.words()) }
@functools.lru_cache()
def word_space_size():
return len(words())
def rnn_cell():
from .rnn import cell
return getattr(cell, FLAGS.rnn_cell)
def float_type():
return getattr(tf, FLAGS.float_type)
| Fix float type flag definition | Fix float type flag definition
| Python | unlicense | raviqqe/tensorflow-extenteten,raviqqe/tensorflow-extenteten | ---
+++
@@ -12,7 +12,7 @@
tf.app.flags.DEFINE_integer("queue-capacity", 2, "")
tf.app.flags.DEFINE_string("length-boundaries", "", "")
tf.app.flags.DEFINE_string("rnn-cell", "ln_lstm", "Default RNN cell")
-tf.app.flags.DEFINE_string("float32", "", "")
+tf.app.flags.DEFINE_string("float-type", "float32", "")
@functools.lru_cache() |
10db5e8b893a84e765162535f64e1ede81d48b47 | empty_check.py | empty_check.py | from django.core.exceptions import ValidationError
class EmptyCheck(object):
def __call__(self, value):
if len(value.strip()) == 0:
raise ValidationError("Value cannot be empty")
| from django.core.exceptions import ValidationError
# Usage example in a custom form
# firstname = forms.CharField(validators = [EmptyCheck()])
class EmptyCheck(object):
def __call__(self, value):
if len(value.strip()) == 0:
raise ValidationError("Value cannot be empty")
| Add comment to show usage example | Add comment to show usage example | Python | mit | vishalsodani/django-empty-check-validator | ---
+++
@@ -1,5 +1,7 @@
from django.core.exceptions import ValidationError
+# Usage example in a custom form
+# firstname = forms.CharField(validators = [EmptyCheck()])
class EmptyCheck(object):
def __call__(self, value):
if len(value.strip()) == 0: |
34812fe2deec64229efd4119640f3c2ddf0ed415 | visualize.py | visualize.py | '''
Create a visual representation of the various DAGs defined
'''
import sys
import requests
import networkx as nx
import matplotlib.pyplot as plt
if __name__ == '__main__':
g = nx.DiGraph()
labels = {
'edges': {},
'nodes': {},
}
nodes = {}
for routeKey, routeMap in requests.get(sys.argv[1]).json().iteritems():
for i, node in enumerate(routeMap['Path']):
g.add_node(node['Name'])
labels['nodes'][node['Name']] = node['Name']
if i - 1 >= 0:
g.add_edge(routeMap['Path'][i-1]['Name'], routeMap['Path'][i]['Name'])
labels['edges'][(routeMap['Path'][i-1]['Name'], routeMap['Path'][i]['Name'])] = (routeMap['Path'][i-1]['Name'], routeMap['Path'][i]['Name'])
nx.draw_networkx(g, with_labels=True)
# add labels
#nx.draw_networkx_labels(g, pos, labels['nodes'])
#nx.draw_networkx_edge_labels(g, pos, labels['edges'])
# write out the graph
plt.savefig(
'topology.png',
dpi=400.0,
)
plt.show() # in case people have the required libraries to make it happen
| '''
Create a visual representation of the various DAGs defined
'''
import sys
import requests
import networkx as nx
import matplotlib.pyplot as plt
if __name__ == '__main__':
g = nx.DiGraph()
labels = {
'edges': {},
'nodes': {},
}
for routeKey, routeMap in requests.get(sys.argv[1]).json().iteritems():
for i, node in enumerate(routeMap['Path']):
g.add_node(node['Name'])
labels['nodes'][node['Name']] = node['Name']
if i - 1 >= 0:
g.add_edge(routeMap['Path'][i-1]['Name'], routeMap['Path'][i]['Name'])
labels['edges'][(routeMap['Path'][i-1]['Name'], routeMap['Path'][i]['Name'])] = (routeMap['Path'][i-1]['Name'], routeMap['Path'][i]['Name'])
pos = nx.drawing.spring_layout(
g,
scale=10.0,
)
nx.draw_networkx(
g,
pos=pos,
with_labels=True,
font_size=8,
)
# write out the graph
plt.savefig(
'topology.png',
dpi=400.0,
)
plt.show() # in case people have the required libraries to make it happen
| Make the sprint layout a bit easier to look at | Make the sprint layout a bit easier to look at
| Python | mit | jacksontj/dnms,jacksontj/dnms | ---
+++
@@ -15,8 +15,6 @@
'nodes': {},
}
- nodes = {}
-
for routeKey, routeMap in requests.get(sys.argv[1]).json().iteritems():
for i, node in enumerate(routeMap['Path']):
g.add_node(node['Name'])
@@ -26,11 +24,16 @@
labels['edges'][(routeMap['Path'][i-1]['Name'], routeMap['Path'][i]['Name'])] = (routeMap['Path'][i-1]['Name'], routeMap['Path'][i]['Name'])
- nx.draw_networkx(g, with_labels=True)
-
- # add labels
- #nx.draw_networkx_labels(g, pos, labels['nodes'])
- #nx.draw_networkx_edge_labels(g, pos, labels['edges'])
+ pos = nx.drawing.spring_layout(
+ g,
+ scale=10.0,
+ )
+ nx.draw_networkx(
+ g,
+ pos=pos,
+ with_labels=True,
+ font_size=8,
+ )
# write out the graph
plt.savefig( |
67a230dd5673601f2e1f1a8c3deb8597f29287db | src/tmlib/workflow/align/args.py | src/tmlib/workflow/align/args.py | from tmlib.workflow.args import BatchArguments
from tmlib.workflow.args import SubmissionArguments
from tmlib.workflow.args import Argument
from tmlib.workflow import register_batch_args
from tmlib.workflow import register_submission_args
@register_batch_args('align')
class AlignBatchArguments(BatchArguments):
ref_cycle = Argument(
type=int, required=True, flag='c',
help='''zero-based index of the cycle whose sites should be used
as reference
'''
)
ref_wavelength = Argument(
type=str, required=True, flag='w',
help='name of the wavelength whose images should be used as reference'
)
batch_size = Argument(
type=int, default=10, flag='b',
help='number of image files that should be processed per job'
)
@register_submission_args('align')
class AlignSubmissionArguments(SubmissionArguments):
pass
| from tmlib.workflow.args import BatchArguments
from tmlib.workflow.args import SubmissionArguments
from tmlib.workflow.args import Argument
from tmlib.workflow import register_batch_args
from tmlib.workflow import register_submission_args
@register_batch_args('align')
class AlignBatchArguments(BatchArguments):
ref_cycle = Argument(
type=int, required=True, flag='c',
help='''zero-based index of the cycle whose sites should be used
as reference
'''
)
ref_wavelength = Argument(
type=str, required=True, flag='w',
help='name of the wavelength whose images should be used as reference'
)
batch_size = Argument(
type=int, default=100, flag='b',
help='number of image files that should be processed per job'
)
@register_submission_args('align')
class AlignSubmissionArguments(SubmissionArguments):
pass
| Increase default batch size for align step | Increase default batch size for align step
| Python | agpl-3.0 | TissueMAPS/TmLibrary,TissueMAPS/TmLibrary,TissueMAPS/TmLibrary,TissueMAPS/TmLibrary,TissueMAPS/TmLibrary | ---
+++
@@ -21,7 +21,7 @@
)
batch_size = Argument(
- type=int, default=10, flag='b',
+ type=int, default=100, flag='b',
help='number of image files that should be processed per job'
)
|
057510c78f80c3592c562006413049ab1292d0a3 | ipaqe_provision_hosts/backend/base.py | ipaqe_provision_hosts/backend/base.py | # author: Milan Kubik
NOT_IMPLEMENTED_MSG = "You need to override this method in a subclass"
class IDMBackendException(Exception):
pass
class VMsNotCreatedError(IDMBackendException):
pass
class IDMBackendMissingName(IDMBackendException):
pass
class IDMBackendBase(object):
"""IDMBackendBase class
This class represents a contract between the
idm-prepare-hosts utility and a backend implementation.
"""
def __init__(self, config=None):
self._config = config or {}
self._vms = []
@property
def vms(self):
"""The attribute returns a list of host entries"""
if not self._vms:
raise VMsNotCreatedError("No VMs were provisioned yet")
else:
return self._vms
def provision_resources(self, vm_count):
"""Provision the hosts in a backend"""
raise NotImplementedError(NOT_IMPLEMENTED_MSG)
def delete_resources(self):
"""Delete the resources provisioned by the backend"""
raise NotImplementedError(NOT_IMPLEMENTED_MSG)
| # author: Milan Kubik
NOT_IMPLEMENTED_MSG = "You need to override this method in a subclass"
from ipaqe_provision_hosts.errors import IPAQEProvisionerError
class VMsNotCreatedError(IPAQEProvisionerError):
pass
class IDMBackendBase(object):
"""IDMBackendBase class
This class represents a contract between the
idm-prepare-hosts utility and a backend implementation.
"""
def __init__(self, config=None):
self._config = config or {}
self._vms = []
@property
def vms(self):
"""The attribute returns a list of host entries"""
if not self._vms:
raise VMsNotCreatedError("No VMs were provisioned yet")
else:
return self._vms
def provision_resources(self, vm_count):
"""Provision the hosts in a backend"""
raise NotImplementedError(NOT_IMPLEMENTED_MSG)
def delete_resources(self):
"""Delete the resources provisioned by the backend"""
raise NotImplementedError(NOT_IMPLEMENTED_MSG)
| Replace the exceptions in backend classes | Replace the exceptions in backend classes
| Python | mit | apophys/ipaqe-provision-hosts | ---
+++
@@ -2,15 +2,10 @@
NOT_IMPLEMENTED_MSG = "You need to override this method in a subclass"
-
-class IDMBackendException(Exception):
- pass
+from ipaqe_provision_hosts.errors import IPAQEProvisionerError
-class VMsNotCreatedError(IDMBackendException):
- pass
-
-class IDMBackendMissingName(IDMBackendException):
+class VMsNotCreatedError(IPAQEProvisionerError):
pass
|
ecde3e00de3782024f0dfbab7f2d7912ab62342b | src/birding/__init__.py | src/birding/__init__.py | from __future__ import absolute_import, print_function
import logging
from . import bolt, config, follow, search, spout, twitter_api
from .search import SearchManager
from .twitter_api import Twitter
from .version import VERSION, __version__
from .version import __doc__ as __license__
__all__ = [
'SearchManager',
'Twitter',
'VERSION',
'__license__',
'__version__',
'bolt',
'config',
'follow',
'search',
'spout',
'twitter_api',
]
# Configure the logger. No configuration is exposed by birding itself. A
# project using birding can change the log level after importing `birding`
# with:
#
# logging.getLogger('birding').setLevel(logging.DEBUG)
#
logger = logging.getLogger('birding')
logger.addHandler(logging.StreamHandler())
logger.setLevel(logging.WARNING)
| from __future__ import absolute_import, print_function
import logging
from . import bolt, config, follow, search, spout, twitter_api
from .version import VERSION, __version__
from .version import __doc__ as __license__
__all__ = [
'VERSION',
'__license__',
'__version__',
'bolt',
'config',
'follow',
'search',
'spout',
'twitter_api',
]
# Configure the logger. No configuration is exposed by birding itself. A
# project using birding can change the log level after importing `birding`
# with:
#
# logging.getLogger('birding').setLevel(logging.DEBUG)
#
logger = logging.getLogger('birding')
logger.addHandler(logging.StreamHandler())
logger.setLevel(logging.WARNING)
| Remove unnecessary objects from root namespace. | Remove unnecessary objects from root namespace.
| Python | apache-2.0 | Parsely/birding,Parsely/birding | ---
+++
@@ -3,15 +3,11 @@
import logging
from . import bolt, config, follow, search, spout, twitter_api
-from .search import SearchManager
-from .twitter_api import Twitter
from .version import VERSION, __version__
from .version import __doc__ as __license__
__all__ = [
- 'SearchManager',
- 'Twitter',
'VERSION',
'__license__',
'__version__', |
8a43cf58791a665a4fc23bc5d0911af61f7e1fb6 | qipr_approver/approver/views/similar_projects.py | qipr_approver/approver/views/similar_projects.py | from django.shortcuts import redirect
from approver.workflows import project_crud
from approver.decorators import login_required
import approver.utils as utils
from django.core.urlresolvers import reverse
@login_required
def similar_projects(request, project_id=None,from_page=None):
project = project_crud.get_project_or_none(project_id)
if project is None:
utils.dashboard_redirect_and_toast(request, 'Invalid request'.format(project_id))
elif request.method == 'GET':
project_scores = project_crud.get_similar_projects(project)
if (len(project_scores) == 0) :
utils.set_toast(request.session, 'No relevant projects were found!')
if(from_page == "dashboard") :
return redirect(reverse("approver:dashboard"))
else :
return redirect(reverse("approver:approve") + str(project.id) + '/')
context = {
'content': 'approver/similar_projects.html',
'project_scores': project_scores,
'project_id' : project_id,
}
return utils.layout_render(request, context)
elif request.method == 'POST':
return redirect(reverse("approver:approve") + str(project.id) + '/') | from django.shortcuts import redirect
from django.core.urlresolvers import reverse
from django.contrib.auth.decorators import login_required
from approver.workflows import project_crud
import approver.utils as utils
@login_required
def similar_projects(request, project_id=None,from_page=None):
project = project_crud.get_project_or_none(project_id)
if project is None:
utils.dashboard_redirect_and_toast(request, 'Invalid request'.format(project_id))
elif request.method == 'GET':
project_scores = project_crud.get_similar_projects(project)
if (len(project_scores) == 0) :
utils.set_toast(request.session, 'No relevant projects were found!')
if(from_page == "dashboard") :
return redirect(reverse("approver:dashboard"))
else :
return redirect(reverse("approver:approve") + str(project.id) + '/')
context = {
'content': 'approver/similar_projects.html',
'project_scores': project_scores,
'project_id' : project_id,
}
return utils.layout_render(request, context)
elif request.method == 'POST':
return redirect(reverse("approver:approve") + str(project.id) + '/')
| Add shib auth to similar projects page | Add shib auth to similar projects page
| Python | apache-2.0 | DevMattM/qipr_approver,DevMattM/qipr_approver,ctsit/qipr_approver,ctsit/qipr_approver,ctsit/qipr_approver,DevMattM/qipr_approver,DevMattM/qipr_approver,PFWhite/qipr_approver,DevMattM/qipr_approver,PFWhite/qipr_approver,ctsit/qipr_approver,PFWhite/qipr_approver,ctsit/qipr_approver,PFWhite/qipr_approver,PFWhite/qipr_approver | ---
+++
@@ -1,13 +1,14 @@
from django.shortcuts import redirect
+from django.core.urlresolvers import reverse
+from django.contrib.auth.decorators import login_required
from approver.workflows import project_crud
-from approver.decorators import login_required
+
import approver.utils as utils
-from django.core.urlresolvers import reverse
@login_required
def similar_projects(request, project_id=None,from_page=None):
-
+
project = project_crud.get_project_or_none(project_id)
if project is None:
@@ -26,7 +27,7 @@
'content': 'approver/similar_projects.html',
'project_scores': project_scores,
'project_id' : project_id,
- }
+ }
return utils.layout_render(request, context)
elif request.method == 'POST': |
1e574befc82b7df53fe62b9c5089772172f99178 | apps/reactions/serializers.py | apps/reactions/serializers.py | from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField
from django.contrib.auth.models import User
from rest_framework import serializers
from .models import Reaction
from rest_framework.fields import HyperlinkedIdentityField
class ReactionAuthorSerializer(serializers.ModelSerializer):
picture = SorlImageField('userprofile.picture', '90x90', crop='center')
class Meta:
model = User
fields = ('first_name', 'last_name', 'picture')
class ReactionDetailSerializer(serializers.ModelSerializer):
# Read-only fields.
created = serializers.Field()
# Custom fields.
author = ReactionAuthorSerializer()
# TODO: This isn't work with the pattern: api/blogs/<slug>/reactions/<pk>
# Delete or fix this ... we don't really need it so removing it is ok but it's nice to have.
# url = HyperlinkedIdentityField(view_name='reactions:reaction-detail')
class Meta:
model = Reaction
fields = ('created', 'author', 'reaction')
class ReactionListSerializer(ReactionDetailSerializer):
class Meta:
model = Reaction
fields = ('created', 'author', 'reaction')
| from apps.bluebottle_utils.serializers import SorlImageField, SlugHyperlinkedIdentityField
from django.contrib.auth.models import User
from rest_framework import serializers
from .models import Reaction
from rest_framework.fields import HyperlinkedIdentityField
class ReactionAuthorSerializer(serializers.ModelSerializer):
picture = SorlImageField('userprofile.picture', '90x90', crop='center')
class Meta:
model = User
fields = ('first_name', 'last_name', 'picture')
class ReactionDetailSerializer(serializers.ModelSerializer):
# Read-only fields.
created = serializers.Field()
# Custom fields.
author = ReactionAuthorSerializer()
# TODO: This isn't work with the pattern: api/blogs/<slug>/reactions/<pk>
# Delete or fix this ... we don't really need it so removing it is ok but it's nice to have.
# url = HyperlinkedIdentityField(view_name='reactions:reaction-detail')
class Meta:
model = Reaction
fields = ('created', 'author', 'reaction', 'id')
class ReactionListSerializer(ReactionDetailSerializer):
class Meta:
model = Reaction
fields = ('created', 'author', 'reaction', 'id')
| Add id to API reponse for Reaction. | Add id to API reponse for Reaction.
| Python | bsd-3-clause | onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site | ---
+++
@@ -25,11 +25,11 @@
class Meta:
model = Reaction
- fields = ('created', 'author', 'reaction')
+ fields = ('created', 'author', 'reaction', 'id')
class ReactionListSerializer(ReactionDetailSerializer):
class Meta:
model = Reaction
- fields = ('created', 'author', 'reaction')
+ fields = ('created', 'author', 'reaction', 'id') |
896b385f983ecf939bdc2ea938b9949fdc3fdbb8 | colorise/color_tools.py | colorise/color_tools.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Functions for converting and comparing colors."""
import colorsys
import math
import operator
def hls_to_rgb(hue, lightness, saturation):
"""Convert HLS (hue, lightness, saturation) values to RGB."""
return tuple(int(math.ceil(c * 255.))
for c in colorsys.hls_to_rgb(hue, lightness, saturation))
def hsv_to_rgb(hue, saturation, value):
"""Convert HSV (hue, saturation, value) values to RGB."""
return tuple(int(c * 255.)
for c in colorsys.hsv_to_rgb(hue/360.,
saturation/100.,
value/100.))
def color_difference(rgb1, rgb2):
"""Return the sums of component differences between two colors."""
return sum(abs(i - j) for i, j in zip(rgb1, rgb2))
def color_distance(rgb1, rgb2):
"""Compute the Euclidian distance between two colors."""
r1, g1, b1 = rgb1
r2, g2, b2 = rgb2
return math.sqrt((r2 - r1)**2 + (g2 - g1)**2 + (b2 - b1)**2)
def closest_color(rgb, clut):
"""Return the CLUT index of the closest RGB color to a given RGB tuple."""
# Generate a list of tuples of CLUT indices and the color difference value
indexed_diffs = ((idx, color_difference(rgb, clut[idx])) for idx in clut)
return min(indexed_diffs, key=operator.itemgetter(1))[0]
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Functions for converting and comparing colors."""
import colorsys
import math
import operator
def hls_to_rgb(hue, lightness, saturation):
"""Convert HLS (hue, lightness, saturation) values to RGB."""
return tuple(int(math.ceil(c * 255.))
for c in colorsys.hls_to_rgb(hue, lightness, saturation))
def hsv_to_rgb(hue, saturation, value):
"""Convert HSV (hue, saturation, value) values to RGB."""
return tuple(int(c * 255.)
for c in colorsys.hsv_to_rgb(hue/360.,
saturation/100.,
value/100.))
def color_difference(rgb1, rgb2):
"""Return the sums of component differences between two colors."""
return sum(abs(i - j) for i, j in zip(rgb1, rgb2))
def closest_color(rgb, clut):
"""Return the CLUT index of the closest RGB color to a given RGB tuple."""
# Generate a list of tuples of CLUT indices and the color difference value
indexed_diffs = ((idx, color_difference(rgb, clut[idx])) for idx in clut)
return min(indexed_diffs, key=operator.itemgetter(1))[0]
| Remove unused color distance function | Remove unused color distance function
| Python | bsd-3-clause | MisanthropicBit/colorise | ---
+++
@@ -27,14 +27,6 @@
return sum(abs(i - j) for i, j in zip(rgb1, rgb2))
-def color_distance(rgb1, rgb2):
- """Compute the Euclidian distance between two colors."""
- r1, g1, b1 = rgb1
- r2, g2, b2 = rgb2
-
- return math.sqrt((r2 - r1)**2 + (g2 - g1)**2 + (b2 - b1)**2)
-
-
def closest_color(rgb, clut):
"""Return the CLUT index of the closest RGB color to a given RGB tuple."""
# Generate a list of tuples of CLUT indices and the color difference value |
1c0644d5629008c98f8402b95b797beea1a50bc5 | promgen/sender/__init__.py | promgen/sender/__init__.py | import logging
from promgen.models import Project, Service
logger = logging.getLogger(__name__)
class SenderBase(object):
def send(self, data):
sent = 0
for alert in data['alerts']:
if 'project' in alert['labels']:
logger.debug('Checking for projects')
for project in Project.objects.filter(name=alert['labels']['project']):
logger.debug('Checking %s', project)
for sender in project.sender.all():
logger.debug('Sending to %s', sender)
if self._send(sender.value, alert, data):
sent += 1
if 'service' in alert['labels']:
logger.debug('Checking for service')
for service in Service.objects.filter(name=alert['labels']['service']):
logger.debug('Checking %s', service)
for sender in service.sender.all():
logger.debug('Sending to %s', sender)
if self._send(sender.value, alert, data):
sent += 1
if sent == 0:
logger.debug('No senders configured for project or service %s', alert['labels']['project'])
return sent
def test(self, target, alert):
logger.debug('Sending test message to %s', target)
self._send(target, alert, {'externalURL': ''})
| import logging
from promgen.models import Project, Service
logger = logging.getLogger(__name__)
class SenderBase(object):
MAPPING = [
('project', Project),
('service', Service),
]
def send(self, data):
sent = 0
for alert in data['alerts']:
for label, klass in self.MAPPING:
logger.debug('Checking for %s', label)
if label in alert['labels']:
logger.debug('Checking for %s %s', label, klass)
for obj in klass.objects.filter(name=alert['labels'][label]):
for sender in obj.sender.filter(sender=self.__module__):
logger.debug('Sending to %s', sender)
if self._send(sender.value, alert, data):
sent += 1
if sent == 0:
logger.debug('No senders configured for project or service')
return sent
def test(self, target, alert):
logger.debug('Sending test message to %s', target)
self._send(target, alert, {'externalURL': ''})
| Add a MAPPING table to remove duplicated loop logic | Add a MAPPING table to remove duplicated loop logic
| Python | mit | kfdm/promgen,kfdm/promgen,kfdm/promgen,kfdm/promgen | ---
+++
@@ -6,27 +6,25 @@
class SenderBase(object):
+ MAPPING = [
+ ('project', Project),
+ ('service', Service),
+ ]
+
def send(self, data):
sent = 0
for alert in data['alerts']:
- if 'project' in alert['labels']:
- logger.debug('Checking for projects')
- for project in Project.objects.filter(name=alert['labels']['project']):
- logger.debug('Checking %s', project)
- for sender in project.sender.all():
- logger.debug('Sending to %s', sender)
- if self._send(sender.value, alert, data):
- sent += 1
- if 'service' in alert['labels']:
- logger.debug('Checking for service')
- for service in Service.objects.filter(name=alert['labels']['service']):
- logger.debug('Checking %s', service)
- for sender in service.sender.all():
- logger.debug('Sending to %s', sender)
- if self._send(sender.value, alert, data):
- sent += 1
+ for label, klass in self.MAPPING:
+ logger.debug('Checking for %s', label)
+ if label in alert['labels']:
+ logger.debug('Checking for %s %s', label, klass)
+ for obj in klass.objects.filter(name=alert['labels'][label]):
+ for sender in obj.sender.filter(sender=self.__module__):
+ logger.debug('Sending to %s', sender)
+ if self._send(sender.value, alert, data):
+ sent += 1
if sent == 0:
- logger.debug('No senders configured for project or service %s', alert['labels']['project'])
+ logger.debug('No senders configured for project or service')
return sent
def test(self, target, alert): |
e174a898595664ff291cbf8ccda0f1c404a73575 | control/server.py | control/server.py | import asyncore
import socket
from logging import error, info, warning
from client import Client
class Server(asyncore.dispatcher):
def __init__(self, port, host="localhost"):
asyncore.dispatcher.__init__(self)
self.create_socket()
self.set_reuse_addr()
self.bind((host, port))
self.listen(5)
self.connect_fn = None
self.msg_fn = None
self.close_fn = None
self.clients = []
def handle_accepted(self, sock, addr):
new_client = Client(sock)
new_client.msg_fn = self.msg_fn
new_client.close_fn = self.close_fn
self.clients.append(new_client)
if self.connect_fn is not None:
self.connect_fn(new_client)
def broadcast(self, msg):
for client in self.clients:
client.send_msg(msg)
| import asyncore
import socket
from logging import error, info, warning
from client import Client
class Server(asyncore.dispatcher):
def __init__(self, port, connect_fn=None, msg_fn=None, close_fn=None):
asyncore.dispatcher.__init__(self)
self.create_socket()
self.set_reuse_addr()
self.bind(('localhost', port))
self.listen(5)
self.client_connect_fn = connect_fn
self.client_msg_fn = msg_fn
self.client_close_fn = close_fn
self.clients = []
def handle_accepted(self, sock, addr):
client = Client(sock)
client.msg_fn = self.client_msg_fn
client.close_fn = self.client_close
self.clients.append(client)
if self.client_connect_fn:
self.client_connect_fn(client)
def client_close(self, client):
self.clients.remove(client)
if self.client_close_fn:
self.client_close_fn(client)
def broadcast(self, msg):
for client in self.clients:
client.send_msg(msg)
| Remove disconnected client from clients list, allow client callbacks to be set in constructor. | Remove disconnected client from clients list, allow client callbacks to be set in constructor.
| Python | mit | zwarren/morse-car-controller,zwarren/morse-car-controller | ---
+++
@@ -6,28 +6,34 @@
class Server(asyncore.dispatcher):
- def __init__(self, port, host="localhost"):
+ def __init__(self, port, connect_fn=None, msg_fn=None, close_fn=None):
asyncore.dispatcher.__init__(self)
self.create_socket()
self.set_reuse_addr()
- self.bind((host, port))
+ self.bind(('localhost', port))
self.listen(5)
- self.connect_fn = None
- self.msg_fn = None
- self.close_fn = None
+ self.client_connect_fn = connect_fn
+ self.client_msg_fn = msg_fn
+ self.client_close_fn = close_fn
self.clients = []
def handle_accepted(self, sock, addr):
- new_client = Client(sock)
- new_client.msg_fn = self.msg_fn
- new_client.close_fn = self.close_fn
+ client = Client(sock)
+ client.msg_fn = self.client_msg_fn
+ client.close_fn = self.client_close
- self.clients.append(new_client)
+ self.clients.append(client)
- if self.connect_fn is not None:
- self.connect_fn(new_client)
+ if self.client_connect_fn:
+ self.client_connect_fn(client)
+
+ def client_close(self, client):
+ self.clients.remove(client)
+
+ if self.client_close_fn:
+ self.client_close_fn(client)
def broadcast(self, msg):
for client in self.clients: |
91720739af3c7b35e331949cdd64a98023e23799 | parkings/api/public/parking_area.py | parkings/api/public/parking_area.py | from rest_framework import viewsets
from rest_framework_gis.pagination import GeoJsonPagination
from rest_framework_gis.serializers import GeoFeatureModelSerializer, GeometrySerializerMethodField
from parkings.models import ParkingArea
class ParkingAreaSerializer(GeoFeatureModelSerializer):
wgs84_areas = GeometrySerializerMethodField()
def get_wgs84_areas(self, area):
return area.areas.transform(4326, clone=True)
class Meta:
model = ParkingArea
geo_field = 'wgs84_areas'
fields = (
'id',
'space_amount_estimate',
)
class PublicAPIParkingAreaViewSet(viewsets.ReadOnlyModelViewSet):
queryset = ParkingArea.objects.all()
serializer_class = ParkingAreaSerializer
pagination_class = GeoJsonPagination
| from rest_framework import viewsets
from rest_framework_gis.pagination import GeoJsonPagination
from rest_framework_gis.serializers import GeoFeatureModelSerializer, GeometrySerializerMethodField
from parkings.models import ParkingArea
from ..common import WGS84InBBoxFilter
class ParkingAreaSerializer(GeoFeatureModelSerializer):
wgs84_areas = GeometrySerializerMethodField()
def get_wgs84_areas(self, area):
return area.areas.transform(4326, clone=True)
class Meta:
model = ParkingArea
geo_field = 'wgs84_areas'
fields = (
'id',
'space_amount_estimate',
)
class PublicAPIParkingAreaViewSet(viewsets.ReadOnlyModelViewSet):
queryset = ParkingArea.objects.all()
serializer_class = ParkingAreaSerializer
pagination_class = GeoJsonPagination
bbox_filter_field = 'areas'
filter_backends = (WGS84InBBoxFilter,)
bbox_filter_include_overlapping = True
| Add bbox to parking area view set | Add bbox to parking area view set
| Python | mit | tuomas777/parkkihubi | ---
+++
@@ -3,6 +3,8 @@
from rest_framework_gis.serializers import GeoFeatureModelSerializer, GeometrySerializerMethodField
from parkings.models import ParkingArea
+
+from ..common import WGS84InBBoxFilter
class ParkingAreaSerializer(GeoFeatureModelSerializer):
@@ -24,3 +26,6 @@
queryset = ParkingArea.objects.all()
serializer_class = ParkingAreaSerializer
pagination_class = GeoJsonPagination
+ bbox_filter_field = 'areas'
+ filter_backends = (WGS84InBBoxFilter,)
+ bbox_filter_include_overlapping = True |
d52034eddeb510acc367c87c88e4277994157338 | githubsetupircnotifications.py | githubsetupircnotifications.py | """
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('--org')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
print('Failed to sign into github')
sys.exit(1)
org = github.organization(args.org)
if org is None:
print('Organization "{}" does not appear to exist'.format(args.org))
sys.exit(1)
| """
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('--org')
parser.add_argument('--channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
print('Failed to sign into github')
sys.exit(1)
org = github.organization(args.org)
if org is None:
print('Organization "{}" does not appear to exist'.format(args.org))
sys.exit(1)
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
for r in org.iter_repos():
r.create_hook('irc', conf)
| Create the hook for each repo | Create the hook for each repo
| Python | mit | kragniz/github-setup-irc-notifications | ---
+++
@@ -15,6 +15,7 @@
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('--org')
+ parser.add_argument('--channel')
args = parser.parse_args()
if args.password is None:
@@ -34,3 +35,21 @@
if org is None:
print('Organization "{}" does not appear to exist'.format(args.org))
sys.exit(1)
+
+ conf = {'nickserv_password': '',
+ 'no_colors': '0',
+ 'password': '',
+ 'branch_regexes': '',
+ 'room': args.channel,
+ 'ssl': '0',
+ 'port': '',
+ 'branches': '',
+ 'server': 'chat.freenode.net',
+ 'long_url': '0',
+ 'notice': '0',
+ 'message_without_join': '1',
+ 'nick': 'github'
+ }
+
+ for r in org.iter_repos():
+ r.create_hook('irc', conf) |
ed4c80aa8e9ee628876c3cc96907ca407ee4ff5d | backend/scripts/ddirdenorm.py | backend/scripts/ddirdenorm.py | #!/usr/bin/env python
import rethinkdb as r
import optparse
if __name__ == "__main__":
parser = optparse.OptionParser()
parser.add_option("-p", "--port", dest="port",
help="rethinkdb port", default=30815)
(options, args) = parser.parse_args()
conn = r.connect('localhost', int(options.port), db='materialscommons')
selection = list(r.table('datadirs').run(conn))
for datadir in selection:
print "Updating datadir %s" % (datadir['name'])
ddir = {}
ddir['id'] = datadir['id']
ddir['name'] = datadir['name']
ddir['owner'] = datadir['owner']
ddir['birthtime'] = datadir['birthtime']
ddir['datafiles'] = []
for dfid in datadir['datafiles']:
datafile = r.table('datafiles').get(dfid).run(conn)
df = {}
df['id'] = datafile['id']
df['name'] = datafile['name']
df['owner'] = datafile['owner']
df['birthtime'] = datafile['birthtime']
df['size'] = datafile['size']
df['checksum'] = datafile['checksum']
ddir['datafiles'].append(df)
r.table('datadirs_denorm').insert(ddir).run(conn)
| #!/usr/bin/env python
import rethinkdb as r
import optparse
if __name__ == "__main__":
parser = optparse.OptionParser()
parser.add_option("-p", "--port", dest="port",
help="rethinkdb port", default=30815)
(options, args) = parser.parse_args()
conn = r.connect('localhost', int(options.port), db='materialscommons')
selection = list(r.table('datadirs').run(conn))
for datadir in selection:
print "Updating datadir %s" % (datadir['name'])
ddir = {}
ddir['id'] = datadir['id']
ddir['name'] = datadir['name']
ddir['owner'] = datadir['owner']
ddir['birthtime'] = datadir['birthtime']
ddir['datafiles'] = []
for dfid in datadir['datafiles']:
datafile = r.table('datafiles').get(dfid).run(conn)
if datafile is None:
continue
df = {}
df['id'] = datafile['id']
df['name'] = datafile['name']
df['owner'] = datafile['owner']
df['birthtime'] = datafile['birthtime']
df['size'] = datafile['size']
df['checksum'] = datafile['checksum']
ddir['datafiles'].append(df)
r.table('datadirs_denorm').insert(ddir).run(conn)
| Handle non-existent files in the database. | Handle non-existent files in the database.
| Python | mit | materials-commons/materialscommons.org,materials-commons/materialscommons.org,materials-commons/materialscommons.org,materials-commons/materialscommons.org,materials-commons/materialscommons.org | ---
+++
@@ -22,6 +22,8 @@
ddir['datafiles'] = []
for dfid in datadir['datafiles']:
datafile = r.table('datafiles').get(dfid).run(conn)
+ if datafile is None:
+ continue
df = {}
df['id'] = datafile['id']
df['name'] = datafile['name'] |
e7627ee439e2e4f17466bf124629ae353460a68d | __init__.py | __init__.py | # -*- coding: utf-8 -*-
##############################################################################
#
# Copyright (C) 2012 OpenERP - Team de Localización Argentina.
# https://launchpad.net/~openerp-l10n-ar-localization
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import test
import afip
import invoice
import config
import partner
import account
import country
import report
import currency
import product
# vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
| # -*- coding: utf-8 -*-
##############################################################################
#
# Copyright (C) 2012 OpenERP - Team de Localización Argentina.
# https://launchpad.net/~openerp-l10n-ar-localization
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import test
import afip
import invoice
import config
import partner
import account
import country
import report
import currency
# vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
| Change product types are really dangerous!!! | [FIX] Change product types are really dangerous!!! | Python | agpl-3.0 | odoo-l10n-ar/l10n_ar_invoice | ---
+++
@@ -28,6 +28,5 @@
import country
import report
import currency
-import product
# vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4: |
abea9151ce3f97033038102b4b01e84b66b8b670 | __init__.py | __init__.py | ###
# Copyright (c) 2007, Max Kanat-Alexander
# All rights reserved.
#
#
###
"""
Interact with Bugzilla installations.
"""
import supybot
import supybot.world as world
# Use this for the version of this plugin. You may wish to put a CVS keyword
# in here if you're keeping the plugin in CVS or some similar system.
__version__ = "3.0.0.1"
# XXX Replace this with an appropriate author or supybot.Author instance.
__author__ = supybot.Author('Max Kanat-Alexander', 'mkanat',
'mkanat@bugzilla.org')
# This is a dictionary mapping supybot.Author instances to lists of
# contributions.
__contributors__ = {}
# This is a url where the most recent plugin package can be downloaded.
__url__ = 'http://supybot.com/Members/mkanat/Bugzilla'
import config
import plugin
reload(plugin) # In case we're being reloaded.
# Add more reloads here if you add third-party modules and want them to be
# reloaded when this plugin is reloaded. Don't forget to import them as well!
if world.testing:
import test
Class = plugin.Class
configure = config.configure
# vim:set shiftwidth=4 tabstop=4 expandtab textwidth=79:
| ###
# Copyright (c) 2007, Max Kanat-Alexander
# All rights reserved.
#
#
###
"""
Interact with Bugzilla installations.
"""
import supybot
import supybot.world as world
# Use this for the version of this plugin. You may wish to put a CVS keyword
# in here if you're keeping the plugin in CVS or some similar system.
__version__ = "3.0.0.1"
# XXX Replace this with an appropriate author or supybot.Author instance.
__author__ = supybot.Author('Max Kanat-Alexander', 'mkanat',
'mkanat@bugzilla.org')
# This is a dictionary mapping supybot.Author instances to lists of
# contributions.
__contributors__ = {}
# This is a url where the most recent plugin package can be downloaded.
__url__ = 'http://supybot.com/Members/mkanat/Bugzilla'
import config
import plugin
reload(plugin) # In case we're being reloaded.
reload(bugmail)
reload(traceparser)
# Add more reloads here if you add third-party modules and want them to be
# reloaded when this plugin is reloaded. Don't forget to import them as well!
if world.testing:
import test
Class = plugin.Class
configure = config.configure
# vim:set shiftwidth=4 tabstop=4 expandtab textwidth=79:
| Make the bot reload its secondary modules when you reload the plugin. | Make the bot reload its secondary modules when you reload the plugin.
| Python | bsd-3-clause | aleb/supybot-bugzilla | ---
+++
@@ -30,6 +30,9 @@
import config
import plugin
reload(plugin) # In case we're being reloaded.
+reload(bugmail)
+reload(traceparser)
+
# Add more reloads here if you add third-party modules and want them to be
# reloaded when this plugin is reloaded. Don't forget to import them as well!
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.