commit
stringlengths 40
40
| old_file
stringlengths 4
150
| new_file
stringlengths 4
150
| old_contents
stringlengths 0
3.26k
| new_contents
stringlengths 1
4.43k
| subject
stringlengths 15
501
| message
stringlengths 15
4.06k
| lang
stringclasses 4
values | license
stringclasses 13
values | repos
stringlengths 5
91.5k
| diff
stringlengths 0
4.35k
|
|---|---|---|---|---|---|---|---|---|---|---|
ce3af1f615b0979f8fcaf0fcec2efb1d73c13a36
|
api/tests/v2/test_token_update.py
|
api/tests/v2/test_token_update.py
|
from django.core.urlresolvers import reverse
from rest_framework.test import APITestCase, APIRequestFactory, force_authenticate
from api.tests.factories import UserFactory, AnonymousUserFactory, ProviderFactory
from api.v2.views import TokenUpdateViewSet, IdentityViewSet, CredentialViewSet
class TokenUpdateTests(APITestCase):
def setUp(self):
self.anonymous_user = AnonymousUserFactory()
self.user = UserFactory.create()
self.provider = ProviderFactory.create(location="mock location", type__name="mock")
self.view = TokenUpdateViewSet.as_view({'post': 'create'})
self.identity_view = IdentityViewSet.as_view({'get': 'retrieve'})
self.credentials_view = CredentialViewSet.as_view({'get': 'list'})
self.token_uuid = "test-token-1234-debug"
def test_invalid_provider_token_update(self):
factory = APIRequestFactory()
url = reverse('api:v2:token_update-list')
data = {
'username': self.user.username,
'project_name': self.user.username,
'provider': "nopenopenope",
'token': self.token_uuid
}
request = factory.post(url, data)
force_authenticate(request, user=self.user)
response = self.view(request)
self.assertTrue(response.status_code == 400)
self.assertTrue('provider' in response.data)
self.assertTrue("not a valid UUID" in response.data['provider'])
def test_valid_data_token_update(self):
factory = APIRequestFactory()
provider_uuid = str(self.provider.uuid)
url = reverse('api:v2:token_update-list')
data = {
'username': self.user.username,
'project_name': self.user.username,
'provider': provider_uuid,
'token': self.token_uuid
}
request = factory.post(url, data)
force_authenticate(request, user=self.user)
response = self.view(request)
self.assertEquals(response.status_code, 201)
data = response.data
self.assertTrue('identity_uuid' in data)
identity_uuid = data['identity_uuid']
cred_url = reverse('api:v2:credential-list')
cred_request = factory.get(cred_url)
force_authenticate(cred_request, user=self.user)
cred_response = self.credentials_view(cred_request)
self.assertTrue('results' in cred_response.data)
for cred in cred_response.data['results']:
self.assertTrue(cred['identity']['uuid'] == identity_uuid)
if cred['key'] == 'key':
self.assertTrue(cred['value'] == self.user.username)
elif cred['key'] == 'ex_project_name':
self.assertTrue(cred['value'] == self.user.username)
elif cred['key'] == 'ex_force_auth_token':
self.assertTrue(cred['value'] == self.token_uuid)
|
Include test file for token_update API
|
Include test file for token_update API
|
Python
|
apache-2.0
|
CCI-MOC/GUI-Backend,CCI-MOC/GUI-Backend,CCI-MOC/GUI-Backend,CCI-MOC/GUI-Backend
|
---
+++
@@ -0,0 +1,61 @@
+from django.core.urlresolvers import reverse
+from rest_framework.test import APITestCase, APIRequestFactory, force_authenticate
+from api.tests.factories import UserFactory, AnonymousUserFactory, ProviderFactory
+from api.v2.views import TokenUpdateViewSet, IdentityViewSet, CredentialViewSet
+
+class TokenUpdateTests(APITestCase):
+ def setUp(self):
+ self.anonymous_user = AnonymousUserFactory()
+ self.user = UserFactory.create()
+ self.provider = ProviderFactory.create(location="mock location", type__name="mock")
+ self.view = TokenUpdateViewSet.as_view({'post': 'create'})
+ self.identity_view = IdentityViewSet.as_view({'get': 'retrieve'})
+ self.credentials_view = CredentialViewSet.as_view({'get': 'list'})
+ self.token_uuid = "test-token-1234-debug"
+
+ def test_invalid_provider_token_update(self):
+ factory = APIRequestFactory()
+ url = reverse('api:v2:token_update-list')
+ data = {
+ 'username': self.user.username,
+ 'project_name': self.user.username,
+ 'provider': "nopenopenope",
+ 'token': self.token_uuid
+ }
+ request = factory.post(url, data)
+ force_authenticate(request, user=self.user)
+ response = self.view(request)
+ self.assertTrue(response.status_code == 400)
+ self.assertTrue('provider' in response.data)
+ self.assertTrue("not a valid UUID" in response.data['provider'])
+
+ def test_valid_data_token_update(self):
+ factory = APIRequestFactory()
+ provider_uuid = str(self.provider.uuid)
+ url = reverse('api:v2:token_update-list')
+ data = {
+ 'username': self.user.username,
+ 'project_name': self.user.username,
+ 'provider': provider_uuid,
+ 'token': self.token_uuid
+ }
+ request = factory.post(url, data)
+ force_authenticate(request, user=self.user)
+ response = self.view(request)
+ self.assertEquals(response.status_code, 201)
+ data = response.data
+ self.assertTrue('identity_uuid' in data)
+ identity_uuid = data['identity_uuid']
+ cred_url = reverse('api:v2:credential-list')
+ cred_request = factory.get(cred_url)
+ force_authenticate(cred_request, user=self.user)
+ cred_response = self.credentials_view(cred_request)
+ self.assertTrue('results' in cred_response.data)
+ for cred in cred_response.data['results']:
+ self.assertTrue(cred['identity']['uuid'] == identity_uuid)
+ if cred['key'] == 'key':
+ self.assertTrue(cred['value'] == self.user.username)
+ elif cred['key'] == 'ex_project_name':
+ self.assertTrue(cred['value'] == self.user.username)
+ elif cred['key'] == 'ex_force_auth_token':
+ self.assertTrue(cred['value'] == self.token_uuid)
|
|
3c5d21b41c78b87e1567453b4a6f15ca3d97c966
|
pyblogit/blog_model.py
|
pyblogit/blog_model.py
|
"""
pyblogit.blog_model
~~~~~~~~~~~~~~~~~~~
This module contains the data model to represent a blog and methods to
manipulate it.
"""
class blog(object):
"""The blog data model"""
def __init__(self, blog_id, name, url, desc, posts, pages):
self._blog_id = blog_id
self._name = name
self._url = url
self._desc = desc
self._posts = posts
self._pages = pages
@property
def blog_id(self):
return self._blog_id
@property
def name(self):
return self._name
@property
def url(self):
return self._url
@property
def desc(self):
return self._desc
@property
def posts(self):
return self._posts
@property
delf pages(self):
return self._pages
|
Add class to represent a blog
|
Add class to represent a blog
|
Python
|
mit
|
jamalmoir/pyblogit
|
---
+++
@@ -0,0 +1,42 @@
+"""
+pyblogit.blog_model
+~~~~~~~~~~~~~~~~~~~
+
+This module contains the data model to represent a blog and methods to
+manipulate it.
+"""
+
+class blog(object):
+ """The blog data model"""
+
+ def __init__(self, blog_id, name, url, desc, posts, pages):
+ self._blog_id = blog_id
+ self._name = name
+ self._url = url
+ self._desc = desc
+ self._posts = posts
+ self._pages = pages
+
+ @property
+ def blog_id(self):
+ return self._blog_id
+
+ @property
+ def name(self):
+ return self._name
+
+ @property
+ def url(self):
+ return self._url
+
+ @property
+ def desc(self):
+ return self._desc
+
+ @property
+ def posts(self):
+ return self._posts
+
+ @property
+ delf pages(self):
+ return self._pages
|
|
9a80f30acc2b57c37726d765977ff471b0033e1a
|
ideascube/conf/idb_jor_zaatari.py
|
ideascube/conf/idb_jor_zaatari.py
|
# -*- coding: utf-8 -*-
"""Ideaxbox for Zaatari, Jordan"""
from .idb import * # noqa
from django.utils.translation import ugettext_lazy as _
IDEASCUBE_NAME = u"Zaatari"
IDEASCUBE_PLACE_NAME = _("city")
COUNTRIES_FIRST = ['SY', 'JO']
TIME_ZONE = 'Asia/Amman'
LANGUAGE_CODE = 'ar'
LOAN_DURATION = 14
MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'user_id', 'birth_year', 'gender']
USER_FORM_FIELDS = (
('Ideasbox', ['serial', 'box_awareness']),
(_('Personal informations'), ['short_name', 'full_name', 'birth_year', 'gender', 'id_card_number']), # noqa
(_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa
(_('In the town'), ['current_occupation', 'school_level']),
(_('Language skills'), ['en_level']),
)
HOME_CARDS = STAFF_HOME_CARDS + [
{
'id': 'blog',
},
{
'id': 'library',
},
{
'id': 'mediacenter',
},
{
'id': 'khanacademy',
},
{
'id': 'wikipedia',
'languages': ['en','ar']
},
{
'id': 'wiktionary',
'languages': ['en', 'ar']
},
{
'id': 'wikiversity',
'languages': ['en', 'ar']
},
{
'id': 'wikibooks',
'languages': ['en', 'ar']
},
{
'id': 'wikisource',
'languages': ['en', 'ar']
},
{
'id': 'wikiquote',
'languages': ['en', 'ar']
},
{
'id': 'ted',
'sessions': [
('tedbusiness.en', 'Business'),
('teddesign.en', 'Design'),
('tedentertainment.en', 'Entertainment'),
('tedglobalissues.en', 'Global Issues'),
('tedscience.en', 'Science'),
('tedtechnology.en', 'Technology'),
('ted-ed', 'Education'),
('tedmed.en', 'Medical'),
]
},
{
'id': 'gutenberg',
'languages': ['en', 'fr']
},
{
'id': 'crashcourse',
'languages': ['en']
},
{
'id': 'vikidia',
'languages': ['en']
},
]
|
Add configuration for Zaatari, Jordan.
|
Add configuration for Zaatari, Jordan.
|
Python
|
agpl-3.0
|
ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube
|
---
+++
@@ -0,0 +1,82 @@
+# -*- coding: utf-8 -*-
+"""Ideaxbox for Zaatari, Jordan"""
+from .idb import * # noqa
+from django.utils.translation import ugettext_lazy as _
+
+IDEASCUBE_NAME = u"Zaatari"
+IDEASCUBE_PLACE_NAME = _("city")
+COUNTRIES_FIRST = ['SY', 'JO']
+TIME_ZONE = 'Asia/Amman'
+LANGUAGE_CODE = 'ar'
+LOAN_DURATION = 14
+MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'user_id', 'birth_year', 'gender']
+USER_FORM_FIELDS = (
+ ('Ideasbox', ['serial', 'box_awareness']),
+ (_('Personal informations'), ['short_name', 'full_name', 'birth_year', 'gender', 'id_card_number']), # noqa
+ (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa
+ (_('In the town'), ['current_occupation', 'school_level']),
+ (_('Language skills'), ['en_level']),
+)
+HOME_CARDS = STAFF_HOME_CARDS + [
+ {
+ 'id': 'blog',
+ },
+ {
+ 'id': 'library',
+ },
+ {
+ 'id': 'mediacenter',
+ },
+ {
+ 'id': 'khanacademy',
+ },
+ {
+ 'id': 'wikipedia',
+ 'languages': ['en','ar']
+ },
+ {
+ 'id': 'wiktionary',
+ 'languages': ['en', 'ar']
+ },
+ {
+ 'id': 'wikiversity',
+ 'languages': ['en', 'ar']
+ },
+ {
+ 'id': 'wikibooks',
+ 'languages': ['en', 'ar']
+ },
+ {
+ 'id': 'wikisource',
+ 'languages': ['en', 'ar']
+ },
+ {
+ 'id': 'wikiquote',
+ 'languages': ['en', 'ar']
+ },
+ {
+ 'id': 'ted',
+ 'sessions': [
+ ('tedbusiness.en', 'Business'),
+ ('teddesign.en', 'Design'),
+ ('tedentertainment.en', 'Entertainment'),
+ ('tedglobalissues.en', 'Global Issues'),
+ ('tedscience.en', 'Science'),
+ ('tedtechnology.en', 'Technology'),
+ ('ted-ed', 'Education'),
+ ('tedmed.en', 'Medical'),
+ ]
+ },
+ {
+ 'id': 'gutenberg',
+ 'languages': ['en', 'fr']
+ },
+ {
+ 'id': 'crashcourse',
+ 'languages': ['en']
+ },
+ {
+ 'id': 'vikidia',
+ 'languages': ['en']
+ },
+]
|
|
b0e6556a908c1c520b060ea868fb18ea00bea4d9
|
moderation/migrations/0007_auto_20141207_1025.py
|
moderation/migrations/0007_auto_20141207_1025.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
import django.utils.timezone
class Migration(migrations.Migration):
dependencies = [
('moderation', '0006_auto_20141104_1434'),
]
operations = [
migrations.AlterModelOptions(
name='moderationlogmsg',
options={'verbose_name': 'log entry', 'verbose_name_plural': 'log entries'},
),
migrations.AlterField(
model_name='moderationlogmsg',
name='comment',
field=models.TextField(verbose_name='log comment'),
),
migrations.AlterField(
model_name='moderationlogmsg',
name='logged_by',
field=models.ForeignKey(related_name='log_messages_by', to=settings.AUTH_USER_MODEL, verbose_name='logged by', help_text='Moderator who created the log'),
),
migrations.AlterField(
model_name='moderationlogmsg',
name='msg_datetime',
field=models.DateTimeField(verbose_name='date and time recorded', default=django.utils.timezone.now),
),
migrations.AlterField(
model_name='moderationlogmsg',
name='msg_type',
field=models.CharField(verbose_name='message type', choices=[('ALL', 'All'), ('INVITATION', 'Invitation'), ('REINVITATION', 'Invitation Resent'), ('APPROVAL', 'Application Approved'), ('REJECTION', 'Application Rejected'), ('DISMISSAL', 'Abuse Report Dismissed'), ('WARNING', 'Official Warning'), ('BANNING', 'Ban User')], max_length=20),
),
migrations.AlterField(
model_name='moderationlogmsg',
name='pertains_to',
field=models.ForeignKey(related_name='log_messages_about', to=settings.AUTH_USER_MODEL, verbose_name='pertains to', help_text='User who moderation log is about'),
),
]
|
Add migration for translation cleanup
|
Add migration for translation cleanup
|
Python
|
bsd-3-clause
|
f3r3nc/connect,nlhkabu/connect,f3r3nc/connect,nlhkabu/connect,f3r3nc/connect,nlhkabu/connect,nlhkabu/connect,f3r3nc/connect
|
---
+++
@@ -0,0 +1,45 @@
+# -*- coding: utf-8 -*-
+from __future__ import unicode_literals
+
+from django.db import models, migrations
+from django.conf import settings
+import django.utils.timezone
+
+
+class Migration(migrations.Migration):
+
+ dependencies = [
+ ('moderation', '0006_auto_20141104_1434'),
+ ]
+
+ operations = [
+ migrations.AlterModelOptions(
+ name='moderationlogmsg',
+ options={'verbose_name': 'log entry', 'verbose_name_plural': 'log entries'},
+ ),
+ migrations.AlterField(
+ model_name='moderationlogmsg',
+ name='comment',
+ field=models.TextField(verbose_name='log comment'),
+ ),
+ migrations.AlterField(
+ model_name='moderationlogmsg',
+ name='logged_by',
+ field=models.ForeignKey(related_name='log_messages_by', to=settings.AUTH_USER_MODEL, verbose_name='logged by', help_text='Moderator who created the log'),
+ ),
+ migrations.AlterField(
+ model_name='moderationlogmsg',
+ name='msg_datetime',
+ field=models.DateTimeField(verbose_name='date and time recorded', default=django.utils.timezone.now),
+ ),
+ migrations.AlterField(
+ model_name='moderationlogmsg',
+ name='msg_type',
+ field=models.CharField(verbose_name='message type', choices=[('ALL', 'All'), ('INVITATION', 'Invitation'), ('REINVITATION', 'Invitation Resent'), ('APPROVAL', 'Application Approved'), ('REJECTION', 'Application Rejected'), ('DISMISSAL', 'Abuse Report Dismissed'), ('WARNING', 'Official Warning'), ('BANNING', 'Ban User')], max_length=20),
+ ),
+ migrations.AlterField(
+ model_name='moderationlogmsg',
+ name='pertains_to',
+ field=models.ForeignKey(related_name='log_messages_about', to=settings.AUTH_USER_MODEL, verbose_name='pertains to', help_text='User who moderation log is about'),
+ ),
+ ]
|
|
50ed5c6566fb684bfea125d7084a5972edbe5c2a
|
test/unit/test_log.py
|
test/unit/test_log.py
|
# :coding: utf-8
# :copyright: Copyright (c) 2013 Martin Pengelly-Phillips
# :license: See LICENSE.txt.
import pytest
from bark.log import Log
def test_create():
'''Test creating a Log instance.'''
log = Log(name='bark.test.log')
assert log.items() == [('name', 'bark.test.log')]
def test_string_represenation():
'''Test string representation of Log instance.'''
log = Log(name='bark.test.log')
assert str(log) == "{'name': 'bark.test.log'}"
def test_length():
'''Test len method returns number of current keys.'''
log = Log(name='bark.test.log')
assert len(log) == 1
log['message'] = 'A message'
assert len(log) == 2
def test_setting_and_getting_item():
'''Test setting and getting key value pair.'''
log = Log()
assert len(log) == 0
log['message'] = 'A message'
assert len(log) == 1
assert log['message'] == 'A message'
def test_delete_item():
'''Test removing an item.'''
log = Log()
assert len(log) == 0
log['message'] = 'A message'
assert len(log) == 1
assert log['message'] == 'A message'
del log['message']
assert len(log) == 0
with pytest.raises(KeyError):
log['message']
|
Add basic unit tests for Log.
|
Add basic unit tests for Log.
|
Python
|
apache-2.0
|
4degrees/mill,4degrees/sawmill
|
---
+++
@@ -0,0 +1,51 @@
+# :coding: utf-8
+# :copyright: Copyright (c) 2013 Martin Pengelly-Phillips
+# :license: See LICENSE.txt.
+
+import pytest
+
+from bark.log import Log
+
+
+def test_create():
+ '''Test creating a Log instance.'''
+ log = Log(name='bark.test.log')
+ assert log.items() == [('name', 'bark.test.log')]
+
+
+def test_string_represenation():
+ '''Test string representation of Log instance.'''
+ log = Log(name='bark.test.log')
+ assert str(log) == "{'name': 'bark.test.log'}"
+
+
+def test_length():
+ '''Test len method returns number of current keys.'''
+ log = Log(name='bark.test.log')
+ assert len(log) == 1
+ log['message'] = 'A message'
+ assert len(log) == 2
+
+
+def test_setting_and_getting_item():
+ '''Test setting and getting key value pair.'''
+ log = Log()
+ assert len(log) == 0
+ log['message'] = 'A message'
+ assert len(log) == 1
+ assert log['message'] == 'A message'
+
+
+def test_delete_item():
+ '''Test removing an item.'''
+ log = Log()
+ assert len(log) == 0
+ log['message'] = 'A message'
+ assert len(log) == 1
+ assert log['message'] == 'A message'
+
+ del log['message']
+ assert len(log) == 0
+ with pytest.raises(KeyError):
+ log['message']
+
|
|
b5a5f381d69260df65248d0e2ff99b9a7c47ffe7
|
tests/test_general.py
|
tests/test_general.py
|
import os
import unittest
from mpower.opr import OPR
from mpower.store import Store
from . import MP_ACCESS_TOKENS
class TestGeneral(unittest.TestCase):
"""General/Miscellaneous tests"""
def setUp(self):
# Your MPower developer tokens
self.store = Store({"name":"FooBar store"})
self.opr_data = {'total_amount': 345, 'description': "Hello World",
"account_alias":"0266636984"}
self.opr = OPR(self.opr_data, self.store)
def tearDown(self):
self.opr = None
self.store = None
self.opr_data = None
def test_runtime_configs(self):
self.assertEqual(MP_ACCESS_TOKENS,
self.opr.runtime_configs)
def test_system_configs_env(self):
os.environ['MP-Master-Key'] = "5b9f531a-fbb8-487a-8045-3b4c7ac5acee"
os.environ['MP-Private-Key'] = "test_private_oGslgmzSNL3RSkjlsnPOsZZg9IA"
os.environ['MP-Token'] = "ff1d576409b2587cc1c2"
self.assertTrue(self.opr.config)
def test_rsc_endpoints(self):
endpoint = 'checkout-invoice/confirm/test_98567JGF'
url= self.opr.get_rsc_endpoint(endpoint)
self.assertTrue(url.startswith('https') and url.endswith(endpoint))
def test_add_headers(self):
header = {'Foo':'Bar'}
self.opr.add_header(header)
self.assertTrue("Foo" in self.opr.headers.keys())
self.assertFalse('FooBar' in self.opr.headers.keys())
if __name__ == '__main__':
unittest.main()
|
Add a 'general' unit tests file
|
Add a 'general' unit tests file
This 'general' unit tests is responsible for all
miscellaneous/general functionalities
|
Python
|
mit
|
mawuli/mpower-python,rpip/mpower-python
|
---
+++
@@ -0,0 +1,46 @@
+import os
+import unittest
+from mpower.opr import OPR
+from mpower.store import Store
+from . import MP_ACCESS_TOKENS
+
+
+class TestGeneral(unittest.TestCase):
+ """General/Miscellaneous tests"""
+ def setUp(self):
+ # Your MPower developer tokens
+ self.store = Store({"name":"FooBar store"})
+ self.opr_data = {'total_amount': 345, 'description': "Hello World",
+ "account_alias":"0266636984"}
+ self.opr = OPR(self.opr_data, self.store)
+
+ def tearDown(self):
+ self.opr = None
+ self.store = None
+ self.opr_data = None
+
+ def test_runtime_configs(self):
+ self.assertEqual(MP_ACCESS_TOKENS,
+ self.opr.runtime_configs)
+
+ def test_system_configs_env(self):
+ os.environ['MP-Master-Key'] = "5b9f531a-fbb8-487a-8045-3b4c7ac5acee"
+ os.environ['MP-Private-Key'] = "test_private_oGslgmzSNL3RSkjlsnPOsZZg9IA"
+ os.environ['MP-Token'] = "ff1d576409b2587cc1c2"
+ self.assertTrue(self.opr.config)
+
+
+ def test_rsc_endpoints(self):
+ endpoint = 'checkout-invoice/confirm/test_98567JGF'
+ url= self.opr.get_rsc_endpoint(endpoint)
+ self.assertTrue(url.startswith('https') and url.endswith(endpoint))
+
+ def test_add_headers(self):
+ header = {'Foo':'Bar'}
+ self.opr.add_header(header)
+ self.assertTrue("Foo" in self.opr.headers.keys())
+ self.assertFalse('FooBar' in self.opr.headers.keys())
+
+
+if __name__ == '__main__':
+ unittest.main()
|
|
5ae52b9e16b073322550f0a7ed9d560f5f823847
|
tests/test_hackage.py
|
tests/test_hackage.py
|
from tests.helper import ExternalVersionTestCase
class HackageTest(ExternalVersionTestCase):
def test_hackage(self):
self.assertEqual(self.sync_get_version("sessions", {"hackage": None}), "2008.7.18")
|
Add a testcase for Hackage
|
Add a testcase for Hackage
|
Python
|
mit
|
lilydjwg/nvchecker
|
---
+++
@@ -0,0 +1,6 @@
+from tests.helper import ExternalVersionTestCase
+
+
+class HackageTest(ExternalVersionTestCase):
+ def test_hackage(self):
+ self.assertEqual(self.sync_get_version("sessions", {"hackage": None}), "2008.7.18")
|
|
9320338b7edebdf864bfe8d3dd6cfb1a5c2f868d
|
util/calc_ir_table.py
|
util/calc_ir_table.py
|
#! /usr/bin/env python
import sys
it = iter(l)
for i,j in zip(it, it):
aij.append(i/2**39+j/2**32),
for i in range(0, len(aij)):
sys.stdout.write("{0:.10e}f, ".format(aij[i]))
if (i+1)%8==0:
print()
|
Add MLX IR table generator.
|
Add MLX IR table generator.
|
Python
|
mit
|
iabdalkader/openmv,iabdalkader/openmv,kwagyeman/openmv,openmv/openmv,kwagyeman/openmv,openmv/openmv,openmv/openmv,openmv/openmv,kwagyeman/openmv,iabdalkader/openmv,kwagyeman/openmv,iabdalkader/openmv
|
---
+++
@@ -0,0 +1,12 @@
+#! /usr/bin/env python
+import sys
+
+it = iter(l)
+
+for i,j in zip(it, it):
+ aij.append(i/2**39+j/2**32),
+
+for i in range(0, len(aij)):
+ sys.stdout.write("{0:.10e}f, ".format(aij[i]))
+ if (i+1)%8==0:
+ print()
|
|
e29dc1788292930d4d9585b5ef764ffdf567ade3
|
show_usbcamera.py
|
show_usbcamera.py
|
#! /usr/bin/env python
#
# Show the USB camera
#
#
# External dependencies
#
import sys
from PySide import QtGui
import VisionToolkit as vtk
#
# Main application
#
if __name__ == '__main__' :
application = QtGui.QApplication( sys.argv )
widget = vtk.UsbCameraWidget()
widget.show()
sys.exit( application.exec_() )
|
Add a script to display a USB camera.
|
Add a script to display a USB camera.
|
Python
|
mit
|
microy/VisionToolkit,microy/PyStereoVisionToolkit,microy/VisionToolkit,microy/StereoVision,microy/StereoVision,microy/PyStereoVisionToolkit
|
---
+++
@@ -0,0 +1,25 @@
+#! /usr/bin/env python
+
+
+#
+# Show the USB camera
+#
+
+
+#
+# External dependencies
+#
+import sys
+from PySide import QtGui
+import VisionToolkit as vtk
+
+
+#
+# Main application
+#
+if __name__ == '__main__' :
+
+ application = QtGui.QApplication( sys.argv )
+ widget = vtk.UsbCameraWidget()
+ widget.show()
+ sys.exit( application.exec_() )
|
|
db9772f5cd856f4fa66625d229982e2546574f59
|
avalonstar/apps/subscribers/migrations/0005_count.py
|
avalonstar/apps/subscribers/migrations/0005_count.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
import datetime
from django.utils.timezone import utc
class Migration(migrations.Migration):
dependencies = [
('subscribers', '0004_auto_20150224_1454'),
]
operations = [
migrations.CreateModel(
name='Count',
fields=[
('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),
('total', models.IntegerField()),
('timestamp', models.DateTimeField(default=datetime.datetime(2015, 3, 4, 1, 52, 0, 594138, tzinfo=utc))),
],
options={
'ordering': ['timestamp'],
},
bases=(models.Model,),
),
]
|
Add a migration for Count.
|
Add a migration for Count.
|
Python
|
apache-2.0
|
bryanveloso/avalonstar-tv,bryanveloso/avalonstar-tv,bryanveloso/avalonstar-tv
|
---
+++
@@ -0,0 +1,28 @@
+# -*- coding: utf-8 -*-
+from __future__ import unicode_literals
+
+from django.db import models, migrations
+import datetime
+from django.utils.timezone import utc
+
+
+class Migration(migrations.Migration):
+
+ dependencies = [
+ ('subscribers', '0004_auto_20150224_1454'),
+ ]
+
+ operations = [
+ migrations.CreateModel(
+ name='Count',
+ fields=[
+ ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),
+ ('total', models.IntegerField()),
+ ('timestamp', models.DateTimeField(default=datetime.datetime(2015, 3, 4, 1, 52, 0, 594138, tzinfo=utc))),
+ ],
+ options={
+ 'ordering': ['timestamp'],
+ },
+ bases=(models.Model,),
+ ),
+ ]
|
|
1d94290cdf5b0a111924ff9aa73d69b9e0f41165
|
scripts/update_thanks.py
|
scripts/update_thanks.py
|
#!/usr/bin/env python
# Usage: git log --format="%an <%ae>" | python update_thanks.py
# You will get a result.txt file, you can work with the file (update, remove, ...)
#
# Install
# =======
# pip install validate_email pyDNS
#
from __future__ import print_function
import os
import sys
from validate_email import validate_email
from email.utils import parseaddr
import DNS.Base
addresses = set()
bad_addresses = set()
collection = []
lines = list(reversed(sys.stdin.readlines()))
for author in map(str.strip, lines):
realname, email_address = parseaddr(author)
if email_address not in addresses:
if email_address in bad_addresses:
continue
else:
try:
value = validate_email(email_address)
if value:
addresses.add(email_address)
collection.append(author)
else:
bad_addresses.add(email_address)
except DNS.Base.TimeoutError:
bad_addresses.add(email_address)
with open('result.txt', 'w') as output:
output.write('\n'.join(collection))
|
Add the script for the THANKS file
|
Add the script for the THANKS file
|
Python
|
mit
|
prezi/gunicorn,malept/gunicorn,jamesblunt/gunicorn,mvaled/gunicorn,prezi/gunicorn,harrisonfeng/gunicorn,GitHublong/gunicorn,keakon/gunicorn,zhoucen/gunicorn,malept/gunicorn,zhoucen/gunicorn,ccl0326/gunicorn,jamesblunt/gunicorn,mvaled/gunicorn,ccl0326/gunicorn,gtrdotmcs/gunicorn,1stvamp/gunicorn,WSDC-NITWarangal/gunicorn,zhoucen/gunicorn,tempbottle/gunicorn,jamesblunt/gunicorn,mvaled/gunicorn,ccl0326/gunicorn,1stvamp/gunicorn,tejasmanohar/gunicorn,1stvamp/gunicorn,elelianghh/gunicorn,z-fork/gunicorn,beni55/gunicorn,gtrdotmcs/gunicorn,gtrdotmcs/gunicorn,MrKiven/gunicorn,prezi/gunicorn,malept/gunicorn,ephes/gunicorn
|
---
+++
@@ -0,0 +1,42 @@
+#!/usr/bin/env python
+# Usage: git log --format="%an <%ae>" | python update_thanks.py
+# You will get a result.txt file, you can work with the file (update, remove, ...)
+#
+# Install
+# =======
+# pip install validate_email pyDNS
+#
+from __future__ import print_function
+import os
+import sys
+
+from validate_email import validate_email
+from email.utils import parseaddr
+import DNS.Base
+
+addresses = set()
+bad_addresses = set()
+collection = []
+
+lines = list(reversed(sys.stdin.readlines()))
+
+for author in map(str.strip, lines):
+ realname, email_address = parseaddr(author)
+
+ if email_address not in addresses:
+ if email_address in bad_addresses:
+ continue
+ else:
+ try:
+ value = validate_email(email_address)
+ if value:
+ addresses.add(email_address)
+ collection.append(author)
+ else:
+ bad_addresses.add(email_address)
+ except DNS.Base.TimeoutError:
+ bad_addresses.add(email_address)
+
+
+with open('result.txt', 'w') as output:
+ output.write('\n'.join(collection))
|
|
d766bfc19ce627e5141e6ab355957eb1fa5d716d
|
sii/utils.py
|
sii/utils.py
|
import ssl
def fix_ssl_verify():
try:
_create_unverified_https_context = ssl._create_unverified_context
except AttributeError:
# Legacy Python that doesn't verify HTTPS certificates by default
pass
else:
# Handle target environment that doesn't support HTTPS verification
ssl._create_default_https_context = _create_unverified_https_context
|
Add function to not verify ssl
|
Add function to not verify ssl
|
Python
|
mit
|
gisce/sii
|
---
+++
@@ -0,0 +1,12 @@
+import ssl
+
+
+def fix_ssl_verify():
+ try:
+ _create_unverified_https_context = ssl._create_unverified_context
+ except AttributeError:
+ # Legacy Python that doesn't verify HTTPS certificates by default
+ pass
+ else:
+ # Handle target environment that doesn't support HTTPS verification
+ ssl._create_default_https_context = _create_unverified_https_context
|
|
964a0f848786cdbbf0b41c46cb66d3ec1a255120
|
hacks.py
|
hacks.py
|
import ConfigParser
import requests
import urllib
def get_settings(_config_cache=[]):
config = ConfigParser.RawConfigParser()
config.read(['settings.ini'])
return config
def meetup_urls(method='groups.json'):
base_url = 'http://api.meetup.com/'
url = (base_url + method)
return (url, {'key': get_settings().get('api_keys', 'meetup')})
def top_python_groups():
url, data = meetup_urls()
data['topic']='python'
data['order'] = 'members'
data['page'] = '200'
response = requests.get(url + '?' + urllib.urlencode(data))
return response
|
Add minimal script to get top 200 Python-related Meetup groups
|
Add minimal script to get top 200 Python-related Meetup groups
|
Python
|
apache-2.0
|
paulproteus/pug-meta-organizing
|
---
+++
@@ -0,0 +1,21 @@
+import ConfigParser
+import requests
+import urllib
+
+def get_settings(_config_cache=[]):
+ config = ConfigParser.RawConfigParser()
+ config.read(['settings.ini'])
+ return config
+
+def meetup_urls(method='groups.json'):
+ base_url = 'http://api.meetup.com/'
+ url = (base_url + method)
+ return (url, {'key': get_settings().get('api_keys', 'meetup')})
+
+def top_python_groups():
+ url, data = meetup_urls()
+ data['topic']='python'
+ data['order'] = 'members'
+ data['page'] = '200'
+ response = requests.get(url + '?' + urllib.urlencode(data))
+ return response
|
|
1bef42bceaa3b188f3f54d97d6927dc9c86a8598
|
text-and-point.py
|
text-and-point.py
|
# On some systems, need to import like this:
# import Image
# import ImageDraw
# On others, import like this:
from PIL import Image, ImageDraw
im = Image.new('RGB', (100, 100))
draw = ImageDraw.Draw(im)
red = (255, 0, 0)
draw.text((5,5), 'Hello', red)
draw.point((50, 50), red)
im.save('f.png')
|
Add simple script to print text and a point
|
Add simple script to print text and a point
|
Python
|
mit
|
redpig2/pilhacks
|
---
+++
@@ -0,0 +1,16 @@
+# On some systems, need to import like this:
+# import Image
+# import ImageDraw
+
+# On others, import like this:
+from PIL import Image, ImageDraw
+
+im = Image.new('RGB', (100, 100))
+
+draw = ImageDraw.Draw(im)
+red = (255, 0, 0)
+draw.text((5,5), 'Hello', red)
+draw.point((50, 50), red)
+
+im.save('f.png')
+
|
|
8e3f9cdc55fae71634559c83637e8ab3d69b732d
|
src/main/resources/script_templates/Hadim_Scripts/ROI/Circle_ROI_Builder.py
|
src/main/resources/script_templates/Hadim_Scripts/ROI/Circle_ROI_Builder.py
|
# @Float(label="Diameter of the circle ROI (pixel)", value=7) circle_diam
from ij.plugin.frame import RoiManager
from ij.gui import OvalRoi
rm = RoiManager.getInstance()
new_rois = []
for roi in rm.getRoisAsArray():
assert roi.getTypeAsString() == 'Point', "ROI needs to be a point"
x_center = roi.getContainedPoints()[0].x - (circle_diam / 2) + 0.5
y_center = roi.getContainedPoints()[0].y - (circle_diam / 2) + 0.5
new_roi = OvalRoi(x_center, y_center, circle_diam, circle_diam)
new_rois.append(new_roi)
rm.reset()
for new_roi in new_rois:
rm.addRoi(new_roi)
print("Done")
|
Add circle roi builder script
|
Add circle roi builder script
|
Python
|
bsd-3-clause
|
hadim/fiji_scripts,hadim/fiji_tools,hadim/fiji_scripts,hadim/fiji_scripts,hadim/fiji_tools
|
---
+++
@@ -0,0 +1,21 @@
+# @Float(label="Diameter of the circle ROI (pixel)", value=7) circle_diam
+
+from ij.plugin.frame import RoiManager
+from ij.gui import OvalRoi
+
+rm = RoiManager.getInstance()
+
+new_rois = []
+for roi in rm.getRoisAsArray():
+ assert roi.getTypeAsString() == 'Point', "ROI needs to be a point"
+ x_center = roi.getContainedPoints()[0].x - (circle_diam / 2) + 0.5
+ y_center = roi.getContainedPoints()[0].y - (circle_diam / 2) + 0.5
+ new_roi = OvalRoi(x_center, y_center, circle_diam, circle_diam)
+ new_rois.append(new_roi)
+
+rm.reset()
+
+for new_roi in new_rois:
+ rm.addRoi(new_roi)
+
+print("Done")
|
|
43eddc3663f64e92a673c09ce52ddcd50b935842
|
ipywidgets/widgets/tests/test_widget_float.py
|
ipywidgets/widgets/tests/test_widget_float.py
|
from unittest import TestCase
from traitlets import TraitError
from ipywidgets import FloatSlider
class TestFloatSlider(TestCase):
def test_construction(self):
FloatSlider()
def test_construction_readout_format(self):
slider = FloatSlider(readout_format='$.1f')
assert slider.get_state()['readout_format'] == '$.1f'
def test_construction_invalid_readout_format(self):
with self.assertRaises(TraitError):
FloatSlider(readout_format='broken')
|
Test that the float slider uses the NumberFormat traitlet
|
Test that the float slider uses the NumberFormat traitlet
|
Python
|
bsd-3-clause
|
ipython/ipywidgets,ipython/ipywidgets,ipython/ipywidgets,ipython/ipywidgets,SylvainCorlay/ipywidgets,jupyter-widgets/ipywidgets,SylvainCorlay/ipywidgets,jupyter-widgets/ipywidgets,SylvainCorlay/ipywidgets,jupyter-widgets/ipywidgets,jupyter-widgets/ipywidgets,ipython/ipywidgets,SylvainCorlay/ipywidgets
|
---
+++
@@ -0,0 +1,20 @@
+
+from unittest import TestCase
+
+from traitlets import TraitError
+
+from ipywidgets import FloatSlider
+
+
+class TestFloatSlider(TestCase):
+
+ def test_construction(self):
+ FloatSlider()
+
+ def test_construction_readout_format(self):
+ slider = FloatSlider(readout_format='$.1f')
+ assert slider.get_state()['readout_format'] == '$.1f'
+
+ def test_construction_invalid_readout_format(self):
+ with self.assertRaises(TraitError):
+ FloatSlider(readout_format='broken')
|
|
db3749beefb5d8d33dae2044aa4ac09b1d3a0d80
|
tests/test_movingfiles.py
|
tests/test_movingfiles.py
|
#!/usr/bin/env python
#encoding:utf-8
#author:dbr/Ben
#project:tvnamer
#repository:http://github.com/dbr/tvnamer
#license:Creative Commons GNU GPL v2
# http://creativecommons.org/licenses/GPL/2.0/
"""Tests moving renamed files
"""
from functional_runner import run_tvnamer, verify_out_data
def test_simple_realtive_move():
"""
"""
conf = """
{"move_files_enable": true,
"move_files_desination": "test/",
"batch": true}
"""
out_data = run_tvnamer(
with_files = ['scrubs.s01e01.avi'],
with_config = conf,
with_input = "")
expected_files = ['test/Scrubs - [01x01] - My First Day.avi']
verify_out_data(out_data, expected_files)
|
Add (currently failing) test for file moving
|
Add (currently failing) test for file moving
|
Python
|
unlicense
|
dbr/tvnamer,lahwaacz/tvnamer,m42e/tvnamer
|
---
+++
@@ -0,0 +1,32 @@
+#!/usr/bin/env python
+#encoding:utf-8
+#author:dbr/Ben
+#project:tvnamer
+#repository:http://github.com/dbr/tvnamer
+#license:Creative Commons GNU GPL v2
+# http://creativecommons.org/licenses/GPL/2.0/
+
+"""Tests moving renamed files
+"""
+
+from functional_runner import run_tvnamer, verify_out_data
+
+
+def test_simple_realtive_move():
+ """
+ """
+
+ conf = """
+ {"move_files_enable": true,
+ "move_files_desination": "test/",
+ "batch": true}
+ """
+
+ out_data = run_tvnamer(
+ with_files = ['scrubs.s01e01.avi'],
+ with_config = conf,
+ with_input = "")
+
+ expected_files = ['test/Scrubs - [01x01] - My First Day.avi']
+
+ verify_out_data(out_data, expected_files)
|
|
f4f6b1af7c11389666cf5ec562fef874388eb5a3
|
loess.py
|
loess.py
|
from pylab import *
def loessInternal( x, h, xp, yp ):
w = exp( -0.5*( ((x-xp)/(2*h))**2 )/sqrt(2*pi*h**2) )
b = sum(w*xp)*sum(w*yp) - sum(w)*sum(w*xp*yp)
b /= sum(w*xp)**2 - sum(w)*sum(w*xp**2)
a = ( sum(w*yp) - b*sum(w*xp) )/sum(w)
return a + b*x
def loess(x,y,h):
"""LOESS model free bandwidth reduction.
See "Data Analysis with Open Source Tools" by P. K. Janert for fdetails.
Watch out that x and y do not become too small,
microseconds don't work. h is bandwidth in units of x"""
out = []
for k in x:
out.append( loessInternal(k, h, x, y))
return out
|
Add LOESS module for model free bandwidth reduction.
|
Add LOESS module for model free bandwidth reduction.
|
Python
|
mit
|
bennomeier/pyNMR,kourk0am/pyNMR
|
---
+++
@@ -0,0 +1,23 @@
+from pylab import *
+
+def loessInternal( x, h, xp, yp ):
+ w = exp( -0.5*( ((x-xp)/(2*h))**2 )/sqrt(2*pi*h**2) )
+ b = sum(w*xp)*sum(w*yp) - sum(w)*sum(w*xp*yp)
+
+ b /= sum(w*xp)**2 - sum(w)*sum(w*xp**2)
+ a = ( sum(w*yp) - b*sum(w*xp) )/sum(w)
+
+ return a + b*x
+
+def loess(x,y,h):
+ """LOESS model free bandwidth reduction.
+
+ See "Data Analysis with Open Source Tools" by P. K. Janert for fdetails.
+
+ Watch out that x and y do not become too small,
+ microseconds don't work. h is bandwidth in units of x"""
+ out = []
+ for k in x:
+ out.append( loessInternal(k, h, x, y))
+ return out
+
|
|
eb25f8bd1ab344d951f78cfe1e655e4f755bebd6
|
bin/verify-identity.py
|
bin/verify-identity.py
|
"""verify-identity.py <participant_id>, <country_code>
"""
from __future__ import absolute_import, division, print_function, unicode_literals
import sys
from gratipay import wireup
from gratipay.models.participant import Participant
from gratipay.models.country import Country
wireup.db(wireup.env())
participant = Participant.from_id(int(sys.argv[1]))
country = Country.from_code(sys.argv[2])
participant.set_identity_verification(country.id, True)
|
Write a script to verify identity
|
Write a script to verify identity
|
Python
|
mit
|
gratipay/gratipay.com,gratipay/gratipay.com,gratipay/gratipay.com,gratipay/gratipay.com
|
---
+++
@@ -0,0 +1,15 @@
+"""verify-identity.py <participant_id>, <country_code>
+"""
+from __future__ import absolute_import, division, print_function, unicode_literals
+
+import sys
+
+from gratipay import wireup
+from gratipay.models.participant import Participant
+from gratipay.models.country import Country
+
+wireup.db(wireup.env())
+
+participant = Participant.from_id(int(sys.argv[1]))
+country = Country.from_code(sys.argv[2])
+participant.set_identity_verification(country.id, True)
|
|
0eddc59a97fca233b5979d61a19f58179c57bc08
|
CodeFights/twoLines.py
|
CodeFights/twoLines.py
|
#!/usr/local/bin/python
# Code Fights Two Lines Problem
from functools import partial
def line_y(m, b, x):
return m * x + b
def twoLines(line1, line2, l, r):
line1_y = partial(line_y, *line1)
line2_y = partial(line_y, *line2)
balance = 0
for x in range(l, r + 1):
y1 = line1_y(x)
y2 = line2_y(x)
if y1 > y2:
balance += 1
elif y1 < y2:
balance -= 1
if balance > 0:
return "first"
if balance < 0:
return "second"
return "any"
def main():
tests = [
[[1, 2], [2, 1], 0, 2, "any"],
[[1, 2], [2, 1], -1, 2, "first"],
[[1, 2], [2, 1], 0, 3, "second"],
[[1, 2], [1, 0], -1000, 1000, "first"],
[[1, 0], [-1, 0], -239, 239, "any"],
[[1, 0], [-1, 0], -999, 998, "second"]
]
for t in tests:
res = twoLines(t[0], t[1], t[2], t[3])
ans = t[4]
if ans == res:
print("PASSED: twoLines({}, {}, {}, {}) returned {}"
.format(t[0], t[1], t[2], t[3], res))
else:
print(("FAILED: twoLines({}, {}, {}, {}) returned {},"
"answer: {}").format(t[0], t[1], t[2], t[3], res, ans))
if __name__ == '__main__':
main()
|
Solve Code Fights two lines problem
|
Solve Code Fights two lines problem
|
Python
|
mit
|
HKuz/Test_Code
|
---
+++
@@ -0,0 +1,51 @@
+#!/usr/local/bin/python
+# Code Fights Two Lines Problem
+
+from functools import partial
+
+
+def line_y(m, b, x):
+ return m * x + b
+
+
+def twoLines(line1, line2, l, r):
+ line1_y = partial(line_y, *line1)
+ line2_y = partial(line_y, *line2)
+ balance = 0
+ for x in range(l, r + 1):
+ y1 = line1_y(x)
+ y2 = line2_y(x)
+ if y1 > y2:
+ balance += 1
+ elif y1 < y2:
+ balance -= 1
+ if balance > 0:
+ return "first"
+ if balance < 0:
+ return "second"
+ return "any"
+
+
+def main():
+ tests = [
+ [[1, 2], [2, 1], 0, 2, "any"],
+ [[1, 2], [2, 1], -1, 2, "first"],
+ [[1, 2], [2, 1], 0, 3, "second"],
+ [[1, 2], [1, 0], -1000, 1000, "first"],
+ [[1, 0], [-1, 0], -239, 239, "any"],
+ [[1, 0], [-1, 0], -999, 998, "second"]
+ ]
+
+ for t in tests:
+ res = twoLines(t[0], t[1], t[2], t[3])
+ ans = t[4]
+ if ans == res:
+ print("PASSED: twoLines({}, {}, {}, {}) returned {}"
+ .format(t[0], t[1], t[2], t[3], res))
+ else:
+ print(("FAILED: twoLines({}, {}, {}, {}) returned {},"
+ "answer: {}").format(t[0], t[1], t[2], t[3], res, ans))
+
+
+if __name__ == '__main__':
+ main()
|
|
2594b559722efa75322e669792cf8b9ba14f5014
|
Data-Structures/Trees/Binary_Trees/binary_tree.py
|
Data-Structures/Trees/Binary_Trees/binary_tree.py
|
"""Implementation of a Binary Tree in Python."""
class BinaryTree:
def __init__(self, root_node):
self.key = root_node
self.left_node = None
self.right_node = None
def insert_left(self, new_node):
if self.left_node == None:
self.left_node = BinaryTree(new_node)
else:
tree = BinaryTree(new_node)
tree.left_node = self.left_node
self.left_node = tree
def insert_right(self, new_node):
if self.right_node == None:
self.right_node = BinaryTree(new_node)
else:
tree = BinaryTree(new_node)
tree.right_node = self.right_node
self.right_node = tree
def get_left_node(self):
return self.left_node
def get_right_node(self):
return self.right_node
def get_root_value(self):
return self.key
def set_root_node(self, new_root):
self.key = new_root
|
Add python binary tree data structure
|
Add python binary tree data structure
|
Python
|
mit
|
salman-bhai/DS-Algo-Handbook,salman-bhai/DS-Algo-Handbook,salman-bhai/DS-Algo-Handbook,salman-bhai/DS-Algo-Handbook
|
---
+++
@@ -0,0 +1,35 @@
+"""Implementation of a Binary Tree in Python."""
+
+class BinaryTree:
+ def __init__(self, root_node):
+ self.key = root_node
+ self.left_node = None
+ self.right_node = None
+
+ def insert_left(self, new_node):
+ if self.left_node == None:
+ self.left_node = BinaryTree(new_node)
+ else:
+ tree = BinaryTree(new_node)
+ tree.left_node = self.left_node
+ self.left_node = tree
+
+ def insert_right(self, new_node):
+ if self.right_node == None:
+ self.right_node = BinaryTree(new_node)
+ else:
+ tree = BinaryTree(new_node)
+ tree.right_node = self.right_node
+ self.right_node = tree
+
+ def get_left_node(self):
+ return self.left_node
+
+ def get_right_node(self):
+ return self.right_node
+
+ def get_root_value(self):
+ return self.key
+
+ def set_root_node(self, new_root):
+ self.key = new_root
|
|
ff81d21d5e68e916282b61b3c65bf0af41a1bad8
|
app/scripts/po_stats.py
|
app/scripts/po_stats.py
|
#! /usr/bin/env python
import argparse
import glob
import json
import os
import subprocess
import sys
# Import local libraries
library_path = os.path.abspath(os.path.join(
os.path.dirname(__file__), os.pardir, 'libraries'))
# Polib library (https://bitbucket.org/izi/polib)
polib_path = os.path.join(library_path, 'polib')
if not os.path.isdir(polib_path):
try:
print 'Cloning polib...'
cmd_status = subprocess.check_output(
'hg clone https://bitbucket.org/izi/polib/ %s -u 1.0.7' % polib_path,
stderr=subprocess.STDOUT,
shell=True)
print cmd_status
except Exception as e:
print e
sys.path.append(os.path.join(polib_path))
try:
import polib
except ImportError:
print 'Error importing polib library'
sys.exit(1)
def create_file_list(repo_folder, locale, source_pattern):
''' Search for files to analyze '''
# Get a list of all reference files, since source_pattern can use wildcards
locale_files = glob.glob(
os.path.join(repo_folder, locale, source_pattern)
)
locale_files.sort()
return locale_files
def analyze_files(repo_folder, locale, source_pattern):
''' Analyze files, returning an array with stats and errors '''
global_stats = {}
# Get a list of all files for the reference locale
locale_files = create_file_list(repo_folder, locale, source_pattern)
for locale_file in locale_files:
fuzzy = 0
total = 0
translated = 0
untranslated = 0
try:
po = polib.pofile(locale_file)
fuzzy = len(po.fuzzy_entries())
translated = len(po.translated_entries())
untranslated = len(po.untranslated_entries())
except Exception as e:
print e
sys.exit(1)
total = translated + untranslated + fuzzy
source_index = os.path.basename(locale_file)
global_stats[source_index] = {
'fuzzy': fuzzy,
'total': total,
'translated': translated,
'untranslated': untranslated
}
return global_stats
def main():
parser = argparse.ArgumentParser()
parser.add_argument('repo_folder', help='Path to repository')
parser.add_argument(
'source_pattern', help='Source file pattern (wildcards are supported)')
parser.add_argument('locale', help='Locale code to analyze')
parser.add_argument('--pretty', action='store_true',
help='export indented and more readable JSON')
args = parser.parse_args()
global_stats = analyze_files(
args.repo_folder, args.locale, args.source_pattern)
if args.pretty:
print json.dumps(global_stats, sort_keys=True, indent=2)
else:
print json.dumps(global_stats)
if __name__ == '__main__':
main()
|
Create script using polib to generate stats for gettext file
|
Create script using polib to generate stats for gettext file
|
Python
|
mpl-2.0
|
flodolo/webstatus,mozilla-l10n/webstatus,mozilla-l10n/webstatus,flodolo/webstatus,flodolo/webstatus,mozilla-l10n/webstatus,mozilla-l10n/webstatus,flodolo/webstatus,mozilla-l10n/webstatus,flodolo/webstatus
|
---
+++
@@ -0,0 +1,99 @@
+#! /usr/bin/env python
+
+import argparse
+import glob
+import json
+import os
+import subprocess
+import sys
+
+# Import local libraries
+library_path = os.path.abspath(os.path.join(
+ os.path.dirname(__file__), os.pardir, 'libraries'))
+
+# Polib library (https://bitbucket.org/izi/polib)
+polib_path = os.path.join(library_path, 'polib')
+if not os.path.isdir(polib_path):
+ try:
+ print 'Cloning polib...'
+ cmd_status = subprocess.check_output(
+ 'hg clone https://bitbucket.org/izi/polib/ %s -u 1.0.7' % polib_path,
+ stderr=subprocess.STDOUT,
+ shell=True)
+ print cmd_status
+ except Exception as e:
+ print e
+sys.path.append(os.path.join(polib_path))
+try:
+ import polib
+except ImportError:
+ print 'Error importing polib library'
+ sys.exit(1)
+
+
+def create_file_list(repo_folder, locale, source_pattern):
+ ''' Search for files to analyze '''
+
+ # Get a list of all reference files, since source_pattern can use wildcards
+ locale_files = glob.glob(
+ os.path.join(repo_folder, locale, source_pattern)
+ )
+ locale_files.sort()
+
+ return locale_files
+
+
+def analyze_files(repo_folder, locale, source_pattern):
+ ''' Analyze files, returning an array with stats and errors '''
+
+ global_stats = {}
+
+ # Get a list of all files for the reference locale
+ locale_files = create_file_list(repo_folder, locale, source_pattern)
+ for locale_file in locale_files:
+ fuzzy = 0
+ total = 0
+ translated = 0
+ untranslated = 0
+ try:
+ po = polib.pofile(locale_file)
+ fuzzy = len(po.fuzzy_entries())
+ translated = len(po.translated_entries())
+ untranslated = len(po.untranslated_entries())
+ except Exception as e:
+ print e
+ sys.exit(1)
+
+ total = translated + untranslated + fuzzy
+ source_index = os.path.basename(locale_file)
+ global_stats[source_index] = {
+ 'fuzzy': fuzzy,
+ 'total': total,
+ 'translated': translated,
+ 'untranslated': untranslated
+ }
+
+ return global_stats
+
+
+def main():
+ parser = argparse.ArgumentParser()
+ parser.add_argument('repo_folder', help='Path to repository')
+ parser.add_argument(
+ 'source_pattern', help='Source file pattern (wildcards are supported)')
+ parser.add_argument('locale', help='Locale code to analyze')
+ parser.add_argument('--pretty', action='store_true',
+ help='export indented and more readable JSON')
+ args = parser.parse_args()
+
+ global_stats = analyze_files(
+ args.repo_folder, args.locale, args.source_pattern)
+
+ if args.pretty:
+ print json.dumps(global_stats, sort_keys=True, indent=2)
+ else:
+ print json.dumps(global_stats)
+
+
+if __name__ == '__main__':
+ main()
|
|
20b3b6415f68a25de82e5f6a3e2a3ded93d1620f
|
cal_pipe/plot_scans.py
|
cal_pipe/plot_scans.py
|
import numpy as np
import re
ms_active = raw_input("MS? : ")
field_str = raw_input("Field? : ")
tb.open(vis+"/FIELD")
names = tb.getcol('NAME')
matches = [string for string in names if re.match(field_str, string)]
posn_matches = \
[i for i, string in enumerate(names) if re.match(field_str, string)]
if len(matches) == 0:
raise TypeError("No matches found for the given field string")
tb.open(ms_active)
scanNums = sorted(np.unique(tb.getcol('SCAN_NUMBER')))
field_scans = []
for ii in range(numFields):
subtable = tb.query('FIELD_ID==%s'%ii)
field_scans.append(list(np.unique(subtable.getcol('SCAN_NUMBER'))))
tb.close()
field_scans = [scans for i, scans in field_scans if i in posn_matches]
for ii in range(len(field_scans)):
for jj in range(len(field_scans[ii])):
default('plotms')
vis = ms_active
xaxis = 'time'
yaxis = 'amp'
ydatacolumn = 'corrected'
selectdata = True
field = ii
scan = jj
correlation = "RR,LL"
averagedata = True
avgbaseline = True
transform = False
extendflag = False
plotrange = []
title = 'Amp vs Time: Field'+matches[ii]+' Scan'+str(jj)
xlabel = ''
ylabel = ''
showmajorgrid = False
showminorgrid = False
plotfile = 'field_'+matches[ii]+'_scan_'+str(jj)+'.png'
overwrite = True
showgui = False
async = False
plotms()
|
Save plots of amp vs time for each scan
|
Save plots of amp vs time for each scan
|
Python
|
mit
|
e-koch/canfar_scripts,e-koch/canfar_scripts
|
---
+++
@@ -0,0 +1,55 @@
+
+import numpy as np
+import re
+
+
+ms_active = raw_input("MS? : ")
+field_str = raw_input("Field? : ")
+
+
+tb.open(vis+"/FIELD")
+names = tb.getcol('NAME')
+matches = [string for string in names if re.match(field_str, string)]
+posn_matches = \
+ [i for i, string in enumerate(names) if re.match(field_str, string)]
+
+if len(matches) == 0:
+ raise TypeError("No matches found for the given field string")
+
+tb.open(ms_active)
+scanNums = sorted(np.unique(tb.getcol('SCAN_NUMBER')))
+field_scans = []
+for ii in range(numFields):
+ subtable = tb.query('FIELD_ID==%s'%ii)
+ field_scans.append(list(np.unique(subtable.getcol('SCAN_NUMBER'))))
+tb.close()
+
+field_scans = [scans for i, scans in field_scans if i in posn_matches]
+
+for ii in range(len(field_scans)):
+ for jj in range(len(field_scans[ii])):
+
+ default('plotms')
+ vis = ms_active
+ xaxis = 'time'
+ yaxis = 'amp'
+ ydatacolumn = 'corrected'
+ selectdata = True
+ field = ii
+ scan = jj
+ correlation = "RR,LL"
+ averagedata = True
+ avgbaseline = True
+ transform = False
+ extendflag = False
+ plotrange = []
+ title = 'Amp vs Time: Field'+matches[ii]+' Scan'+str(jj)
+ xlabel = ''
+ ylabel = ''
+ showmajorgrid = False
+ showminorgrid = False
+ plotfile = 'field_'+matches[ii]+'_scan_'+str(jj)+'.png'
+ overwrite = True
+ showgui = False
+ async = False
+ plotms()
|
|
9b46d66f89dba1f5bd2507f0a3dcddfd2758fe2b
|
cluster/update_jobs.py
|
cluster/update_jobs.py
|
from django.contrib.auth.models import User
from models import Job
from interface import get_all_jobs
def run_all():
for user in User.objects.all():
creds = user.credentials.all()
for i, cluster in enumerate(get_all_jobs(user)):
cred = creds[i]
jobs = {}
for job in cluster["jobs"]:
status = job[-1]
job_id = job[0]
if status in jobs:
jobs[status].append(job_id)
else:
jobs[status] = [job_id]
Job.update_states(cred, jobs)
if __name__ == "__main__":
run_all()
|
Add start for the job status updater
|
Add start for the job status updater
|
Python
|
mit
|
crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp
|
---
+++
@@ -0,0 +1,24 @@
+from django.contrib.auth.models import User
+
+from models import Job
+from interface import get_all_jobs
+
+
+def run_all():
+ for user in User.objects.all():
+ creds = user.credentials.all()
+ for i, cluster in enumerate(get_all_jobs(user)):
+ cred = creds[i]
+ jobs = {}
+ for job in cluster["jobs"]:
+ status = job[-1]
+ job_id = job[0]
+ if status in jobs:
+ jobs[status].append(job_id)
+ else:
+ jobs[status] = [job_id]
+ Job.update_states(cred, jobs)
+
+
+if __name__ == "__main__":
+ run_all()
|
|
013f2e526c862ed8e2c9b79aba43618b381d4bd3
|
test/test_session_getchatserver.py
|
test/test_session_getchatserver.py
|
import mock
import pytest
import requests
from pytwitcherapi import session
@pytest.fixture(scope="function")
def ts(mock_session):
"""Return a :class:`session.TwitchSession`
and mock the request of :class:`Session`
"""
return session.TwitchSession()
@pytest.fixture(scope='function')
def mock_chatpropresponse(servers, mock_session):
chatservers = [s.address for s in servers]
channelprop = {"chat_servers": chatservers}
chatpropresponse = mock.Mock()
chatpropresponse.json.return_value = channelprop
return chatpropresponse
@pytest.fixture(scope='function')
def mock_serverstatresponse(servers_json, mock_session):
serverstatresponse = mock.Mock()
serverstatresponse.json.return_value = servers_json
return serverstatresponse
@pytest.fixture(scope='function')
def mock_chatserverresponse(mock_serverstatresponse, mock_chatpropresponse,
servers_json):
requests.Session.request.side_effect = [mock_chatpropresponse,
mock_serverstatresponse]
# if serverstatresponse is successful return the best
s = servers_json[2]
return s['ip'], s['port']
@pytest.fixture(scope='function')
def mock_failchatserverresponse(mock_chatpropresponse, servers_json):
serverstatresponse = mock.Mock()
serverstatresponse.raise_for_status.side_effect = requests.HTTPError()
requests.Session.request.side_effect = [mock_chatpropresponse,
serverstatresponse]
# if serverstatresponse fails just return the first
s = servers_json[0]
return s['ip'], s['port']
@pytest.fixture(scope='function')
def mock_nochatserverresponse(mock_serverstatresponse):
# random server status that will not be in the available ones
chatprop = {"chat_servers": ['0.16.64.11:80', '0.16.24.11:123']}
chatpropresponse = mock.Mock()
chatpropresponse.json.return_value = chatprop
requests.Session.request.side_effect = [chatpropresponse,
mock_serverstatresponse]
# if no server stat for the chat servers can be found just return the first
return '0.16.64.11', 80
@pytest.mark.parametrize('fix', ['mock_chatserverresponse',
'mock_failchatserverresponse',
'mock_nochatserverresponse'])
def test_get_chat_server(ts, channel1, fix, request):
expected = request.getfuncargvalue(fix)
server, port = ts.get_chat_server(channel1)
assert (server, port) == expected
|
Add missing test for getchatserver
|
Add missing test for getchatserver
Forgot to track it -.-
|
Python
|
bsd-3-clause
|
Pytwitcher/pytwitcherapi,Pytwitcher/pytwitcherapi
|
---
+++
@@ -0,0 +1,71 @@
+import mock
+import pytest
+import requests
+
+from pytwitcherapi import session
+
+
+@pytest.fixture(scope="function")
+def ts(mock_session):
+ """Return a :class:`session.TwitchSession`
+ and mock the request of :class:`Session`
+ """
+ return session.TwitchSession()
+
+
+@pytest.fixture(scope='function')
+def mock_chatpropresponse(servers, mock_session):
+ chatservers = [s.address for s in servers]
+ channelprop = {"chat_servers": chatservers}
+ chatpropresponse = mock.Mock()
+ chatpropresponse.json.return_value = channelprop
+ return chatpropresponse
+
+
+@pytest.fixture(scope='function')
+def mock_serverstatresponse(servers_json, mock_session):
+ serverstatresponse = mock.Mock()
+ serverstatresponse.json.return_value = servers_json
+ return serverstatresponse
+
+
+@pytest.fixture(scope='function')
+def mock_chatserverresponse(mock_serverstatresponse, mock_chatpropresponse,
+ servers_json):
+ requests.Session.request.side_effect = [mock_chatpropresponse,
+ mock_serverstatresponse]
+ # if serverstatresponse is successful return the best
+ s = servers_json[2]
+ return s['ip'], s['port']
+
+
+@pytest.fixture(scope='function')
+def mock_failchatserverresponse(mock_chatpropresponse, servers_json):
+ serverstatresponse = mock.Mock()
+ serverstatresponse.raise_for_status.side_effect = requests.HTTPError()
+ requests.Session.request.side_effect = [mock_chatpropresponse,
+ serverstatresponse]
+ # if serverstatresponse fails just return the first
+ s = servers_json[0]
+ return s['ip'], s['port']
+
+
+@pytest.fixture(scope='function')
+def mock_nochatserverresponse(mock_serverstatresponse):
+ # random server status that will not be in the available ones
+ chatprop = {"chat_servers": ['0.16.64.11:80', '0.16.24.11:123']}
+ chatpropresponse = mock.Mock()
+ chatpropresponse.json.return_value = chatprop
+ requests.Session.request.side_effect = [chatpropresponse,
+ mock_serverstatresponse]
+ # if no server stat for the chat servers can be found just return the first
+ return '0.16.64.11', 80
+
+
+@pytest.mark.parametrize('fix', ['mock_chatserverresponse',
+ 'mock_failchatserverresponse',
+ 'mock_nochatserverresponse'])
+def test_get_chat_server(ts, channel1, fix, request):
+ expected = request.getfuncargvalue(fix)
+ server, port = ts.get_chat_server(channel1)
+ assert (server, port) == expected
|
|
aaea0100aed4ff33c4f67518d7859b78baffd2b6
|
setup.py
|
setup.py
|
# -*- coding: utf-8 -*-
import os
from setuptools import setup
def read(fname):
try:
return open(os.path.join(os.path.dirname(__file__), fname)).read()
except:
return ''
setup(
name='todoist-python',
version='0.2.10',
packages=['todoist', 'todoist.managers'],
author='Doist Team',
author_email='info@todoist.com',
license='BSD',
description='todoist-python - The official Todoist Python API library',
long_description = read('README.md'),
install_requires=[
'requests',
],
# see here for complete list of classifiers
# http://pypi.python.org/pypi?%3Aaction=list_classifiers
classifiers=(
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
),
)
|
# -*- coding: utf-8 -*-
import os
from setuptools import setup
def read(fname):
try:
return open(os.path.join(os.path.dirname(__file__), fname)).read()
except:
return ''
setup(
name='todoist-python',
version='0.2.11',
packages=['todoist', 'todoist.managers'],
author='Doist Team',
author_email='info@todoist.com',
license='BSD',
description='todoist-python - The official Todoist Python API library',
long_description = read('README.md'),
install_requires=[
'requests',
],
# see here for complete list of classifiers
# http://pypi.python.org/pypi?%3Aaction=list_classifiers
classifiers=(
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
),
)
|
Update the PyPI version to 0.2.11
|
Update the PyPI version to 0.2.11
|
Python
|
mit
|
electronick1/todoist-python,Doist/todoist-python
|
---
+++
@@ -10,7 +10,7 @@
setup(
name='todoist-python',
- version='0.2.10',
+ version='0.2.11',
packages=['todoist', 'todoist.managers'],
author='Doist Team',
author_email='info@todoist.com',
|
a99b2cb06acf5e4018f203c4298c62dc69655281
|
setup.py
|
setup.py
|
from setuptools import setup
setup(
name='XBlock',
version='0.1',
description='XBlock Core Library',
packages=['xblock'],
entry_points={
'xblock.v1': [
'helloworld = xblock.content:HelloWorldBlock',
'html = xblock.content:HtmlBlock',
'sequence = xblock.structure:Sequence',
'vertical = xblock.structure:VerticalBlock',
'sidebar = xblock.structure:SidebarBlock',
'problem = xblock.problem:ProblemBlock',
'textinput = xblock.problem:TextInputBlock',
'equality = xblock.problem:EqualityCheckerBlock',
'attempts_scoreboard = xblock.problem:AttemptsScoreboardBlock',
'thumbs = xblock.thumbs:ThumbsBlock',
'slider = xblock.slider:Slider',
'progress_slider = xblock.slider:ProgressSlider',
]
}
)
|
from setuptools import setup
setup(
name='XBlock',
version='0.1',
description='XBlock Core Library',
packages=['xblock'],
requires=[
'webob',
],
entry_points={
'xblock.v1': [
'helloworld = xblock.content:HelloWorldBlock',
'html = xblock.content:HtmlBlock',
'sequence = xblock.structure:Sequence',
'vertical = xblock.structure:VerticalBlock',
'sidebar = xblock.structure:SidebarBlock',
'problem = xblock.problem:ProblemBlock',
'textinput = xblock.problem:TextInputBlock',
'equality = xblock.problem:EqualityCheckerBlock',
'attempts_scoreboard = xblock.problem:AttemptsScoreboardBlock',
'thumbs = xblock.thumbs:ThumbsBlock',
'slider = xblock.slider:Slider',
'progress_slider = xblock.slider:ProgressSlider',
]
}
)
|
Add webob as a requirement.
|
Add webob as a requirement.
|
Python
|
apache-2.0
|
open-craft/XBlock,Lyla-Fischer/xblock-sdk,nagyistoce/edx-xblock-sdk,cpennington/XBlock,nagyistoce/edx-XBlock,4eek/XBlock,edx/XBlock,Lyla-Fischer/xblock-sdk,stvstnfrd/xblock-sdk,edx-solutions/xblock-sdk,edx-solutions/XBlock,dcadams/xblock-sdk,stvstnfrd/xblock-sdk,Pilou81715/hackathon_edX,edx-solutions/XBlock,cpennington/XBlock,stvstnfrd/xblock-sdk,edx/xblock-sdk,dcadams/xblock-sdk,mitodl/XBlock,jamiefolsom/xblock-sdk,Pilou81715/hackathon_edX,jamiefolsom/xblock-sdk,lovehhf/xblock-sdk,Pilou81715/hackathon_edX,dcadams/xblock-sdk,lovehhf/XBlock,edx/XBlock,nagyistoce/edx-xblock-sdk,nagyistoce/edx-xblock-sdk,lovehhf/xblock-sdk,EDUlib/XBlock,lovehhf/XBlock,cpennington/XBlock,lovehhf/xblock-sdk,jamiefolsom/xblock-sdk,edx-solutions/xblock-sdk,mitodl/XBlock,nagyistoce/edx-xblock-sdk,Pilou81715/hackathon_edX,edx/xblock-sdk,jamiefolsom/xblock-sdk,4eek/XBlock,edx-solutions/xblock-sdk,Lyla-Fischer/xblock-sdk,open-craft/XBlock,edx-solutions/xblock-sdk,edx/xblock-sdk,lovehhf/xblock-sdk,EDUlib/XBlock,nagyistoce/edx-XBlock
|
---
+++
@@ -5,6 +5,9 @@
version='0.1',
description='XBlock Core Library',
packages=['xblock'],
+ requires=[
+ 'webob',
+ ],
entry_points={
'xblock.v1': [
'helloworld = xblock.content:HelloWorldBlock',
|
955c4584a304c3b6a6dbbcf12eae3eed5e9a4cf5
|
scripts/python/cleanSimulation.py
|
scripts/python/cleanSimulation.py
|
#!/usr/bin/env/ python
# file: cleanSimulation.py
# author: Olivier Mesnard (mesnardo@gwu.edu)
# description: Clean a cuIBM simulation.
import os
import argparse
def read_inputs():
"""Parses the command-line."""
# create parser
parser = argparse.ArgumentParser(description='Clean PetIBM case',
formatter_class= argparse.ArgumentDefaultsHelpFormatter)
# fill parser with arguments
parser.add_argument('--case', dest='case_directory', type=str,
default=os.getcwd(),
help='directory of the PetIBM simulation')
parser.add_argument('--no-images', dest='images', action='store_false',
help='does not remove the images folder')
parser.add_argument('--no-grid', dest='grid', action='store_false',
help='does not remove the grid file')
parser.add_argument('--no-solutions', dest='solutions', action='store_false',
help='does not remove the numrical solution folders')
parser.add_argument('--no-forces', dest='forces', action='store_false',
help='does not remove the forces data file')
parser.add_argument('--no-logs', dest='logs', action='store_false',
help='does not remove log files '
'(iterations, run.info)')
parser.set_defaults(images=True, grid=True, solutions=True,
forces=True, logs=True)
return parser.parse_args()
def main():
"""Cleans a cuIBM simulation."""
# parser command-line
args = read_inputs()
# get different parts to clean
parts = {}
if args.images:
parts['images'] = '%s/images' % args.case_directory
if args.grid:
parts['grid'] = '%s/grid' % args.case_directory
if args.solutions:
parts['solutions'] = '%s/0*' % args.case_directory
if args.forces:
parts['forces'] = '%s/forces' % args.case_directory
if args.logs:
parts['logs'] = ('%s/iterations %s/run.info'
% (args.case_directory, args.case_directory))
# remove appropriate files/folders
print '[case-directory] %s' % args.case_directory
for key, part in parts.iteritems():
print '\t--> removing %s ...' % key
os.system('rm -rf %s' % part)
if __name__ == '__main__':
main()
|
Add script to clean a simulation
|
Add script to clean a simulation
|
Python
|
mit
|
barbagroup/cuIBM,barbagroup/cuIBM,barbagroup/cuIBM
|
---
+++
@@ -0,0 +1,64 @@
+#!/usr/bin/env/ python
+
+# file: cleanSimulation.py
+# author: Olivier Mesnard (mesnardo@gwu.edu)
+# description: Clean a cuIBM simulation.
+
+
+import os
+import argparse
+
+
+def read_inputs():
+ """Parses the command-line."""
+ # create parser
+ parser = argparse.ArgumentParser(description='Clean PetIBM case',
+ formatter_class= argparse.ArgumentDefaultsHelpFormatter)
+ # fill parser with arguments
+ parser.add_argument('--case', dest='case_directory', type=str,
+ default=os.getcwd(),
+ help='directory of the PetIBM simulation')
+ parser.add_argument('--no-images', dest='images', action='store_false',
+ help='does not remove the images folder')
+ parser.add_argument('--no-grid', dest='grid', action='store_false',
+ help='does not remove the grid file')
+ parser.add_argument('--no-solutions', dest='solutions', action='store_false',
+ help='does not remove the numrical solution folders')
+ parser.add_argument('--no-forces', dest='forces', action='store_false',
+ help='does not remove the forces data file')
+ parser.add_argument('--no-logs', dest='logs', action='store_false',
+ help='does not remove log files '
+ '(iterations, run.info)')
+ parser.set_defaults(images=True, grid=True, solutions=True,
+ forces=True, logs=True)
+ return parser.parse_args()
+
+
+def main():
+ """Cleans a cuIBM simulation."""
+ # parser command-line
+ args = read_inputs()
+
+ # get different parts to clean
+ parts = {}
+ if args.images:
+ parts['images'] = '%s/images' % args.case_directory
+ if args.grid:
+ parts['grid'] = '%s/grid' % args.case_directory
+ if args.solutions:
+ parts['solutions'] = '%s/0*' % args.case_directory
+ if args.forces:
+ parts['forces'] = '%s/forces' % args.case_directory
+ if args.logs:
+ parts['logs'] = ('%s/iterations %s/run.info'
+ % (args.case_directory, args.case_directory))
+
+ # remove appropriate files/folders
+ print '[case-directory] %s' % args.case_directory
+ for key, part in parts.iteritems():
+ print '\t--> removing %s ...' % key
+ os.system('rm -rf %s' % part)
+
+
+if __name__ == '__main__':
+ main()
|
|
048939778e5637eff997d674395d0f6df860a3eb
|
tests/integration/test_debugger.py
|
tests/integration/test_debugger.py
|
# -*- coding: utf8 -*-
from __future__ import unicode_literals
from future.standard_library import install_aliases
install_aliases()
import sys
import time
import threading
from urllib import parse as urlparse
import pytest
from bugbuzz import BugBuzz
from bugbuzz.packages import requests
# just a dummy unicode string, to see if we can handle unicode correctly
DUMMY_STR = u'除錯'
@pytest.fixture(scope='session')
def bugbuzz_dbg(
base_url='https://bugbuzz-api.herokuapp.com',
dashboard_url='http://dashboard.bugbuzz.io/',
):
return BugBuzz(base_url, dashboard_url)
def test_set_trace(mocker, bugbuzz_dbg):
mocker.patch('webbrowser.open_new_tab')
# post continue command
def post_continue():
time.sleep(3)
url = urlparse.urljoin(
bugbuzz_dbg.base_url,
'/sessions/{}/actions/continue'.format(
bugbuzz_dbg.client.session_id
),
)
requests.post(url)
thread = threading.Thread(target=post_continue)
thread.daemon = True
thread.start()
# TODO: set a timeout here?
bugbuzz_dbg.set_trace(sys._getframe())
url = urlparse.urljoin(
bugbuzz_dbg.base_url,
'/sessions/{}'.format(bugbuzz_dbg.client.session_id),
)
resp = requests.get(url)
session = resp.json()['session']
assert len(session['files']) == 1
assert len(session['breaks']) == 1
|
Add test to reproduce bug
|
Add test to reproduce bug
|
Python
|
mit
|
victorlin/bugbuzz-python,victorlin/bugbuzz-python
|
---
+++
@@ -0,0 +1,57 @@
+# -*- coding: utf8 -*-
+from __future__ import unicode_literals
+
+from future.standard_library import install_aliases
+install_aliases()
+
+import sys
+import time
+import threading
+from urllib import parse as urlparse
+
+import pytest
+
+from bugbuzz import BugBuzz
+from bugbuzz.packages import requests
+
+# just a dummy unicode string, to see if we can handle unicode correctly
+DUMMY_STR = u'除錯'
+
+
+@pytest.fixture(scope='session')
+def bugbuzz_dbg(
+ base_url='https://bugbuzz-api.herokuapp.com',
+ dashboard_url='http://dashboard.bugbuzz.io/',
+):
+ return BugBuzz(base_url, dashboard_url)
+
+
+def test_set_trace(mocker, bugbuzz_dbg):
+ mocker.patch('webbrowser.open_new_tab')
+
+ # post continue command
+ def post_continue():
+ time.sleep(3)
+ url = urlparse.urljoin(
+ bugbuzz_dbg.base_url,
+ '/sessions/{}/actions/continue'.format(
+ bugbuzz_dbg.client.session_id
+ ),
+ )
+ requests.post(url)
+
+ thread = threading.Thread(target=post_continue)
+ thread.daemon = True
+ thread.start()
+
+ # TODO: set a timeout here?
+ bugbuzz_dbg.set_trace(sys._getframe())
+
+ url = urlparse.urljoin(
+ bugbuzz_dbg.base_url,
+ '/sessions/{}'.format(bugbuzz_dbg.client.session_id),
+ )
+ resp = requests.get(url)
+ session = resp.json()['session']
+ assert len(session['files']) == 1
+ assert len(session['breaks']) == 1
|
|
d7ab04186f3b8c7c58b654a7372b1d4f3ffad64e
|
tests/unit/test_domain_commands.py
|
tests/unit/test_domain_commands.py
|
from caspy.domain import command, models
class TestBook:
def test_prepare_new_book(self):
empty_book = models.Book()
result = command.prepare_book(empty_book, 'now')
assert isinstance(result, models.Book)
assert result.created_at == 'now'
def test_prepare_old_book(self):
dated_book = models.Book(created_at='last week')
result = command.prepare_book(dated_book, 'now')
assert isinstance(result, models.Book)
assert result.created_at == 'last week'
|
Add unit tests for prepare_book
|
Add unit tests for prepare_book
|
Python
|
bsd-3-clause
|
altaurog/django-caspy,altaurog/django-caspy,altaurog/django-caspy
|
---
+++
@@ -0,0 +1,14 @@
+from caspy.domain import command, models
+
+class TestBook:
+ def test_prepare_new_book(self):
+ empty_book = models.Book()
+ result = command.prepare_book(empty_book, 'now')
+ assert isinstance(result, models.Book)
+ assert result.created_at == 'now'
+
+ def test_prepare_old_book(self):
+ dated_book = models.Book(created_at='last week')
+ result = command.prepare_book(dated_book, 'now')
+ assert isinstance(result, models.Book)
+ assert result.created_at == 'last week'
|
|
eac5962530542e1732326c1c0173294682d6256b
|
models/ras_220_genes/pmc_ids_venn.py
|
models/ras_220_genes/pmc_ids_venn.py
|
"""Plot a Venn diagram showing the IDs associated with articles in PMC."""
import matplotlib_venn as mv
import csv
from matplotlib import pyplot as plt
import plot_formatting as pf
pf.set_fig_params()
all_pmcids = set([])
has_doi = set([])
has_pmid = set([])
with open('PMC-ids.csv') as f:
csvreader = csv.reader(f, delimiter=',')
for row in csvreader:
pmcid = row[8].strip()
pmid = row[9].strip()
doi = row[7].strip()
all_pmcids.add(pmcid)
if doi:
has_doi.add(pmcid)
if pmid:
has_pmid.add(pmcid)
print len(all_pmcids)
plt.figure(figsize=(4, 4), dpi=150)
res = mv.venn2([has_doi, has_pmid],
("DOI", "PMID"))
plt.title('IDs for articles in PMC')
num_neither = len(all_pmcids.difference(has_doi).difference(has_pmid))
def commafy(text):
text_with_commas = ''
for ix, char in enumerate(reversed(str(text))):
if ix % 3 == 0 and ix != 0:
text_with_commas += ','
text_with_commas += char
return text_with_commas[::-1]
plt.text(-0.55, -0.8, '(plus %s with no DOI or PMID)' % commafy(num_neither))
# Add commas
for label in res.subset_labels:
text = str(label.get_text())
label.set_text(commafy(text))
plt.show()
plt.savefig('pmc_ids_venn.png', dpi=150)
plt.savefig('pmc_ids_venn.pdf')
|
Make Venn diagram of IDs in PMC
|
Make Venn diagram of IDs in PMC
|
Python
|
bsd-2-clause
|
jmuhlich/indra,johnbachman/belpy,pvtodorov/indra,sorgerlab/indra,pvtodorov/indra,johnbachman/indra,sorgerlab/indra,johnbachman/belpy,pvtodorov/indra,bgyori/indra,pvtodorov/indra,sorgerlab/belpy,johnbachman/belpy,jmuhlich/indra,sorgerlab/belpy,sorgerlab/belpy,sorgerlab/indra,johnbachman/indra,bgyori/indra,jmuhlich/indra,bgyori/indra,johnbachman/indra
|
---
+++
@@ -0,0 +1,51 @@
+"""Plot a Venn diagram showing the IDs associated with articles in PMC."""
+
+import matplotlib_venn as mv
+import csv
+from matplotlib import pyplot as plt
+import plot_formatting as pf
+
+pf.set_fig_params()
+
+all_pmcids = set([])
+has_doi = set([])
+has_pmid = set([])
+
+with open('PMC-ids.csv') as f:
+ csvreader = csv.reader(f, delimiter=',')
+ for row in csvreader:
+ pmcid = row[8].strip()
+ pmid = row[9].strip()
+ doi = row[7].strip()
+ all_pmcids.add(pmcid)
+ if doi:
+ has_doi.add(pmcid)
+ if pmid:
+ has_pmid.add(pmcid)
+print len(all_pmcids)
+
+plt.figure(figsize=(4, 4), dpi=150)
+res = mv.venn2([has_doi, has_pmid],
+ ("DOI", "PMID"))
+plt.title('IDs for articles in PMC')
+num_neither = len(all_pmcids.difference(has_doi).difference(has_pmid))
+
+def commafy(text):
+ text_with_commas = ''
+ for ix, char in enumerate(reversed(str(text))):
+ if ix % 3 == 0 and ix != 0:
+ text_with_commas += ','
+ text_with_commas += char
+ return text_with_commas[::-1]
+
+plt.text(-0.55, -0.8, '(plus %s with no DOI or PMID)' % commafy(num_neither))
+
+# Add commas
+for label in res.subset_labels:
+ text = str(label.get_text())
+ label.set_text(commafy(text))
+
+plt.show()
+plt.savefig('pmc_ids_venn.png', dpi=150)
+plt.savefig('pmc_ids_venn.pdf')
+
|
|
2ca84d16ff355a3e5a271ba800acb7c3cf4f8441
|
librisxl-tools/scripts/split_auth_source.py
|
librisxl-tools/scripts/split_auth_source.py
|
#!/usr/bin/env python
from __future__ import unicode_literals, print_function
import re
from os import makedirs, path as P
find_token = re.compile(r'{"(100|110|111|130|148|150|151|155|162|180|181|182|185)":').findall
def split_auth_source(sourcefile, outdir):
name_parts = P.basename(sourcefile).split('.', 1)
if not P.exists(outdir):
makedirs(outdir)
outfiles = {}
try:
source = open(sourcefile)
for i, l in enumerate(source):
for token in find_token(l):
outfp = outfiles.get(token)
if not outfp:
outfile = P.join(outdir,
'%s-%s.%s' % (name_parts[0], token, name_parts[1]))
print("Opening %s for writing..." % outfile)
outfp = outfiles[token] = open(outfile, 'w')
print(l, end="", file=outfp)
break
finally:
source.close()
for outfp in outfiles.values():
outfp.close()
if __name__ == '__main__':
import sys
args = sys.argv[1:]
sourcefile = args.pop(0)
outdir = args.pop(0)
split_auth_source(sourcefile, outdir)
|
Add script for splitting an auth dump on primary field presence
|
Add script for splitting an auth dump on primary field presence
|
Python
|
apache-2.0
|
libris/librisxl,libris/librisxl,libris/librisxl
|
---
+++
@@ -0,0 +1,35 @@
+#!/usr/bin/env python
+from __future__ import unicode_literals, print_function
+import re
+from os import makedirs, path as P
+
+find_token = re.compile(r'{"(100|110|111|130|148|150|151|155|162|180|181|182|185)":').findall
+
+def split_auth_source(sourcefile, outdir):
+ name_parts = P.basename(sourcefile).split('.', 1)
+ if not P.exists(outdir):
+ makedirs(outdir)
+ outfiles = {}
+ try:
+ source = open(sourcefile)
+ for i, l in enumerate(source):
+ for token in find_token(l):
+ outfp = outfiles.get(token)
+ if not outfp:
+ outfile = P.join(outdir,
+ '%s-%s.%s' % (name_parts[0], token, name_parts[1]))
+ print("Opening %s for writing..." % outfile)
+ outfp = outfiles[token] = open(outfile, 'w')
+ print(l, end="", file=outfp)
+ break
+ finally:
+ source.close()
+ for outfp in outfiles.values():
+ outfp.close()
+
+if __name__ == '__main__':
+ import sys
+ args = sys.argv[1:]
+ sourcefile = args.pop(0)
+ outdir = args.pop(0)
+ split_auth_source(sourcefile, outdir)
|
|
d55b5aa49d2ba1c98d568660b1a91b4b552872f0
|
numpy/core/tests/test_scalarprint.py
|
numpy/core/tests/test_scalarprint.py
|
# -*- coding: utf-8 -*-
""" Test printing of scalar types.
"""
import numpy as np
from numpy.testing import TestCase, assert_, run_module_suite
class TestRealScalars(TestCase):
def test_str(self):
svals = [0.0, -0.0, 1, -1, np.inf, -np.inf, np.nan]
styps = [np.float16, np.float32, np.float64, np.longdouble]
actual = [str(f(c)) for c in svals for f in styps]
wanted = [
'0.0', '0.0', '0.0', '0.0',
'-0.0', '-0.0', '-0.0', '-0.0',
'1.0', '1.0', '1.0', '1.0',
'-1.0', '-1.0', '-1.0', '-1.0',
'inf', 'inf', 'inf', 'inf',
'-inf', '-inf', '-inf', '-inf',
'nan', 'nan', 'nan', 'nan']
for res, val in zip(actual, wanted):
assert_(res == val)
if __name__ == "__main__":
run_module_suite()
|
Add test for printing of scalar values.
|
TST: Add test for printing of scalar values.
|
Python
|
bsd-3-clause
|
ViralLeadership/numpy,githubmlai/numpy,bringingheavendown/numpy,sigma-random/numpy,groutr/numpy,anntzer/numpy,drasmuss/numpy,chatcannon/numpy,pdebuyl/numpy,ChristopherHogan/numpy,dato-code/numpy,ViralLeadership/numpy,tacaswell/numpy,tdsmith/numpy,rudimeier/numpy,leifdenby/numpy,jakirkham/numpy,solarjoe/numpy,mwiebe/numpy,simongibbons/numpy,bmorris3/numpy,jonathanunderwood/numpy,rudimeier/numpy,Anwesh43/numpy,immerrr/numpy,hainm/numpy,yiakwy/numpy,dwillmer/numpy,has2k1/numpy,Srisai85/numpy,leifdenby/numpy,MSeifert04/numpy,jorisvandenbossche/numpy,joferkington/numpy,immerrr/numpy,mingwpy/numpy,larsmans/numpy,mathdd/numpy,utke1/numpy,dimasad/numpy,dimasad/numpy,shoyer/numpy,ewmoore/numpy,BMJHayward/numpy,mattip/numpy,madphysicist/numpy,rgommers/numpy,SunghanKim/numpy,GrimDerp/numpy,mhvk/numpy,cowlicks/numpy,skwbc/numpy,mathdd/numpy,simongibbons/numpy,jankoslavic/numpy,brandon-rhodes/numpy,njase/numpy,SiccarPoint/numpy,stuarteberg/numpy,andsor/numpy,pdebuyl/numpy,shoyer/numpy,mindw/numpy,pbrod/numpy,naritta/numpy,MSeifert04/numpy,stuarteberg/numpy,rajathkumarmp/numpy,ssanderson/numpy,githubmlai/numpy,sigma-random/numpy,WarrenWeckesser/numpy,maniteja123/numpy,jorisvandenbossche/numpy,ogrisel/numpy,embray/numpy,githubmlai/numpy,ahaldane/numpy,jschueller/numpy,ssanderson/numpy,AustereCuriosity/numpy,dato-code/numpy,ajdawson/numpy,WillieMaddox/numpy,Srisai85/numpy,hainm/numpy,ChristopherHogan/numpy,gmcastil/numpy,skwbc/numpy,embray/numpy,utke1/numpy,mhvk/numpy,cowlicks/numpy,rudimeier/numpy,tynn/numpy,Yusa95/numpy,mwiebe/numpy,Eric89GXL/numpy,maniteja123/numpy,ahaldane/numpy,dato-code/numpy,Anwesh43/numpy,ahaldane/numpy,argriffing/numpy,larsmans/numpy,Linkid/numpy,GaZ3ll3/numpy,rherault-insa/numpy,skymanaditya1/numpy,ContinuumIO/numpy,bringingheavendown/numpy,ssanderson/numpy,sinhrks/numpy,nbeaver/numpy,dwillmer/numpy,solarjoe/numpy,dwillmer/numpy,pbrod/numpy,numpy/numpy,pizzathief/numpy,GrimDerp/numpy,GrimDerp/numpy,rherault-insa/numpy,githubmlai/numpy,dwillmer/numpy,BabeNovelty/numpy,pizzathief/numpy,ekalosak/numpy,ddasilva/numpy,jakirkham/numpy,numpy/numpy,seberg/numpy,bringingheavendown/numpy,ESSS/numpy,jakirkham/numpy,rmcgibbo/numpy,behzadnouri/numpy,tacaswell/numpy,nbeaver/numpy,KaelChen/numpy,charris/numpy,jorisvandenbossche/numpy,ContinuumIO/numpy,ChanderG/numpy,mathdd/numpy,BabeNovelty/numpy,NextThought/pypy-numpy,Dapid/numpy,mortada/numpy,empeeu/numpy,nguyentu1602/numpy,rherault-insa/numpy,shoyer/numpy,tdsmith/numpy,ChanderG/numpy,moreati/numpy,seberg/numpy,mindw/numpy,musically-ut/numpy,SunghanKim/numpy,madphysicist/numpy,pyparallel/numpy,sonnyhu/numpy,MSeifert04/numpy,dch312/numpy,felipebetancur/numpy,rhythmsosad/numpy,skwbc/numpy,njase/numpy,empeeu/numpy,larsmans/numpy,b-carter/numpy,rmcgibbo/numpy,seberg/numpy,MichaelAquilina/numpy,SiccarPoint/numpy,Yusa95/numpy,NextThought/pypy-numpy,Srisai85/numpy,jorisvandenbossche/numpy,ChanderG/numpy,rgommers/numpy,ChanderG/numpy,ESSS/numpy,kiwifb/numpy,felipebetancur/numpy,Eric89GXL/numpy,kirillzhuravlev/numpy,mortada/numpy,mingwpy/numpy,trankmichael/numpy,groutr/numpy,bmorris3/numpy,endolith/numpy,empeeu/numpy,anntzer/numpy,kiwifb/numpy,kirillzhuravlev/numpy,jakirkham/numpy,CMartelLML/numpy,joferkington/numpy,argriffing/numpy,nguyentu1602/numpy,ahaldane/numpy,rajathkumarmp/numpy,ekalosak/numpy,has2k1/numpy,SiccarPoint/numpy,ddasilva/numpy,jakirkham/numpy,immerrr/numpy,pyparallel/numpy,NextThought/pypy-numpy,ChristopherHogan/numpy,jschueller/numpy,mortada/numpy,jorisvandenbossche/numpy,mattip/numpy,shoyer/numpy,BMJHayward/numpy,SiccarPoint/numpy,numpy/numpy,grlee77/numpy,sigma-random/numpy,NextThought/pypy-numpy,GaZ3ll3/numpy,brandon-rhodes/numpy,ewmoore/numpy,mattip/numpy,tynn/numpy,CMartelLML/numpy,Anwesh43/numpy,brandon-rhodes/numpy,ahaldane/numpy,drasmuss/numpy,GrimDerp/numpy,chiffa/numpy,SunghanKim/numpy,stuarteberg/numpy,grlee77/numpy,behzadnouri/numpy,bertrand-l/numpy,yiakwy/numpy,trankmichael/numpy,b-carter/numpy,pbrod/numpy,MSeifert04/numpy,WarrenWeckesser/numpy,BMJHayward/numpy,rhythmsosad/numpy,tacaswell/numpy,Anwesh43/numpy,charris/numpy,behzadnouri/numpy,WarrenWeckesser/numpy,jankoslavic/numpy,BabeNovelty/numpy,simongibbons/numpy,madphysicist/numpy,mhvk/numpy,anntzer/numpy,sonnyhu/numpy,moreati/numpy,grlee77/numpy,charris/numpy,sigma-random/numpy,chiffa/numpy,rajathkumarmp/numpy,naritta/numpy,sinhrks/numpy,sonnyhu/numpy,BMJHayward/numpy,has2k1/numpy,jschueller/numpy,njase/numpy,mingwpy/numpy,simongibbons/numpy,ajdawson/numpy,yiakwy/numpy,abalkin/numpy,ajdawson/numpy,KaelChen/numpy,felipebetancur/numpy,b-carter/numpy,BabeNovelty/numpy,Yusa95/numpy,ContinuumIO/numpy,andsor/numpy,pizzathief/numpy,ESSS/numpy,joferkington/numpy,rgommers/numpy,naritta/numpy,chatcannon/numpy,jankoslavic/numpy,ogrisel/numpy,grlee77/numpy,tdsmith/numpy,skymanaditya1/numpy,nguyentu1602/numpy,mingwpy/numpy,embray/numpy,rgommers/numpy,simongibbons/numpy,mhvk/numpy,rhythmsosad/numpy,Linkid/numpy,nguyentu1602/numpy,MaPePeR/numpy,MaPePeR/numpy,MichaelAquilina/numpy,mortada/numpy,rmcgibbo/numpy,utke1/numpy,skymanaditya1/numpy,joferkington/numpy,AustereCuriosity/numpy,cjermain/numpy,dch312/numpy,GaZ3ll3/numpy,CMartelLML/numpy,pdebuyl/numpy,gfyoung/numpy,cowlicks/numpy,mindw/numpy,rmcgibbo/numpy,ViralLeadership/numpy,grlee77/numpy,anntzer/numpy,AustereCuriosity/numpy,empeeu/numpy,dimasad/numpy,pbrod/numpy,stuarteberg/numpy,musically-ut/numpy,sonnyhu/numpy,solarjoe/numpy,bertrand-l/numpy,embray/numpy,kirillzhuravlev/numpy,rudimeier/numpy,argriffing/numpy,leifdenby/numpy,jschueller/numpy,rajathkumarmp/numpy,cjermain/numpy,endolith/numpy,gfyoung/numpy,pizzathief/numpy,dch312/numpy,felipebetancur/numpy,KaelChen/numpy,mindw/numpy,numpy/numpy,kirillzhuravlev/numpy,brandon-rhodes/numpy,ekalosak/numpy,nbeaver/numpy,hainm/numpy,maniteja123/numpy,has2k1/numpy,musically-ut/numpy,drasmuss/numpy,bmorris3/numpy,pbrod/numpy,Eric89GXL/numpy,MSeifert04/numpy,KaelChen/numpy,dimasad/numpy,ogrisel/numpy,CMartelLML/numpy,musically-ut/numpy,GaZ3ll3/numpy,andsor/numpy,bertrand-l/numpy,mathdd/numpy,madphysicist/numpy,pdebuyl/numpy,jonathanunderwood/numpy,bmorris3/numpy,ewmoore/numpy,gmcastil/numpy,pyparallel/numpy,ekalosak/numpy,ChristopherHogan/numpy,abalkin/numpy,MichaelAquilina/numpy,WarrenWeckesser/numpy,Srisai85/numpy,ddasilva/numpy,gfyoung/numpy,dch312/numpy,ogrisel/numpy,naritta/numpy,Linkid/numpy,andsor/numpy,shoyer/numpy,mhvk/numpy,cjermain/numpy,ajdawson/numpy,Dapid/numpy,MaPePeR/numpy,sinhrks/numpy,hainm/numpy,moreati/numpy,MaPePeR/numpy,kiwifb/numpy,MichaelAquilina/numpy,Linkid/numpy,trankmichael/numpy,jankoslavic/numpy,ogrisel/numpy,endolith/numpy,immerrr/numpy,rhythmsosad/numpy,tdsmith/numpy,tynn/numpy,chiffa/numpy,madphysicist/numpy,yiakwy/numpy,gmcastil/numpy,mwiebe/numpy,larsmans/numpy,mattip/numpy,WarrenWeckesser/numpy,endolith/numpy,ewmoore/numpy,seberg/numpy,chatcannon/numpy,pizzathief/numpy,jonathanunderwood/numpy,ewmoore/numpy,cjermain/numpy,abalkin/numpy,Eric89GXL/numpy,WillieMaddox/numpy,Yusa95/numpy,charris/numpy,sinhrks/numpy,groutr/numpy,cowlicks/numpy,Dapid/numpy,dato-code/numpy,trankmichael/numpy,skymanaditya1/numpy,SunghanKim/numpy,WillieMaddox/numpy,embray/numpy
|
---
+++
@@ -0,0 +1,29 @@
+# -*- coding: utf-8 -*-
+""" Test printing of scalar types.
+
+"""
+
+import numpy as np
+from numpy.testing import TestCase, assert_, run_module_suite
+
+
+class TestRealScalars(TestCase):
+ def test_str(self):
+ svals = [0.0, -0.0, 1, -1, np.inf, -np.inf, np.nan]
+ styps = [np.float16, np.float32, np.float64, np.longdouble]
+ actual = [str(f(c)) for c in svals for f in styps]
+ wanted = [
+ '0.0', '0.0', '0.0', '0.0',
+ '-0.0', '-0.0', '-0.0', '-0.0',
+ '1.0', '1.0', '1.0', '1.0',
+ '-1.0', '-1.0', '-1.0', '-1.0',
+ 'inf', 'inf', 'inf', 'inf',
+ '-inf', '-inf', '-inf', '-inf',
+ 'nan', 'nan', 'nan', 'nan']
+
+ for res, val in zip(actual, wanted):
+ assert_(res == val)
+
+
+if __name__ == "__main__":
+ run_module_suite()
|
|
f293d3d6aff79f424ab290347f99b85ca993e196
|
obfsproxy/common/transport_config.py
|
obfsproxy/common/transport_config.py
|
# -*- coding: utf-8 -*-
"""
Provides a class which represents a pluggable transport's configuration.
"""
class TransportConfig( object ):
"""
This class embeds configuration options for pluggable transport modules.
The options are set by obfsproxy and then passed to the transport's class
constructor. The pluggable transport might want to use these options but
does not have to. An example of such an option is the state location which
can be used by the pluggable transport to store persistent information.
"""
def __init__( self ):
"""
Initialise a `TransportConfig' object.
"""
self.stateLocation = None
def setStateLocation( self, stateLocation ):
"""
Set the given `stateLocation'.
"""
self.stateLocation = stateLocation
def getStateLocation( self ):
"""
Return the stored `stateLocation'.
"""
return self.stateLocation
def __str__( self ):
"""
Return a string representation of the `TransportConfig' instance.
"""
return str(vars(self))
|
Add a `TransportConfig' class which should contain configuration options (such as the state location) meant for pluggable transport modules.
|
Add a `TransportConfig' class which should contain configuration options (such
as the state location) meant for pluggable transport modules.
|
Python
|
bsd-3-clause
|
david415/obfsproxy,catinred2/obfsproxy,infinity0/obfsproxy,Yawning/obfsproxy-wfpadtools,sunsong/obfsproxy,isislovecruft/obfsproxy,qdzheng/obfsproxy,masterkorp/obfsproxy,Yawning/obfsproxy,NullHypothesis/obfsproxy
|
---
+++
@@ -0,0 +1,44 @@
+# -*- coding: utf-8 -*-
+
+"""
+Provides a class which represents a pluggable transport's configuration.
+"""
+
+class TransportConfig( object ):
+
+ """
+ This class embeds configuration options for pluggable transport modules.
+
+ The options are set by obfsproxy and then passed to the transport's class
+ constructor. The pluggable transport might want to use these options but
+ does not have to. An example of such an option is the state location which
+ can be used by the pluggable transport to store persistent information.
+ """
+
+ def __init__( self ):
+ """
+ Initialise a `TransportConfig' object.
+ """
+
+ self.stateLocation = None
+
+ def setStateLocation( self, stateLocation ):
+ """
+ Set the given `stateLocation'.
+ """
+
+ self.stateLocation = stateLocation
+
+ def getStateLocation( self ):
+ """
+ Return the stored `stateLocation'.
+ """
+
+ return self.stateLocation
+
+ def __str__( self ):
+ """
+ Return a string representation of the `TransportConfig' instance.
+ """
+
+ return str(vars(self))
|
|
1ddae2ddab2f0681d52f46baf2f0ec3926508d1b
|
bin/cwtv_json_to_srt.py
|
bin/cwtv_json_to_srt.py
|
#!/usr/bin/env python3
# encoding: utf-8
'''
This script will convert a close captioning subtitle .json file as found on the
CWTV streaming website, to a regular .srt for use in common media players.
.json example:
{
"endTime": 10.04,
"guid": "ffffffff-0000-1111-2222-aaaaaaaaaaaa",
"hitType": "tag",
"id": "abc123abc123abc123abc123",
"metadata": {
"ID": "",
"Language": "en",
"Text": "All that glitters"
},
"segmentTypeId": "abc123abc123abc123abc123",
"startTime": 1.002999,
"subTrack": null,
"tags": [],
"track": "Closed Captioning"
}
.srt example:
1
00:00:10,500 --> 00:00:13,000
All that glitters
'''
from __future__ import print_function
import json
import sys
def seconds_to_hms(seconds):
whole_seconds, microseconds = str(seconds).split('.')
whole_seconds = int(whole_seconds)
# The seconds part is just an int, but we need to turn the microseconds part
# into milliseconds. Subtracting seconds from this value will give us a
# float between 0.000000 and 0.999999, and multiplying that by a million is
# the number of microseconds as an int, from which we can have milliseconds.
microseconds = int((float(seconds) - whole_seconds) * 1000000)
milliseconds = int(microseconds / 1000) + whole_seconds * 1000
seconds, milliseconds = divmod(milliseconds, 1000)
minutes, seconds = divmod(seconds, 60)
hours, minutes = divmod(minutes, 60)
return '{:02d}:{:02d}:{:02d},{}'.format(hours, minutes, seconds,
str(milliseconds * 10).ljust(3, '0'))
if len(sys.argv) != 2:
print('You must provide a .json input file.')
sys.exit(1)
with open(sys.argv[1], 'rU') as f:
for index, item in enumerate(json.load(f)):
text = item['metadata']['Text']
start = seconds_to_hms(item['startTime'])
end = seconds_to_hms(item['endTime'])
print('{}\n{} --> {}\n{}\n'.format(index + 1, start, end, text))
|
Add CWTV .json to .srt script.
|
Add CWTV .json to .srt script.
|
Python
|
mit
|
alimony/dotfiles,alimony/dotfiles
|
---
+++
@@ -0,0 +1,67 @@
+#!/usr/bin/env python3
+# encoding: utf-8
+
+'''
+This script will convert a close captioning subtitle .json file as found on the
+CWTV streaming website, to a regular .srt for use in common media players.
+
+.json example:
+
+{
+ "endTime": 10.04,
+ "guid": "ffffffff-0000-1111-2222-aaaaaaaaaaaa",
+ "hitType": "tag",
+ "id": "abc123abc123abc123abc123",
+ "metadata": {
+ "ID": "",
+ "Language": "en",
+ "Text": "All that glitters"
+ },
+ "segmentTypeId": "abc123abc123abc123abc123",
+ "startTime": 1.002999,
+ "subTrack": null,
+ "tags": [],
+ "track": "Closed Captioning"
+}
+
+.srt example:
+
+1
+00:00:10,500 --> 00:00:13,000
+All that glitters
+
+'''
+
+from __future__ import print_function
+import json
+import sys
+
+
+def seconds_to_hms(seconds):
+ whole_seconds, microseconds = str(seconds).split('.')
+ whole_seconds = int(whole_seconds)
+
+ # The seconds part is just an int, but we need to turn the microseconds part
+ # into milliseconds. Subtracting seconds from this value will give us a
+ # float between 0.000000 and 0.999999, and multiplying that by a million is
+ # the number of microseconds as an int, from which we can have milliseconds.
+ microseconds = int((float(seconds) - whole_seconds) * 1000000)
+ milliseconds = int(microseconds / 1000) + whole_seconds * 1000
+
+ seconds, milliseconds = divmod(milliseconds, 1000)
+ minutes, seconds = divmod(seconds, 60)
+ hours, minutes = divmod(minutes, 60)
+
+ return '{:02d}:{:02d}:{:02d},{}'.format(hours, minutes, seconds,
+ str(milliseconds * 10).ljust(3, '0'))
+
+if len(sys.argv) != 2:
+ print('You must provide a .json input file.')
+ sys.exit(1)
+
+with open(sys.argv[1], 'rU') as f:
+ for index, item in enumerate(json.load(f)):
+ text = item['metadata']['Text']
+ start = seconds_to_hms(item['startTime'])
+ end = seconds_to_hms(item['endTime'])
+ print('{}\n{} --> {}\n{}\n'.format(index + 1, start, end, text))
|
|
9502c0e816097cf65fa92c6dd255c3356cf20964
|
test/api_class_repr_test.py
|
test/api_class_repr_test.py
|
# Copyright (c) 2012 - 2015 Lars Hupfeldt Nielsen, Hupfeldt IT
# All rights reserved. This work is under a BSD license, see LICENSE.TXT.
from __future__ import print_function
import pytest
from .. import jenkins_api
from .framework import api_select
from .cfg import ApiType
@pytest.mark.not_apis(ApiType.MOCK, ApiType.SCRIPT)
def test_api_class_repr_job(api_type):
api = api_select.api(__file__, api_type, login=True)
job = jenkins_api.ApiJob(api, {}, 'my-job')
jrd = eval(repr(job))
assert jrd == {'name': 'my-job', 'dct': {}}
invocation = jenkins_api.Invocation(job, "http://dummy", 'hello')
assert repr(invocation) == "Invocation: 'http://dummy' None None"
|
Test jenkis_api ApiJob and Invocation classes __repr__ methods
|
Test jenkis_api ApiJob and Invocation classes __repr__ methods
|
Python
|
bsd-3-clause
|
lhupfeldt/jenkinsflow,lhupfeldt/jenkinsflow,lechat/jenkinsflow,lhupfeldt/jenkinsflow,lechat/jenkinsflow,lechat/jenkinsflow,lechat/jenkinsflow,lhupfeldt/jenkinsflow
|
---
+++
@@ -0,0 +1,23 @@
+# Copyright (c) 2012 - 2015 Lars Hupfeldt Nielsen, Hupfeldt IT
+# All rights reserved. This work is under a BSD license, see LICENSE.TXT.
+
+from __future__ import print_function
+
+import pytest
+
+from .. import jenkins_api
+from .framework import api_select
+from .cfg import ApiType
+
+
+@pytest.mark.not_apis(ApiType.MOCK, ApiType.SCRIPT)
+def test_api_class_repr_job(api_type):
+ api = api_select.api(__file__, api_type, login=True)
+ job = jenkins_api.ApiJob(api, {}, 'my-job')
+
+ jrd = eval(repr(job))
+ assert jrd == {'name': 'my-job', 'dct': {}}
+
+ invocation = jenkins_api.Invocation(job, "http://dummy", 'hello')
+ assert repr(invocation) == "Invocation: 'http://dummy' None None"
+
|
|
9655b7349f48fe57a72897800d353aa1df4d5783
|
Problems/fibMemoization.py
|
Problems/fibMemoization.py
|
#!/usr/local/bin/python3
def main():
# Test suite
tests = [
[None, None], # Should throw a TypeError
[4, 3],
[7, 13]
]
for item in tests:
try:
temp_result = fib_memoization(item[0])
if temp_result == item[1]:
print('PASSED: fib_memoization({}) returned {}'.format(item[0], temp_result))
else:
print('FAILED: fib_memoization({}) returned {}, should have returned {}'.format(item[0], temp_result, item[1]))
except TypeError:
print('PASSED TypeError test')
return 0
def fib_memoization(n, d={1:1, 2:1}):
'''
Calculates the nth Fibonacci number
Input: n is integer, d is dictionary of previously calculated Fibonacci numbers
Output: integer, the nth Fibonacci number
'''
# Input checks
if type(n) is not int or type(d) is not dict:
raise TypeError('n must be an integer, d must be a dictionary')
if n in d:
return d[n]
else:
result = fib_memoization(n-1, d) + fib_memoization(n-2, d)
d[n] = result
return result
if __name__ == '__main__':
main()
|
Add memoization version to find nth Fibonacci number
|
Add memoization version to find nth Fibonacci number
|
Python
|
mit
|
HKuz/Test_Code
|
---
+++
@@ -0,0 +1,47 @@
+#!/usr/local/bin/python3
+
+
+def main():
+ # Test suite
+ tests = [
+ [None, None], # Should throw a TypeError
+ [4, 3],
+ [7, 13]
+ ]
+
+ for item in tests:
+ try:
+ temp_result = fib_memoization(item[0])
+ if temp_result == item[1]:
+ print('PASSED: fib_memoization({}) returned {}'.format(item[0], temp_result))
+ else:
+ print('FAILED: fib_memoization({}) returned {}, should have returned {}'.format(item[0], temp_result, item[1]))
+
+ except TypeError:
+ print('PASSED TypeError test')
+
+ return 0
+
+
+def fib_memoization(n, d={1:1, 2:1}):
+ '''
+ Calculates the nth Fibonacci number
+ Input: n is integer, d is dictionary of previously calculated Fibonacci numbers
+ Output: integer, the nth Fibonacci number
+ '''
+ # Input checks
+ if type(n) is not int or type(d) is not dict:
+ raise TypeError('n must be an integer, d must be a dictionary')
+
+
+ if n in d:
+ return d[n]
+ else:
+ result = fib_memoization(n-1, d) + fib_memoization(n-2, d)
+ d[n] = result
+ return result
+
+
+
+if __name__ == '__main__':
+ main()
|
|
8f117d62a60699b3d6b87e15962fca43339616a6
|
new_src/plot_multi_curves.py
|
new_src/plot_multi_curves.py
|
from __future__ import print_function
import os
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
parent_dir = os.path.dirname(os.getcwd())
models_dir = os.path.join(parent_dir, "models")
csv_name = "learning_curve.csv"
def plot_multi_curves(dfs, labels,
figure_name="learning_curves",
alphas=[0.6, 0.6, 0.6, 1.0]):
metrics = ["acc", "loss", "val_acc", "val_loss"]
for metric in metrics:
plt.figure(num=figure_name + "_" + metric)
for df, label, alpha in zip(dfs, labels, alphas):
curve = df[metric].values.tolist()
num = len(curve)
x = np.arange(1, num + 1)
if alpha == 1.0:
plt.plot(x, curve, color="k", label=label, alpha=alpha)
else:
plt.plot(x, curve, label=label, alpha=alpha)
if "loss" in metric:
legend_loc = 1
ylim = [0.0, 2.5]
ylabel = "Loss"
else:
legend_loc = 4
ylim = [0.4, 1.0]
ylabel = "Accuracy"
plt.ylim(ylim)
plt.xlim([0, num])
plt.xticks(fontsize=16)
plt.yticks(fontsize=16)
plt.xlabel("Epochs", fontsize=18)
plt.ylabel(ylabel, fontsize=18)
plt.legend(fontsize=16, loc=legend_loc, ncol=2)
plt.grid("on", linestyle="--", linewidth=0.5, alpha=0.5)
plt.tight_layout()
plt.show()
return
lr_models_names = ["model-test-lr-2", "model-test-lr-3",
"model-test-lr-4", "model-afm-max-adam-5-5"]
lr_labels = ["1e-3", "1e-4", "1e-5", "1e-3~1e-4~1e-5"]
lr_dfs = [pd.read_csv(os.path.join(models_dir, model_name, csv_name))
for model_name in lr_models_names]
plot_multi_curves(lr_dfs, lr_labels, "lr")
bs_models_names = ["model-test-bs-1", "model-test-bs-2",
"model-test-bs-3", "model-afm-max-adam-5-5"]
bs_labels = ["4", "8", "12", "16"]
bs_dfs = [pd.read_csv(os.path.join(models_dir, model_name, csv_name))
for model_name in bs_models_names]
plot_multi_curves(bs_dfs, bs_labels, "bs")
|
Add script to plot metrics of dataset
|
Add script to plot metrics of dataset
|
Python
|
mit
|
quqixun/BrainTumorClassification,quqixun/BrainTumorClassification
|
---
+++
@@ -0,0 +1,65 @@
+from __future__ import print_function
+import os
+import numpy as np
+import pandas as pd
+import matplotlib.pyplot as plt
+
+
+parent_dir = os.path.dirname(os.getcwd())
+models_dir = os.path.join(parent_dir, "models")
+csv_name = "learning_curve.csv"
+
+
+def plot_multi_curves(dfs, labels,
+ figure_name="learning_curves",
+ alphas=[0.6, 0.6, 0.6, 1.0]):
+
+ metrics = ["acc", "loss", "val_acc", "val_loss"]
+ for metric in metrics:
+ plt.figure(num=figure_name + "_" + metric)
+ for df, label, alpha in zip(dfs, labels, alphas):
+ curve = df[metric].values.tolist()
+ num = len(curve)
+ x = np.arange(1, num + 1)
+ if alpha == 1.0:
+ plt.plot(x, curve, color="k", label=label, alpha=alpha)
+ else:
+ plt.plot(x, curve, label=label, alpha=alpha)
+
+ if "loss" in metric:
+ legend_loc = 1
+ ylim = [0.0, 2.5]
+ ylabel = "Loss"
+ else:
+ legend_loc = 4
+ ylim = [0.4, 1.0]
+ ylabel = "Accuracy"
+ plt.ylim(ylim)
+ plt.xlim([0, num])
+ plt.xticks(fontsize=16)
+ plt.yticks(fontsize=16)
+ plt.xlabel("Epochs", fontsize=18)
+ plt.ylabel(ylabel, fontsize=18)
+ plt.legend(fontsize=16, loc=legend_loc, ncol=2)
+ plt.grid("on", linestyle="--", linewidth=0.5, alpha=0.5)
+ plt.tight_layout()
+ plt.show()
+ return
+
+
+lr_models_names = ["model-test-lr-2", "model-test-lr-3",
+ "model-test-lr-4", "model-afm-max-adam-5-5"]
+
+lr_labels = ["1e-3", "1e-4", "1e-5", "1e-3~1e-4~1e-5"]
+lr_dfs = [pd.read_csv(os.path.join(models_dir, model_name, csv_name))
+ for model_name in lr_models_names]
+plot_multi_curves(lr_dfs, lr_labels, "lr")
+
+
+bs_models_names = ["model-test-bs-1", "model-test-bs-2",
+ "model-test-bs-3", "model-afm-max-adam-5-5"]
+
+bs_labels = ["4", "8", "12", "16"]
+bs_dfs = [pd.read_csv(os.path.join(models_dir, model_name, csv_name))
+ for model_name in bs_models_names]
+plot_multi_curves(bs_dfs, bs_labels, "bs")
|
|
43bba1633233a03be5d585a2341ba56860b93c6b
|
tests/test_finance_model.py
|
tests/test_finance_model.py
|
# Copyright (c) 2012 The Pycroft Authors. See the AUTHORS file.
# This file is part of the Pycroft project and licensed under the terms of
# the Apache License, Version 2.0. See the LICENSE file for details.
import unittest
from pycroft import model
from pycroft.model import session, user, finance, _all
class Test_010_TransactionSplits(unittest.TestCase):
@classmethod
def setUpClass(cls):
session.reinit_session("sqlite://")
model.drop_db_model()
model.create_db_model()
cls.account = finance.FinanceAccount(name="Testaccount", type="EXPENSE")
cls.s = session.session
cls.s.add(cls.account)
cls.s.commit()
def tearDown(self):
super(Test_010_TransactionSplits, self).tearDown()
print "bla"
self.s.remove()
def test_0010_empty_transaction(self):
tr = finance.Transaction(message="Transaction1")
self.s.add(tr)
self.s.commit()
self.assertEqual(finance.Transaction.q.filter_by(message="Transaction1").count(), 1)
def test_0020_fail_on_unbalanced(self):
tr = finance.Transaction(message="Transaction2")
self.s.add(tr)
self.s.commit()
sp1 = finance.Split(amount=100, account=self.account, transaction=tr)
self.s.add(sp1)
self.assertRaisesRegexp(Exception, 'Transaction "Transaction2" is not balanced!', self.s.commit)
#self.s.rollback()
def test_0030_insert_balanced(self):
tr = finance.Transaction(message="Transaction2")
self.s.add(tr)
self.s.commit()
sp1 = finance.Split(amount=100, account=self.account, transaction=tr)
sp2 = finance.Split(amount=-100, account=self.account, transaction=tr)
self.s.add(sp1)
self.s.add(sp2)
self.s.commit()
|
Add a test to assert that transactions are balanced
|
Add a test to assert that transactions are balanced
|
Python
|
apache-2.0
|
lukasjuhrich/pycroft,lukasjuhrich/pycroft,lukasjuhrich/pycroft,agdsn/pycroft,agdsn/pycroft,lukasjuhrich/pycroft,agdsn/pycroft,agdsn/pycroft,agdsn/pycroft
|
---
+++
@@ -0,0 +1,49 @@
+# Copyright (c) 2012 The Pycroft Authors. See the AUTHORS file.
+# This file is part of the Pycroft project and licensed under the terms of
+# the Apache License, Version 2.0. See the LICENSE file for details.
+import unittest
+
+from pycroft import model
+from pycroft.model import session, user, finance, _all
+
+
+class Test_010_TransactionSplits(unittest.TestCase):
+ @classmethod
+ def setUpClass(cls):
+ session.reinit_session("sqlite://")
+ model.drop_db_model()
+ model.create_db_model()
+ cls.account = finance.FinanceAccount(name="Testaccount", type="EXPENSE")
+ cls.s = session.session
+ cls.s.add(cls.account)
+ cls.s.commit()
+
+ def tearDown(self):
+ super(Test_010_TransactionSplits, self).tearDown()
+ print "bla"
+ self.s.remove()
+
+ def test_0010_empty_transaction(self):
+ tr = finance.Transaction(message="Transaction1")
+ self.s.add(tr)
+ self.s.commit()
+ self.assertEqual(finance.Transaction.q.filter_by(message="Transaction1").count(), 1)
+
+ def test_0020_fail_on_unbalanced(self):
+ tr = finance.Transaction(message="Transaction2")
+ self.s.add(tr)
+ self.s.commit()
+ sp1 = finance.Split(amount=100, account=self.account, transaction=tr)
+ self.s.add(sp1)
+ self.assertRaisesRegexp(Exception, 'Transaction "Transaction2" is not balanced!', self.s.commit)
+ #self.s.rollback()
+
+ def test_0030_insert_balanced(self):
+ tr = finance.Transaction(message="Transaction2")
+ self.s.add(tr)
+ self.s.commit()
+ sp1 = finance.Split(amount=100, account=self.account, transaction=tr)
+ sp2 = finance.Split(amount=-100, account=self.account, transaction=tr)
+ self.s.add(sp1)
+ self.s.add(sp2)
+ self.s.commit()
|
|
0507bf58c7d73f91e645d64c69fcccb35542703a
|
VolumeUtilities.py
|
VolumeUtilities.py
|
#!/usr/bin/python
from snappy import *
from math import log
# Global constants:
PARI_PRECISION = 100
LINDEP_PRECISION = 50
EPSILON = 1e-12
# *** ATTENTION USER ***
# If you want to dynamicallly change the above constants for all sessions,
# or just want to keep the default values so you can revert,
# do so here! That way the values will both be used and tested in the warning system below
# User definitions:
# Enviroment setup code:
pari.set_real_precision(PARI_PRECISION)
# Test code; makes sure the constants are are sanely set:
# PARI & LINDEP PRECISION
if .9*PARI_PRECISION <= LINDEP_PRECISION or PARI_PRECISION - 3 <= LINDEP_PRECISION:
print 'WARNING: You set PARI to use '+str(PARI_PRECISION)+' places by default, with lindep calls at '+str(LINDEP_PRECISION)+' places;'
print 'This will probably read to rounding errors messing things up when lindep is used.'
print 'You should probably make sure LINDEP_PRECISION < both .9*PARI_PRECISION and PARI_PRECISION - 3 to avoid this.'
# EPSILON (vs. PARI_PRECISION)
if EPSILON <= 0:
print 'WARNING: You set EPSILON to '+str(EPSILON)+', but it must be positive.'
print 'Try setting EPSILON=abs(EPSILON)'
if EPSILON > .01:
print 'WARNING: You set EPSILON to '+str(EPSILON)+', which is really big.'
print 'PARI is capable of computing to hundreds of places of PRECISION, currently '+str(PARI_PRECISION)+', and at worst will use over 10.'
print 'You should set EPSILON to something smaller.'
if log(EPSILON) >= .9*log(PARI_PRECISION):
print 'WARNING: You set EPSILON to '+str(EPSILON)+', which is small compared to PARI\'s PRECISION limit of '+str(float(10)**(-PARI_PRECISION))+'.'
print 'Maybe you should set EPSILON a little bigger, so rounding errors from PARI don\'t get through.'
|
Add a new file for constants and other utilities to be thrown into.
|
Add a new file for constants and other utilities to be thrown into.
|
Python
|
mit
|
s-gilles/maps-reu-code
|
---
+++
@@ -0,0 +1,37 @@
+#!/usr/bin/python
+
+from snappy import *
+from math import log
+
+# Global constants:
+PARI_PRECISION = 100
+LINDEP_PRECISION = 50
+EPSILON = 1e-12
+
+# *** ATTENTION USER ***
+# If you want to dynamicallly change the above constants for all sessions,
+# or just want to keep the default values so you can revert,
+# do so here! That way the values will both be used and tested in the warning system below
+
+# User definitions:
+
+# Enviroment setup code:
+pari.set_real_precision(PARI_PRECISION)
+
+# Test code; makes sure the constants are are sanely set:
+# PARI & LINDEP PRECISION
+if .9*PARI_PRECISION <= LINDEP_PRECISION or PARI_PRECISION - 3 <= LINDEP_PRECISION:
+ print 'WARNING: You set PARI to use '+str(PARI_PRECISION)+' places by default, with lindep calls at '+str(LINDEP_PRECISION)+' places;'
+ print 'This will probably read to rounding errors messing things up when lindep is used.'
+ print 'You should probably make sure LINDEP_PRECISION < both .9*PARI_PRECISION and PARI_PRECISION - 3 to avoid this.'
+# EPSILON (vs. PARI_PRECISION)
+if EPSILON <= 0:
+ print 'WARNING: You set EPSILON to '+str(EPSILON)+', but it must be positive.'
+ print 'Try setting EPSILON=abs(EPSILON)'
+if EPSILON > .01:
+ print 'WARNING: You set EPSILON to '+str(EPSILON)+', which is really big.'
+ print 'PARI is capable of computing to hundreds of places of PRECISION, currently '+str(PARI_PRECISION)+', and at worst will use over 10.'
+ print 'You should set EPSILON to something smaller.'
+if log(EPSILON) >= .9*log(PARI_PRECISION):
+ print 'WARNING: You set EPSILON to '+str(EPSILON)+', which is small compared to PARI\'s PRECISION limit of '+str(float(10)**(-PARI_PRECISION))+'.'
+ print 'Maybe you should set EPSILON a little bigger, so rounding errors from PARI don\'t get through.'
|
|
26a847a9b5f9db3279849c6cc7505d41653887c9
|
setup.py
|
setup.py
|
from setuptools import setup, find_packages
setup(
name='typesystem',
version='0.1',
description="An abstract type system",
long_description="",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7'
],
keywords='types rdf',
author='Friedrich Lindenberg',
author_email='friedrich@pudo.org',
url='http://github.com/granoproject/typesystem',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'test']),
namespace_packages=[],
package_data={},
include_package_data=True,
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'python-dateutil==1.5',
'normality>=0.2.2'
],
tests_require=[],
entry_points={}
)
|
Make it a python package
|
Make it a python package
|
Python
|
mit
|
pudo/typecast,influencemapping/typesystem
|
---
+++
@@ -0,0 +1,34 @@
+from setuptools import setup, find_packages
+
+
+setup(
+ name='typesystem',
+ version='0.1',
+ description="An abstract type system",
+ long_description="",
+ classifiers=[
+ "Development Status :: 3 - Alpha",
+ "Intended Audience :: Developers",
+ "License :: OSI Approved :: MIT License",
+ "Operating System :: OS Independent",
+ 'Programming Language :: Python :: 2.6',
+ 'Programming Language :: Python :: 2.7'
+ ],
+ keywords='types rdf',
+ author='Friedrich Lindenberg',
+ author_email='friedrich@pudo.org',
+ url='http://github.com/granoproject/typesystem',
+ license='MIT',
+ packages=find_packages(exclude=['ez_setup', 'examples', 'test']),
+ namespace_packages=[],
+ package_data={},
+ include_package_data=True,
+ zip_safe=False,
+ test_suite='nose.collector',
+ install_requires=[
+ 'python-dateutil==1.5',
+ 'normality>=0.2.2'
+ ],
+ tests_require=[],
+ entry_points={}
+)
|
|
f698ca84bf01ea36dafa11a9e4937d733737c08b
|
fmn/lib/tests/test_regexes.py
|
fmn/lib/tests/test_regexes.py
|
import fmn.lib
class MockContext(object):
def __init__(self, name):
self.name = name
email = MockContext('email')
irc = MockContext('irc')
class TestRegexes(fmn.lib.tests.Base):
def test_valid_emails(self):
values = [
'awesome@fedoraproject.org',
'foo+fedora.org@bar.baz',
]
for value in values:
# None of these should raise exceptions
fmn.lib.validate_detail_value(email, value)
def test_invalid_emails(self):
values = [
'wat',
'not@anemail.org?',
]
for value in values:
# All of these should raise exceptions
try:
fmn.lib.validate_detail_value(email, value)
except ValueError:
pass
else:
raise ValueError("Invalid email %r did not fail" % value)
def test_valid_ircnicks(self):
values = [
'threebean',
'awesome|guy',
]
for value in values:
# None of these should raise exceptions
fmn.lib.validate_detail_value(irc, value)
def test_invalid_ircnicks(self):
values = [
'?',
]
for value in values:
# All of these should raise exceptions
try:
fmn.lib.validate_detail_value(irc, value)
except ValueError:
pass
else:
raise ValueError("Invalid ircnick %r did not fail" % value)
|
Add some tests for our detail value validator(s).
|
Add some tests for our detail value validator(s).
|
Python
|
lgpl-2.1
|
jeremycline/fmn,jeremycline/fmn,jeremycline/fmn
|
---
+++
@@ -0,0 +1,61 @@
+import fmn.lib
+
+
+class MockContext(object):
+ def __init__(self, name):
+ self.name = name
+
+
+email = MockContext('email')
+irc = MockContext('irc')
+
+
+class TestRegexes(fmn.lib.tests.Base):
+ def test_valid_emails(self):
+ values = [
+ 'awesome@fedoraproject.org',
+ 'foo+fedora.org@bar.baz',
+ ]
+
+ for value in values:
+ # None of these should raise exceptions
+ fmn.lib.validate_detail_value(email, value)
+
+ def test_invalid_emails(self):
+ values = [
+ 'wat',
+ 'not@anemail.org?',
+ ]
+
+ for value in values:
+ # All of these should raise exceptions
+ try:
+ fmn.lib.validate_detail_value(email, value)
+ except ValueError:
+ pass
+ else:
+ raise ValueError("Invalid email %r did not fail" % value)
+
+ def test_valid_ircnicks(self):
+ values = [
+ 'threebean',
+ 'awesome|guy',
+ ]
+
+ for value in values:
+ # None of these should raise exceptions
+ fmn.lib.validate_detail_value(irc, value)
+
+ def test_invalid_ircnicks(self):
+ values = [
+ '?',
+ ]
+
+ for value in values:
+ # All of these should raise exceptions
+ try:
+ fmn.lib.validate_detail_value(irc, value)
+ except ValueError:
+ pass
+ else:
+ raise ValueError("Invalid ircnick %r did not fail" % value)
|
|
96c4e54ed7bde9e41c6c235ff1654f47da2e23f3
|
cms_lab_data/cms_app.py
|
cms_lab_data/cms_app.py
|
from cms.app_base import CMSApp
from cms.apphook_pool import apphook_pool
class DataApp(CMSApp):
name = 'Data App'
urls = ['cms_lab_data.urls']
app_name = 'cms_lab_data'
apphook_pool.register(DataApp)
|
Create DataApp app hook for CMS
|
Create DataApp app hook for CMS
|
Python
|
bsd-3-clause
|
mfcovington/djangocms-lab-data,mfcovington/djangocms-lab-data,mfcovington/djangocms-lab-data
|
---
+++
@@ -0,0 +1,10 @@
+from cms.app_base import CMSApp
+from cms.apphook_pool import apphook_pool
+
+
+class DataApp(CMSApp):
+ name = 'Data App'
+ urls = ['cms_lab_data.urls']
+ app_name = 'cms_lab_data'
+
+apphook_pool.register(DataApp)
|
|
9801764a4b60a4ca6630936cdc1e8f85beb6020b
|
cea/tests/test_inputs_setup_workflow.py
|
cea/tests/test_inputs_setup_workflow.py
|
import os
import unittest
import cea.config
from cea.utilities import create_polygon
from cea.datamanagement import zone_helper, surroundings_helper, terrain_helper, streets_helper, data_initializer, \
archetypes_mapper
# Zug site coordinates
POLYGON_COORDINATES = [(8.513465734818856, 47.178027239429234), (8.515472027162078, 47.177895971877604),
(8.515214535096632, 47.175496635565885), (8.513139577193424, 47.175600066313542),
(8.513465734818856, 47.178027239429234)]
class TestInputSetupWorkflowCase(unittest.TestCase):
def setUp(self):
self.config = cea.config.Configuration(cea.config.DEFAULT_CONFIG)
self.config.project = os.path.expandvars("${TEMP}/reference-case-open")
def test_input_setup_workflow(self):
self.config.create_polygon.coordinates = POLYGON_COORDINATES
self.config.create_polygon.filename = 'site'
data_initializer.main(self.config)
create_polygon.main(self.config)
# TODO: Mock osmnx.create_footprints_download
zone_helper.main(self.config)
surroundings_helper.main(self.config)
terrain_helper.main(self.config)
streets_helper.main(self.config)
archetypes_mapper.main(self.config)
if __name__ == '__main__':
unittest.main()
|
Add test for input setup workflow
|
Add test for input setup workflow
|
Python
|
mit
|
architecture-building-systems/CEAforArcGIS,architecture-building-systems/CEAforArcGIS
|
---
+++
@@ -0,0 +1,35 @@
+import os
+import unittest
+
+import cea.config
+from cea.utilities import create_polygon
+from cea.datamanagement import zone_helper, surroundings_helper, terrain_helper, streets_helper, data_initializer, \
+ archetypes_mapper
+
+# Zug site coordinates
+POLYGON_COORDINATES = [(8.513465734818856, 47.178027239429234), (8.515472027162078, 47.177895971877604),
+ (8.515214535096632, 47.175496635565885), (8.513139577193424, 47.175600066313542),
+ (8.513465734818856, 47.178027239429234)]
+
+
+class TestInputSetupWorkflowCase(unittest.TestCase):
+ def setUp(self):
+ self.config = cea.config.Configuration(cea.config.DEFAULT_CONFIG)
+ self.config.project = os.path.expandvars("${TEMP}/reference-case-open")
+
+ def test_input_setup_workflow(self):
+ self.config.create_polygon.coordinates = POLYGON_COORDINATES
+ self.config.create_polygon.filename = 'site'
+
+ data_initializer.main(self.config)
+ create_polygon.main(self.config)
+ # TODO: Mock osmnx.create_footprints_download
+ zone_helper.main(self.config)
+ surroundings_helper.main(self.config)
+ terrain_helper.main(self.config)
+ streets_helper.main(self.config)
+ archetypes_mapper.main(self.config)
+
+
+if __name__ == '__main__':
+ unittest.main()
|
|
4fe20a71ef3a432a8e53bf498d847d9a66b099e9
|
migrations/002_add_month_start.py
|
migrations/002_add_month_start.py
|
"""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
record = Record(document)
collection.save(record.to_mongo())
|
Add migration for monthly data
|
Add migration for monthly data
|
Python
|
mit
|
alphagov/backdrop,alphagov/backdrop,alphagov/backdrop
|
---
+++
@@ -0,0 +1,23 @@
+"""
+Add _week_start_at field to all documents in all collections
+"""
+from backdrop.core.bucket import utc
+from backdrop.core.records import Record
+import logging
+
+log = logging.getLogger(__name__)
+
+
+def up(db):
+ for name in db.collection_names():
+ log.info("Migrating collection: {0}".format(name))
+ collection = db[name]
+ query = {
+ "_timestamp": {"$exists": True},
+ "_month_start_at": {"$exists": False}
+ }
+ for document in collection.find(query):
+ document['_timestamp'] = utc(document['_timestamp'])
+ record = Record(document)
+
+ collection.save(record.to_mongo())
|
|
b737cbbd3425d2c661ffa73bff39b18b30d8f914
|
cwappy/libcwap_test.py
|
cwappy/libcwap_test.py
|
#!/usr/bin/env python
import libcwap
def reader(size):
print "Got read"
return 'T' * size
def callback(name):
def actual_callback(*args, **kwargs):
print name, 'got args', args, 'and kwargs', kwargs
return actual_callback
actions = (
callback("time_request_function"),
callback("speaking_clock_request_function"),
callback("noise_request_function"),
callback("alarms_request_function"),
callback("time_set_function"),
callback("record_data"),
callback("record_data"),
callback("record_data"),
callback("record_data"),
callback("record_data"),
callback("delete_data"),
)
if __name__ == '__main__':
libcwap.register(actions)
libcwap.action(reader)
|
Add file for testing libcwap from the python side
|
Add file for testing libcwap from the python side
|
Python
|
mit
|
xim/tsoc,xim/tsoc,xim/tsoc,xim/tsoc
|
---
+++
@@ -0,0 +1,31 @@
+#!/usr/bin/env python
+
+import libcwap
+
+def reader(size):
+ print "Got read"
+ return 'T' * size
+
+def callback(name):
+ def actual_callback(*args, **kwargs):
+ print name, 'got args', args, 'and kwargs', kwargs
+ return actual_callback
+
+actions = (
+ callback("time_request_function"),
+ callback("speaking_clock_request_function"),
+ callback("noise_request_function"),
+ callback("alarms_request_function"),
+
+ callback("time_set_function"),
+ callback("record_data"),
+ callback("record_data"),
+ callback("record_data"),
+ callback("record_data"),
+ callback("record_data"),
+ callback("delete_data"),
+)
+
+if __name__ == '__main__':
+ libcwap.register(actions)
+ libcwap.action(reader)
|
|
8533068444eacf4a7731d3543b9308cd6d41b51d
|
Core/Color.py
|
Core/Color.py
|
# -*- coding:utf-8 -*-
# ***************************************************************************
# Color.py
# -------------------
# update : 2013-11-21
# copyright : (C) 2013 by Michaël Roy
# email : microygh@gmail.com
# ***************************************************************************
# ***************************************************************************
# * *
# * This program is free software; you can redistribute it and/or modify *
# * it under the terms of the GNU General Public License as published by *
# * the Free Software Foundation; either version 2 of the License, or *
# * (at your option) any later version. *
# * *
# ***************************************************************************
#-
#
# External dependencies
#
#-
from numpy import empty
from numpy.linalg import norm
#--
#
# Double2Color
#
#--
#
# Convert a value in range [ 0.0, 1.0 ] to a pseudo-color
#
def Value2Color( value ) :
if( value < 0.0 ) : return [ 0.0, 0.0 , 1.0 ]
if( value < 0.25 ) : return [ 0.0, value * 4.0, 1.0 ]
if( value < 0.50 ) : return [ 0.0, 1.0, 1.0 - (value - 0.25) * 4.0 ]
if( value < 0.75 ) : return [ (value - 0.5) * 4.0, 1.0, 0.0 ]
if( value < 1.0 ) : return [ 1.0, 1.0 - (value - 0.75) * 4.0, 0.0 ]
return [ 1.0, 0.0, 0.0 ]
#--
#
# Array2Color
#
#--
#
# Convert an array to a pseudo-color
#
def Array2Color( values ) :
# Initialize variables
colors = empty( (len(values), 3) )
value_lengths = empty( (len(values), 1) )
# Compute value vector length
for i in range( len(values) ) :
value_lengths[i] = norm( values[i] )
# Compute minimum and maximum value
min_value = value_lengths.min()
max_value = value_lengths.max()
# Compute the range of the values
value_range = max_value - min_value
# Convert each value to a pseudo-color
for i in range( len(values) ) :
colors[i] = Value2Color( (value_lengths[i] - min_value) / value_range )
# Return result
return colors
|
Add some color utility functions.
|
Add some color utility functions.
|
Python
|
mit
|
microy/MeshToolkit,microy/PyMeshToolkit,microy/MeshToolkit,microy/PyMeshToolkit
|
---
+++
@@ -0,0 +1,81 @@
+# -*- coding:utf-8 -*-
+
+# ***************************************************************************
+# Color.py
+# -------------------
+# update : 2013-11-21
+# copyright : (C) 2013 by Michaël Roy
+# email : microygh@gmail.com
+# ***************************************************************************
+
+# ***************************************************************************
+# * *
+# * This program is free software; you can redistribute it and/or modify *
+# * it under the terms of the GNU General Public License as published by *
+# * the Free Software Foundation; either version 2 of the License, or *
+# * (at your option) any later version. *
+# * *
+# ***************************************************************************
+
+
+#-
+#
+# External dependencies
+#
+#-
+from numpy import empty
+from numpy.linalg import norm
+
+
+#--
+#
+# Double2Color
+#
+#--
+#
+# Convert a value in range [ 0.0, 1.0 ] to a pseudo-color
+#
+def Value2Color( value ) :
+
+ if( value < 0.0 ) : return [ 0.0, 0.0 , 1.0 ]
+ if( value < 0.25 ) : return [ 0.0, value * 4.0, 1.0 ]
+ if( value < 0.50 ) : return [ 0.0, 1.0, 1.0 - (value - 0.25) * 4.0 ]
+ if( value < 0.75 ) : return [ (value - 0.5) * 4.0, 1.0, 0.0 ]
+ if( value < 1.0 ) : return [ 1.0, 1.0 - (value - 0.75) * 4.0, 0.0 ]
+ return [ 1.0, 0.0, 0.0 ]
+
+
+
+
+#--
+#
+# Array2Color
+#
+#--
+#
+# Convert an array to a pseudo-color
+#
+def Array2Color( values ) :
+
+ # Initialize variables
+ colors = empty( (len(values), 3) )
+ value_lengths = empty( (len(values), 1) )
+
+ # Compute value vector length
+ for i in range( len(values) ) :
+ value_lengths[i] = norm( values[i] )
+
+ # Compute minimum and maximum value
+ min_value = value_lengths.min()
+ max_value = value_lengths.max()
+
+ # Compute the range of the values
+ value_range = max_value - min_value
+
+ # Convert each value to a pseudo-color
+ for i in range( len(values) ) :
+ colors[i] = Value2Color( (value_lengths[i] - min_value) / value_range )
+
+ # Return result
+ return colors
+
|
|
70294b332920d23f97c93c720e6078e6358b7272
|
prjxray/xyaml.py
|
prjxray/xyaml.py
|
#!/usr/bin/env python3
import io
import re
import yaml
import json
import unittest
def load(f):
data = f.read()
# Strip out of !<tags>
data = re.sub("!<[^>]*>", "", data)
return yaml.load(io.StringIO(data))
def tojson(f):
d = load(f)
return json.dumps(d, sort_keys=True, indent=4)
class XYamlTest(unittest.TestCase):
def test(self):
s = io.StringIO("""\
!<xilinx/xc7series/part>
idcode: 0x362d093
global_clock_regions:
top: !<xilinx/xc7series/global_clock_region>
rows:
0: !<xilinx/xc7series/row>
configuration_buses:
CLB_IO_CLK: !<xilinx/xc7series/configuration_bus>
configuration_columns:
0: !<xilinx/xc7series/configuration_column>
frame_count: 42
""")
djson = tojson(s)
self.assertMultiLineEqual(djson, """\
{
"global_clock_regions": {
"top": {
"rows": {
"0": {
"configuration_buses": {
"CLB_IO_CLK": {
"configuration_columns": {
"0": {
"frame_count": 42
}
}
}
}
}
}
}
},
"idcode": 56807571
}""")
if __name__ == "__main__":
import sys
if len(sys.argv) == 1:
unittest.main()
else:
assert len(sys.argv) == 2
print(tojson(open(sys.argv[1])))
|
Add wrapper to allow easy loading of yaml files.
|
Add wrapper to allow easy loading of yaml files.
Fixes #327.
Signed-off-by: Tim 'mithro' Ansell <b1c1d8736f20db3fb6c1c66bb1455ed43909f0d8@mith.ro>
|
Python
|
isc
|
SymbiFlow/prjxray,SymbiFlow/prjxray,SymbiFlow/prjxray,SymbiFlow/prjxray,SymbiFlow/prjxray
|
---
+++
@@ -0,0 +1,65 @@
+#!/usr/bin/env python3
+import io
+import re
+import yaml
+import json
+import unittest
+
+
+def load(f):
+ data = f.read()
+ # Strip out of !<tags>
+ data = re.sub("!<[^>]*>", "", data)
+ return yaml.load(io.StringIO(data))
+
+
+def tojson(f):
+ d = load(f)
+ return json.dumps(d, sort_keys=True, indent=4)
+
+
+class XYamlTest(unittest.TestCase):
+ def test(self):
+ s = io.StringIO("""\
+!<xilinx/xc7series/part>
+idcode: 0x362d093
+global_clock_regions:
+ top: !<xilinx/xc7series/global_clock_region>
+ rows:
+ 0: !<xilinx/xc7series/row>
+ configuration_buses:
+ CLB_IO_CLK: !<xilinx/xc7series/configuration_bus>
+ configuration_columns:
+ 0: !<xilinx/xc7series/configuration_column>
+ frame_count: 42
+""")
+ djson = tojson(s)
+ self.assertMultiLineEqual(djson, """\
+{
+ "global_clock_regions": {
+ "top": {
+ "rows": {
+ "0": {
+ "configuration_buses": {
+ "CLB_IO_CLK": {
+ "configuration_columns": {
+ "0": {
+ "frame_count": 42
+ }
+ }
+ }
+ }
+ }
+ }
+ }
+ },
+ "idcode": 56807571
+}""")
+
+if __name__ == "__main__":
+ import sys
+ if len(sys.argv) == 1:
+ unittest.main()
+ else:
+ assert len(sys.argv) == 2
+ print(tojson(open(sys.argv[1])))
|
|
8479da007990d561147951440f1f520ebbcfdadc
|
spyder_unittest/tests/test_unittestplugin.py
|
spyder_unittest/tests/test_unittestplugin.py
|
# -*- coding: utf-8 -*-
#
# Copyright © 2017 Spyder Project Contributors
# Licensed under the terms of the MIT License
# (see LICENSE.txt for details)
"""Tests for unittestplugin.py"""
# Third party imports
from qtpy.QtWidgets import QWidget
import pytest
# Local imports
from spyder_unittest.unittestplugin import UnitTestPlugin
try:
from unittest.mock import Mock
except ImportError:
from mock import Mock # Python 2
@pytest.fixture
def plugin(qtbot):
"""Set up the unittest plugin."""
res = UnitTestPlugin(None)
qtbot.addWidget(res)
res.main = Mock()
res.main.run_menu_actions = [42]
res.main.editor.pythonfile_dependent_actions = [42]
res.register_plugin()
return res
def test_initialization(plugin):
"""Check that plugin initialization does not yield an error."""
plugin.show()
|
Add baseline test for initializing plugin
|
Add baseline test for initializing plugin
|
Python
|
mit
|
jitseniesen/spyder-unittest
|
---
+++
@@ -0,0 +1,35 @@
+# -*- coding: utf-8 -*-
+#
+# Copyright © 2017 Spyder Project Contributors
+# Licensed under the terms of the MIT License
+# (see LICENSE.txt for details)
+"""Tests for unittestplugin.py"""
+
+# Third party imports
+from qtpy.QtWidgets import QWidget
+import pytest
+
+# Local imports
+from spyder_unittest.unittestplugin import UnitTestPlugin
+
+try:
+ from unittest.mock import Mock
+except ImportError:
+ from mock import Mock # Python 2
+
+
+@pytest.fixture
+def plugin(qtbot):
+ """Set up the unittest plugin."""
+ res = UnitTestPlugin(None)
+ qtbot.addWidget(res)
+ res.main = Mock()
+ res.main.run_menu_actions = [42]
+ res.main.editor.pythonfile_dependent_actions = [42]
+ res.register_plugin()
+ return res
+
+
+def test_initialization(plugin):
+ """Check that plugin initialization does not yield an error."""
+ plugin.show()
|
|
fbb07f1448d7759c13db735f34d30a86938b8bf4
|
tests/test_postgresql.py
|
tests/test_postgresql.py
|
import pytest
from mock import MagicMock
import zmon_aws_agent.postgresql as postgresql
def test_get_databases_from_clusters():
pgclusters = [
{
'id': 'test-1',
'dnsname': 'test-1.db.zalan.do'
}
]
acc = '1234567890'
region = 'eu-xxx-1'
postgresql.list_postgres_databases = MagicMock()
postgresql.list_postgres_databases.return_value = ['db1', 'db2']
databases = postgresql.get_databases_from_clusters(pgclusters, acc, region,
'pguser', 'pgpass')
assert databases == [
{
'id': 'db1-test-1.db.zalan.do[aws:1234567890:eu-xxx-1]',
'type': 'postgresql_database',
'created_by': 'agent',
'infrastructure_account': acc,
'region': region,
'postgresql_cluster': 'test-1',
'database_name': 'db1',
'shards': {
'db1': 'test-1.db.zalan.do:5432/db1'
}
},
{
'id': 'db2-test-1.db.zalan.do[aws:1234567890:eu-xxx-1]',
'type': 'postgresql_database',
'created_by': 'agent',
'infrastructure_account': acc,
'region': region,
'postgresql_cluster': 'test-1',
'database_name': 'db2',
'shards': {
'db2': 'test-1.db.zalan.do:5432/db2'
}
}
]
|
Add minimal test for new postgresql.py module
|
Add minimal test for new postgresql.py module
|
Python
|
apache-2.0
|
zalando/zmon-aws-agent,zalando/zmon-aws-agent
|
---
+++
@@ -0,0 +1,47 @@
+import pytest
+from mock import MagicMock
+
+import zmon_aws_agent.postgresql as postgresql
+
+
+def test_get_databases_from_clusters():
+ pgclusters = [
+ {
+ 'id': 'test-1',
+ 'dnsname': 'test-1.db.zalan.do'
+ }
+ ]
+ acc = '1234567890'
+ region = 'eu-xxx-1'
+
+ postgresql.list_postgres_databases = MagicMock()
+ postgresql.list_postgres_databases.return_value = ['db1', 'db2']
+
+ databases = postgresql.get_databases_from_clusters(pgclusters, acc, region,
+ 'pguser', 'pgpass')
+ assert databases == [
+ {
+ 'id': 'db1-test-1.db.zalan.do[aws:1234567890:eu-xxx-1]',
+ 'type': 'postgresql_database',
+ 'created_by': 'agent',
+ 'infrastructure_account': acc,
+ 'region': region,
+ 'postgresql_cluster': 'test-1',
+ 'database_name': 'db1',
+ 'shards': {
+ 'db1': 'test-1.db.zalan.do:5432/db1'
+ }
+ },
+ {
+ 'id': 'db2-test-1.db.zalan.do[aws:1234567890:eu-xxx-1]',
+ 'type': 'postgresql_database',
+ 'created_by': 'agent',
+ 'infrastructure_account': acc,
+ 'region': region,
+ 'postgresql_cluster': 'test-1',
+ 'database_name': 'db2',
+ 'shards': {
+ 'db2': 'test-1.db.zalan.do:5432/db2'
+ }
+ }
+ ]
|
|
2bbffd5eec2e42897969e504551c16d4abbf5ba9
|
tests/test_test_utils.py
|
tests/test_test_utils.py
|
from django.test import TestCase
from constance import config
from constance.test import override_config
class OverrideConfigFunctionDecoratorTestCase(TestCase):
"""Test that the override_config decorator works correctly.
Test usage of override_config on test method and as context manager.
"""
def test_default_value_is_true(self):
"""Assert that the default value of config.BOOL_VALUE is True."""
self.assertTrue(config.BOOL_VALUE)
@override_config(BOOL_VALUE=False)
def test_override_config_on_method_changes_config_value(self):
"""Assert that the method decorator changes config.BOOL_VALUE."""
self.assertFalse(config.BOOL_VALUE)
def test_override_config_as_context_manager_changes_config_value(self):
"""Assert that the context manager changes config.BOOL_VALUE."""
with override_config(BOOL_VALUE=False):
self.assertFalse(config.BOOL_VALUE)
self.assertTrue(config.BOOL_VALUE)
@override_config(BOOL_VALUE=False)
class OverrideConfigClassDecoratorTestCase(TestCase):
"""Test that the override_config decorator works on classes."""
def test_override_config_on_class_changes_config_value(self):
"""Asser that the class decorator changes config.BOOL_VALUE."""
self.assertFalse(config.BOOL_VALUE)
|
Add test cases for override_config
|
Add test cases for override_config
Test usage of override_config in different forms
Ensure flexibility between decorator and context manager
|
Python
|
bsd-3-clause
|
pombredanne/django-constance,winzard/django-constance,jonzlin95/django-constance,pombredanne/django-constance,jonzlin95/django-constance,jazzband/django-constance,dmugtasimov/django-constance,jezdez/django-constance,APSL/django-constance,jazzband/django-constance,dmugtasimov/django-constance,winzard/django-constance,thorgate/django-constance,jazzband/django-constance,jezdez/django-constance,APSL/django-constance,thorgate/django-constance
|
---
+++
@@ -0,0 +1,34 @@
+from django.test import TestCase
+
+from constance import config
+from constance.test import override_config
+
+
+class OverrideConfigFunctionDecoratorTestCase(TestCase):
+ """Test that the override_config decorator works correctly.
+
+ Test usage of override_config on test method and as context manager.
+ """
+ def test_default_value_is_true(self):
+ """Assert that the default value of config.BOOL_VALUE is True."""
+ self.assertTrue(config.BOOL_VALUE)
+
+ @override_config(BOOL_VALUE=False)
+ def test_override_config_on_method_changes_config_value(self):
+ """Assert that the method decorator changes config.BOOL_VALUE."""
+ self.assertFalse(config.BOOL_VALUE)
+
+ def test_override_config_as_context_manager_changes_config_value(self):
+ """Assert that the context manager changes config.BOOL_VALUE."""
+ with override_config(BOOL_VALUE=False):
+ self.assertFalse(config.BOOL_VALUE)
+
+ self.assertTrue(config.BOOL_VALUE)
+
+
+@override_config(BOOL_VALUE=False)
+class OverrideConfigClassDecoratorTestCase(TestCase):
+ """Test that the override_config decorator works on classes."""
+ def test_override_config_on_class_changes_config_value(self):
+ """Asser that the class decorator changes config.BOOL_VALUE."""
+ self.assertFalse(config.BOOL_VALUE)
|
|
bc69916b71a04a4f54ef7c8ca2fb7142260634f2
|
tests/test_validators.py
|
tests/test_validators.py
|
"""
test_validators
~~~~~~~~~~~~~~
Unittests for bundled validators.
:copyright: 2007-2008 by James Crasta, Thomas Johansson.
:license: MIT, see LICENSE.txt for details.
"""
from py.test import raises
from wtforms.validators import ValidationError, length, url, not_empty, email, ip_address
class DummyForm(object):
pass
class DummyField(object):
def __init__(self, data):
self.data = data
form = DummyForm()
def test_email():
assert email(form, DummyField('foo@bar.dk')) == None
assert email(form, DummyField('123@bar.dk')) == None
assert email(form, DummyField('foo@456.dk')) == None
assert email(form, DummyField('foo@bar456.info')) == None
raises(ValidationError, email, form, DummyField('foo')) == None
raises(ValidationError, email, form, DummyField('bar.dk')) == None
raises(ValidationError, email, form, DummyField('foo@')) == None
raises(ValidationError, email, form, DummyField('@bar.dk')) == None
raises(ValidationError, email, form, DummyField('foo@bar')) == None
raises(ValidationError, email, form, DummyField('foo@bar.ab12')) == None
raises(ValidationError, email, form, DummyField('foo@bar.abcde')) == None
def test_length():
field = DummyField('foobar')
assert length(min=2, max=6)(form, field) == None
raises(ValidationError, length(min=7), form, field)
raises(ValidationError, length(max=5), form, field)
def test_url():
assert url()(form, DummyField('http://foobar.dk')) == None
assert url()(form, DummyField('http://foobar.dk/')) == None
assert url()(form, DummyField('http://foobar.dk/foobar')) == None
raises(ValidationError, url(), form, DummyField('http://foobar'))
raises(ValidationError, url(), form, DummyField('foobar.dk'))
raises(ValidationError, url(), form, DummyField('http://foobar.12'))
def test_not_empty():
assert not_empty()(form, DummyField('foobar')) == None
raises(ValidationError, not_empty(), form, DummyField(''))
raises(ValidationError, not_empty(), form, DummyField(' '))
def test_ip_address():
assert ip_address(form, DummyField('127.0.0.1')) == None
raises(ValidationError, ip_address, form, DummyField('abc.0.0.1'))
raises(ValidationError, ip_address, form, DummyField('1278.0.0.1'))
raises(ValidationError, ip_address, form, DummyField('127.0.0.abc'))
|
Add first basic unittests using py.test
|
Add first basic unittests using py.test
|
Python
|
bsd-3-clause
|
jmagnusson/wtforms,cklein/wtforms,pawl/wtforms,pawl/wtforms,subyraman/wtforms,crast/wtforms,Aaron1992/wtforms,Xender/wtforms,wtforms/wtforms,hsum/wtforms,Aaron1992/wtforms,skytreader/wtforms
|
---
+++
@@ -0,0 +1,60 @@
+"""
+ test_validators
+ ~~~~~~~~~~~~~~
+
+ Unittests for bundled validators.
+
+ :copyright: 2007-2008 by James Crasta, Thomas Johansson.
+ :license: MIT, see LICENSE.txt for details.
+"""
+
+from py.test import raises
+from wtforms.validators import ValidationError, length, url, not_empty, email, ip_address
+
+class DummyForm(object):
+ pass
+
+class DummyField(object):
+ def __init__(self, data):
+ self.data = data
+
+form = DummyForm()
+
+def test_email():
+ assert email(form, DummyField('foo@bar.dk')) == None
+ assert email(form, DummyField('123@bar.dk')) == None
+ assert email(form, DummyField('foo@456.dk')) == None
+ assert email(form, DummyField('foo@bar456.info')) == None
+ raises(ValidationError, email, form, DummyField('foo')) == None
+ raises(ValidationError, email, form, DummyField('bar.dk')) == None
+ raises(ValidationError, email, form, DummyField('foo@')) == None
+ raises(ValidationError, email, form, DummyField('@bar.dk')) == None
+ raises(ValidationError, email, form, DummyField('foo@bar')) == None
+ raises(ValidationError, email, form, DummyField('foo@bar.ab12')) == None
+ raises(ValidationError, email, form, DummyField('foo@bar.abcde')) == None
+
+def test_length():
+ field = DummyField('foobar')
+ assert length(min=2, max=6)(form, field) == None
+ raises(ValidationError, length(min=7), form, field)
+ raises(ValidationError, length(max=5), form, field)
+
+def test_url():
+ assert url()(form, DummyField('http://foobar.dk')) == None
+ assert url()(form, DummyField('http://foobar.dk/')) == None
+ assert url()(form, DummyField('http://foobar.dk/foobar')) == None
+ raises(ValidationError, url(), form, DummyField('http://foobar'))
+ raises(ValidationError, url(), form, DummyField('foobar.dk'))
+ raises(ValidationError, url(), form, DummyField('http://foobar.12'))
+
+def test_not_empty():
+ assert not_empty()(form, DummyField('foobar')) == None
+ raises(ValidationError, not_empty(), form, DummyField(''))
+ raises(ValidationError, not_empty(), form, DummyField(' '))
+
+def test_ip_address():
+ assert ip_address(form, DummyField('127.0.0.1')) == None
+ raises(ValidationError, ip_address, form, DummyField('abc.0.0.1'))
+ raises(ValidationError, ip_address, form, DummyField('1278.0.0.1'))
+ raises(ValidationError, ip_address, form, DummyField('127.0.0.abc'))
+
|
|
b189e7043115f788093b129815ff0bde5895ee0b
|
glanguage/__init__.py
|
glanguage/__init__.py
|
import httplib2
try:
import simplejson as json
except:
import json
from googleapiclient import discovery
from oauth2client.client import GoogleCredentials
OAUTH_SCOPE = 'https://www.googleapis.com/auth/cloud-platform'
DISCOVERY_URL = 'https://{api}.googleapis.com/$discovery/rest?version={apiVersion}'
class LanguageClient(object):
def __init__(self):
credentials = GoogleCredentials.get_application_default().create_scoped(OAUTH_SCOPE)
http = httplib2.Http()
credentials.authorize(http)
self.service = discovery.build('language','v1beta1',http=http,discoveryServiceUrl=DISCOVERY_URL)
def get_entities(self,document):
"""Get named entites in document."""
request_body={'document':{'type':'PLAIN_TEXT','content':document.encode('utf-8')},'encodingType':'UTF8'}
service_request = self.service.documents().analyzeEntities(body=request_body)
response_body = service_request.execute()
return response_body['entities']
def get_sentiment(self,document):
"""Get sentiment in document as polarity and magnitude."""
request_body={'document':{'type':'PLAIN_TEXT','content':document.encode('utf-8')}}
service_request = self.service.documents().analyzeSentiment(body=request_body)
response_body = service_request.execute()
return response_body['documentSentiment']
|
Add Natural Language API wrapper.
|
Add Natural Language API wrapper.
|
Python
|
apache-2.0
|
alexcchan/google
|
---
+++
@@ -0,0 +1,36 @@
+import httplib2
+try:
+ import simplejson as json
+except:
+ import json
+
+
+from googleapiclient import discovery
+from oauth2client.client import GoogleCredentials
+
+
+OAUTH_SCOPE = 'https://www.googleapis.com/auth/cloud-platform'
+DISCOVERY_URL = 'https://{api}.googleapis.com/$discovery/rest?version={apiVersion}'
+
+
+class LanguageClient(object):
+
+ def __init__(self):
+ credentials = GoogleCredentials.get_application_default().create_scoped(OAUTH_SCOPE)
+ http = httplib2.Http()
+ credentials.authorize(http)
+ self.service = discovery.build('language','v1beta1',http=http,discoveryServiceUrl=DISCOVERY_URL)
+
+ def get_entities(self,document):
+ """Get named entites in document."""
+ request_body={'document':{'type':'PLAIN_TEXT','content':document.encode('utf-8')},'encodingType':'UTF8'}
+ service_request = self.service.documents().analyzeEntities(body=request_body)
+ response_body = service_request.execute()
+ return response_body['entities']
+
+ def get_sentiment(self,document):
+ """Get sentiment in document as polarity and magnitude."""
+ request_body={'document':{'type':'PLAIN_TEXT','content':document.encode('utf-8')}}
+ service_request = self.service.documents().analyzeSentiment(body=request_body)
+ response_body = service_request.execute()
+ return response_body['documentSentiment']
|
|
809781cc832d79e1f746df385317ce4dae6223b3
|
tests/test_GetCalls.py
|
tests/test_GetCalls.py
|
#!/usr/bin/env python3
import codecs, json
from PokeFacts import RedditBot
from PokeFacts import DataPulls
def getDataPullsObject():
store = DataPulls.ItemStore({'term_property': 'term'})
with codecs.open('tests/test_data.json', "r", "utf-8") as data_file:
store.addItems(json.load(data_file))
return DataPulls.DataPulls(store=store)
def test_GetCalls():
main = RedditBot.CallResponse(reddit=False, data=getDataPullsObject())
calls = main.get_calls("{charizard} {charzard} { charizard }")
assert len(calls) == 1
assert calls[0].term == 'charizard'
|
Add unit test for get_calls
|
Add unit test for get_calls
|
Python
|
mit
|
rpokemon/PokeFacts
|
---
+++
@@ -0,0 +1,20 @@
+#!/usr/bin/env python3
+
+import codecs, json
+from PokeFacts import RedditBot
+from PokeFacts import DataPulls
+
+def getDataPullsObject():
+ store = DataPulls.ItemStore({'term_property': 'term'})
+
+ with codecs.open('tests/test_data.json', "r", "utf-8") as data_file:
+ store.addItems(json.load(data_file))
+
+ return DataPulls.DataPulls(store=store)
+
+def test_GetCalls():
+ main = RedditBot.CallResponse(reddit=False, data=getDataPullsObject())
+
+ calls = main.get_calls("{charizard} {charzard} { charizard }")
+ assert len(calls) == 1
+ assert calls[0].term == 'charizard'
|
|
47f4e738cc11ec40d3410332106163b0235f5da4
|
tests/python/tests/test_result.py
|
tests/python/tests/test_result.py
|
import unittest
import librepo
from librepo import LibrepoException
class TestCaseResult(unittest.TestCase):
def test_result_getinfo(self):
r = librepo.Result()
self.assertTrue(r)
self.assertRaises(ValueError, r.getinfo, 99999999)
self.assertFalse(r.getinfo(librepo.LRR_YUM_REPO))
self.assertFalse(r.getinfo(librepo.LRR_YUM_REPOMD))
self.assertRaises(LibrepoException, r.getinfo, librepo.LRR_YUM_TIMESTAMP)
def test_result_attrs(self):
r = librepo.Result()
self.assertTrue(r)
self.assertRaises(AttributeError, getattr, r, 'foobar_attr')
# Attrs should not be filled (that's why None or
# LibrepoException is expected), but they definitelly
# should exists (not AttributeError should be raised)
self.assertFalse(r.yum_repo)
self.assertFalse(r.yum_repomd)
self.assertRaises(LibrepoException, getattr, r, 'yum_timestamp')
|
Add tests for Result object
|
Tests: Add tests for Result object
|
Python
|
lgpl-2.1
|
Conan-Kudo/librepo,bgamari/librepo,Tojaj/librepo,rholy/librepo,rholy/librepo,cgwalters/librepo,rpm-software-management/librepo,Tojaj/librepo,Conan-Kudo/librepo,rholy/librepo,cgwalters/librepo,cgwalters/librepo,rholy/librepo,Conan-Kudo/librepo,cgwalters/librepo,rpm-software-management/librepo,Tojaj/librepo,rpm-software-management/librepo,bgamari/librepo,bgamari/librepo
|
---
+++
@@ -0,0 +1,27 @@
+import unittest
+import librepo
+from librepo import LibrepoException
+
+class TestCaseResult(unittest.TestCase):
+
+ def test_result_getinfo(self):
+ r = librepo.Result()
+ self.assertTrue(r)
+
+ self.assertRaises(ValueError, r.getinfo, 99999999)
+ self.assertFalse(r.getinfo(librepo.LRR_YUM_REPO))
+ self.assertFalse(r.getinfo(librepo.LRR_YUM_REPOMD))
+ self.assertRaises(LibrepoException, r.getinfo, librepo.LRR_YUM_TIMESTAMP)
+
+ def test_result_attrs(self):
+ r = librepo.Result()
+ self.assertTrue(r)
+
+ self.assertRaises(AttributeError, getattr, r, 'foobar_attr')
+
+ # Attrs should not be filled (that's why None or
+ # LibrepoException is expected), but they definitelly
+ # should exists (not AttributeError should be raised)
+ self.assertFalse(r.yum_repo)
+ self.assertFalse(r.yum_repomd)
+ self.assertRaises(LibrepoException, getattr, r, 'yum_timestamp')
|
|
4736ed07ea8b83ca8c32c2d675f67883050b8c26
|
tests/test_provider_lawrenceks.py
|
tests/test_provider_lawrenceks.py
|
import busbus
from busbus.provider.lawrenceks import LawrenceTransitProvider
import arrow
import pytest
@pytest.fixture(scope='module')
def lawrenceks_provider():
return LawrenceTransitProvider()
def test_43_to_eaton_hall(lawrenceks_provider):
stop = lawrenceks_provider.get(busbus.Stop, u'15TH_SPAHR_WB')
route = lawrenceks_provider.get(busbus.Route, u'RT_43')
assert len(list(lawrenceks_provider.arrivals.where(
stop=stop, route=route,
start_time=arrow.get('2015-03-10T14:00:00-05:00'),
end_time=arrow.get('2015-03-10T16:00:00-05:00')))) == 13
|
Add a simple test case for LawrenceTransitProvider
|
Add a simple test case for LawrenceTransitProvider
|
Python
|
mit
|
spaceboats/busbus
|
---
+++
@@ -0,0 +1,19 @@
+import busbus
+from busbus.provider.lawrenceks import LawrenceTransitProvider
+
+import arrow
+import pytest
+
+
+@pytest.fixture(scope='module')
+def lawrenceks_provider():
+ return LawrenceTransitProvider()
+
+
+def test_43_to_eaton_hall(lawrenceks_provider):
+ stop = lawrenceks_provider.get(busbus.Stop, u'15TH_SPAHR_WB')
+ route = lawrenceks_provider.get(busbus.Route, u'RT_43')
+ assert len(list(lawrenceks_provider.arrivals.where(
+ stop=stop, route=route,
+ start_time=arrow.get('2015-03-10T14:00:00-05:00'),
+ end_time=arrow.get('2015-03-10T16:00:00-05:00')))) == 13
|
|
15d782aaddf1e8a4215df2fa3ef60b8801fe382a
|
tests_tf/test_utils.py
|
tests_tf/test_utils.py
|
from __future__ import absolute_import, division, print_function
import unittest
import numpy as np
from cleverhans.utils_tf import kl_with_logits, l2_batch_normalize
def numpy_kl_with_logits(q_logits, p_logits):
def numpy_softmax(logits):
exp_logits = np.exp(logits)
return exp_logits / np.sum(exp_logits, axis=1, keepdims=True)
q = numpy_softmax(q_logits)
p = numpy_softmax(p_logits)
return (q * (np.log(q) - np.log(p))).sum(axis=1).mean()
class TestUtils(unittest.TestCase):
def test_l2_batch_normalize(self):
import tensorflow as tf
with tf.Session() as sess:
x = tf.random_normal((100, 1000))
x_norm = sess.run(l2_batch_normalize(x))
self.assertTrue(
np.allclose(np.sum(x_norm**2, axis=1), 1, atol=1e-6))
def test_kl_with_logits(self):
import tensorflow as tf
q_logits = tf.placeholder(tf.float32, shape=(100, 20))
p_logits = tf.placeholder(tf.float32, shape=(100, 20))
q_logits_np = np.random.normal(0, 10, size=(100, 20))
p_logits_np = np.random.normal(0, 10, size=(100, 20))
with tf.Session() as sess:
kl_div_tf = sess.run(kl_with_logits(q_logits, p_logits),
feed_dict={q_logits: q_logits_np,
p_logits: p_logits_np})
kl_div_ref = numpy_kl_with_logits(q_logits_np, p_logits_np)
self.assertTrue(np.allclose(kl_div_ref, kl_div_tf))
if __name__ == '__main__':
unittest.main()
|
Add tests for vat utils functions
|
Add tests for vat utils functions
|
Python
|
mit
|
fartashf/cleverhans,carlini/cleverhans,openai/cleverhans,cleverhans-lab/cleverhans,cleverhans-lab/cleverhans,cihangxie/cleverhans,carlini/cleverhans,cleverhans-lab/cleverhans
|
---
+++
@@ -0,0 +1,44 @@
+from __future__ import absolute_import, division, print_function
+
+import unittest
+
+import numpy as np
+
+from cleverhans.utils_tf import kl_with_logits, l2_batch_normalize
+
+
+def numpy_kl_with_logits(q_logits, p_logits):
+ def numpy_softmax(logits):
+ exp_logits = np.exp(logits)
+ return exp_logits / np.sum(exp_logits, axis=1, keepdims=True)
+
+ q = numpy_softmax(q_logits)
+ p = numpy_softmax(p_logits)
+ return (q * (np.log(q) - np.log(p))).sum(axis=1).mean()
+
+
+class TestUtils(unittest.TestCase):
+ def test_l2_batch_normalize(self):
+ import tensorflow as tf
+ with tf.Session() as sess:
+ x = tf.random_normal((100, 1000))
+ x_norm = sess.run(l2_batch_normalize(x))
+ self.assertTrue(
+ np.allclose(np.sum(x_norm**2, axis=1), 1, atol=1e-6))
+
+ def test_kl_with_logits(self):
+ import tensorflow as tf
+ q_logits = tf.placeholder(tf.float32, shape=(100, 20))
+ p_logits = tf.placeholder(tf.float32, shape=(100, 20))
+ q_logits_np = np.random.normal(0, 10, size=(100, 20))
+ p_logits_np = np.random.normal(0, 10, size=(100, 20))
+ with tf.Session() as sess:
+ kl_div_tf = sess.run(kl_with_logits(q_logits, p_logits),
+ feed_dict={q_logits: q_logits_np,
+ p_logits: p_logits_np})
+ kl_div_ref = numpy_kl_with_logits(q_logits_np, p_logits_np)
+ self.assertTrue(np.allclose(kl_div_ref, kl_div_tf))
+
+
+if __name__ == '__main__':
+ unittest.main()
|
|
810eeddaff32f9b608b0b61cfcb48826ec1b15bf
|
various/Crop_Big_ROIs.py
|
various/Crop_Big_ROIs.py
|
# @DatasetService datasetservice
# @ImageDisplayService displayservice
# @ImageJ ij
# @AbstractLogService log
# @DefaultLegacyService legacyservice
from ij import IJ
from ij import Macro
from ij.plugin.frame import RoiManager
from io.scif.img import ImgSaver
from net.imagej import DefaultDataset
from loci.plugins import BF
from loci.plugins import LociExporter
from loci.plugins.out import Exporter
from loci.plugins.in import ImporterOptions
from loci.common import Region
import os
import sys
import glob
sys.path.append(os.path.join(IJ.getDirectory('plugins'), "Scripts", "Plugins"))
from libtools import crop
from libtools.utils import get_dt
def main():
# Get image path
fname = "/home/hadim/local/data/microscopy_data/zurick/movies/metC.ome.tif"
basename = os.path.basename(fname)
dir_path = os.path.dirname(fname)
if not fname:
IJ.showMessage('Source image needs to match a file on the system.')
return
# Open ROIs
roi_path = os.path.join(dir_path, "RoiSet.zip")
if not os.path.isfile(roi_path):
try:
roi_path = glob.glob(os.path.join(dir_path, "*.roi"))[0]
except:
roi_path = None
if not roi_path:
IJ.showMessage('No ROIs. Please use Analyze > Tools > ROI Manager...')
return
rois = RoiManager(True)
rois.reset()
rois.runCommand("Open", roi_path)
IJ.log('Image filename is %s' % fname)
rois_array = rois.getRoisAsArray()
for i, roi in enumerate(rois_array):
crop_id = i +1
IJ.log("Open cropped region %i / %i" % (crop_id, len(rois_array)))
# Get filename and basename of the current cropped image
crop_basename = "crop%i_%s" % (crop_id, basename)
crop_fname = os.path.join(os.path.dirname(fname), crop_basename)
# Get bounds and crop
bounds = roi.getBounds()
x = bounds.x
y = bounds.y
w = bounds.width
h = bounds.height
# Import only cropped region of the image
options = ImporterOptions()
options.setCrop(True)
options.setCropRegion(0, Region(x, y, w, h))
options.setId(fname)
#options.setVirtual(True)
imps = BF.openImagePlus(options)
imp = imps[0]
# Save cropped image
bfExporter = LociExporter()
macroOpts = "save=[" + crop_fname + "]"
bfExporter.setup(None, imp)
Macro.setOptions(macroOpts)
bfExporter.run(None)
imp.close()
IJ.log('Done')
main()
|
Add crop multi roi alternative script
|
Add crop multi roi alternative script
|
Python
|
bsd-3-clause
|
hadim/fiji_tools,hadim/fiji_scripts,hadim/fiji_scripts,hadim/fiji_scripts,hadim/fiji_tools
|
---
+++
@@ -0,0 +1,97 @@
+# @DatasetService datasetservice
+# @ImageDisplayService displayservice
+# @ImageJ ij
+# @AbstractLogService log
+# @DefaultLegacyService legacyservice
+
+from ij import IJ
+from ij import Macro
+from ij.plugin.frame import RoiManager
+
+from io.scif.img import ImgSaver
+from net.imagej import DefaultDataset
+
+from loci.plugins import BF
+from loci.plugins import LociExporter
+from loci.plugins.out import Exporter
+from loci.plugins.in import ImporterOptions
+from loci.common import Region
+
+import os
+import sys
+import glob
+
+sys.path.append(os.path.join(IJ.getDirectory('plugins'), "Scripts", "Plugins"))
+from libtools import crop
+from libtools.utils import get_dt
+
+def main():
+
+ # Get image path
+ fname = "/home/hadim/local/data/microscopy_data/zurick/movies/metC.ome.tif"
+
+ basename = os.path.basename(fname)
+ dir_path = os.path.dirname(fname)
+
+ if not fname:
+ IJ.showMessage('Source image needs to match a file on the system.')
+ return
+
+ # Open ROIs
+ roi_path = os.path.join(dir_path, "RoiSet.zip")
+ if not os.path.isfile(roi_path):
+ try:
+ roi_path = glob.glob(os.path.join(dir_path, "*.roi"))[0]
+ except:
+ roi_path = None
+
+ if not roi_path:
+ IJ.showMessage('No ROIs. Please use Analyze > Tools > ROI Manager...')
+ return
+
+ rois = RoiManager(True)
+ rois.reset()
+ rois.runCommand("Open", roi_path)
+
+ IJ.log('Image filename is %s' % fname)
+
+ rois_array = rois.getRoisAsArray()
+ for i, roi in enumerate(rois_array):
+
+ crop_id = i +1
+ IJ.log("Open cropped region %i / %i" % (crop_id, len(rois_array)))
+
+ # Get filename and basename of the current cropped image
+ crop_basename = "crop%i_%s" % (crop_id, basename)
+ crop_fname = os.path.join(os.path.dirname(fname), crop_basename)
+
+ # Get bounds and crop
+ bounds = roi.getBounds()
+
+ x = bounds.x
+ y = bounds.y
+ w = bounds.width
+ h = bounds.height
+
+ # Import only cropped region of the image
+ options = ImporterOptions()
+ options.setCrop(True)
+ options.setCropRegion(0, Region(x, y, w, h))
+ options.setId(fname)
+ #options.setVirtual(True)
+ imps = BF.openImagePlus(options)
+
+ imp = imps[0]
+
+ # Save cropped image
+ bfExporter = LociExporter()
+ macroOpts = "save=[" + crop_fname + "]"
+ bfExporter.setup(None, imp)
+ Macro.setOptions(macroOpts)
+ bfExporter.run(None)
+
+ imp.close()
+
+ IJ.log('Done')
+
+main()
|
|
8f1e94f79ddd398112ed33485bfc3d735e1edda2
|
maint/scripts/download_wheels.py
|
maint/scripts/download_wheels.py
|
#!/usr/bin/env python3
import asyncio
import json
import pathlib
import sys
from tornado.httpclient import AsyncHTTPClient
BASE_URL = "https://ci.appveyor.com/api"
async def fetch_job(directory, job):
http = AsyncHTTPClient()
artifacts = await http.fetch(f"{BASE_URL}/buildjobs/{job}/artifacts")
paths = [pathlib.PurePosixPath(a["fileName"]) for a in json.loads(artifacts.body)]
for path in paths:
artifact = await http.fetch(f"{BASE_URL}/buildjobs/{job}/artifacts/{path}")
with open(directory.joinpath(path.name), "wb") as f:
f.write(artifact.body)
async def main():
http = AsyncHTTPClient()
try:
_, version = sys.argv
except ValueError:
print("usage: maint/scripts/download_wheels.py v6.0.1", file=sys.stderr)
sys.exit(1)
directory = pathlib.Path(f"downloads-{version}")
directory.mkdir(exist_ok=True)
build = await http.fetch(f"{BASE_URL}/projects/bdarnell/tornado/branch/{version}")
jobs = [job["jobId"] for job in json.loads(build.body)["build"]["jobs"]]
await asyncio.gather(*(fetch_job(directory, job) for job in jobs))
if __name__ == "__main__":
asyncio.run(main())
|
Add script to download wheels from appveyor
|
Add script to download wheels from appveyor
|
Python
|
apache-2.0
|
mivade/tornado,tornadoweb/tornado,bdarnell/tornado,mivade/tornado,dongpinglai/my_tornado,mivade/tornado,dongpinglai/my_tornado,tornadoweb/tornado,NoyaInRain/tornado,NoyaInRain/tornado,NoyaInRain/tornado,allenl203/tornado,bdarnell/tornado,dongpinglai/my_tornado,lilydjwg/tornado,tornadoweb/tornado,dongpinglai/my_tornado,lilydjwg/tornado,allenl203/tornado,wujuguang/tornado,allenl203/tornado,lilydjwg/tornado,lilydjwg/tornado,wujuguang/tornado,bdarnell/tornado,bdarnell/tornado,wujuguang/tornado,mivade/tornado,mivade/tornado,NoyaInRain/tornado,wujuguang/tornado,NoyaInRain/tornado,bdarnell/tornado,allenl203/tornado,dongpinglai/my_tornado,dongpinglai/my_tornado,NoyaInRain/tornado,allenl203/tornado,tornadoweb/tornado,wujuguang/tornado
|
---
+++
@@ -0,0 +1,41 @@
+#!/usr/bin/env python3
+
+import asyncio
+import json
+import pathlib
+import sys
+from tornado.httpclient import AsyncHTTPClient
+
+BASE_URL = "https://ci.appveyor.com/api"
+
+
+async def fetch_job(directory, job):
+ http = AsyncHTTPClient()
+ artifacts = await http.fetch(f"{BASE_URL}/buildjobs/{job}/artifacts")
+ paths = [pathlib.PurePosixPath(a["fileName"]) for a in json.loads(artifacts.body)]
+
+ for path in paths:
+ artifact = await http.fetch(f"{BASE_URL}/buildjobs/{job}/artifacts/{path}")
+ with open(directory.joinpath(path.name), "wb") as f:
+ f.write(artifact.body)
+
+
+async def main():
+ http = AsyncHTTPClient()
+ try:
+ _, version = sys.argv
+ except ValueError:
+ print("usage: maint/scripts/download_wheels.py v6.0.1", file=sys.stderr)
+ sys.exit(1)
+
+ directory = pathlib.Path(f"downloads-{version}")
+ directory.mkdir(exist_ok=True)
+
+ build = await http.fetch(f"{BASE_URL}/projects/bdarnell/tornado/branch/{version}")
+ jobs = [job["jobId"] for job in json.loads(build.body)["build"]["jobs"]]
+
+ await asyncio.gather(*(fetch_job(directory, job) for job in jobs))
+
+
+if __name__ == "__main__":
+ asyncio.run(main())
|
|
1846e6fe7f6b6a31f7921303556393e7f6fd9845
|
dev_tools/src/d1_dev/src-print-redbaron-tree.py
|
dev_tools/src/d1_dev/src-print-redbaron-tree.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# This work was created by participants in the DataONE project, and is
# jointly copyrighted by participating institutions in DataONE. For
# more information on DataONE, see our web site at http://dataone.org.
#
# Copyright 2009-2016 DataONE
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import argparse
import sys
import d1_dev.util
def main():
"""Print the RedBaron syntax tree for a Python module
"""
parser = argparse.ArgumentParser(
description=__doc__, formatter_class=argparse.RawDescriptionHelpFormatter
)
parser.add_argument("path", help="Python module path")
args = parser.parse_args()
r = d1_dev.util.redbaron_module_path_to_tree(args.path)
print(r.help(True))
if __name__ == "__main__":
sys.exit(main())
|
Add command to print the syntax tree for a script
|
Add command to print the syntax tree for a script
|
Python
|
apache-2.0
|
DataONEorg/d1_python,DataONEorg/d1_python,DataONEorg/d1_python,DataONEorg/d1_python
|
---
+++
@@ -0,0 +1,41 @@
+#!/usr/bin/env python
+# -*- coding: utf-8 -*-
+
+# This work was created by participants in the DataONE project, and is
+# jointly copyrighted by participating institutions in DataONE. For
+# more information on DataONE, see our web site at http://dataone.org.
+#
+# Copyright 2009-2016 DataONE
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import argparse
+import sys
+
+import d1_dev.util
+
+
+def main():
+ """Print the RedBaron syntax tree for a Python module
+ """
+ parser = argparse.ArgumentParser(
+ description=__doc__, formatter_class=argparse.RawDescriptionHelpFormatter
+ )
+ parser.add_argument("path", help="Python module path")
+ args = parser.parse_args()
+
+ r = d1_dev.util.redbaron_module_path_to_tree(args.path)
+ print(r.help(True))
+
+if __name__ == "__main__":
+ sys.exit(main())
|
|
e9f30ec92520f8caa4f5d08fdf43b08ced84fd6b
|
CodeFights/leastFactorial.py
|
CodeFights/leastFactorial.py
|
#!/usr/local/bin/python
# Code Fights Least Factorial (Core) Problem
def leastFactorial(n):
def factGen():
m, res = 1, 1
while True:
res *= m
yield res
m += 1
for f in factGen():
if f >= n:
return f
def main():
tests = [
[17, 24],
[1, 1],
[5, 6]
]
for t in tests:
res = leastFactorial(t[0])
if t[1] == res:
print("PASSED: leastFactorial({}) returned {}"
.format(t[0], res))
else:
print(("FAILED: leastFactorial({}) returned {},"
"answer: {}").format(t[0], res, t[1]))
if __name__ == '__main__':
main()
|
Solve Code Fights least factorial problem
|
Solve Code Fights least factorial problem
|
Python
|
mit
|
HKuz/Test_Code
|
---
+++
@@ -0,0 +1,36 @@
+#!/usr/local/bin/python
+# Code Fights Least Factorial (Core) Problem
+
+
+def leastFactorial(n):
+ def factGen():
+ m, res = 1, 1
+ while True:
+ res *= m
+ yield res
+ m += 1
+
+ for f in factGen():
+ if f >= n:
+ return f
+
+
+def main():
+ tests = [
+ [17, 24],
+ [1, 1],
+ [5, 6]
+ ]
+
+ for t in tests:
+ res = leastFactorial(t[0])
+ if t[1] == res:
+ print("PASSED: leastFactorial({}) returned {}"
+ .format(t[0], res))
+ else:
+ print(("FAILED: leastFactorial({}) returned {},"
+ "answer: {}").format(t[0], res, t[1]))
+
+
+if __name__ == '__main__':
+ main()
|
|
77b6e4995743bca4036e5b5dc498cfcdd4e2908e
|
jarviscli/tests/test_wifi_password_getter.py
|
jarviscli/tests/test_wifi_password_getter.py
|
import unittest
from tests import PluginTest
from plugins import wifi_password_getter
from colorama import Fore
class TestWifiPasswordGetter(PluginTest):
"""
A test class that contains test cases for the methods of
the wifi_password_getter plugin for Windows.
"""
def setUp(self):
self.test = self.load_plugin(
wifi_password_getter.WifiPasswordGetterWINDOWS)
def test_show_options_last_text(self):
self.queue_input("2")
profiles = ["profile_1", "profile_2", "profile_3"]
self.test.show_options(self.jarvis_api, profiles)
self.assertEqual(self.history_say().last_text(), "4: Exit")
def test_get_choice_valid(self):
self.queue_input("2")
input_text = "Please select a number or Exit: "
max_valid_value = 3
self.assertEqual(
self.test.get_choice(
self.jarvis_api,
input_text,
max_valid_value),
2)
def test_get_choice_terminator(self):
self.queue_input("3")
input_text = "Please select a number or Exit: "
max_valid_value = 3
self.assertEqual(
self.test.get_choice(
self.jarvis_api, input_text, max_valid_value), -1)
def test_get_choice_invalid(self):
self.queue_input("7")
self.queue_input("2")
input_text = "Please select a number or Exit: "
max_valid_value = 3
self.test.get_choice(self.jarvis_api, input_text, max_valid_value)
self.assertEqual(
self.history_say().last_text(),
"Invalid input! Enter a number from the choices provided.")
self.assertEqual(self.history_say().last_color(), Fore.YELLOW)
def test_get_choice_exception(self):
self.queue_input("wrong_input")
self.queue_input("2")
input_text = "Please select a number or Exit: "
max_valid_value = 3
self.test.get_choice(self.jarvis_api, input_text, max_valid_value)
self.assertEqual(
self.history_say().last_text(),
"Invalid input! Enter a number from the choices provided.")
self.assertEqual(self.history_say().last_color(), Fore.YELLOW)
if __name__ == '__main__':
unittest.main()
|
Create test cases for the wifi_password_getter plugin
|
Create test cases for the wifi_password_getter plugin
|
Python
|
mit
|
sukeesh/Jarvis,sukeesh/Jarvis,sukeesh/Jarvis,sukeesh/Jarvis
|
---
+++
@@ -0,0 +1,66 @@
+import unittest
+from tests import PluginTest
+from plugins import wifi_password_getter
+from colorama import Fore
+
+
+class TestWifiPasswordGetter(PluginTest):
+ """
+ A test class that contains test cases for the methods of
+ the wifi_password_getter plugin for Windows.
+ """
+
+ def setUp(self):
+ self.test = self.load_plugin(
+ wifi_password_getter.WifiPasswordGetterWINDOWS)
+
+ def test_show_options_last_text(self):
+ self.queue_input("2")
+ profiles = ["profile_1", "profile_2", "profile_3"]
+ self.test.show_options(self.jarvis_api, profiles)
+ self.assertEqual(self.history_say().last_text(), "4: Exit")
+
+ def test_get_choice_valid(self):
+ self.queue_input("2")
+ input_text = "Please select a number or Exit: "
+ max_valid_value = 3
+ self.assertEqual(
+ self.test.get_choice(
+ self.jarvis_api,
+ input_text,
+ max_valid_value),
+ 2)
+
+ def test_get_choice_terminator(self):
+ self.queue_input("3")
+ input_text = "Please select a number or Exit: "
+ max_valid_value = 3
+ self.assertEqual(
+ self.test.get_choice(
+ self.jarvis_api, input_text, max_valid_value), -1)
+
+ def test_get_choice_invalid(self):
+ self.queue_input("7")
+ self.queue_input("2")
+ input_text = "Please select a number or Exit: "
+ max_valid_value = 3
+ self.test.get_choice(self.jarvis_api, input_text, max_valid_value)
+ self.assertEqual(
+ self.history_say().last_text(),
+ "Invalid input! Enter a number from the choices provided.")
+ self.assertEqual(self.history_say().last_color(), Fore.YELLOW)
+
+ def test_get_choice_exception(self):
+ self.queue_input("wrong_input")
+ self.queue_input("2")
+ input_text = "Please select a number or Exit: "
+ max_valid_value = 3
+ self.test.get_choice(self.jarvis_api, input_text, max_valid_value)
+ self.assertEqual(
+ self.history_say().last_text(),
+ "Invalid input! Enter a number from the choices provided.")
+ self.assertEqual(self.history_say().last_color(), Fore.YELLOW)
+
+
+if __name__ == '__main__':
+ unittest.main()
|
|
374f32a1d5feaf2e912d901b9398f50f00e7d481
|
scripts/most_recent.py
|
scripts/most_recent.py
|
from datetime import datetime
from optparse import OptionParser
from urllib2 import urlopen
from BeautifulSoup import BeautifulSoup
if __name__ == '__main__':
usage = "%prog <USERNAME> <WEB SERVER>"
parser = OptionParser(usage=usage)
opts,args = parser.parse_args()
if len(args) != 2:
parser.error("All arguments are required.")
username,server = args
if server[:7] != "http://":
server = "http://%s" % (server,)
soup = BeautifulSoup(urlopen('%s/user/%s/recent-tracks'
% (server, username)))
gobbles_list = soup.find('ul', {'class': 'gobbles'})
timestamp = gobbles_list.find('li')['about'].split('#')[1].split('.')[0]
print datetime.fromtimestamp(float(timestamp))
|
Add script to print time of most recent gobble.
|
Add script to print time of most recent gobble.
|
Python
|
agpl-3.0
|
foocorp/gnu-fm,foocorp/gnu-fm,foocorp/gnu-fm,foocorp/gnu-fm,foocorp/gnu-fm,foocorp/gnu-fm,foocorp/gnu-fm,foocorp/gnu-fm,foocorp/gnu-fm
|
---
+++
@@ -0,0 +1,22 @@
+from datetime import datetime
+from optparse import OptionParser
+from urllib2 import urlopen
+
+from BeautifulSoup import BeautifulSoup
+
+
+if __name__ == '__main__':
+ usage = "%prog <USERNAME> <WEB SERVER>"
+ parser = OptionParser(usage=usage)
+ opts,args = parser.parse_args()
+ if len(args) != 2:
+ parser.error("All arguments are required.")
+
+ username,server = args
+ if server[:7] != "http://":
+ server = "http://%s" % (server,)
+ soup = BeautifulSoup(urlopen('%s/user/%s/recent-tracks'
+ % (server, username)))
+ gobbles_list = soup.find('ul', {'class': 'gobbles'})
+ timestamp = gobbles_list.find('li')['about'].split('#')[1].split('.')[0]
+ print datetime.fromtimestamp(float(timestamp))
|
|
b47dbd6b6f2e19632e90036f14cd85bbf3f8cbd1
|
utils/get_collection_object_count.py
|
utils/get_collection_object_count.py
|
#!/usr/bin/env python
# -*- coding: utf8 -*-
import sys, os
import argparse
from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo
def main(argv=None):
parser = argparse.ArgumentParser(description='Print count of objects for a given collection.')
parser.add_argument('path', help="Nuxeo path to collection")
parser.add_argument('--pynuxrc', default='~/.pynuxrc-prod', help="rcfile for use with pynux utils")
if argv is None:
argv = parser.parse_args()
dh = DeepHarvestNuxeo(argv.path, 'barbarahui_test_bucket', argv.pynuxrc)
print "about to fetch objects for path {}".format(dh.path)
objects = dh.fetch_objects()
print "finished"
print "len(objects): {}".format(len(objects))
if __name__ == "__main__":
sys.exit(main())
|
Add script to get count of objects in a collection.
|
Add script to get count of objects in a collection.
|
Python
|
bsd-3-clause
|
barbarahui/nuxeo-calisphere,barbarahui/nuxeo-calisphere
|
---
+++
@@ -0,0 +1,23 @@
+#!/usr/bin/env python
+# -*- coding: utf8 -*-
+
+import sys, os
+import argparse
+from deepharvest.deepharvest_nuxeo import DeepHarvestNuxeo
+
+def main(argv=None):
+
+ parser = argparse.ArgumentParser(description='Print count of objects for a given collection.')
+ parser.add_argument('path', help="Nuxeo path to collection")
+ parser.add_argument('--pynuxrc', default='~/.pynuxrc-prod', help="rcfile for use with pynux utils")
+ if argv is None:
+ argv = parser.parse_args()
+
+ dh = DeepHarvestNuxeo(argv.path, 'barbarahui_test_bucket', argv.pynuxrc)
+ print "about to fetch objects for path {}".format(dh.path)
+ objects = dh.fetch_objects()
+ print "finished"
+ print "len(objects): {}".format(len(objects))
+
+if __name__ == "__main__":
+ sys.exit(main())
|
|
0798e457957b3db8f5de1891900d639961d78a0f
|
emgapimetadata/management/commands/test-data.py
|
emgapimetadata/management/commands/test-data.py
|
#!/usr/bin/python
# -*- coding: utf-8 -*-
import os
import csv
from django.core.management.base import BaseCommand
from emgapimetadata import models as m_models
class Command(BaseCommand):
def add_arguments(self, parser):
parser.add_argument('importpath', type=str)
def handle(self, *args, **options):
self.populate(options)
def populate(self, options):
# check if path is valid
_path = options.get('importpath', None)
if os.path.exists(_path):
if os.path.isdir(_path):
for root, dirs, files in os.walk(_path, topdown=False):
for name in files:
accession = name.split("_")[0]
f = os.path.join(root, name)
if name.endswith("go"):
self.import_go(f, accession)
# TODO: is file get dir:
elif os.path.isfile(_path):
raise NotImplemented("Give path to directory.")
else:
raise NotImplemented("Path doesn't exist.")
def import_go(self, f, accession):
with open(f, newline='') as fcsv:
reader = csv.reader(fcsv)
run = m_models.Run()
run.accession = "ERR700147"
run.pipeline_version = "1.0"
for row in reader:
try:
ann = m_models.Annotation(
accession=row[0],
description=row[1],
lineage=row[2],
).save()
except:
ann = m_models.Annotation.objects.get(accession=row[0])
rann = m_models.RunAnnotation()
rann.count = row[3]
rann.annotation = ann
run.annotations.append(rann)
# ranns = m_models.RunAnnotation.objects.insert(ranns)
run.save()
|
Add command line tool to import metadata
|
Add command line tool to import metadata
|
Python
|
apache-2.0
|
EBI-Metagenomics/emgapi,EBI-Metagenomics/emgapi,EBI-Metagenomics/emgapi,EBI-Metagenomics/emgapi,EBI-Metagenomics/emgapi
|
---
+++
@@ -0,0 +1,57 @@
+#!/usr/bin/python
+# -*- coding: utf-8 -*-
+
+import os
+import csv
+
+from django.core.management.base import BaseCommand
+
+from emgapimetadata import models as m_models
+
+
+class Command(BaseCommand):
+
+ def add_arguments(self, parser):
+ parser.add_argument('importpath', type=str)
+
+ def handle(self, *args, **options):
+ self.populate(options)
+
+ def populate(self, options):
+ # check if path is valid
+ _path = options.get('importpath', None)
+ if os.path.exists(_path):
+ if os.path.isdir(_path):
+ for root, dirs, files in os.walk(_path, topdown=False):
+ for name in files:
+ accession = name.split("_")[0]
+ f = os.path.join(root, name)
+ if name.endswith("go"):
+ self.import_go(f, accession)
+ # TODO: is file get dir:
+ elif os.path.isfile(_path):
+ raise NotImplemented("Give path to directory.")
+ else:
+ raise NotImplemented("Path doesn't exist.")
+
+ def import_go(self, f, accession):
+ with open(f, newline='') as fcsv:
+ reader = csv.reader(fcsv)
+ run = m_models.Run()
+ run.accession = "ERR700147"
+ run.pipeline_version = "1.0"
+ for row in reader:
+ try:
+ ann = m_models.Annotation(
+ accession=row[0],
+ description=row[1],
+ lineage=row[2],
+ ).save()
+ except:
+ ann = m_models.Annotation.objects.get(accession=row[0])
+ rann = m_models.RunAnnotation()
+ rann.count = row[3]
+ rann.annotation = ann
+ run.annotations.append(rann)
+ # ranns = m_models.RunAnnotation.objects.insert(ranns)
+ run.save()
|
|
b3f185033ee758e9407240243e263e07c8a28e35
|
services/imu-logger.py
|
services/imu-logger.py
|
#!/usr/bin/env python3
from sense_hat import SenseHat
from pymongo import MongoClient
import time
DELAY = 1 # in seconds
sense = SenseHat()
client = MongoClient("mongodb://192.168.0.128:27017")
db = client.g2x
while True:
orientation = sense.get_orientation_degrees()
print(orientation)
acceleration = sense.get_accelerometer()
compass = sense.get_compass()
temperature_from_humidity = sense.get_temperature()
temperature_from_pressure = sense.get_temperature_from_pressure()
db.gyroscope.insert_one({
"pitch": orientation["pitch"],
"roll": orientation["roll"],
"yaw": orientation["yaw"]
})
db.accelerometer.insert_one({
"pitch": acceleration["pitch"],
"roll": acceleration["roll"],
"yaw": acceleration["yaw"]
})
db.compass.insert_one({"angle": compass})
db.temperature.insert_one({
"from_humidity": temperature_from_humidity,
"from_pressure": temperature_from_pressure
})
time.sleep(DELAY)
|
Create script to log imu values
|
Create script to log imu values
|
Python
|
bsd-3-clause
|
gizmo-cda/g2x-submarine-v2,gizmo-cda/g2x-submarine-v2,gizmo-cda/g2x-submarine-v2,gizmo-cda/g2x-submarine-v2
|
---
+++
@@ -0,0 +1,38 @@
+#!/usr/bin/env python3
+
+from sense_hat import SenseHat
+from pymongo import MongoClient
+import time
+
+
+DELAY = 1 # in seconds
+
+sense = SenseHat()
+client = MongoClient("mongodb://192.168.0.128:27017")
+db = client.g2x
+
+while True:
+ orientation = sense.get_orientation_degrees()
+ print(orientation)
+ acceleration = sense.get_accelerometer()
+ compass = sense.get_compass()
+ temperature_from_humidity = sense.get_temperature()
+ temperature_from_pressure = sense.get_temperature_from_pressure()
+
+ db.gyroscope.insert_one({
+ "pitch": orientation["pitch"],
+ "roll": orientation["roll"],
+ "yaw": orientation["yaw"]
+ })
+ db.accelerometer.insert_one({
+ "pitch": acceleration["pitch"],
+ "roll": acceleration["roll"],
+ "yaw": acceleration["yaw"]
+ })
+ db.compass.insert_one({"angle": compass})
+ db.temperature.insert_one({
+ "from_humidity": temperature_from_humidity,
+ "from_pressure": temperature_from_pressure
+ })
+
+ time.sleep(DELAY)
|
|
a24b3122176b7435469b5275264dd6f53ff78165
|
demo_a3c_continuous.py
|
demo_a3c_continuous.py
|
import argparse
import chainer
from chainer import serializers
import gym
import numpy as np
import random_seed
from train_a3c_continuous import phi, A3CLSTMGaussian
import env_modifiers
def eval_single_run(env, model, phi):
model.reset_state()
test_r = 0
obs = env.reset()
done = False
while not done:
s = chainer.Variable(np.expand_dims(phi(obs), 0))
pout = model.pi_and_v(s)[0]
model.unchain_backward()
a = pout.sampled_actions.data[0]
obs, r, done, info = env.step(a)
test_r += r
return test_r
def main():
import logging
logging.basicConfig(level=logging.DEBUG)
parser = argparse.ArgumentParser()
parser.add_argument('model', type=str)
parser.add_argument('--env', type=str, default='Pendulum-v0')
parser.add_argument('--seed', type=int, default=0)
parser.add_argument('--n-runs', type=int, default=10)
parser.add_argument('--window-visible', action='store_true')
parser.add_argument('--render', action='store_true')
parser.set_defaults(render=False)
args = parser.parse_args()
random_seed.set_random_seed(args.seed)
env = gym.make(args.env)
timestep_limit = env.spec.timestep_limit
env_modifiers.make_timestep_limited(env, timestep_limit)
if args.render:
env_modifiers.make_rendered(env)
obs_size = np.asarray(env.observation_space.shape).prod()
action_size = np.asarray(env.action_space.shape).prod()
model = A3CLSTMGaussian(obs_size, action_size)
serializers.load_hdf5(args.model, model)
scores = []
for i in range(args.n_runs):
score = eval_single_run(env, model, phi)
print('Run {}: {}'.format(i, score))
scores.append(score)
print('Average: {}'.format(sum(scores) / args.n_runs))
if __name__ == '__main__':
main()
|
Add a demo script for gym continous tasks
|
Add a demo script for gym continous tasks
|
Python
|
mit
|
toslunar/chainerrl,toslunar/chainerrl
|
---
+++
@@ -0,0 +1,66 @@
+import argparse
+
+import chainer
+from chainer import serializers
+import gym
+import numpy as np
+
+import random_seed
+from train_a3c_continuous import phi, A3CLSTMGaussian
+import env_modifiers
+
+
+def eval_single_run(env, model, phi):
+ model.reset_state()
+ test_r = 0
+ obs = env.reset()
+ done = False
+ while not done:
+ s = chainer.Variable(np.expand_dims(phi(obs), 0))
+ pout = model.pi_and_v(s)[0]
+ model.unchain_backward()
+ a = pout.sampled_actions.data[0]
+ obs, r, done, info = env.step(a)
+ test_r += r
+ return test_r
+
+
+def main():
+ import logging
+ logging.basicConfig(level=logging.DEBUG)
+
+ parser = argparse.ArgumentParser()
+ parser.add_argument('model', type=str)
+ parser.add_argument('--env', type=str, default='Pendulum-v0')
+ parser.add_argument('--seed', type=int, default=0)
+ parser.add_argument('--n-runs', type=int, default=10)
+ parser.add_argument('--window-visible', action='store_true')
+ parser.add_argument('--render', action='store_true')
+ parser.set_defaults(render=False)
+ args = parser.parse_args()
+
+ random_seed.set_random_seed(args.seed)
+
+ env = gym.make(args.env)
+ timestep_limit = env.spec.timestep_limit
+ env_modifiers.make_timestep_limited(env, timestep_limit)
+ if args.render:
+ env_modifiers.make_rendered(env)
+
+ obs_size = np.asarray(env.observation_space.shape).prod()
+ action_size = np.asarray(env.action_space.shape).prod()
+
+ model = A3CLSTMGaussian(obs_size, action_size)
+ serializers.load_hdf5(args.model, model)
+
+ scores = []
+
+ for i in range(args.n_runs):
+ score = eval_single_run(env, model, phi)
+ print('Run {}: {}'.format(i, score))
+ scores.append(score)
+ print('Average: {}'.format(sum(scores) / args.n_runs))
+
+
+if __name__ == '__main__':
+ main()
|
|
50e16f8212c87ceb1f3fd1c896149b626e6e4178
|
indra/util/__init__.py
|
indra/util/__init__.py
|
def has_str(obj):
if type(obj) == str:
return True
# Check for an iterable
if hasattr(obj, '__iter__'):
for item in obj:
item_has_str = has_str(item)
if item_has_str:
return True
if hasattr(obj, '__dict__'):
for item in obj.__dict__.values():
item_has_str = has_str(item)
if item_has_str:
return True
return False
|
Test for objs with strs rather than unicode
|
Test for objs with strs rather than unicode
|
Python
|
bsd-2-clause
|
jmuhlich/indra,sorgerlab/indra,johnbachman/indra,pvtodorov/indra,sorgerlab/belpy,pvtodorov/indra,bgyori/indra,pvtodorov/indra,sorgerlab/indra,jmuhlich/indra,johnbachman/belpy,johnbachman/indra,jmuhlich/indra,johnbachman/belpy,sorgerlab/belpy,johnbachman/indra,pvtodorov/indra,sorgerlab/belpy,bgyori/indra,bgyori/indra,johnbachman/belpy,sorgerlab/indra
|
---
+++
@@ -0,0 +1,17 @@
+def has_str(obj):
+ if type(obj) == str:
+ return True
+ # Check for an iterable
+ if hasattr(obj, '__iter__'):
+ for item in obj:
+ item_has_str = has_str(item)
+ if item_has_str:
+ return True
+ if hasattr(obj, '__dict__'):
+ for item in obj.__dict__.values():
+ item_has_str = has_str(item)
+ if item_has_str:
+ return True
+ return False
+
+
|
|
9bce480b245e20f6c6ef93d34e92c27cea9f6d77
|
tests/test_settings.py
|
tests/test_settings.py
|
import pytest
from isort import exceptions
from isort.settings import Config
class TestConfig:
def test_init(self):
assert Config()
def test_invalid_pyversion(self):
with pytest.raises(ValueError):
Config(py_version=10)
def test_invalid_profile(self):
with pytest.raises(exceptions.ProfileDoesNotExist):
Config(profile="blackandwhitestylemixedwithpep8")
|
Add initial settings test file
|
Add initial settings test file
|
Python
|
mit
|
PyCQA/isort,PyCQA/isort
|
---
+++
@@ -0,0 +1,17 @@
+import pytest
+
+from isort import exceptions
+from isort.settings import Config
+
+
+class TestConfig:
+ def test_init(self):
+ assert Config()
+
+ def test_invalid_pyversion(self):
+ with pytest.raises(ValueError):
+ Config(py_version=10)
+
+ def test_invalid_profile(self):
+ with pytest.raises(exceptions.ProfileDoesNotExist):
+ Config(profile="blackandwhitestylemixedwithpep8")
|
|
be70b1528f51385c8221b7337cdc8669f53fa1c6
|
textblob/decorators.py
|
textblob/decorators.py
|
# -*- coding: utf-8 -*-
'''Custom decorators.'''
from __future__ import absolute_import
from textblob.exceptions import MissingCorpusException
class cached_property(object):
'''A property that is only computed once per instance and then replaces
itself with an ordinary attribute. Deleting the attribute resets the
property.
Credit to Marcel Hellkamp, author of bottle.py.
'''
def __init__(self, func):
self.__doc__ = getattr(func, '__doc__')
self.func = func
def __get__(self, obj, cls):
if obj is None:
return self
value = obj.__dict__[self.func.__name__] = self.func(obj)
return value
def requires_nltk_corpus(func):
'''Wraps a function that requires an NLTK corpus. If the corpus isn't found,
raise a MissingCorpusException.
'''
def decorated(*args, **kwargs):
try:
return func(*args, **kwargs)
except LookupError as err:
print(err)
raise MissingCorpusException()
return decorated
|
# -*- coding: utf-8 -*-
'''Custom decorators.'''
from __future__ import absolute_import
from functools import wraps
from textblob.exceptions import MissingCorpusException
class cached_property(object):
'''A property that is only computed once per instance and then replaces
itself with an ordinary attribute. Deleting the attribute resets the
property.
Credit to Marcel Hellkamp, author of bottle.py.
'''
def __init__(self, func):
self.__doc__ = getattr(func, '__doc__')
self.func = func
def __get__(self, obj, cls):
if obj is None:
return self
value = obj.__dict__[self.func.__name__] = self.func(obj)
return value
def requires_nltk_corpus(func):
'''Wraps a function that requires an NLTK corpus. If the corpus isn't found,
raise a MissingCorpusException.
'''
@wraps(func)
def decorated(*args, **kwargs):
try:
return func(*args, **kwargs)
except LookupError as err:
print(err)
raise MissingCorpusException()
return decorated
|
Use wraps decorator for requires_nltk_corpus
|
Use wraps decorator for requires_nltk_corpus
|
Python
|
mit
|
jcalbert/TextBlob,freakynit/TextBlob,nvoron23/TextBlob,IrisSteenhout/TextBlob,adelq/TextBlob,beni55/TextBlob,jonmcoe/TextBlob,dipeshtech/TextBlob,sargam111/python,sloria/TextBlob,Windy-Ground/TextBlob,laugustyniak/TextBlob
|
---
+++
@@ -2,6 +2,7 @@
'''Custom decorators.'''
from __future__ import absolute_import
+from functools import wraps
from textblob.exceptions import MissingCorpusException
@@ -28,6 +29,7 @@
'''Wraps a function that requires an NLTK corpus. If the corpus isn't found,
raise a MissingCorpusException.
'''
+ @wraps(func)
def decorated(*args, **kwargs):
try:
return func(*args, **kwargs)
|
a86525047658bf5adcd2133f71fe392a11883916
|
tools/unicode_tests.py
|
tools/unicode_tests.py
|
# coding: utf-8
"""These tests have to be run separately from the main test suite (iptest),
because that sets the default encoding to utf-8, and it cannot be changed after
the interpreter is up and running. The default encoding in a Python 2.x
environment is ASCII."""
import unittest, sys
from IPython.core import compilerop
assert sys.getdefaultencoding() == "ascii"
class CompileropTest(unittest.TestCase):
def test_accept_unicode(self):
cp = compilerop.CachingCompiler()
cp(u"t = 'žćčšđ'", "single")
if __name__ == "__main__":
unittest.main()
|
Test case for the failure to compile code including unicode characters.
|
Test case for the failure to compile code including unicode characters.
|
Python
|
bsd-3-clause
|
ipython/ipython,ipython/ipython
|
---
+++
@@ -0,0 +1,18 @@
+# coding: utf-8
+"""These tests have to be run separately from the main test suite (iptest),
+because that sets the default encoding to utf-8, and it cannot be changed after
+the interpreter is up and running. The default encoding in a Python 2.x
+environment is ASCII."""
+import unittest, sys
+
+from IPython.core import compilerop
+
+assert sys.getdefaultencoding() == "ascii"
+
+class CompileropTest(unittest.TestCase):
+ def test_accept_unicode(self):
+ cp = compilerop.CachingCompiler()
+ cp(u"t = 'žćčšđ'", "single")
+
+if __name__ == "__main__":
+ unittest.main()
|
|
eb1daa3edfaa72cad2cb39507b2db0bf95204561
|
markitup/renderers.py
|
markitup/renderers.py
|
from __future__ import unicode_literals
try:
from docutils.core import publish_parts
def render_rest(markup, **docutils_settings):
parts = publish_parts(source=markup, writer_name="html4css1", settings_overrides=docutils_settings)
return parts["html_body"]
except ImportError:
pass
|
from __future__ import unicode_literals
try:
from docutils.core import publish_parts
def render_rest(markup, **docutils_settings):
docutils_settings.update({
'raw_enabled': False,
'file_insertion_enabled': False,
})
parts = publish_parts(
source=markup,
writer_name="html4css1",
settings_overrides=docutils_settings,
)
return parts["html_body"]
except ImportError:
pass
|
Enforce better security in sample ReST renderer.
|
Enforce better security in sample ReST renderer.
|
Python
|
bsd-3-clause
|
WimpyAnalytics/django-markitup,carljm/django-markitup,WimpyAnalytics/django-markitup,zsiciarz/django-markitup,zsiciarz/django-markitup,carljm/django-markitup,WimpyAnalytics/django-markitup,carljm/django-markitup,zsiciarz/django-markitup
|
---
+++
@@ -2,8 +2,18 @@
try:
from docutils.core import publish_parts
+
def render_rest(markup, **docutils_settings):
- parts = publish_parts(source=markup, writer_name="html4css1", settings_overrides=docutils_settings)
+ docutils_settings.update({
+ 'raw_enabled': False,
+ 'file_insertion_enabled': False,
+ })
+
+ parts = publish_parts(
+ source=markup,
+ writer_name="html4css1",
+ settings_overrides=docutils_settings,
+ )
return parts["html_body"]
except ImportError:
pass
|
c332b83cc543f8e43405cf72e6e7c80b0cafba80
|
datasets/online_products_dataset.py
|
datasets/online_products_dataset.py
|
# -*- coding: utf-8 -*-
"""
Created on Tue Feb 14 17:30:49 2017
@author: sakurai
"""
from fuel.datasets import H5PYDataset
from fuel.utils import find_in_data_path
from fuel.schemes import SequentialScheme
from fuel.streams import DataStream
class OnlineProductsDataset(H5PYDataset):
_filename = 'online_products/online_products.hdf5'
def __init__(self, which_sets, **kwargs):
super(OnlineProductsDataset, self).__init__(
file_or_path=find_in_data_path(self._filename),
which_sets=which_sets, **kwargs)
def load_as_ndarray(which_sets=['train', 'test']):
datasets = []
for split in which_sets:
data = OnlineProductsDataset([split], load_in_memory=True).data_sources
datasets.append(data)
return datasets
if __name__ == '__main__':
dataset = OnlineProductsDataset(['train'])
st = DataStream(
dataset, iteration_scheme=SequentialScheme(dataset.num_examples, 1))
it = st.get_epoch_iterator()
it.next()
|
Add the dataset for online products dataset
|
Add the dataset for online products dataset
|
Python
|
mit
|
ronekko/deep_metric_learning
|
---
+++
@@ -0,0 +1,38 @@
+# -*- coding: utf-8 -*-
+"""
+Created on Tue Feb 14 17:30:49 2017
+
+@author: sakurai
+"""
+
+from fuel.datasets import H5PYDataset
+from fuel.utils import find_in_data_path
+from fuel.schemes import SequentialScheme
+from fuel.streams import DataStream
+
+
+class OnlineProductsDataset(H5PYDataset):
+
+ _filename = 'online_products/online_products.hdf5'
+
+ def __init__(self, which_sets, **kwargs):
+ super(OnlineProductsDataset, self).__init__(
+ file_or_path=find_in_data_path(self._filename),
+ which_sets=which_sets, **kwargs)
+
+
+def load_as_ndarray(which_sets=['train', 'test']):
+ datasets = []
+ for split in which_sets:
+ data = OnlineProductsDataset([split], load_in_memory=True).data_sources
+ datasets.append(data)
+ return datasets
+
+
+if __name__ == '__main__':
+ dataset = OnlineProductsDataset(['train'])
+
+ st = DataStream(
+ dataset, iteration_scheme=SequentialScheme(dataset.num_examples, 1))
+ it = st.get_epoch_iterator()
+ it.next()
|
|
f82c43f3cc1fc74cd23b7ae4b957c464e09fe179
|
AttributeExploration.py
|
AttributeExploration.py
|
#-------------------------------------------------------------------------------
# Name: AttributeExploration.py
# Purpose: class for attribute exploration
#
# Author: Jakob Kogler
#-------------------------------------------------------------------------------
class AttributeExploration:
def __init__(self, attributes, objects):
self.attributes = attributes
self.attributeCount = len(attributes)
self.objects = objects
self.B = 0
self.implicationsBasis = []
self.waitForResponse = False
def getNextImplication(self):
def Lstar(X):
B = [b for (a,b) in self.implicationsBasis if a & X == a and a != X]
for b in B:
X |= b
return X
if not self.waitForResponse:
for i in reversed(range(self.attributeCount)):
j = self.attributeCount - 1 - i
# m = {m_1, m_2, ..., m_i-1}
m = 2**self.attributeCount - 2 * 2**j
# P = (B cut {m_1, m_2, ..., m_i-1}) union m_i
P = (self.B & m) | 2**j
# L*-operator
LstarP, P = P, -1
while LstarP != P:
LstarP, P = Lstar(LstarP), LstarP
# B <_i L*
if (P & ~self.B & 2**j == 0) or (self.B & m != P & m):
continue
# P**
Pstar = [obj for obj in self.objects if obj & P == P]
Pstarstar = 2**self.attributeCount - 1
for obj in Pstar:
Pstarstar &= obj
if P == Pstarstar:
# P => P, not interesting
self.B = P
return self.getNextImplication()
else:
# interesting implication found
self.implication = (P, Pstarstar)
self.waitForResponse = True
return self.implication
return None
def acceptImplication(self):
if self.waitForResponse:
self.waitForResponse = False
self.implicationsBasis.append(self.implication)
self.B = self.implication[0]
def rejectImplication(self, counterExample):
if self.waitForResponse:
self.waitForResponse = False
self.objects.append(counterExample)
|
ADD attributeExploration algorithm wrapped in a class
|
ADD attributeExploration algorithm wrapped in a class
|
Python
|
mit
|
jakobkogler/AttributeExploration,jakobkogler/AttributeExploration
|
---
+++
@@ -0,0 +1,69 @@
+#-------------------------------------------------------------------------------
+# Name: AttributeExploration.py
+# Purpose: class for attribute exploration
+#
+# Author: Jakob Kogler
+#-------------------------------------------------------------------------------
+
+class AttributeExploration:
+ def __init__(self, attributes, objects):
+ self.attributes = attributes
+ self.attributeCount = len(attributes)
+ self.objects = objects
+ self.B = 0
+ self.implicationsBasis = []
+ self.waitForResponse = False
+
+ def getNextImplication(self):
+ def Lstar(X):
+ B = [b for (a,b) in self.implicationsBasis if a & X == a and a != X]
+ for b in B:
+ X |= b
+ return X
+
+ if not self.waitForResponse:
+ for i in reversed(range(self.attributeCount)):
+ j = self.attributeCount - 1 - i
+
+ # m = {m_1, m_2, ..., m_i-1}
+ m = 2**self.attributeCount - 2 * 2**j
+
+ # P = (B cut {m_1, m_2, ..., m_i-1}) union m_i
+ P = (self.B & m) | 2**j
+
+ # L*-operator
+ LstarP, P = P, -1
+ while LstarP != P:
+ LstarP, P = Lstar(LstarP), LstarP
+
+ # B <_i L*
+ if (P & ~self.B & 2**j == 0) or (self.B & m != P & m):
+ continue
+
+ # P**
+ Pstar = [obj for obj in self.objects if obj & P == P]
+ Pstarstar = 2**self.attributeCount - 1
+ for obj in Pstar:
+ Pstarstar &= obj
+
+ if P == Pstarstar:
+ # P => P, not interesting
+ self.B = P
+ return self.getNextImplication()
+ else:
+ # interesting implication found
+ self.implication = (P, Pstarstar)
+ self.waitForResponse = True
+ return self.implication
+ return None
+
+ def acceptImplication(self):
+ if self.waitForResponse:
+ self.waitForResponse = False
+ self.implicationsBasis.append(self.implication)
+ self.B = self.implication[0]
+
+ def rejectImplication(self, counterExample):
+ if self.waitForResponse:
+ self.waitForResponse = False
+ self.objects.append(counterExample)
|
|
6eb358fbbe7351c65885d63726d895335832cf3c
|
tests/inheritance/test_multi_level_inheritance.py
|
tests/inheritance/test_multi_level_inheritance.py
|
import sqlalchemy as sa
from sqlalchemy_continuum import version_class
from tests import TestCase
class TestCommonBaseClass(TestCase):
def create_models(self):
class BaseModel(self.Model):
__tablename__ = 'base_model'
__versioned__ = {}
id = sa.Column(sa.Integer, primary_key=True)
discriminator = sa.Column(sa.String(50), index=True)
__mapper_args__ = {
'polymorphic_on': discriminator,
'polymorphic_identity': 'product'
}
class FirstLevel(BaseModel):
__tablename__ = 'first_level'
id = sa.Column(sa.Integer, sa.ForeignKey('base_model.id'), primary_key=True)
__mapper_args__ = {
'polymorphic_identity': 'first_level'
}
class SecondLevel(FirstLevel):
__mapper_args__ = {
'polymorphic_identity': 'second_level'
}
self.BaseModel = BaseModel
self.FirstLevel = FirstLevel
self.SecondLevel = SecondLevel
def test_sa_inheritance_with_no_distinct_table_has_right_translation_class(self):
class_ = version_class(self.BaseModel)
assert class_.__name__ == 'BaseModelVersion'
assert class_.__table__.name == 'base_model_version'
class_ = version_class(self.FirstLevel)
assert class_.__name__ == 'FirstLevelVersion'
assert class_.__table__.name == 'first_level_version'
class_ = version_class(self.SecondLevel)
assert class_.__name__ == 'SecondLevelVersion'
assert class_.__table__.name == 'first_level_version'
|
Add test for inheritance case.
|
Add test for inheritance case.
|
Python
|
bsd-3-clause
|
kvesteri/sqlalchemy-continuum
|
---
+++
@@ -0,0 +1,47 @@
+import sqlalchemy as sa
+from sqlalchemy_continuum import version_class
+from tests import TestCase
+
+
+class TestCommonBaseClass(TestCase):
+ def create_models(self):
+ class BaseModel(self.Model):
+ __tablename__ = 'base_model'
+ __versioned__ = {}
+
+ id = sa.Column(sa.Integer, primary_key=True)
+ discriminator = sa.Column(sa.String(50), index=True)
+
+ __mapper_args__ = {
+ 'polymorphic_on': discriminator,
+ 'polymorphic_identity': 'product'
+ }
+
+ class FirstLevel(BaseModel):
+ __tablename__ = 'first_level'
+
+ id = sa.Column(sa.Integer, sa.ForeignKey('base_model.id'), primary_key=True)
+
+ __mapper_args__ = {
+ 'polymorphic_identity': 'first_level'
+ }
+
+ class SecondLevel(FirstLevel):
+ __mapper_args__ = {
+ 'polymorphic_identity': 'second_level'
+ }
+
+ self.BaseModel = BaseModel
+ self.FirstLevel = FirstLevel
+ self.SecondLevel = SecondLevel
+
+ def test_sa_inheritance_with_no_distinct_table_has_right_translation_class(self):
+ class_ = version_class(self.BaseModel)
+ assert class_.__name__ == 'BaseModelVersion'
+ assert class_.__table__.name == 'base_model_version'
+ class_ = version_class(self.FirstLevel)
+ assert class_.__name__ == 'FirstLevelVersion'
+ assert class_.__table__.name == 'first_level_version'
+ class_ = version_class(self.SecondLevel)
+ assert class_.__name__ == 'SecondLevelVersion'
+ assert class_.__table__.name == 'first_level_version'
|
|
c7b57a9fcff6741869394ee2e6619db684e3d522
|
scripts/fix_weather_timestamp_errors.py
|
scripts/fix_weather_timestamp_errors.py
|
#!/usr/bin/env python3
"""This is a script for fixing odd weather timestamp values in the database. Sometimes
the timestamp is off by many hours from the previous and this script fixes those."""
from datetime import timedelta
import psycopg
def main():
"""Module main function."""
# pylint: disable=invalid-name
THRESHOLD_HOURS = 3
# Change these as necessary
DB_NAME = 'env_logger'
ROW_LIMIT = 4000
# pylint: disable=not-context-manager
with psycopg.connect(f'dbname={DB_NAME}') as conn:
with conn.cursor() as cursor:
cursor.execute('SELECT id, time FROM weather_data ORDER by id DESC LIMIT %s',
(ROW_LIMIT,))
rows = cursor.fetchall()
i = 0
while i < len(rows):
diff_hours = (rows[i][1] - rows[i + 1][1]).total_seconds() / 3600
if abs(diff_hours) >= THRESHOLD_HOURS:
print(f'The difference {int(diff_hours)} hours of {rows[i]} and {rows[i + 1]} '
f'exceeds {THRESHOLD_HOURS} hours')
if diff_hours > 0:
corr_index = i + 1
corrected = rows[corr_index][1] + timedelta(hours=int(diff_hours))
else:
corr_index = i
corrected = rows[corr_index][1] + timedelta(hours=int(abs(diff_hours)) + 1)
print(f'Correcting timestamp of row ID {rows[corr_index][0]} to {corrected}')
cursor.execute('UPDATE weather_data SET time = %s WHERE id = %s',
(corrected, rows[corr_index][0]))
i += 2
if __name__ == '__main__':
main()
|
Add FMI weather table timestamp fixing script
|
Add FMI weather table timestamp fixing script
|
Python
|
mit
|
terop/env-logger,terop/env-logger,terop/env-logger,terop/env-logger,terop/env-logger,terop/env-logger,terop/env-logger
|
---
+++
@@ -0,0 +1,46 @@
+#!/usr/bin/env python3
+
+"""This is a script for fixing odd weather timestamp values in the database. Sometimes
+the timestamp is off by many hours from the previous and this script fixes those."""
+
+from datetime import timedelta
+
+import psycopg
+
+
+def main():
+ """Module main function."""
+ # pylint: disable=invalid-name
+ THRESHOLD_HOURS = 3
+ # Change these as necessary
+ DB_NAME = 'env_logger'
+ ROW_LIMIT = 4000
+
+ # pylint: disable=not-context-manager
+ with psycopg.connect(f'dbname={DB_NAME}') as conn:
+ with conn.cursor() as cursor:
+ cursor.execute('SELECT id, time FROM weather_data ORDER by id DESC LIMIT %s',
+ (ROW_LIMIT,))
+ rows = cursor.fetchall()
+ i = 0
+
+ while i < len(rows):
+ diff_hours = (rows[i][1] - rows[i + 1][1]).total_seconds() / 3600
+ if abs(diff_hours) >= THRESHOLD_HOURS:
+ print(f'The difference {int(diff_hours)} hours of {rows[i]} and {rows[i + 1]} '
+ f'exceeds {THRESHOLD_HOURS} hours')
+ if diff_hours > 0:
+ corr_index = i + 1
+ corrected = rows[corr_index][1] + timedelta(hours=int(diff_hours))
+ else:
+ corr_index = i
+ corrected = rows[corr_index][1] + timedelta(hours=int(abs(diff_hours)) + 1)
+
+ print(f'Correcting timestamp of row ID {rows[corr_index][0]} to {corrected}')
+ cursor.execute('UPDATE weather_data SET time = %s WHERE id = %s',
+ (corrected, rows[corr_index][0]))
+ i += 2
+
+
+if __name__ == '__main__':
+ main()
|
|
31d67e804ba44645b701c2624ee30c31023e994e
|
changed_options.py
|
changed_options.py
|
#!/usr/bin/env python3
# file: changed_options.py
# vim:fileencoding=utf-8:fdm=marker:ft=python
#
# Author: R.F. Smith <rsmith@xs4all.nl>
# Created: 2018-03-26 20:53:13 +0200
# Last modified: 2018-03-26 23:03:02 +0200
"""
Get a list of installed packages. For each package, determine if the options
have been changed compared to the default options, and print that.
* The ‘pkg query’ command is used to retrieve the options that are set.
* For determining the default options, ‘make -V OPTIONS_DEFAULT’ is called
from the port directory.
This program requires pkg(8) and the ports tree to be installed.
So this program will run on FreeBSD and maybe DragonflyBSD.
"""
# Imports {{{1
import concurrent.futures as cf
import os
import subprocess as sp
import sys
def run(args): # {{{1
"""
Run a subprocess and return the standard output.
Arguments:
args (list): List of argument strings. Typically a command name
followed by options.
Returns:
Standard output of the program, converted to UTF-8 string.
"""
comp = sp.run(args, stdout=sp.PIPE, stderr=sp.DEVNULL)
return comp.stdout.decode('utf-8')
def check(line): # {{{1
"""
Check of a given package uses the default options or
if options have been changed.
Arguments:
line (str): A line of text containing the package name and origin,
separated by whitespace.
Returns:
A string containing the package name and either [CHANGED] or [default].
"""
pkg, origin = line.split()
optionlines = run(['pkg', 'query', '%Ok %Ov', pkg]).splitlines()
options_set = set(opt.split()[0] for opt in optionlines if opt.endswith('on'))
try:
os.chdir('/usr/ports/{}'.format(origin))
except FileNotFoundError:
return ('{}: undetermined'.format(pkg))
default = run(['make', '-V', 'OPTIONS_DEFAULT'])
options_default = set(default.split())
if options_default == options_set:
v = 'default'
else:
v = 'CHANGED'
return '{}: [{}]'.format(pkg, v)
def main(argv): # {{{1
"""
Entry point for changed_options.py.
Arguments:
argv: command line arguments
"""
data = run(['pkg', 'info', '-a', '-o'])
packagelines = data.splitlines()
with cf.ThreadPoolExecutor(max_workers=os.cpu_count()) as tp:
for rv in tp.map(check, packagelines):
print(rv)
if __name__ == '__main__':
main(sys.argv[1:])
|
Add script to detect changed options.
|
Add script to detect changed options.
|
Python
|
mit
|
rsmith-nl/scripts,rsmith-nl/scripts
|
---
+++
@@ -0,0 +1,84 @@
+#!/usr/bin/env python3
+# file: changed_options.py
+# vim:fileencoding=utf-8:fdm=marker:ft=python
+#
+# Author: R.F. Smith <rsmith@xs4all.nl>
+# Created: 2018-03-26 20:53:13 +0200
+# Last modified: 2018-03-26 23:03:02 +0200
+"""
+Get a list of installed packages. For each package, determine if the options
+have been changed compared to the default options, and print that.
+
+* The ‘pkg query’ command is used to retrieve the options that are set.
+* For determining the default options, ‘make -V OPTIONS_DEFAULT’ is called
+ from the port directory.
+
+This program requires pkg(8) and the ports tree to be installed.
+So this program will run on FreeBSD and maybe DragonflyBSD.
+"""
+# Imports {{{1
+import concurrent.futures as cf
+import os
+import subprocess as sp
+import sys
+
+
+def run(args): # {{{1
+ """
+ Run a subprocess and return the standard output.
+
+ Arguments:
+ args (list): List of argument strings. Typically a command name
+ followed by options.
+
+ Returns:
+ Standard output of the program, converted to UTF-8 string.
+ """
+ comp = sp.run(args, stdout=sp.PIPE, stderr=sp.DEVNULL)
+ return comp.stdout.decode('utf-8')
+
+
+def check(line): # {{{1
+ """
+ Check of a given package uses the default options or
+ if options have been changed.
+
+ Arguments:
+ line (str): A line of text containing the package name and origin,
+ separated by whitespace.
+
+ Returns:
+ A string containing the package name and either [CHANGED] or [default].
+ """
+ pkg, origin = line.split()
+ optionlines = run(['pkg', 'query', '%Ok %Ov', pkg]).splitlines()
+ options_set = set(opt.split()[0] for opt in optionlines if opt.endswith('on'))
+ try:
+ os.chdir('/usr/ports/{}'.format(origin))
+ except FileNotFoundError:
+ return ('{}: undetermined'.format(pkg))
+ default = run(['make', '-V', 'OPTIONS_DEFAULT'])
+ options_default = set(default.split())
+ if options_default == options_set:
+ v = 'default'
+ else:
+ v = 'CHANGED'
+ return '{}: [{}]'.format(pkg, v)
+
+
+def main(argv): # {{{1
+ """
+ Entry point for changed_options.py.
+
+ Arguments:
+ argv: command line arguments
+ """
+ data = run(['pkg', 'info', '-a', '-o'])
+ packagelines = data.splitlines()
+ with cf.ThreadPoolExecutor(max_workers=os.cpu_count()) as tp:
+ for rv in tp.map(check, packagelines):
+ print(rv)
+
+
+if __name__ == '__main__':
+ main(sys.argv[1:])
|
|
bfe5ebae0261e49045e468bc183f54bcd7fbeafc
|
openfisca_core/scripts/xml_to_json/xml_to_json_extension_template.py
|
openfisca_core/scripts/xml_to_json/xml_to_json_extension_template.py
|
# -*- coding: utf-8 -*-
''' xml_to_json_extension_template.py : Parse XML parameter files for Extension-Template and convert them to YAML files. Comments are NOT transformed.
Usage :
`python xml_to_json_extension_template.py output_dir`
or just (output is written in a directory called `yaml_parameters`):
`python xml_to_json_extension_template.py`
'''
import sys
import os
from openfisca_core.scripts.xml_to_json import xml_to_json
import openfisca_extension_template
if len(sys.argv) > 1:
target_path = sys.argv[1]
else:
target_path = 'yaml_parameters'
param_dir = os.path.dirname(openfisca_extension_template.__file__)
param_files = [
'parameters.xml',
]
legislation_xml_info_list = [
(os.path.join(param_dir, param_file), [])
for param_file in param_files
]
xml_to_json.write_legislation(legislation_xml_info_list, target_path)
|
Add script to transform ExtensionTemplate legilation
|
Add script to transform ExtensionTemplate legilation
|
Python
|
agpl-3.0
|
openfisca/openfisca-core,openfisca/openfisca-core
|
---
+++
@@ -0,0 +1,31 @@
+# -*- coding: utf-8 -*-
+
+''' xml_to_json_extension_template.py : Parse XML parameter files for Extension-Template and convert them to YAML files. Comments are NOT transformed.
+
+Usage :
+ `python xml_to_json_extension_template.py output_dir`
+or just (output is written in a directory called `yaml_parameters`):
+ `python xml_to_json_extension_template.py`
+'''
+
+import sys
+import os
+
+from openfisca_core.scripts.xml_to_json import xml_to_json
+import openfisca_extension_template
+
+if len(sys.argv) > 1:
+ target_path = sys.argv[1]
+else:
+ target_path = 'yaml_parameters'
+
+param_dir = os.path.dirname(openfisca_extension_template.__file__)
+param_files = [
+ 'parameters.xml',
+ ]
+legislation_xml_info_list = [
+ (os.path.join(param_dir, param_file), [])
+ for param_file in param_files
+ ]
+
+xml_to_json.write_legislation(legislation_xml_info_list, target_path)
|
|
18f983cf035704588f904d966f8bf10ca4a16b01
|
src/mmw/apps/modeling/migrations/0040_clear_nlcd2019_tr55_results.py
|
src/mmw/apps/modeling/migrations/0040_clear_nlcd2019_tr55_results.py
|
# Generated by Django 3.2.13 on 2022-04-20 23:35
from django.db import migrations
def clear_nlcd2019_tr55_results(apps, schema_editor):
"""
Clear the results For all scenarios belonging to TR-55 projects made after
the release of 1.33.0, which switched NLCD19 2019 to be the default on
2022-01-17:
https://github.com/WikiWatershed/model-my-watershed/releases/tag/1.33.0
These will be recalculated with NLCD11 2011 the next time it is opened.
"""
Scenario = apps.get_model('modeling', 'Scenario')
Scenario.objects.filter(
project__model_package='tr-55',
project__created_at__gte='2022-01-17'
).update(
results='[]',
modification_hash=''
)
class Migration(migrations.Migration):
dependencies = [
('modeling', '0039_override_sedaadjust_for_old_scenarios'),
]
operations = [
migrations.RunPython(clear_nlcd2019_tr55_results)
]
|
Clear all NLCD19 2019 TR-55 results
|
Clear all NLCD19 2019 TR-55 results
Since we're switching TR-55 projects back to NLCD11 2011,
this migration clears the results for all TR-55 projects
made since 1.33.0, which switched the default to NLCD19 2019.
These results will be recalculated with NLCD11 2011 whenever
they are opened next in the UI.
|
Python
|
apache-2.0
|
WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed,WikiWatershed/model-my-watershed
|
---
+++
@@ -0,0 +1,34 @@
+# Generated by Django 3.2.13 on 2022-04-20 23:35
+
+from django.db import migrations
+
+
+def clear_nlcd2019_tr55_results(apps, schema_editor):
+ """
+ Clear the results For all scenarios belonging to TR-55 projects made after
+ the release of 1.33.0, which switched NLCD19 2019 to be the default on
+ 2022-01-17:
+ https://github.com/WikiWatershed/model-my-watershed/releases/tag/1.33.0
+
+ These will be recalculated with NLCD11 2011 the next time it is opened.
+ """
+ Scenario = apps.get_model('modeling', 'Scenario')
+
+ Scenario.objects.filter(
+ project__model_package='tr-55',
+ project__created_at__gte='2022-01-17'
+ ).update(
+ results='[]',
+ modification_hash=''
+ )
+
+
+class Migration(migrations.Migration):
+
+ dependencies = [
+ ('modeling', '0039_override_sedaadjust_for_old_scenarios'),
+ ]
+
+ operations = [
+ migrations.RunPython(clear_nlcd2019_tr55_results)
+ ]
|
|
727e57b8f639a471423c4b5a87af594632ae609d
|
scripts/tools/make_manhole.py
|
scripts/tools/make_manhole.py
|
#!/usr/bin/env python
# Copyright (c) 2011 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""Generate a .manhole for all masters."""
import getpass
import os
import optparse
import subprocess
import sys
def check_output(cmd):
p = subprocess.Popen(cmd, stdout=subprocess.PIPE)
stdout = p.communicate(None)[0]
if p.returncode:
raise subprocess.CalledProcessError(p.returncode, cmd)
return stdout
def main():
parser = optparse.OptionParser()
parser.add_option('-u', '--user', default=getpass.getuser())
parser.add_option('-p', '--port', type='int', help='Base port')
parser.add_option('-r', '--root', default=os.getcwd(), help='Path to masters')
options, args = parser.parse_args(None)
if args:
parser.error('Have you tried not using the wrong argument?')
if not options.port:
parser.error('Use --port to specify a base port')
if not os.path.basename(options.root) == 'masters':
parser.error('Use --root or cd into the masters directory')
try:
check_output(['apg', '-q', '-n', '1'])
except subprocess.CalledProcessError:
parser.error('Run sudo apt-get install apg')
for i in os.listdir(options.root):
dirpath = os.path.join(options.root, i)
if not os.path.isdir(dirpath):
continue
filepath = os.path.join(dirpath, '.manhole')
if os.path.isfile(filepath):
print '%-30s already had .manhole' % i
continue
print '%-30s Generating password' % i
password = check_output(['apg', '-q', '-n', '1']).strip()
content = "user='%s'\npassword='/!%s'\nport=%d\n" % (
options.user, password, options.port)
options.port += 1
open(filepath, 'w').write(content)
return 0
if __name__ == '__main__':
sys.exit(main())
|
Add tool to generate .manhole files.
|
Add tool to generate .manhole files.
This is cleaner than creating them one by one.
R=cmp@chromium.org
BUG=
TEST=
Review URL: http://codereview.chromium.org/8347006
git-svn-id: 239fca9b83025a0b6f823aeeca02ba5be3d9fd76@106120 0039d316-1c4b-4281-b951-d872f2087c98
|
Python
|
bsd-3-clause
|
eunchong/build,eunchong/build,eunchong/build,eunchong/build
|
---
+++
@@ -0,0 +1,60 @@
+#!/usr/bin/env python
+# Copyright (c) 2011 The Chromium Authors. All rights reserved.
+# Use of this source code is governed by a BSD-style license that can be
+# found in the LICENSE file.
+
+"""Generate a .manhole for all masters."""
+
+import getpass
+import os
+import optparse
+import subprocess
+import sys
+
+
+def check_output(cmd):
+ p = subprocess.Popen(cmd, stdout=subprocess.PIPE)
+ stdout = p.communicate(None)[0]
+ if p.returncode:
+ raise subprocess.CalledProcessError(p.returncode, cmd)
+ return stdout
+
+
+def main():
+ parser = optparse.OptionParser()
+ parser.add_option('-u', '--user', default=getpass.getuser())
+ parser.add_option('-p', '--port', type='int', help='Base port')
+ parser.add_option('-r', '--root', default=os.getcwd(), help='Path to masters')
+ options, args = parser.parse_args(None)
+
+ if args:
+ parser.error('Have you tried not using the wrong argument?')
+ if not options.port:
+ parser.error('Use --port to specify a base port')
+ if not os.path.basename(options.root) == 'masters':
+ parser.error('Use --root or cd into the masters directory')
+
+ try:
+ check_output(['apg', '-q', '-n', '1'])
+ except subprocess.CalledProcessError:
+ parser.error('Run sudo apt-get install apg')
+
+ for i in os.listdir(options.root):
+ dirpath = os.path.join(options.root, i)
+ if not os.path.isdir(dirpath):
+ continue
+ filepath = os.path.join(dirpath, '.manhole')
+ if os.path.isfile(filepath):
+ print '%-30s already had .manhole' % i
+ continue
+ print '%-30s Generating password' % i
+ password = check_output(['apg', '-q', '-n', '1']).strip()
+ content = "user='%s'\npassword='/!%s'\nport=%d\n" % (
+ options.user, password, options.port)
+ options.port += 1
+ open(filepath, 'w').write(content)
+ return 0
+
+
+if __name__ == '__main__':
+ sys.exit(main())
|
|
e471fa49409d45b2b76b12ac63fb6487466be174
|
csunplugged/utils/retrieve_query_parameter.py
|
csunplugged/utils/retrieve_query_parameter.py
|
"""Module for retrieving a GET request query parameter."""
from django.http import Http404
def retrieve_query_parameter(request, parameter, valid_options=None):
"""Retrieve the query parameter.
If the parameter cannot be found, or is not found in the list of
valid options, then a 404 error is raised.
Args:
request: Request object (Request).
parameter: Parameter to retrieve (str).
valid_options: If provided, a list of valid options (list of str).
Returns:
String value of parameter.
"""
value = request.get(parameter, None)
if value is None:
raise Http404("{} parameter not specified.".format(parameter))
if valid_options and value not in valid_options:
raise Http404("{} parameter not valid.".format(parameter))
return value
|
Add utility function for retrieving get parameters
|
Add utility function for retrieving get parameters
|
Python
|
mit
|
uccser/cs-unplugged,uccser/cs-unplugged,uccser/cs-unplugged,uccser/cs-unplugged
|
---
+++
@@ -0,0 +1,25 @@
+"""Module for retrieving a GET request query parameter."""
+
+from django.http import Http404
+
+
+def retrieve_query_parameter(request, parameter, valid_options=None):
+ """Retrieve the query parameter.
+
+ If the parameter cannot be found, or is not found in the list of
+ valid options, then a 404 error is raised.
+
+ Args:
+ request: Request object (Request).
+ parameter: Parameter to retrieve (str).
+ valid_options: If provided, a list of valid options (list of str).
+
+ Returns:
+ String value of parameter.
+ """
+ value = request.get(parameter, None)
+ if value is None:
+ raise Http404("{} parameter not specified.".format(parameter))
+ if valid_options and value not in valid_options:
+ raise Http404("{} parameter not valid.".format(parameter))
+ return value
|
|
2865af9eba55f1b3bbf14d26fd9691925fde8f5e
|
py/reconstruct-original-digits-from-english.py
|
py/reconstruct-original-digits-from-english.py
|
from collections import Counter
class Solution(object):
def originalDigits(self, s):
"""
:type s: str
:rtype: str
"""
phase1 = dict(
g=(8, 'eight'),
u=(4, 'four'),
w=(2, 'two'),
x=(6, 'six'),
z=(0, 'zero')
)
phase2 = dict(
t=(3, 'thre'),
s=(7, 'sevn'),
o=(1, 'one'),
f=(5, 'five')
)
phase3 = dict(
i=(9, 'nie')
)
counter = Counter(s)
ansCounter = Counter()
for k, (n, w) in phase1.iteritems():
ck = counter[k]
ansCounter[n] += ck
for c in w:
counter[c] -= ck
for k, (n, w) in phase2.iteritems():
ck = counter[k]
ansCounter[n] += ck
for c in w:
counter[c] -= ck
for k, (n, w) in phase3.iteritems():
ck = counter[k]
ansCounter[n] += ck
for c in w:
counter[c] -= ck
return ''.join(str(i) * ansCounter[i] for i in xrange(10))
|
Add py solution for 423. Reconstruct Original Digits from English
|
Add py solution for 423. Reconstruct Original Digits from English
423. Reconstruct Original Digits from English: https://leetcode.com/problems/reconstruct-original-digits-from-english/
|
Python
|
apache-2.0
|
ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode
|
---
+++
@@ -0,0 +1,46 @@
+from collections import Counter
+class Solution(object):
+ def originalDigits(self, s):
+ """
+ :type s: str
+ :rtype: str
+ """
+ phase1 = dict(
+ g=(8, 'eight'),
+ u=(4, 'four'),
+ w=(2, 'two'),
+ x=(6, 'six'),
+ z=(0, 'zero')
+ )
+ phase2 = dict(
+ t=(3, 'thre'),
+ s=(7, 'sevn'),
+ o=(1, 'one'),
+ f=(5, 'five')
+ )
+ phase3 = dict(
+ i=(9, 'nie')
+ )
+
+ counter = Counter(s)
+ ansCounter = Counter()
+ for k, (n, w) in phase1.iteritems():
+ ck = counter[k]
+ ansCounter[n] += ck
+ for c in w:
+ counter[c] -= ck
+
+ for k, (n, w) in phase2.iteritems():
+ ck = counter[k]
+ ansCounter[n] += ck
+ for c in w:
+ counter[c] -= ck
+
+ for k, (n, w) in phase3.iteritems():
+ ck = counter[k]
+ ansCounter[n] += ck
+ for c in w:
+ counter[c] -= ck
+
+ return ''.join(str(i) * ansCounter[i] for i in xrange(10))
+
|
|
09911a0fe8135fa2534c9d6e708e688fcfe54ca7
|
analysis/check_calls.py
|
analysis/check_calls.py
|
import os
import click
import cv2
from AppKit import NSScreen
def check_image(filename, height):
image = cv2.imread(filename)
image_name = os.path.basename(filename)
is_positive = _check(image, image_name, height)
return image_name, is_positive
def _check(image, image_name, target_height):
cv2.namedWindow(image_name, cv2.WINDOW_NORMAL)
while True:
height, width = image.shape[:2]
scaling_f = height / target_height
small = cv2.resize(image, None, fx=1/scaling_f, fy=1/scaling_f)
# display the image and wait for a keypress
cv2.imshow(image_name, small)
key = cv2.waitKey(1)
# if the 'y' key is pressed return True
if key == ord("y"):
cv2.destroyWindow(image_name)
print('True')
return True
# if the 'n' key is pressed return False
elif key == ord("n"):
print('False')
cv2.destroyWindow(image_name)
return False
@click.command()
@click.argument('files',
nargs=-1, type=click.Path(exists=True))
@click.option('--output_path',
help='Write result here',
default=None,
type=click.Path(exists=False))
def check_files(files, output_path):
height = NSScreen.mainScreen().frame().size.height
with open(output_path, 'w') as out:
for file in files:
image_name, is_positive = check_image(file, height)
template = "%s\t%s\n"
out.write(template % (image_name, is_positive))
if __name__ == "__main__":
check_files()
|
Add a small script for rapidly verifyinh IGV screenshots
|
Add a small script for rapidly verifyinh IGV screenshots
|
Python
|
mit
|
bardin-lab/readtagger,bardin-lab/readtagger
|
---
+++
@@ -0,0 +1,54 @@
+import os
+import click
+import cv2
+from AppKit import NSScreen
+
+
+def check_image(filename, height):
+ image = cv2.imread(filename)
+ image_name = os.path.basename(filename)
+ is_positive = _check(image, image_name, height)
+ return image_name, is_positive
+
+
+def _check(image, image_name, target_height):
+ cv2.namedWindow(image_name, cv2.WINDOW_NORMAL)
+ while True:
+ height, width = image.shape[:2]
+ scaling_f = height / target_height
+ small = cv2.resize(image, None, fx=1/scaling_f, fy=1/scaling_f)
+ # display the image and wait for a keypress
+ cv2.imshow(image_name, small)
+ key = cv2.waitKey(1)
+
+ # if the 'y' key is pressed return True
+ if key == ord("y"):
+ cv2.destroyWindow(image_name)
+ print('True')
+ return True
+
+ # if the 'n' key is pressed return False
+ elif key == ord("n"):
+ print('False')
+ cv2.destroyWindow(image_name)
+ return False
+
+
+@click.command()
+@click.argument('files',
+ nargs=-1, type=click.Path(exists=True))
+@click.option('--output_path',
+ help='Write result here',
+ default=None,
+ type=click.Path(exists=False))
+def check_files(files, output_path):
+ height = NSScreen.mainScreen().frame().size.height
+ with open(output_path, 'w') as out:
+ for file in files:
+ image_name, is_positive = check_image(file, height)
+ template = "%s\t%s\n"
+ out.write(template % (image_name, is_positive))
+
+
+if __name__ == "__main__":
+ check_files()
|
|
10782310cfee0d2c2938748056f6549b5918b969
|
src/sentry/debug/utils/patch_context.py
|
src/sentry/debug/utils/patch_context.py
|
from __future__ import absolute_import
from sentry.utils.imports import import_string
class PatchContext(object):
def __init__(self, target, callback):
target, attr = target.rsplit('.', 1)
target = import_string(target)
self.func = getattr(target, attr)
self.target = target
self.attr = attr
self.callback = callback
def __enter__(self):
self.patch()
return self
def __exit__(self, exc_type, exc_value, traceback):
self.unpatch()
def patch(self):
func = getattr(self.target, self.attr)
def wrapped(*args, **kwargs):
__traceback_hide__ = True # NOQA
return self.callback(self.func, *args, **kwargs)
wrapped.__name__ = func.__name__
if hasattr(func, '__doc__'):
wrapped.__doc__ = func.__doc__
if hasattr(func, '__module__'):
wrapped.__module__ = func.__module__
setattr(self.target, self.attr, wrapped)
def unpatch(self):
setattr(self.target, self.attr, self.func)
|
from __future__ import absolute_import
from threading import Lock
from sentry.utils.imports import import_string
class PatchContext(object):
def __init__(self, target, callback):
target, attr = target.rsplit('.', 1)
target = import_string(target)
self.target = target
self.attr = attr
self.callback = callback
self._lock = Lock()
with self._lock:
self.func = getattr(target, attr)
def __enter__(self):
self.patch()
return self
def __exit__(self, exc_type, exc_value, traceback):
self.unpatch()
def patch(self):
with self._lock:
func = getattr(self.target, self.attr)
def wrapped(*args, **kwargs):
__traceback_hide__ = True # NOQA
return self.callback(self.func, *args, **kwargs)
wrapped.__name__ = func.__name__
if hasattr(func, '__doc__'):
wrapped.__doc__ = func.__doc__
if hasattr(func, '__module__'):
wrapped.__module__ = func.__module__
setattr(self.target, self.attr, wrapped)
def unpatch(self):
with self._lock:
setattr(self.target, self.attr, self.func)
|
Use a thread lock to patch contexts.
|
Use a thread lock to patch contexts.
This fixes #3185
|
Python
|
bsd-3-clause
|
looker/sentry,zenefits/sentry,mvaled/sentry,alexm92/sentry,alexm92/sentry,looker/sentry,gencer/sentry,ifduyue/sentry,jean/sentry,JackDanger/sentry,JackDanger/sentry,ifduyue/sentry,BuildingLink/sentry,gencer/sentry,beeftornado/sentry,BuildingLink/sentry,mvaled/sentry,JamesMura/sentry,jean/sentry,zenefits/sentry,zenefits/sentry,mvaled/sentry,mvaled/sentry,jean/sentry,JamesMura/sentry,zenefits/sentry,mvaled/sentry,JamesMura/sentry,fotinakis/sentry,ifduyue/sentry,gencer/sentry,JackDanger/sentry,mitsuhiko/sentry,jean/sentry,zenefits/sentry,mvaled/sentry,gencer/sentry,BuildingLink/sentry,fotinakis/sentry,JamesMura/sentry,mitsuhiko/sentry,beeftornado/sentry,beeftornado/sentry,fotinakis/sentry,ifduyue/sentry,alexm92/sentry,BuildingLink/sentry,JamesMura/sentry,looker/sentry,gencer/sentry,looker/sentry,fotinakis/sentry,looker/sentry,BuildingLink/sentry,ifduyue/sentry,jean/sentry
|
---
+++
@@ -1,5 +1,6 @@
from __future__ import absolute_import
+from threading import Lock
from sentry.utils.imports import import_string
@@ -7,10 +8,12 @@
def __init__(self, target, callback):
target, attr = target.rsplit('.', 1)
target = import_string(target)
- self.func = getattr(target, attr)
self.target = target
self.attr = attr
self.callback = callback
+ self._lock = Lock()
+ with self._lock:
+ self.func = getattr(target, attr)
def __enter__(self):
self.patch()
@@ -20,19 +23,21 @@
self.unpatch()
def patch(self):
- func = getattr(self.target, self.attr)
+ with self._lock:
+ func = getattr(self.target, self.attr)
- def wrapped(*args, **kwargs):
- __traceback_hide__ = True # NOQA
- return self.callback(self.func, *args, **kwargs)
+ def wrapped(*args, **kwargs):
+ __traceback_hide__ = True # NOQA
+ return self.callback(self.func, *args, **kwargs)
- wrapped.__name__ = func.__name__
- if hasattr(func, '__doc__'):
- wrapped.__doc__ = func.__doc__
- if hasattr(func, '__module__'):
- wrapped.__module__ = func.__module__
+ wrapped.__name__ = func.__name__
+ if hasattr(func, '__doc__'):
+ wrapped.__doc__ = func.__doc__
+ if hasattr(func, '__module__'):
+ wrapped.__module__ = func.__module__
- setattr(self.target, self.attr, wrapped)
+ setattr(self.target, self.attr, wrapped)
def unpatch(self):
- setattr(self.target, self.attr, self.func)
+ with self._lock:
+ setattr(self.target, self.attr, self.func)
|
a353ad76774c44004256fef8b076f74b6b639ca4
|
tests/remove_stale_string.py
|
tests/remove_stale_string.py
|
import re
import json
import glob
from collections import OrderedDict
locale_folder = "../locales/"
locale_files = glob.glob(locale_folder + "*.json")
locale_files = [filename.split("/")[-1] for filename in locale_files]
locale_files.remove("en.json")
reference = json.loads(open(locale_folder + "en.json").read())
for locale_file in locale_files:
print(locale_file)
this_locale = json.loads(open(locale_folder + locale_file).read(), object_pairs_hook=OrderedDict)
this_locale_fixed = {k:v for k, v in this_locale.items() if k in reference}
json.dump(this_locale_fixed, open(locale_folder + locale_file, "w"), indent=4, ensure_ascii=False)
|
Add script to remove stale i18n strings
|
Add script to remove stale i18n strings
|
Python
|
agpl-3.0
|
YunoHost/moulinette-yunohost,YunoHost/yunohost,YunoHost/moulinette-yunohost,YunoHost/yunohost,YunoHost/moulinette-yunohost,YunoHost/yunohost,YunoHost/moulinette-yunohost,YunoHost/moulinette-yunohost,YunoHost/yunohost
|
---
+++
@@ -0,0 +1,19 @@
+import re
+import json
+import glob
+from collections import OrderedDict
+
+locale_folder = "../locales/"
+locale_files = glob.glob(locale_folder + "*.json")
+locale_files = [filename.split("/")[-1] for filename in locale_files]
+locale_files.remove("en.json")
+
+reference = json.loads(open(locale_folder + "en.json").read())
+
+for locale_file in locale_files:
+
+ print(locale_file)
+ this_locale = json.loads(open(locale_folder + locale_file).read(), object_pairs_hook=OrderedDict)
+ this_locale_fixed = {k:v for k, v in this_locale.items() if k in reference}
+
+ json.dump(this_locale_fixed, open(locale_folder + locale_file, "w"), indent=4, ensure_ascii=False)
|
|
4cc2861ed79d54c5f59a29b5d56dde5aae9c0c81
|
examples/ex_rofi.py
|
examples/ex_rofi.py
|
#!/usr/bin/env python
import string
from dynmen import Menu
rofi = Menu(command=('rofi', '-fullscreen', '-dmenu', '-i'))
d_string = vars(string)
d_string = {k:v for k,v in d_string.items() if not k.startswith('_')}
print('Launching rofi - given a dict')
output = rofi(d_string)
print(output, '\n')
print('Launching rofi - first sorting entries of dict')
output2 = rofi.sort(d_string)
print(output2, '\n')
print('Launching rofi - given a list')
output3 = rofi(list(d_string))
print(output3, '\n')
|
Add a script showing usage in examples
|
Add a script showing usage in examples
|
Python
|
mit
|
frostidaho/dynmen
|
---
+++
@@ -0,0 +1,19 @@
+#!/usr/bin/env python
+import string
+from dynmen import Menu
+rofi = Menu(command=('rofi', '-fullscreen', '-dmenu', '-i'))
+
+d_string = vars(string)
+d_string = {k:v for k,v in d_string.items() if not k.startswith('_')}
+
+print('Launching rofi - given a dict')
+output = rofi(d_string)
+print(output, '\n')
+
+print('Launching rofi - first sorting entries of dict')
+output2 = rofi.sort(d_string)
+print(output2, '\n')
+
+print('Launching rofi - given a list')
+output3 = rofi(list(d_string))
+print(output3, '\n')
|
|
00596dbc602a3e555cef0a3453d83d475c28fc52
|
tests/test_azure_publish_tools.py
|
tests/test_azure_publish_tools.py
|
from argparse import Namespace
from unittest import TestCase
from azure_publish_tools import (
DELETE,
get_option_parser,
LIST,
PUBLISH,
)
class TestOptionParser(TestCase):
def parse_args(self, args):
parser = get_option_parser()
return parser.parse_args(args)
def test_list(self):
args = self.parse_args(['list', 'mypurpose'])
self.assertEqual(Namespace(
command=LIST, purpose='mypurpose', dry_run=False, verbose=False,
path=[]), args)
def test_list_dry_run(self):
args = self.parse_args(['list', 'mypurpose', '--dry-run'])
self.assertIs(True, args.dry_run)
def test_list_verbose(self):
args = self.parse_args(['list', 'mypurpose', '--verbose'])
self.assertIs(True, args.verbose)
def test_publish(self):
args = self.parse_args(['publish', 'mypurpose'])
self.assertEqual(Namespace(
command=PUBLISH, purpose='mypurpose', dry_run=False, verbose=False,
path=[]), args)
def test_publish_dry_run(self):
args = self.parse_args(['publish', 'mypurpose', '--dry-run'])
self.assertIs(True, args.dry_run)
def test_publish_verbose(self):
args = self.parse_args(['publish', 'mypurpose', '--verbose'])
self.assertIs(True, args.verbose)
def test_publish_path(self):
args = self.parse_args(['publish', 'mypurpose', 'mypath', 'mypath2'])
self.assertEqual(['mypath', 'mypath2'], args.path)
def test_delete(self):
args = self.parse_args(['delete', 'mypurpose'])
self.assertEqual(Namespace(
command=DELETE, purpose='mypurpose', dry_run=False, verbose=False,
path=[]), args)
def test_delete_dry_run(self):
args = self.parse_args(['delete', 'mypurpose', '--dry-run'])
self.assertIs(True, args.dry_run)
def test_delete_verbose(self):
args = self.parse_args(['delete', 'mypurpose', '--verbose'])
self.assertIs(True, args.verbose)
def test_delete_path(self):
args = self.parse_args(['delete', 'mypurpose', 'mypath', 'mypath2'])
self.assertEqual(['mypath', 'mypath2'], args.path)
|
Add tests for current arg parsing.
|
Add tests for current arg parsing.
|
Python
|
agpl-3.0
|
mjs/juju,mjs/juju,mjs/juju,mjs/juju,mjs/juju,mjs/juju,mjs/juju
|
---
+++
@@ -0,0 +1,65 @@
+from argparse import Namespace
+from unittest import TestCase
+
+from azure_publish_tools import (
+ DELETE,
+ get_option_parser,
+ LIST,
+ PUBLISH,
+ )
+
+class TestOptionParser(TestCase):
+
+ def parse_args(self, args):
+ parser = get_option_parser()
+ return parser.parse_args(args)
+
+ def test_list(self):
+ args = self.parse_args(['list', 'mypurpose'])
+ self.assertEqual(Namespace(
+ command=LIST, purpose='mypurpose', dry_run=False, verbose=False,
+ path=[]), args)
+
+ def test_list_dry_run(self):
+ args = self.parse_args(['list', 'mypurpose', '--dry-run'])
+ self.assertIs(True, args.dry_run)
+
+ def test_list_verbose(self):
+ args = self.parse_args(['list', 'mypurpose', '--verbose'])
+ self.assertIs(True, args.verbose)
+
+ def test_publish(self):
+ args = self.parse_args(['publish', 'mypurpose'])
+ self.assertEqual(Namespace(
+ command=PUBLISH, purpose='mypurpose', dry_run=False, verbose=False,
+ path=[]), args)
+
+ def test_publish_dry_run(self):
+ args = self.parse_args(['publish', 'mypurpose', '--dry-run'])
+ self.assertIs(True, args.dry_run)
+
+ def test_publish_verbose(self):
+ args = self.parse_args(['publish', 'mypurpose', '--verbose'])
+ self.assertIs(True, args.verbose)
+
+ def test_publish_path(self):
+ args = self.parse_args(['publish', 'mypurpose', 'mypath', 'mypath2'])
+ self.assertEqual(['mypath', 'mypath2'], args.path)
+
+ def test_delete(self):
+ args = self.parse_args(['delete', 'mypurpose'])
+ self.assertEqual(Namespace(
+ command=DELETE, purpose='mypurpose', dry_run=False, verbose=False,
+ path=[]), args)
+
+ def test_delete_dry_run(self):
+ args = self.parse_args(['delete', 'mypurpose', '--dry-run'])
+ self.assertIs(True, args.dry_run)
+
+ def test_delete_verbose(self):
+ args = self.parse_args(['delete', 'mypurpose', '--verbose'])
+ self.assertIs(True, args.verbose)
+
+ def test_delete_path(self):
+ args = self.parse_args(['delete', 'mypurpose', 'mypath', 'mypath2'])
+ self.assertEqual(['mypath', 'mypath2'], args.path)
|
|
8a9329a5c2b97d32a1fd32ae16c21222fb10b6b2
|
cms_lab_carousel/migrations/0003_auto_20150417_1240.py
|
cms_lab_carousel/migrations/0003_auto_20150417_1240.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.utils.timezone import utc
import django.core.validators
import datetime
class Migration(migrations.Migration):
dependencies = [
('cms_lab_carousel', '0002_auto_20150417_1219'),
]
operations = [
migrations.RemoveField(
model_name='carousel',
name='footer_height',
),
migrations.RemoveField(
model_name='carousel',
name='header_height',
),
migrations.AlterField(
model_name='carousel',
name='slide_limit',
field=models.IntegerField(default=10, verbose_name='slide limit', validators=[django.core.validators.MinValueValidator(1)], help_text='Specify the maximum # of slides to display.'),
preserve_default=True,
),
migrations.AlterField(
model_name='slide',
name='other_url_color',
field=models.CharField(default='default', choices=[('default', 'White'), ('primary', 'Blue'), ('info', 'Light Blue'), ('success', 'Green'), ('warning', 'Orange'), ('danger', 'Red')], blank=True, max_length=7, verbose_name='other URL color', help_text="If there is another relevant URL for this slide, choose the color for it's button."),
preserve_default=True,
),
migrations.AlterField(
model_name='slide',
name='publish_datetime',
field=models.DateTimeField(default=datetime.datetime(2015, 4, 17, 19, 40, 36, 175156, tzinfo=utc), verbose_name='date/time slide published', help_text='Choose date/time to publish slide. Slides are displayed in reverse-chronological order, so this can be used to control their order. A future date will be hide a slide until that date.'),
preserve_default=True,
),
]
|
Make migrations for changes to models
|
Make migrations for changes to models
|
Python
|
bsd-3-clause
|
mfcovington/djangocms-lab-carousel,mfcovington/djangocms-lab-carousel
|
---
+++
@@ -0,0 +1,43 @@
+# -*- coding: utf-8 -*-
+from __future__ import unicode_literals
+
+from django.db import models, migrations
+from django.utils.timezone import utc
+import django.core.validators
+import datetime
+
+
+class Migration(migrations.Migration):
+
+ dependencies = [
+ ('cms_lab_carousel', '0002_auto_20150417_1219'),
+ ]
+
+ operations = [
+ migrations.RemoveField(
+ model_name='carousel',
+ name='footer_height',
+ ),
+ migrations.RemoveField(
+ model_name='carousel',
+ name='header_height',
+ ),
+ migrations.AlterField(
+ model_name='carousel',
+ name='slide_limit',
+ field=models.IntegerField(default=10, verbose_name='slide limit', validators=[django.core.validators.MinValueValidator(1)], help_text='Specify the maximum # of slides to display.'),
+ preserve_default=True,
+ ),
+ migrations.AlterField(
+ model_name='slide',
+ name='other_url_color',
+ field=models.CharField(default='default', choices=[('default', 'White'), ('primary', 'Blue'), ('info', 'Light Blue'), ('success', 'Green'), ('warning', 'Orange'), ('danger', 'Red')], blank=True, max_length=7, verbose_name='other URL color', help_text="If there is another relevant URL for this slide, choose the color for it's button."),
+ preserve_default=True,
+ ),
+ migrations.AlterField(
+ model_name='slide',
+ name='publish_datetime',
+ field=models.DateTimeField(default=datetime.datetime(2015, 4, 17, 19, 40, 36, 175156, tzinfo=utc), verbose_name='date/time slide published', help_text='Choose date/time to publish slide. Slides are displayed in reverse-chronological order, so this can be used to control their order. A future date will be hide a slide until that date.'),
+ preserve_default=True,
+ ),
+ ]
|
|
3eefb913a11a91cfe543b5efe926e233953e6b0c
|
tests/test_usfirst_event_type_parser.py
|
tests/test_usfirst_event_type_parser.py
|
import unittest2
from models.event import Event
from helpers.event_helper import EventHelper
class TestUsfirstEventTypeParser(unittest2.TestCase):
def test_parse(self):
self.assertEqual(EventHelper.parseEventType("Regional"), Event.REGIONAL)
self.assertEqual(EventHelper.parseEventType("regional"), Event.REGIONAL)
self.assertEqual(EventHelper.parseEventType("District"), Event.DISTRICT)
self.assertEqual(EventHelper.parseEventType("district"), Event.DISTRICT)
self.assertEqual(EventHelper.parseEventType("MI District"), Event.DISTRICT)
self.assertEqual(EventHelper.parseEventType("District Event"), Event.DISTRICT)
self.assertEqual(EventHelper.parseEventType("Qualifying Event"), Event.DISTRICT)
self.assertEqual(EventHelper.parseEventType("Qualifier"), Event.DISTRICT)
self.assertEqual(EventHelper.parseEventType("District Championship"), Event.DISTRICT_CMP)
self.assertEqual(EventHelper.parseEventType("MI FRC State Championship"), Event.DISTRICT_CMP)
self.assertEqual(EventHelper.parseEventType("Qualifying Championship"), Event.DISTRICT_CMP)
self.assertEqual(EventHelper.parseEventType("Championship Division"), Event.CMP_DIVISION)
self.assertEqual(EventHelper.parseEventType("Division"), Event.CMP_DIVISION)
self.assertEqual(EventHelper.parseEventType("Championship Finals"), Event.CMP_FINALS)
self.assertEqual(EventHelper.parseEventType("Championship"), Event.CMP_FINALS)
|
Add test for event type parser
|
Add test for event type parser
|
Python
|
mit
|
josephbisch/the-blue-alliance,synth3tk/the-blue-alliance,verycumbersome/the-blue-alliance,fangeugene/the-blue-alliance,bdaroz/the-blue-alliance,fangeugene/the-blue-alliance,1fish2/the-blue-alliance,phil-lopreiato/the-blue-alliance,phil-lopreiato/the-blue-alliance,tsteward/the-blue-alliance,tsteward/the-blue-alliance,1fish2/the-blue-alliance,phil-lopreiato/the-blue-alliance,josephbisch/the-blue-alliance,bdaroz/the-blue-alliance,bdaroz/the-blue-alliance,jaredhasenklein/the-blue-alliance,josephbisch/the-blue-alliance,bvisness/the-blue-alliance,nwalters512/the-blue-alliance,1fish2/the-blue-alliance,bvisness/the-blue-alliance,fangeugene/the-blue-alliance,phil-lopreiato/the-blue-alliance,nwalters512/the-blue-alliance,jaredhasenklein/the-blue-alliance,verycumbersome/the-blue-alliance,jaredhasenklein/the-blue-alliance,1fish2/the-blue-alliance,phil-lopreiato/the-blue-alliance,tsteward/the-blue-alliance,tsteward/the-blue-alliance,bvisness/the-blue-alliance,the-blue-alliance/the-blue-alliance,tsteward/the-blue-alliance,nwalters512/the-blue-alliance,synth3tk/the-blue-alliance,synth3tk/the-blue-alliance,fangeugene/the-blue-alliance,synth3tk/the-blue-alliance,the-blue-alliance/the-blue-alliance,1fish2/the-blue-alliance,josephbisch/the-blue-alliance,phil-lopreiato/the-blue-alliance,bdaroz/the-blue-alliance,bdaroz/the-blue-alliance,josephbisch/the-blue-alliance,bvisness/the-blue-alliance,the-blue-alliance/the-blue-alliance,josephbisch/the-blue-alliance,synth3tk/the-blue-alliance,verycumbersome/the-blue-alliance,jaredhasenklein/the-blue-alliance,jaredhasenklein/the-blue-alliance,bdaroz/the-blue-alliance,nwalters512/the-blue-alliance,the-blue-alliance/the-blue-alliance,the-blue-alliance/the-blue-alliance,verycumbersome/the-blue-alliance,verycumbersome/the-blue-alliance,bvisness/the-blue-alliance,verycumbersome/the-blue-alliance,fangeugene/the-blue-alliance,jaredhasenklein/the-blue-alliance,tsteward/the-blue-alliance,nwalters512/the-blue-alliance,bvisness/the-blue-alliance,fangeugene/the-blue-alliance,nwalters512/the-blue-alliance,the-blue-alliance/the-blue-alliance,synth3tk/the-blue-alliance,1fish2/the-blue-alliance
|
---
+++
@@ -0,0 +1,26 @@
+import unittest2
+
+from models.event import Event
+from helpers.event_helper import EventHelper
+
+class TestUsfirstEventTypeParser(unittest2.TestCase):
+ def test_parse(self):
+ self.assertEqual(EventHelper.parseEventType("Regional"), Event.REGIONAL)
+ self.assertEqual(EventHelper.parseEventType("regional"), Event.REGIONAL)
+
+ self.assertEqual(EventHelper.parseEventType("District"), Event.DISTRICT)
+ self.assertEqual(EventHelper.parseEventType("district"), Event.DISTRICT)
+ self.assertEqual(EventHelper.parseEventType("MI District"), Event.DISTRICT)
+ self.assertEqual(EventHelper.parseEventType("District Event"), Event.DISTRICT)
+ self.assertEqual(EventHelper.parseEventType("Qualifying Event"), Event.DISTRICT)
+ self.assertEqual(EventHelper.parseEventType("Qualifier"), Event.DISTRICT)
+
+ self.assertEqual(EventHelper.parseEventType("District Championship"), Event.DISTRICT_CMP)
+ self.assertEqual(EventHelper.parseEventType("MI FRC State Championship"), Event.DISTRICT_CMP)
+ self.assertEqual(EventHelper.parseEventType("Qualifying Championship"), Event.DISTRICT_CMP)
+
+ self.assertEqual(EventHelper.parseEventType("Championship Division"), Event.CMP_DIVISION)
+ self.assertEqual(EventHelper.parseEventType("Division"), Event.CMP_DIVISION)
+
+ self.assertEqual(EventHelper.parseEventType("Championship Finals"), Event.CMP_FINALS)
+ self.assertEqual(EventHelper.parseEventType("Championship"), Event.CMP_FINALS)
|
|
bc6a9324c00909a62dc26477224bbfc58def9eb2
|
external_tools/src/main/python/omero56/scripts/omerok8s_impc_config.py
|
external_tools/src/main/python/omero56/scripts/omerok8s_impc_config.py
|
"""Configure Omero on a k8s server for the way it is used in IMPC
"""
import sys
import argparse
import omero.cli # For some reason if I do not do this next line throws error
from omero import ApiUsageException
from omero.cli import NonZeroReturnCode
from omeroservice import OmeroService
from utils import get_properties_from_configuration_file
argument_parser = argparse.ArgumentParser(
description="Configure Omero on a k8s server for use in IMPC"
)
argument_parser.add_argument("--omero-props-path", required=True,
help="Path to configuration file"
)
args = argument_parser.parse_args()
omero_props = get_properties_from_configuration_file(args.omero_props_path)
omero_host = omero_props['omerohost']
omero_port = omero_props['omeroport']
omero_root_user = omero_props['omerouser']
omero_root_pass = omero_props['omeropass']
omero_public_user = omero_props['omeropublicuser']
omero_public_group = omero_props['omerogroup']
omero_public_pass = omero_props['omeropublicpass']
omero_system_group = omero_props['omerosystemgroup']
omero_service = OmeroService(omero_host, omero_port, omero_root_user, omero_root_pass, omero_system_group)
cli = omero_service.cli
def run_command(cli, cmd):
"""Run a command in the omero cli. Exit if error
"""
try:
cli.invoke(cmd, strict=True)
except (ApiUsageException, NonZeroReturnCode,) as e:
print(f"Error running command {cmd}.\nError was: {e}")
print("\nExiting")
sys.exit(-1)
# Create the public group
cmd = f"group add --type read-only --ignore-existing {omero_public_group}"
run_command(cli, cmd)
# Create the public user
cmd = f"user add --ignore-existing -i EBI " + \
f"--group-name {omero_public_group} -P {omero_public_pass} " + \
f"{omero_public_user} public user"
run_command(cli, cmd)
# Ensure the webclient allows login by the public user without any password
# Ensure the root user is part of the public group - so can import/export images
cmd = f"user joingroup --name {omero_public_user} --group-name {omero_public_group} --as-owner"
run_command(cli, cmd)
|
Add script to configure omerok8s for public_group and public_user
|
Add script to configure omerok8s for public_group and public_user
|
Python
|
apache-2.0
|
mpi2/PhenotypeData,mpi2/PhenotypeData,mpi2/PhenotypeData,mpi2/PhenotypeData,mpi2/PhenotypeData,mpi2/PhenotypeData
|
---
+++
@@ -0,0 +1,61 @@
+"""Configure Omero on a k8s server for the way it is used in IMPC
+
+"""
+import sys
+import argparse
+
+import omero.cli # For some reason if I do not do this next line throws error
+from omero import ApiUsageException
+from omero.cli import NonZeroReturnCode
+
+from omeroservice import OmeroService
+from utils import get_properties_from_configuration_file
+
+argument_parser = argparse.ArgumentParser(
+ description="Configure Omero on a k8s server for use in IMPC"
+)
+argument_parser.add_argument("--omero-props-path", required=True,
+ help="Path to configuration file"
+)
+
+args = argument_parser.parse_args()
+
+omero_props = get_properties_from_configuration_file(args.omero_props_path)
+omero_host = omero_props['omerohost']
+omero_port = omero_props['omeroport']
+omero_root_user = omero_props['omerouser']
+omero_root_pass = omero_props['omeropass']
+omero_public_user = omero_props['omeropublicuser']
+omero_public_group = omero_props['omerogroup']
+omero_public_pass = omero_props['omeropublicpass']
+omero_system_group = omero_props['omerosystemgroup']
+
+omero_service = OmeroService(omero_host, omero_port, omero_root_user, omero_root_pass, omero_system_group)
+cli = omero_service.cli
+
+def run_command(cli, cmd):
+ """Run a command in the omero cli. Exit if error
+
+ """
+ try:
+ cli.invoke(cmd, strict=True)
+ except (ApiUsageException, NonZeroReturnCode,) as e:
+ print(f"Error running command {cmd}.\nError was: {e}")
+ print("\nExiting")
+ sys.exit(-1)
+
+# Create the public group
+cmd = f"group add --type read-only --ignore-existing {omero_public_group}"
+run_command(cli, cmd)
+
+# Create the public user
+cmd = f"user add --ignore-existing -i EBI " + \
+ f"--group-name {omero_public_group} -P {omero_public_pass} " + \
+ f"{omero_public_user} public user"
+run_command(cli, cmd)
+
+# Ensure the webclient allows login by the public user without any password
+
+# Ensure the root user is part of the public group - so can import/export images
+cmd = f"user joingroup --name {omero_public_user} --group-name {omero_public_group} --as-owner"
+run_command(cli, cmd)
|
|
9cd5678fbeb3ad5a26bf9578a1f562c46a2de26e
|
example_iterator_with_custom_order.py
|
example_iterator_with_custom_order.py
|
# -*- coding: utf-8 -*-
"""
Created on Sat Jan 21 21:16:43 2017
@author: ryuhei
"""
import numpy as np
from chainer.datasets import TupleDataset
from sklearn.preprocessing import LabelEncoder
from my_iterators import SerialIterator
class NPairMCIndexesSampler(object):
def __init__(self, labels, batch_size, num_batches):
assert len(labels) >= (batch_size * num_batches), (
"batch_size * num_batches must not exceed the number of examples")
self._labels = np.array(labels).flatten()
self._label_encoder = LabelEncoder().fit(self._labels)
self._classes = self._label_encoder.classes_
self.num_classes = len(self._classes)
assert batch_size % 2 == 0, ("batch_size must be even number.")
assert batch_size <= self.num_classes * 2, (
"batch_size must not exceed twice the number of classes"
"(i.e. set batch_size <= {}).".format(self.num_classes * 2))
self.batch_size = batch_size
self.num_batches = num_batches
self._class_to_indexes = []
for c in self._classes:
self._class_to_indexes.append(
np.argwhere(self._labels == c).ravel())
def __call__(self):
indexes = []
for _ in range(self.num_batches):
random_classes = np.random.choice(
self.num_classes, self.batch_size / 2, False)
anchor_indexes = []
positive_indexes = []
for c in random_classes:
a, p = np.random.choice(self._class_to_indexes[c], 2, False)
anchor_indexes.append(a)
positive_indexes.append(p)
indexes.append(anchor_indexes)
indexes.append(positive_indexes)
return np.concatenate(indexes)
if __name__ == '__main__':
batch_size = 10
num_batches = 5
repeat = True
labels = np.array(sum([[i]*10 for i in range(10)], []))
num_examples = len(labels)
x = np.arange(num_examples)
dataset = TupleDataset(x, labels)
indexes_sampler = NPairMCIndexesSampler(labels, batch_size, num_batches)
it = SerialIterator(dataset, batch_size, repeat=repeat,
order_sampler=indexes_sampler)
for i in range(num_batches*2):
batch = next(it)
print len(batch)
print batch[:batch_size/2]
print batch[batch_size/2:]
print
|
Implement an example of the modified SerialIterator
|
Implement an example of the modified SerialIterator
|
Python
|
mit
|
ronekko/deep_metric_learning
|
---
+++
@@ -0,0 +1,70 @@
+# -*- coding: utf-8 -*-
+"""
+Created on Sat Jan 21 21:16:43 2017
+
+@author: ryuhei
+"""
+
+import numpy as np
+
+from chainer.datasets import TupleDataset
+from sklearn.preprocessing import LabelEncoder
+from my_iterators import SerialIterator
+
+
+class NPairMCIndexesSampler(object):
+ def __init__(self, labels, batch_size, num_batches):
+ assert len(labels) >= (batch_size * num_batches), (
+ "batch_size * num_batches must not exceed the number of examples")
+ self._labels = np.array(labels).flatten()
+ self._label_encoder = LabelEncoder().fit(self._labels)
+ self._classes = self._label_encoder.classes_
+ self.num_classes = len(self._classes)
+ assert batch_size % 2 == 0, ("batch_size must be even number.")
+ assert batch_size <= self.num_classes * 2, (
+ "batch_size must not exceed twice the number of classes"
+ "(i.e. set batch_size <= {}).".format(self.num_classes * 2))
+ self.batch_size = batch_size
+ self.num_batches = num_batches
+
+ self._class_to_indexes = []
+ for c in self._classes:
+ self._class_to_indexes.append(
+ np.argwhere(self._labels == c).ravel())
+
+ def __call__(self):
+ indexes = []
+ for _ in range(self.num_batches):
+ random_classes = np.random.choice(
+ self.num_classes, self.batch_size / 2, False)
+ anchor_indexes = []
+ positive_indexes = []
+ for c in random_classes:
+ a, p = np.random.choice(self._class_to_indexes[c], 2, False)
+ anchor_indexes.append(a)
+ positive_indexes.append(p)
+ indexes.append(anchor_indexes)
+ indexes.append(positive_indexes)
+ return np.concatenate(indexes)
+
+
+if __name__ == '__main__':
+ batch_size = 10
+ num_batches = 5
+ repeat = True
+
+ labels = np.array(sum([[i]*10 for i in range(10)], []))
+ num_examples = len(labels)
+ x = np.arange(num_examples)
+ dataset = TupleDataset(x, labels)
+
+ indexes_sampler = NPairMCIndexesSampler(labels, batch_size, num_batches)
+ it = SerialIterator(dataset, batch_size, repeat=repeat,
+ order_sampler=indexes_sampler)
+
+ for i in range(num_batches*2):
+ batch = next(it)
+ print len(batch)
+ print batch[:batch_size/2]
+ print batch[batch_size/2:]
+ print
|
|
8768b66ae982d19964f85feb325a1b0f35ed1c87
|
odo/backends/dask_array.py
|
odo/backends/dask_array.py
|
from __future__ import absolute_import, division, print_function
import numpy as np
from datashape import discover
from toolz import merge, accumulate
from datashape.dispatch import dispatch
from datashape import DataShape
from operator import add
import itertools
from dask.array.core import rec_concatenate, Array, getem, get, names, from_array
from dask.core import flatten
from dask.compatibility import long
from ..append import append
from ..convert import convert
from ..resource import resource
@discover.register(Array)
def discover_dask_array(a, **kwargs):
block = a._get_block(*([0] * a.ndim))
return DataShape(*(a.shape + (discover(block).measure,)))
arrays = [np.ndarray]
try:
import h5py
arrays.append(h5py.Dataset)
@dispatch(h5py.Dataset, (int, long))
def resize(x, size):
s = list(x.shape)
s[0] = size
return resize(x, tuple(s))
@dispatch(h5py.Dataset, tuple)
def resize(x, shape):
return x.resize(shape)
except ImportError:
pass
try:
import bcolz
arrays.append(bcolz.carray)
@dispatch(bcolz.carray, (int, long))
def resize(x, size):
return x.resize(size)
except ImportError:
pass
@convert.register(Array, tuple(arrays), cost=0.01)
def array_to_dask(x, name=None, blockshape=None, **kwargs):
return from_array(x, blockshape=blockshape, name=name, **kwargs)
@convert.register(np.ndarray, Array, cost=0.5)
def dask_to_numpy(x, **kwargs):
return rec_concatenate(get(x.dask, x._keys(), **kwargs))
@convert.register(float, Array, cost=0.5)
def dask_to_float(x, **kwargs):
return x.compute()
@append.register(tuple(arrays), Array)
def store_Array_in_ooc_data(out, arr, inplace=False, **kwargs):
if not inplace:
# Resize output dataset to accept new data
assert out.shape[1:] == arr.shape[1:]
resize(out, out.shape[0] + arr.shape[0]) # elongate
return arr.store(out)
|
Migrate dask array odo backend from dask.
|
Migrate dask array odo backend from dask.
|
Python
|
bsd-3-clause
|
cpcloud/odo,ContinuumIO/odo,ywang007/odo,cpcloud/odo,alexmojaki/odo,quantopian/odo,ContinuumIO/odo,Dannnno/odo,cowlicks/odo,quantopian/odo,blaze/odo,cowlicks/odo,alexmojaki/odo,Dannnno/odo,ywang007/odo,blaze/odo
|
---
+++
@@ -0,0 +1,74 @@
+from __future__ import absolute_import, division, print_function
+
+import numpy as np
+from datashape import discover
+from toolz import merge, accumulate
+from datashape.dispatch import dispatch
+from datashape import DataShape
+from operator import add
+import itertools
+
+from dask.array.core import rec_concatenate, Array, getem, get, names, from_array
+from dask.core import flatten
+from dask.compatibility import long
+
+from ..append import append
+from ..convert import convert
+from ..resource import resource
+
+@discover.register(Array)
+def discover_dask_array(a, **kwargs):
+ block = a._get_block(*([0] * a.ndim))
+ return DataShape(*(a.shape + (discover(block).measure,)))
+
+
+arrays = [np.ndarray]
+try:
+ import h5py
+ arrays.append(h5py.Dataset)
+
+ @dispatch(h5py.Dataset, (int, long))
+ def resize(x, size):
+ s = list(x.shape)
+ s[0] = size
+ return resize(x, tuple(s))
+
+ @dispatch(h5py.Dataset, tuple)
+ def resize(x, shape):
+ return x.resize(shape)
+except ImportError:
+ pass
+try:
+ import bcolz
+ arrays.append(bcolz.carray)
+
+ @dispatch(bcolz.carray, (int, long))
+ def resize(x, size):
+ return x.resize(size)
+except ImportError:
+ pass
+
+
+
+@convert.register(Array, tuple(arrays), cost=0.01)
+def array_to_dask(x, name=None, blockshape=None, **kwargs):
+ return from_array(x, blockshape=blockshape, name=name, **kwargs)
+
+
+@convert.register(np.ndarray, Array, cost=0.5)
+def dask_to_numpy(x, **kwargs):
+ return rec_concatenate(get(x.dask, x._keys(), **kwargs))
+
+
+@convert.register(float, Array, cost=0.5)
+def dask_to_float(x, **kwargs):
+ return x.compute()
+
+
+@append.register(tuple(arrays), Array)
+def store_Array_in_ooc_data(out, arr, inplace=False, **kwargs):
+ if not inplace:
+ # Resize output dataset to accept new data
+ assert out.shape[1:] == arr.shape[1:]
+ resize(out, out.shape[0] + arr.shape[0]) # elongate
+ return arr.store(out)
|
|
63ffa531eebfba19344dee67b3f417072012a7f4
|
CodeFights/rangeBitCount.py
|
CodeFights/rangeBitCount.py
|
#!/usr/local/bin/python
# Code Fights Range Bit Count (Core) Problem
def rangeBitCount(a, b):
return (''.join([bin(n) for n in range(a, b + 1)])).count('1')
def main():
tests = [
[2, 7, 11],
[0, 1, 1],
[1, 10, 17],
[8, 9, 3],
[9, 10, 4]
]
for t in tests:
res = rangeBitCount(t[0], t[1])
if t[2] == res:
print("PASSED: rangeBitCount({}, {}) returned {}"
.format(t[0], t[1], res))
else:
print(("FAILED: rangeBitCount({}, {}) returned {},"
"answer: {}").format(t[0], t[1], res, t[2]))
if __name__ == '__main__':
main()
|
Solve Code Fights range bit count problem
|
Solve Code Fights range bit count problem
|
Python
|
mit
|
HKuz/Test_Code
|
---
+++
@@ -0,0 +1,29 @@
+#!/usr/local/bin/python
+# Code Fights Range Bit Count (Core) Problem
+
+
+def rangeBitCount(a, b):
+ return (''.join([bin(n) for n in range(a, b + 1)])).count('1')
+
+
+def main():
+ tests = [
+ [2, 7, 11],
+ [0, 1, 1],
+ [1, 10, 17],
+ [8, 9, 3],
+ [9, 10, 4]
+ ]
+
+ for t in tests:
+ res = rangeBitCount(t[0], t[1])
+ if t[2] == res:
+ print("PASSED: rangeBitCount({}, {}) returned {}"
+ .format(t[0], t[1], res))
+ else:
+ print(("FAILED: rangeBitCount({}, {}) returned {},"
+ "answer: {}").format(t[0], t[1], res, t[2]))
+
+
+if __name__ == '__main__':
+ main()
|
|
cca072b6cf5b0162e1cf10d6873739d762a7a05e
|
examples/console_logger_signals.py
|
examples/console_logger_signals.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
Log radiation hits or noise signals to the console.
Released under MIT License. See LICENSE file.
By Yoan Tournade <y@yoantournade.com>
"""
import time
from PiPocketGeiger import RadiationWatch
def example_run_context():
example_run_context.hit_flag = False
def onRadiation():
example_run_context.hit_flag = True
print("Ray appeared!")
def onNoise():
print("Noisy and moving around here!")
# Create the RadiationWatch object, specifying the used GPIO pins ...
with RadiationWatch(24, 23) as radiation_watch:
print("Waiting for gamma rays to hit the Pocket Geiger.")
radiation_watch.register_radiation_callback(onRadiation)
radiation_watch.register_noise_callback(onNoise)
while 1:
# ... and print readings after radiation hit.
if example_run_context.hit_flag:
print(radiation_watch.status())
example_run_context.hit_flag = False
time.sleep(3)
if __name__ == "__main__":
example_run_context()
|
Add example using callbacks/IRQs for easier debug
|
Add example using callbacks/IRQs for easier debug
|
Python
|
mit
|
MonsieurV/PiPocketGeiger
|
---
+++
@@ -0,0 +1,36 @@
+#!/usr/bin/env python
+# -*- coding: utf-8 -*-
+"""
+Log radiation hits or noise signals to the console.
+
+Released under MIT License. See LICENSE file.
+
+By Yoan Tournade <y@yoantournade.com>
+"""
+import time
+from PiPocketGeiger import RadiationWatch
+
+def example_run_context():
+ example_run_context.hit_flag = False
+
+ def onRadiation():
+ example_run_context.hit_flag = True
+ print("Ray appeared!")
+
+ def onNoise():
+ print("Noisy and moving around here!")
+
+ # Create the RadiationWatch object, specifying the used GPIO pins ...
+ with RadiationWatch(24, 23) as radiation_watch:
+ print("Waiting for gamma rays to hit the Pocket Geiger.")
+ radiation_watch.register_radiation_callback(onRadiation)
+ radiation_watch.register_noise_callback(onNoise)
+ while 1:
+ # ... and print readings after radiation hit.
+ if example_run_context.hit_flag:
+ print(radiation_watch.status())
+ example_run_context.hit_flag = False
+ time.sleep(3)
+
+if __name__ == "__main__":
+ example_run_context()
|
|
fe3f7ae8eb9390a4fe3f59e6244d4bbd6af7a9cd
|
mojo/services/html_viewer/view_url.py
|
mojo/services/html_viewer/view_url.py
|
#!/usr/bin/env python
# Copyright 2015 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import argparse
import os
import subprocess
import sys
root_path = os.path.realpath(
os.path.join(
os.path.dirname(
os.path.realpath(__file__)),
os.pardir,
os.pardir,
os.pardir))
def _BuildShellCommand(args):
sdk_version = subprocess.check_output(["cat",
"third_party/mojo/src/mojo/public/VERSION"], cwd=root_path)
build_dir = os.path.join(root_path, args.build_dir)
shell_command = [os.path.join(build_dir, "mojo_shell")]
options = []
options.append(
"--origin=https://storage.googleapis.com/mojo/services/linux-x64/%s" %
sdk_version)
options.append("--url-mappings=mojo:html_viewer=file://%s/html_viewer.mojo" %
build_dir)
options.append('--args-for=mojo:kiosk_wm %s' % args.url)
app_to_run = "mojo:kiosk_wm"
return shell_command + options + [app_to_run]
def main():
parser = argparse.ArgumentParser(
description="View a URL with HTMLViewer in the Kiosk window manager. "
"You must have built //mojo/services/html_viewer and "
"//mojo/services/network first. Note that this will "
"currently often fail spectacularly due to lack of binary "
"stability in Mojo.")
parser.add_argument(
"--build-dir",
help="Path to the dir containing the linux-x64 binaries relative to the "
"repo root (default: %(default)s)",
default="out/Release")
parser.add_argument("url",
help="The URL to be viewed")
args = parser.parse_args()
return subprocess.call(_BuildShellCommand(args))
if __name__ == '__main__':
sys.exit(main())
|
Add script to view URL in HTMLViewer
|
Add script to view URL in HTMLViewer
This script takes advantage of the fact that Mojo binaries are published in the
cloud to add functionality for viewing a URL in HTMLViewer embedded in the
kiosk window manager.
Review URL: https://codereview.chromium.org/982523004
Cr-Commit-Position: 972c6d2dc6dd5efdad1377c0d224e03eb8f276f7@{#319085}
|
Python
|
bsd-3-clause
|
TheTypoMaster/chromium-crosswalk,chuan9/chromium-crosswalk,chuan9/chromium-crosswalk,Chilledheart/chromium,ltilve/chromium,hgl888/chromium-crosswalk,hgl888/chromium-crosswalk,fujunwei/chromium-crosswalk,Fireblend/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,chuan9/chromium-crosswalk,Just-D/chromium-1,Just-D/chromium-1,fujunwei/chromium-crosswalk,hgl888/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,hgl888/chromium-crosswalk,fujunwei/chromium-crosswalk,hgl888/chromium-crosswalk,hgl888/chromium-crosswalk,Pluto-tv/chromium-crosswalk,Fireblend/chromium-crosswalk,Chilledheart/chromium,Just-D/chromium-1,Pluto-tv/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,Just-D/chromium-1,chuan9/chromium-crosswalk,fujunwei/chromium-crosswalk,chuan9/chromium-crosswalk,axinging/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,Fireblend/chromium-crosswalk,ltilve/chromium,Pluto-tv/chromium-crosswalk,fujunwei/chromium-crosswalk,Pluto-tv/chromium-crosswalk,Fireblend/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,Fireblend/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,axinging/chromium-crosswalk,Chilledheart/chromium,hgl888/chromium-crosswalk,chuan9/chromium-crosswalk,fujunwei/chromium-crosswalk,Just-D/chromium-1,axinging/chromium-crosswalk,Chilledheart/chromium,axinging/chromium-crosswalk,Fireblend/chromium-crosswalk,Fireblend/chromium-crosswalk,axinging/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,Pluto-tv/chromium-crosswalk,axinging/chromium-crosswalk,Chilledheart/chromium,Fireblend/chromium-crosswalk,ltilve/chromium,Chilledheart/chromium,ltilve/chromium,Pluto-tv/chromium-crosswalk,ltilve/chromium,Chilledheart/chromium,chuan9/chromium-crosswalk,axinging/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,axinging/chromium-crosswalk,ltilve/chromium,ltilve/chromium,Just-D/chromium-1,Pluto-tv/chromium-crosswalk,Just-D/chromium-1,fujunwei/chromium-crosswalk,axinging/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,ltilve/chromium,TheTypoMaster/chromium-crosswalk,fujunwei/chromium-crosswalk,chuan9/chromium-crosswalk,chuan9/chromium-crosswalk,Chilledheart/chromium,Pluto-tv/chromium-crosswalk,fujunwei/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,ltilve/chromium,axinging/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,hgl888/chromium-crosswalk,Pluto-tv/chromium-crosswalk,Just-D/chromium-1,TheTypoMaster/chromium-crosswalk,axinging/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,Chilledheart/chromium,hgl888/chromium-crosswalk,Fireblend/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,Just-D/chromium-1
|
---
+++
@@ -0,0 +1,57 @@
+#!/usr/bin/env python
+# Copyright 2015 The Chromium Authors. All rights reserved.
+# Use of this source code is governed by a BSD-style license that can be
+# found in the LICENSE file.
+
+import argparse
+import os
+import subprocess
+import sys
+
+root_path = os.path.realpath(
+ os.path.join(
+ os.path.dirname(
+ os.path.realpath(__file__)),
+ os.pardir,
+ os.pardir,
+ os.pardir))
+
+def _BuildShellCommand(args):
+ sdk_version = subprocess.check_output(["cat",
+ "third_party/mojo/src/mojo/public/VERSION"], cwd=root_path)
+ build_dir = os.path.join(root_path, args.build_dir)
+
+ shell_command = [os.path.join(build_dir, "mojo_shell")]
+
+ options = []
+ options.append(
+ "--origin=https://storage.googleapis.com/mojo/services/linux-x64/%s" %
+ sdk_version)
+ options.append("--url-mappings=mojo:html_viewer=file://%s/html_viewer.mojo" %
+ build_dir)
+ options.append('--args-for=mojo:kiosk_wm %s' % args.url)
+
+ app_to_run = "mojo:kiosk_wm"
+
+ return shell_command + options + [app_to_run]
+
+def main():
+ parser = argparse.ArgumentParser(
+ description="View a URL with HTMLViewer in the Kiosk window manager. "
+ "You must have built //mojo/services/html_viewer and "
+ "//mojo/services/network first. Note that this will "
+ "currently often fail spectacularly due to lack of binary "
+ "stability in Mojo.")
+ parser.add_argument(
+ "--build-dir",
+ help="Path to the dir containing the linux-x64 binaries relative to the "
+ "repo root (default: %(default)s)",
+ default="out/Release")
+ parser.add_argument("url",
+ help="The URL to be viewed")
+
+ args = parser.parse_args()
+ return subprocess.call(_BuildShellCommand(args))
+
+if __name__ == '__main__':
+ sys.exit(main())
|
|
6a835fd8913cdd3a9dc76530230ae2c73d88b48f
|
tests/name_injection_test.py
|
tests/name_injection_test.py
|
"""Test for the name inject utility."""
from drudge import Drudge
def test_drudge_injects_names():
"""Test the name injection method of drudge."""
dr = Drudge(None) # Dummy drudge.
string_name = 'string_name'
dr.set_name(string_name)
dr.set_name(1, 'one')
dr.inject_names(suffix='_')
assert string_name_ == string_name
assert one_ == 1
|
Add tests for name injection
|
Add tests for name injection
Since the name injection facility could taint the entire global
namespace of the module, its test is put into a separate module.
|
Python
|
mit
|
tschijnmo/drudge,tschijnmo/drudge,tschijnmo/drudge
|
---
+++
@@ -0,0 +1,17 @@
+"""Test for the name inject utility."""
+
+from drudge import Drudge
+
+
+def test_drudge_injects_names():
+ """Test the name injection method of drudge."""
+
+ dr = Drudge(None) # Dummy drudge.
+ string_name = 'string_name'
+ dr.set_name(string_name)
+ dr.set_name(1, 'one')
+
+ dr.inject_names(suffix='_')
+
+ assert string_name_ == string_name
+ assert one_ == 1
|
|
69aa2be4eca4ecfa9a73ad38c34bb7a4e46bef97
|
tests/test_epsilon_greedy.py
|
tests/test_epsilon_greedy.py
|
from __future__ import unicode_literals
from __future__ import print_function
from __future__ import division
from __future__ import absolute_import
from future import standard_library
standard_library.install_aliases()
import logging
import unittest
from explorers import epsilon_greedy
class TestEpsilonGreedy(unittest.TestCase):
def test_linear_decay_epsilon_greedy(self):
random_action_func_count = [0]
greedy_action_func_count = [0]
def random_action_func():
random_action_func_count[0] += 1
return 0
def greedy_action_func():
greedy_action_func_count[0] += 1
return 0
explorer = epsilon_greedy.LinearDecayEpsilonGreedy(1.0, 0.1, 50,
random_action_func)
explorer.logger.addHandler(logging.StreamHandler())
explorer.logger.setLevel(logging.DEBUG)
self.assertAlmostEqual(explorer.epsilon, 1.0)
for t in range(100):
explorer.select_action(t, greedy_action_func)
self.assertAlmostEqual(explorer.epsilon, 0.1)
def test_constant_epsilon_greedy(self):
random_action_func_count = [0]
greedy_action_func_count = [0]
def random_action_func():
random_action_func_count[0] += 1
return 0
def greedy_action_func():
greedy_action_func_count[0] += 1
return 0
explorer = epsilon_greedy.ConstantEpsilonGreedy(0.1,
random_action_func)
explorer.logger.addHandler(logging.StreamHandler())
explorer.logger.setLevel(logging.DEBUG)
self.assertAlmostEqual(explorer.epsilon, 0.1)
for t in range(100):
explorer.select_action(t, greedy_action_func)
self.assertAlmostEqual(explorer.epsilon, 0.1)
|
Add tests for epsilon greedy explorers
|
Add tests for epsilon greedy explorers
|
Python
|
mit
|
toslunar/chainerrl,toslunar/chainerrl
|
---
+++
@@ -0,0 +1,66 @@
+from __future__ import unicode_literals
+from __future__ import print_function
+from __future__ import division
+from __future__ import absolute_import
+from future import standard_library
+standard_library.install_aliases()
+
+import logging
+import unittest
+
+from explorers import epsilon_greedy
+
+
+class TestEpsilonGreedy(unittest.TestCase):
+
+ def test_linear_decay_epsilon_greedy(self):
+
+ random_action_func_count = [0]
+ greedy_action_func_count = [0]
+
+ def random_action_func():
+ random_action_func_count[0] += 1
+ return 0
+
+ def greedy_action_func():
+ greedy_action_func_count[0] += 1
+ return 0
+
+ explorer = epsilon_greedy.LinearDecayEpsilonGreedy(1.0, 0.1, 50,
+ random_action_func)
+
+ explorer.logger.addHandler(logging.StreamHandler())
+ explorer.logger.setLevel(logging.DEBUG)
+
+ self.assertAlmostEqual(explorer.epsilon, 1.0)
+
+ for t in range(100):
+ explorer.select_action(t, greedy_action_func)
+
+ self.assertAlmostEqual(explorer.epsilon, 0.1)
+
+ def test_constant_epsilon_greedy(self):
+
+ random_action_func_count = [0]
+ greedy_action_func_count = [0]
+
+ def random_action_func():
+ random_action_func_count[0] += 1
+ return 0
+
+ def greedy_action_func():
+ greedy_action_func_count[0] += 1
+ return 0
+
+ explorer = epsilon_greedy.ConstantEpsilonGreedy(0.1,
+ random_action_func)
+
+ explorer.logger.addHandler(logging.StreamHandler())
+ explorer.logger.setLevel(logging.DEBUG)
+
+ self.assertAlmostEqual(explorer.epsilon, 0.1)
+
+ for t in range(100):
+ explorer.select_action(t, greedy_action_func)
+
+ self.assertAlmostEqual(explorer.epsilon, 0.1)
|
|
86da129dd4d9665dc15218c1d5b4673ee33780f4
|
factory/tools/cat_logs.py
|
factory/tools/cat_logs.py
|
#!/bin/env python
#
# cat_logs.py
#
# Print out the logs for a certain date
#
# Usage: cat_logs.py <factory> YY/MM/DD [hh:mm:ss]
#
import sys,os,os.path,time
sys.path.append("lib")
sys.path.append("..")
sys.path.append("../../lib")
import gWftArgsHelper,gWftLogParser
import glideFactoryConfig
USAGE="Usage: cat_logs.py <factory> YY/MM/DD [hh:mm:ss]"
# return a GlideinDescript with
# factory_dir, date_arr and time_arr
def parse_args():
if len(sys.argv)<3:
raise ValueError,"Not enough arguments!"
factory_dir=sys.argv[1]
try:
glideFactoryConfig.factoryConfig.glidein_descript_file=os.path.join(factory_dir,glideFactoryConfig.factoryConfig.glidein_descript_file)
glideinDescript=glideFactoryConfig.GlideinDescript()
except:
raise ValueError,"%s is not a factory!"%factory_dir
glideinDescript.factory_dir=factory_dir
glideinDescript.date_arr=gWftArgsHelper.parse_date(sys.argv[2])
if len(sys.argv)>=4:
glideinDescript.time_arr=gWftArgsHelper.parse_time(sys.argv[3])
else:
glideinDescript.time_arr=(0,0,0)
return glideinDescript
def main():
try:
glideinDescript=parse_args()
except ValueError, e:
sys.stderr.write("%s\n\n%s\n"%(e,USAGE))
sys.exit(1)
entries=glideinDescript.data['Entries'].split(',')
log_list=gWftLogParser.get_glidein_logs(glideinDescript.factory_dir,entries,glideinDescript.date_arr,glideinDescript.time_arr,"err")
for fname in log_list:
sys.stdout.write("%s\n"%fname)
sys.stdout.write("===========================================================\n")
fd=open(fname,"r")
sys.stdout.write(fd.read())
fd.close()
sys.stdout.write("\n")
if __name__ == '__main__':
main()
|
Print the logs for a certain date
|
Print the logs for a certain date
|
Python
|
bsd-3-clause
|
holzman/glideinwms-old,bbockelm/glideinWMS,holzman/glideinwms-old,bbockelm/glideinWMS,bbockelm/glideinWMS,holzman/glideinwms-old,bbockelm/glideinWMS
|
---
+++
@@ -0,0 +1,62 @@
+#!/bin/env python
+#
+# cat_logs.py
+#
+# Print out the logs for a certain date
+#
+# Usage: cat_logs.py <factory> YY/MM/DD [hh:mm:ss]
+#
+
+import sys,os,os.path,time
+sys.path.append("lib")
+sys.path.append("..")
+sys.path.append("../../lib")
+import gWftArgsHelper,gWftLogParser
+import glideFactoryConfig
+
+USAGE="Usage: cat_logs.py <factory> YY/MM/DD [hh:mm:ss]"
+
+# return a GlideinDescript with
+# factory_dir, date_arr and time_arr
+def parse_args():
+ if len(sys.argv)<3:
+ raise ValueError,"Not enough arguments!"
+
+ factory_dir=sys.argv[1]
+ try:
+ glideFactoryConfig.factoryConfig.glidein_descript_file=os.path.join(factory_dir,glideFactoryConfig.factoryConfig.glidein_descript_file)
+ glideinDescript=glideFactoryConfig.GlideinDescript()
+ except:
+ raise ValueError,"%s is not a factory!"%factory_dir
+
+ glideinDescript.factory_dir=factory_dir
+ glideinDescript.date_arr=gWftArgsHelper.parse_date(sys.argv[2])
+ if len(sys.argv)>=4:
+ glideinDescript.time_arr=gWftArgsHelper.parse_time(sys.argv[3])
+ else:
+ glideinDescript.time_arr=(0,0,0)
+
+ return glideinDescript
+
+def main():
+ try:
+ glideinDescript=parse_args()
+ except ValueError, e:
+ sys.stderr.write("%s\n\n%s\n"%(e,USAGE))
+ sys.exit(1)
+ entries=glideinDescript.data['Entries'].split(',')
+
+ log_list=gWftLogParser.get_glidein_logs(glideinDescript.factory_dir,entries,glideinDescript.date_arr,glideinDescript.time_arr,"err")
+ for fname in log_list:
+ sys.stdout.write("%s\n"%fname)
+ sys.stdout.write("===========================================================\n")
+ fd=open(fname,"r")
+ sys.stdout.write(fd.read())
+ fd.close()
+ sys.stdout.write("\n")
+
+
+
+if __name__ == '__main__':
+ main()
+
|
|
6b0673334d14dca0e64ab9a760d8652b29e26b07
|
fs/test/test_mkdir.py
|
fs/test/test_mkdir.py
|
from __future__ import with_statement
from nose.tools import (
eq_ as eq,
)
from fs.test.util import (
maketemp,
assert_raises,
)
import errno
import os
import fs
def test_mkdir():
tmp = maketemp()
fs.path(tmp).child('foo').mkdir()
foo = os.path.join(tmp, 'foo')
assert os.path.isdir(foo)
def test_mkdir_bad_exists():
tmp = maketemp()
p = fs.path(tmp).child('foo')
with p.open('w') as f:
f.write('bar')
e = assert_raises(
OSError,
p.mkdir,
)
eq(e.errno, errno.EEXIST)
|
Add more tests for mkdir.
|
Add more tests for mkdir.
|
Python
|
mit
|
tv42/fs,nailor/filesystem
|
---
+++
@@ -0,0 +1,32 @@
+from __future__ import with_statement
+
+from nose.tools import (
+ eq_ as eq,
+ )
+
+from fs.test.util import (
+ maketemp,
+ assert_raises,
+ )
+
+import errno
+import os
+
+import fs
+
+def test_mkdir():
+ tmp = maketemp()
+ fs.path(tmp).child('foo').mkdir()
+ foo = os.path.join(tmp, 'foo')
+ assert os.path.isdir(foo)
+
+def test_mkdir_bad_exists():
+ tmp = maketemp()
+ p = fs.path(tmp).child('foo')
+ with p.open('w') as f:
+ f.write('bar')
+ e = assert_raises(
+ OSError,
+ p.mkdir,
+ )
+ eq(e.errno, errno.EEXIST)
|
|
e3afe5628d42abb109f7e2b3be735ef02941051d
|
data/forms.py
|
data/forms.py
|
from django import forms
from crispy_forms.helper import FormHelper
from crispy_forms.layout import Layout, Div
class JobTemplateForm(forms.Form):
name = forms.CharField(max_length=400)
template = forms.CharField(
widget=forms.Textarea(
attrs={
'cols': 50,
'rows': 26,
}),
required=False,
)
helper = FormHelper()
helper.form_tag = False
helper.layout = Layout(
Div(
Div('name', css_class='col-xs-12'),
css_class='row'),
'template',
)
|
Create a form for the JobTemplate
|
Create a form for the JobTemplate
|
Python
|
mit
|
crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp
|
---
+++
@@ -0,0 +1,24 @@
+from django import forms
+from crispy_forms.helper import FormHelper
+from crispy_forms.layout import Layout, Div
+
+
+class JobTemplateForm(forms.Form):
+ name = forms.CharField(max_length=400)
+ template = forms.CharField(
+ widget=forms.Textarea(
+ attrs={
+ 'cols': 50,
+ 'rows': 26,
+ }),
+ required=False,
+ )
+
+ helper = FormHelper()
+ helper.form_tag = False
+ helper.layout = Layout(
+ Div(
+ Div('name', css_class='col-xs-12'),
+ css_class='row'),
+ 'template',
+ )
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.